0% found this document useful (0 votes)
14 views44 pages

2008.06086

The document discusses the technical considerations for using the Verasonics Research Ultrasound Platform in developing a Photoacoustic Imaging (PAI) system. It outlines the setup, performance parameters, and signal processing protocols necessary for effective US/PA imaging, emphasizing the advantages of linear-array based PAI for clinical applications. The paper also provides detailed guidelines and modifications for using various transducers with the Vantage system, highlighting its capabilities in enhancing diagnostic imaging and therapeutic monitoring.

Uploaded by

huanqiu1123
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
14 views44 pages

2008.06086

The document discusses the technical considerations for using the Verasonics Research Ultrasound Platform in developing a Photoacoustic Imaging (PAI) system. It outlines the setup, performance parameters, and signal processing protocols necessary for effective US/PA imaging, emphasizing the advantages of linear-array based PAI for clinical applications. The paper also provides detailed guidelines and modifications for using various transducers with the Vantage system, highlighting its capabilities in enhancing diagnostic imaging and therapeutic monitoring.

Uploaded by

huanqiu1123
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 44

See discussions, stats, and author profiles for this publication at: https://www.researchgate.

net/publication/343689469

Technical Considerations when using Verasonics Research Ultrasound Platform


for Developing a Photoacoustic Imaging System

Preprint · August 2020


DOI: 10.48550/arXiv.2008.06086

CITATIONS READS

0 5,931

5 authors, including:

Karl Kratkiewicz Moein Mozaffarzadeh


Wayne State University KROHNE
31 PUBLICATIONS 900 CITATIONS 72 PUBLICATIONS 1,308 CITATIONS

SEE PROFILE SEE PROFILE

Kamran Avanaki
University of Illinois Chicago
110 PUBLICATIONS 1,202 CITATIONS

SEE PROFILE

All content following this page was uploaded by Moein Mozaffarzadeh on 07 September 2020.

The user has requested enhancement of the downloaded file.


Technical Considerations when using Verasonics Research
Ultrasound Platform for Developing a Photoacoustic Imaging
System
Karl Kratkiewicza¥, Rayyan Manwara,b¥, Yang Zhoua, Moein Mozaffarzadehc, and Kamran
Avanakia,b,d*
a
Wayne State University, Department of Biomedical Engineering, Detroit, USA
b
University of Illinois at Chicago, Department of Bioengineering, Chicago, USA
c
Tarbiat Modares University, Department of Biomedical Engineering, Tehran, Iran
d
University of Illinois at Chicago, Department of Dermatology, Chicago, USA

Abstract
Significance: Photoacoustic imaging (PAI) is an emerging functional and molecular imaging technology that has
attracted much attention in the past decade. Recently, many researchers have used the Vantage Verasonics® research
system for simultaneous ultrasound (US) and photoacoustic (PA) imaging. This was the motivation to write on the
details of US/PA imaging system implementation and characterization using Verasonics platform.
Aim: We describe the subtle details of US/PA imaging system setup, study the performance parameters of the system,
and explain sequencing of the US/PA signal generation and signal amplification as well as the details required for
efficient use of the hardware of the system and data processing protocols.
Approach: We focused on linear-array based PAI due to its popularity and simple setup, as well as its high potential
for clinical translatability.
Results: We have shown the sequencing of the US/PA signal generation, signal amplification, and related data
processing protocols. A step-by-step guideline to develop and characterize PAI system using Vantage 128 has been
presented. Some of the limitations of the vantage system are also listed.
Conclusions: Photoacoustic imaging is a complement to the already established US imaging technique and may
significantly increase its scope of application in diagnostic imaging and therapeutic monitoring. Combining with
commercial medical US systems, the development of PAI can be accelerated by taking advantage of US image
reconstruction and processing. With the information we presented in the body of this review and the four appendices,
we described most of the experimental considerations one should know when working with the Vantage system for
PAI tests.

Keywords: Photoacoustic imaging, ultrasound, linear array, image reconstruction, Vantage Verasonics®.

*Kamran Avanaki, E-mail: [email protected]


¥
These authors have contributed equally.

1
1 Introduction

Photoacoustic imaging (PAI), also called optoacoustic imaging, is a three-dimensional (3-D)

imaging modality that works based on the photoacoustic (PA) effect 1. The sample (light

absorbent) to be imaged is optically excited, leading to a transient temperature rise, resulting in a

thermoelastic expansion of the absorber followed by emission of acoustic waves. The absorber

could be endogenous such as hemoglobin (oxy- or deoxy-), myoglobin (oxy- or deoxy-), melanin,

lipid, Bilirubin, water, or an exogenous contrast agent such as dyes 2. The absorption spectrum of

some of the endogenous and exogenous absorbers are shown in Figure 1; the absorption spectrum

of two example imaging targets that are commonly used for calibration and performance

evaluation in PAI experiments, i.e., vinyl black tape and carbon pencil lead, are given in Appendix

A.

Fig. 1 Absorption spectrum of (a) endogenous, and (b) exogenous contrast agents. Reproduced with permission from

A =  (  ) Cd  ( ) C
3-5
. According to Beer-Lambert law, absorbance is defined by where, , , and d are molar

absorptivity, concentration, and cross-section thickness of the contrast agent, respectively. Absorption coefficient is

 = ( A d ) log10 e 6
expressed as: .

2
The emitted acoustic waves from the absorber are detected by ultrasound (US) transducers.

The transducer signals are then given to an image reconstruction algorithm to generate the

absorption map of the tissue. The PAI process steps are illustrated in Figure 2.

Fig. 2 Principle of PA signal generation, detection, and image reconstruction. Reprinted with permission from Ref. 7.

A high energy and short pulse laser light illuminates the absorber, leading to a transient temperature rise which results

in a thermoelastic expansion of the absorber, and acoustic (or PA) wave generation. The signals generated from the

waves received by an ultrasound probe are given to a reconstruction algorithm to form a PA image.

Due to strong optical scattering in biological tissues, pure optical imaging modalities have a
1, 8-10
shallow imaging depth . The transport mean free path (i.e., the mean distance after which a

photon’s direction becomes random) in biological tissues is around 1 mm 3. Acoustic waves


11, 12
experience far less tissue scattering, thus they propagate a greater distance . Although

ultrasound imaging can image deep in biological tissues with a high spatial resolution, its acoustic

contrast is incapable of providing certain physiological parameters 13. In PAI, there is no restriction

for photons, thus optical excitation can travel far beyond the diffusion limit and still generate

acoustic waves. Sensitivity of PAI in deep tissues is orders of magnitude higher than that of pure

optical imaging modalities 13, 14; the highest penetration depth reported in PAI is ~12 cm 15-17. PAI

is an ideal modality for measuring or monitoring tissue physiological parameters by imaging the
3
concentration of tissue chromophores 18, which are changed during the course of a disease 19. PAI

has been evaluated in preclinical and recently in clinical applications for disease detection and

monitoring purposes 20-41. For instance, it has been used to study human skin abnormalities 37, 42-

44
, brain disease detection 45, 46, human breast tumor detection 19, 47-50, retina disease diagnosis 51,
52
, and atherosclerosis evaluation of vessel walls 30, 53, 54.

PAI can be implemented in tomography or microscopy configurations with variations in

system design (more details are provided in 7). Among all configurations and variations of PAI,

linear-array based PAI is one of the most commonly used due to its straightforward setup, easy

use, and simple clinical translatability. In addition, since linear-array ultrasound imaging systems

have been well-established as clinical tools, slightly modifying them to develop more capable

tools, i.e., US/PA imaging systems, is not far from reality.

In recent years, many researchers have used the Vantage Verasonics® system for simultaneous

US and PA imaging. Vantage system is a MATLAB-friendly programmable platform that can

produce co-registered US and PA images; the PA image represents the optical absorption map of

the tissue while the US image represents the tissue acoustic impedance map. Here, we describe the

subtle details of US/PA imaging system setup, study the performance parameters of the system,

and explain sequencing of the US/PA signal generation and signal amplification as well as the

details required for efficient use of the hardware of the system and data processing protocols.

2 Vantage system architecture for US/PA sequencing

Vantage hardware is controlled by MATLAB coding. The detailed explanation of the parameters

used in the code for data acquisition and processing, image reconstruction, and display are given

in Appendix B. The Vantage system architecture for US/PA sequencing is shown in Figure 3A.

The structures for US and PA run-time defined in the MATLAB code communicate to runAcq

4
through the activation of the MATLAB loader function VSX. This opens a GUI including imaging

windows of the reconstructed images from RunAcq (US and PA), time gain control, US voltage

transmit, and many other parameters that are used for real-time user modification to the imaging

sequence. These modifications take effect on the hardware between cycles of hand-offs between

VSX and RunAcq. After initial activation of VSX, RunAcq then communicates all of the run-time

transmit and receive structures to the Vantage hardware (silver box) through the PCIe cable.

RunAcq, receives the raw data, stores it in the RcvBuffer variable, and performs image

reconstruction on US and PA data and sends it back to the VSX GUI. The raw data of US and PA

are accessible after completion of the imaging sequence 55.

Fig. 3 Vantage system architecture and data acquisition sequencing. (a) Vantage system architecture for US/PA

sequencing and (b) timing diagram of US/PA sequencing in Vantage system. US: Ultrasound, PA: Photoacoustic,

Recon.: Reconstruction. RcvBuffer is the predefined buffer where the US and PA data are stored, Acq: acquisition,

5
TX: transmit, RX: receive, A/D: analog to digital, VSX: Verasonics script execution, GUI: graphical user interface,

RcvBuffer: receive buffer, CPU: central processing unit.

The timing sequence in US and PA data acquisition, illustrated in Figure 3B, is as follows. US

acquisition requires a user defined number of steered angles. Between each set of US acquisitions,

the system waits for a laser trigger input to begin one PA event. After PA acquisition, US and PA

data are transferred to the host controller and stored in the RcvBuffer variable. RunAcq then

performs image reconstruction on the averaged data of all of the steered angles and also on the PA

data. The reconstructed US and PA images are then displayed in the MATLAB VSX GUI.

3 US/PA experimental setup and system characterization

To perform an US/PA experiment, an Nd:YAG laser (PhocusMobil, Opotek Inc., CA, USA) with

a repetition rate of 10 Hz and a pulse width of 8 ns was used. The laser uses an optical parametric

oscillator (OPO) to tune the wavelength of light between 690 nm to 950 nm. For light delivery, we

used a custom fiber bundle (Newport Corporation, Irvine, CA, USA). For data acquisition, the

Verasonics Vantage 128 system was used. The specifications of the vantage system are listed in

Table 1.

Table 1. High frequency Vantage 128 system specifications (Verified by Verasonics’ technical support).
Parameters Specification
Channels 128 Tx / 128 Rx
Power requirements 100V-240V (50-60 Hz)
Programmable pulser voltage 2 to 190 V p-p
Time delay resolution 4.0 ns
HIFU capability Yes
A/D resolution 14 bits
Max sampling rate 62.5MHz
Fastest external trigger 100 KHz
Maximum trigger voltage level 5.5 V
Single transmit event ~ µsec
Data pre-amplification 24 dB
Second stage Amplification 30 dB

6
Noise Figure 1.5 to 3.0 dB
via 8 lanes PCIe 3.0
Data transfer to host computer
sustained data transfer rates up to 6.6 GB/s

The Vantage system is connected to a host computer using a PCIe express cable. A transducer

is connected to the 260 pin Cannon connector mounted on the Vantage system. On the sample end,

the transducer was placed and held perpendicularly to the sample. There is a coupling media

between the imaging target and the transducer. Water and ultrasound gel are the most common

couplant 56. More details about acoustic couplants can be found in Appendix C.

The Verasonics software package provides a PA MATLAB script,

SetUpL11_4vFlashAngPA.m, for the 128-element linear array L11-4v probe (see Appendix B).

This script can be modified to be used with other probes. Two commonly used ultrasound

transducers with Vantage system are ATL Philips L7-4 and Verasonics L22-14v. The

specifications of these transducers are listed in Table 2.

Table 2. Specifications of linear array transducers ATL Philips L7-4, and Verasonics L22-14v
Model (Unit) L7-4 L22-14v
Lens correction 0.8870 0.6804
Central frequency (MHz) 5.2080 15.6250
Number of Elements [4, 7] [14, 22]
Element width (mm) 128 128
Spacing (mm) 0.2500 0.0800
Max high voltage (V) 0.2980 0.1
Focal length (mm) 50 30

Photograph of the device

Dimension
(11cm × 6.0cm × 3.0cm) (9cm × 3.5cm × 2.5cm)

7
Table 3 lists the modifications on the original code, written for the L11-4v, needed for the L7-

4 and L22-14v. The remainder of the script is compatible as long as the probe is a 128-element

linear array.

Table 3. Modifications to original L11-4 script by Verasonics, for ATL Phillips L7-4 and Verasonics L22-14v probes
Parameters Original Script Modified Script
P(1).startDepth = 2; P(1).startDepth = 0;
P(1).endDepth = 192; P(1).endDepth = 192;
Imaging depth
P(2).startDepth = 0; P(2).startDepth = 0;
P(2).endDepth = 128; P(2).endDepth = 192;
oneway = 0; oneway = 1;
Laser parameters flash2Qdelay = 200; flash2Qdelay = 188;
PA_PRF = 100; PA_PRF = 10;
Trans.name = ‘L11-4’; Trans.name = ‘L7-4’;*
Trans.units = ‘mm’; Trans.units = ‘mm’;
Active transducer Trans = computeTrans(Trans); Trans = computeTrans(Trans);
nElem = Trans.numelements; nElem = Trans.numelements;
Trans.maxHighVoltage = 50; Trans.maxHighVoltage = 50;*
*Trans.name = ‘L22-14v’ and Trans.maxHighVoltage = 30 in the case of the L22-14v.

The Vantage system supports linear array transducers from different companies. The list of

different transducer arrays that can be used with the Vantage system is given in Table 4. If an

unrecognized transducer needs to be used, the transducer attributes must be determined

independently and input into the computTrans.m script under a new custom transducer case. Key

attributes include the central frequency, bandwidth, number of elements, element width, element

spacing, element position, and connector pinout arrangement. In addition to unrecognized linear

arrays not listed in Table 4, single-element-based arrays, ring arrays, and hemispherical arrays can

also be connected to the Cannon connector mounted on Vantage system via a custom-made

converter.

Table 4. List of immediately recognizable transducers for the Vantage system


Parameters Transducer Model Number
L7-4 C8-4V*
Philips L11-5 C8-5*
L12-5 38mm C9-5ICT

8
L12-5 50mm P3-2*
CL10-5 P4-1
CL15-7 P4-2
C4-2 P5-3*
C5-2 P6-3
C7-4* P7-4
L10-4v L22-14vX
L11-4v L22-14vX-LF
L11-5v L35-16vX
Verasonics L12-3v L38-22v
L22-8v C5-2v
L22-14v P4-2v
L22-14vLF
GE9LD GEIC5-9D
GE GEC1-6D GEL3-12D
GE4CD GEM5ScD
Siemens L10-5
Vermon 4DL7*
*These transducers are recognizable by the system, but the system does not have the transducer’s
attributes which must be manually entered into the computeTrans.m MATLAB file.

To demonstrate the PA signal characteristics in both time and frequency domains, a 2 mm

diameter carbon lead phantom in water was imaged at 690 nm. The energy after the fiber bundle

was measured as ~15 mJ/cm2. L22-14v transducer array was used in this experiment. The time

domain signal obtained from the central transducer (i.e., 64th element), is shown in Figure 4A. In

Figure 4B, the magnitude of the single-sided Fast Fourier Transform (FFT) of the time domain

signal is shown. In Figure 4A, we also show that the strength of the US waves reflected back from

the phantom is greater than that of the photoacoustic 57, 58.

9
Fig. 4 Photoacoustic signal in (A) time-domain, and (B) frequency-domain, obtained from a 2 mm diameter carbon

lead phantom, imaged with L22-14v transducer at 690 nm wavelength. The red signal in A shows the overlaid US

signal acquired from the same sample.

We characterized L7-4 and L22-14v for US/PA imaging. We created a resolution phantom of

hair of diameter 54 µm in an open top plastic cubic box filled with deionized water (see Figure

5A). The lateral and axial resolutions were quantified by measuring the full width half maximum
59
(FWHM) of the normalized 1D intensity profile across the hair image . The axial and lateral

resolutions of the L7-4 were measured at depths 0.5 to 5.0 cm with steps of 0.5 cm (see Figure

5B). The axial and lateral resolutions of the L22-14v were measured at depths 0.2 to 2.0 cm with

steps of 0.2 cm (see Figure 5C). In L7-4, the axial resolution remained constant in both US and

PA imaging while the lateral resolution worsened with depth, whereas in L22-14v, the axial and

lateral resolutions stayed almost constant in both US and PA images. This is potentially due to the

focal depth specified by the Vantage system. High frequency transducers have less beam

10
divergence than low frequency transducers, this may have also contributed to worsening the lateral

resolution of the L7-4 and constant lateral resolution of the L22-14v 60.

Fig. 5 US/PA resolution study. (A) Schematic of the experimental setup, including a hair phantom photograph

captured by a 4× objective on a light microscope (SME-F8BH, Amscope, CA, USA). Resolution study when (B) L7-

4 was used, (C) L22-14v was used. (i) US axial resolution versus depth, (ii) US lateral resolution versus depth, (iii)

PA axial resolution versus depth, (iv) PA lateral resolution versus depth.

11
4 Technical Considerations

4.1 Synchronization Optimization for Efficient Averaging

Photoacoustic imaging relies on careful timing of data acquisition and laser firing to accurately

reconstruct photoacoustic depth information. In Q-switched Nd:YAG lasers, the timing involves

triggering of the flash lamps to stimulate the emission medium (Nd:YAG) followed by time for

optical buildup and subsequent opening of the Q-switch for release of the laser beam. A

straightforward triggering method (see Figure 6A (i)) involves using internal flash lamp and Q-

switch triggers built into the laser being used and using the flash lamp out port to trigger data

acquisition on the Vantage trigger in-1 port. Upon receipt of the flash lamp trigger, the Vantage

waits for a user-specified delay time equal to the known optical build-up time for the laser and

then begins recording approximately when the laser fires. The major drawback to this triggering

method is that the optical build-up time for the laser will “jitter” between tens of nanoseconds and

will cause the peak PA signal to fluctuate between a few sample numbers. This will cause issues

in data averaging between subsequent frames. We have implemented a more optimal triggering
61
method following the method described in (see Figure 6A (ii)). In this method the flash lamp

and Q-switch of the laser are both externally triggered. Using a function generator (ATF20B,

Shenzhen Atten Electronics Co., Ltd., Nanshan, China) a trigger of 10Hz, 5Vpp, 2.5V offset, 50%

duty cycle was simultaneously sent to the flash lamp-in of the Nd:YAG laser (PhocusMobil,

Opotek, CA, USA) and the trigger in-1 of the Vantage system. Within the Vantage MATLAB

script, the flash2Qdelay parameter is set to the external Q-switch delay time specified by the laser

retailer (290 µs). This allows the Vantage to then trigger the Q-switch by connecting the Vantage’s

trigger-out to the Q-switch-in on the laser. We have compared the performance of the two

triggering methods by imaging a 2mm diameter carbon lead, when L7-4 US probe was used for

12
PA signal detection and the fiber bundle we described earlier was used for illumination at 690 nm.

Five sequential photoacoustic frames were recorded and overlaid with both methods (see Figures

6B(i) and 6B(ii)). The results show a greater “jitter” in the first triggering method compared to

nearly no “jitter” in the second method. We then averaged the signals acquired with each of these

triggering methods. Averaging improved the signal-to-noise ratio (SNR) of the PA signal probably

because the signal components being averaged were correlated and the noise components were

not. 64th element of the L7-4 probe data was averaged using various numbers of stored frames: 1,

10, and 50, and plotted (see Figure 6B(iii)).

Fig. 6 Photoacoustic signal “jitter” comparison between two triggering methods when a 2 mm carbon lead was imaged.

64th element signals are plotted. (A) (i) Schematic of straightforward triggering method where laser triggers Vantage

system, (ii) schematic of the function generator driven triggering method where function generator triggers laser flash

lamp and Vantage system, which then triggers laser Q-switch, (B) (i) five sequential PA frames showing significant

13
“jitter” between frames, (ii) five sequential PA frames showing nearly no “jitter” between frames, and (iii) averaged

frames using methods described in A(i) and A(ii).

4.2 Illumination Angle Matters

62
Light illumination is one of the main components of a photoacoustic imaging system . The

orientation of the illumination of light can influence the amplitude of the PA signal received by
12
the transducer based on how the light energy is deposited on the object . To investigate this

matter, we imaged a 2 mm diameter carbon lead in a scattering media, i.e., 2% milk, at 690 nm.

We acquired PA data at different illumination angles. The angles were measured as the angle

between the transducer and the illumination plane (see the experimental setup in Figure 7A). In

these experiments, the transducer plane was held constant (always perpendicular to the sample)

and the illumination plane was changed while the illumination spot was at the same location on

the surface of the phantom. Figure 7B (i) shows the data collected from the 64th element of the L7-

4 probe at an angle of  = 40o where the PA signal was weakest. Figure 7B (ii) shows the data

collected from the 64th element of the L7-4 probe at an angle of 48o where the PA signal was

strongest. The difference between the amplitudes of the PA signals is probably related to: (i) the

light illumination path at each angle which depends on the scattering map of the phantom, (ii) the

mechanical property of the tissue as to which directions the thermoelastically generated pressure

waves have the maximum amplitude, and (iii) at greater angles, larger area of the target phantom

is illuminated, hence, more numbers of point sources are excited and collectively a larger pressure

wave is generated. It can be seen that the shape of the two induced PA signals are consistent while

the PA signal amplitude is higher at one angle over another. This demonstrates the importance of

optimizing illumination in PA imaging.

14
Fig. 7 Impact of illumination angle. (A) Experimental setup of illumination angle investigation. (B) PA signal profile

from 64th element of L7-4 probe taken from a 2 mm carbon lead phantom with different illumination angles. (i) 40

degrees, (ii) 48 degrees.

4.3 Improving Reading Accuracy by Upsampling

The Vantage system automatically samples signals at a rate that is 4 times the central frequency of

the transducer being used. This is sufficient to meet the Nyquist limit, however it may be necessary

to analyze PA signals at a higher sampling rate. This requires modifications to the script as well

as the use of an in-built filter tool to change the spectrum of the bandpass filter on the receive end

of the transducer so that the system does not mistake the higher sampled signals as high frequency

noise. First, the Resource.RcvBuffer(1).rowsPerFrame parameter is multiplied by the factor that

the sampling rate will be multiplied. This is to allow the data buffer to hold the increased number

of samples. Similarly, the Receive.decimSampleRate parameter is multiplied by the same factor.

Finally, the parameter Receive.inputFilter is modified by entering filterTool into the MATLAB

command line. This brings up a GUI which outputs the Receive.inputFilter value based on the user

input parameters. Inside the GUI, sampleMode are set to custom with the decimSampleRate set to

the rate of choice. The bandpass filter central frequency and relative bandwidth are then set to the

15
standard values of the probe being used. The modifications to the script can be seen in Table 5 for

using the L7-4 probe and changing the sampling rate from 20.8MHz to 62.4MHz.

Table 5. Script modifications to change the sampling rate of L7-4 from 20.8 MHz to 62.4 MHz. The modifications
are indicated in red.
Modification Original Modified
no.
1 Resource.RcvBuffer(1).rows Resource.RcvBuffer(1).rowsPerFra
PerFrame = 2048*(na + ne); me = 3*2048*(na + ne);

2 Receive = Receive = repmat(struct('Apod',


repmat(struct('Apod', ones(1,Trans.numelements), ...
ones(1,Trans.numelements), 'startDepth', P(1).startDepth,
... ...
'startDepth', 'endDepth', P(1).startDepth +
P(1).startDepth, ... wl4sPer128*ceil(maxAcqLngth2D/wl
'endDepth', 4sPer128), ...
P(1).startDepth + 'TGC', 1, ...
wl4sPer128*ceil(maxAcqLngt 'bufnum', 1, ...
h2D/wl4sPer128), ... 'framenum', 1, ...
'TGC', 1, ... 'acqNum', 1, ...
'bufnum', 1, ... 'sampleMode', 'NS200BW', ...
'framenum', 1, ... 'mode', 0, ...
'acqNum', 1, ... 'callMediaFunc', 0,...
'sampleMode', 'NS200BW', 'inputFilter', [+0.00000
... +0.00122 +0.00400 +0.00687
'mode', 0, ... +0.00677 +0.00378 +0.00323 ...
'callMediaFunc', 0), 1, +0.00931 +0.01569 +0.00861 -
(na+ne)*Resource.RcvBuffer 0.01431 -0.03320 -0.02631 -
(1).numFrames); 0.00491 ...
-0.01303 -0.07413 -0.14075 -
0.12170 +0.02014 +0.20425
+0.28894],...
'decimSampleRate', 20.8333*1),
1,
(na+ne)*Resource.RcvBuffer(1).nu
mFrames);

We imaged a black tape of 18 mm width at 3 sampling rates, 20.8, 41.6, and 62.4MHz. The

tape was imaged in a water bath with the L7-4 probe at the illumination wavelength of 690 nm.

The results for these sampling rates are given in Figure 8A, B, and C, respectively. To demonstrate

the improvement in the reading of the signals with increasing the sampling frequency, we

calculated the difference between the top peaks and also the bottom peaks in the signal. The values

16
are indicated in each figure. One drawback of using higher sampling frequency is slower available

frame rate, therefore, an alternative solution to increase the number of samples without increasing

the sampling frequency is interpolation. In Figure 8D, we showed the interpolated signal (3 times

up sampling) using the ‘spline’ algorithm 63


. We also provided the location (i.e., index) of the

highest peak value normalized by respective sample numbers ( xnor = x ( ymax ) sample no. ) and the ratio

of peak values (both positive and negatives).

Fig. 8 PA signal profile from 64th element of L7-4 probe taken from a black tape phantom of 18mm width with varying

sampling rates: (A) 20.8 MHz, (B) 41.6 MHz, (C) 62.4 MHz, and (D) interpolated signal using ‘spline’ algorithm.

xnor = x ( ymax ) sample no. y r x = Ratio of y values , where y = positive or negative peaks .
Here, and

4.4 Simultaneous Real-time Visualization of US and PA Images

The default MATLAB script provided by Verasonics® utilizes a single imaging window for both

US and PA image display, with the user being able to modify parameters to view one or the other.

With real-time imaging, it is more important to simultaneously view US and PA images side by

side. The following describes how to modify the original example script to simultaneously display

US and PA images in real-time. With a single imaging window, all parameters for the Resource

structure have arguments of 1. Adding a second window requires duplication of all of the

Resource.DisplayWindow parameters with the argument of 2. Specifically, the variables Title,

17
pdelta, Position, ReferencePt, numFrames, AxesUnits, Colormap, and splitPalette are copied from

the already existing Resource.DisplayWindow(1) structure with the argument changed to 2. To

differentiate between US in window 1 and PA in window 2, we set

Resource.DisplayWindow(1).Colormap to gray(256) and Resource.DisplayWindow(2).Colormap

to hot. The other change necessary to keep both windows real-time is modification of the

Process(n).Parameters structure where n is 1 for US and 2 for PA. In the original structure,

mappingMethod is set to lowerHalf for US and upperHalf for PA, but both should now be set to

full. Also, displayWindow was originally set to 1 in both processes and now it should be 1 for US

and 2 for PA. Modifications to the original script to simultaneously display US and PA images, in

real-time are given in Appendix D (the modifications are indicated in red).

4.5 US Image Quality Improvement

Vantage offers beam steering as a method of improving US image quality. This technique modifies

transmit delay times in the beam apodization to steer the beam at different directions into the

imaging field. This provides received echoes from different angles of reflection off of the imaging

targets. The received echoes from multiple steered angles are then averaged to form an improved

image. The variable na in the MATLAB script can be set to any positive integer to determine the
𝑛𝜋
number of steered angles. The transmit waveforms will be emitted at angles 𝑁+1, where N is equal

𝜋
to the variable na. n iterates with each transmit angle from 1 to N, where is normal to the
2

transducer elements. Upon definition of na, these angles and corresponding time delays in the

apodization for each element are calculated, and the time occurs between each US steered angle is

determined. Taking this time into account with the pulse repetition rate of the laser, the number of

the angles occur between each laser pulse is determined. Further, increasing the number of steered

18
angles and the number of stored frames in each buffer are balanced to avoid slowdown of the

system due to memory build up, which can potentially freeze MATLAB if the system RAM is

overflowed. To investigate the effect of the number of steered US transmit angles on the image

quality, we imaged a phantom, with 2 mm, 0.9 mm, and 0.2 mm diameter carbon leads in water

with varying numbers of steered angles (i.e., 8, 16, 32, 64, 128), and quantified the contrast to
12
noise ratio (CNR) of resultant images following the method explained in . Figure 9A

demonstrates the experimental setup. Figures 9(B-F) show the US image for 8, 16, 32, 64, and 128

steered angles, respectively. Figure 9G shows the improvement of CNR with greater number of

steered angles.

Fig. 9 US image quality improvement. (A) Experimental setup showing an ultrasound probe and a 3-carbon lead

phantom. (L1) 2 mm diameter carbon lead, (L2) 0.9 mm diameter carbon lead, (L3) 0.2 mm diameter carbon lead, (B-

F) US images of the 3 carbon lead phantom for 8, 16, 32, 64, and 128 steered angles, respectively. Green dashed circle

encloses object pixels and red dashed box encloses both object and background pixels, (G) CNR versus number of

steered angles

19
4.6 Photoacoustic Signal Amplification

The PA pressure waves received by the linear array transducer are weak (i.e., their SNR is low) 64,
65
. In cases where the imaging media is highly scattering, only weak pressure waves reach the

transducer and produce low level voltage signals. These low level signals may not be accurately

digitized using the default Analog Front End (AFE) settings of the Vantage data acquisition (DAQ)

system. The AFE gain can be increased by up to 12 dB by modifying the settings in the RcvProfile

structure, however in some cases, additional gain may be required. To increase the amplitude of

the received signal, we used a 128-channel amplifier (AMP128-18, Photosound, Houston, TX,

USA). The amplifier has 40 dB gain with -6 dB cut off frequencies of 25 kHz and 35 MHz. The

amplifier was attached to the 260 pin Cannon connector on Vantage system (see Figure 10). The

US probe was then plugged into the front end of the amplifier. In the MATLAB script, i.e.,

computeTrans.m, a custom transducer ID was added. Details of the ultrasound transducer were

copied to the custom ID so that the amplifier can be recognized. This is similar to the addition of

any unrecognized transducer to the system, only the transducer attributes (e.g., central frequency,

bandwidth, number of elements, element width, element spacing, and element position) are

determined independently and input into the computTrans.m script under a new custom transducer

case. To allow for identical wiring path lengths through the amplifier from the transducer end to

Vantage system end, the pins were not connected one to one and their organization were corrected

in the MATLAB script in the Trans.Connector variable. Since the amplifier was programmed to

multiply any signal going in or out, to avoid causing damage to the transducer elements during US

transmit, the transmit beam was disabled by setting TX.Apodization to zeros. This allowed the

Vantage system to only work in the receiving mode with amplification. To demonstrate the utility

of the amplifier, a two-wire phantom (each with 600 µm diameter) was imaged with and without

20
the amplifier at varying concentrations of intralipid, i.e., 0%, 25%, 50% (Sigma Aldrich, USA).

The experimental setup is shown in Figures 10A and B. The setup consists of a 50 Hz, 532 nm

Nd:YAG laser (NL231-50-SH, EKSPLA, Vilnius, Lithuania) with 7 ns pulses, and an L7-4 US

transducer. The laser illuminated the 2-wire phantom through a large fiber optic bundle (1 cm

diameter, FO Lightguide, Edmund Optics, NJ, USA). We observed an SNR improvement in

addition to the signal peak increase in all concentrations of intralipid (see Figures 10B, C, and D).

SNR improvement was probably due to the high pre-amplifier input impedance which shifted the

transducer noise spectrum to low frequencies, which was then filtered out using a high-pass filter
66
.

21
Fig. 10 PA signal amplification. (A) System setup to test the Photosound AMP128-18 amplifier: (i) display monitor,

(ii) Nd:YAG laser head, (iii) fiber optic bundle, (iv) laser power supply, (v) laser chiller. (B) Wire imaging results:

wire phantom suspended in 0% intralipid imaged (i) without and (iv) with amplifier, wire phantom suspended in 25%

intralipid imaged (ii) without and (v) with amplifier, and wire phantom suspended in 50% intralipid image (iii) without

22
and (vi) with amplifier. (C) Raw data from the 64 th channel of the linear array when imaging a single wire in 50%

intralipid concentration to demonstrate SNR improvement from (i) without amplifier to (ii) with amplifier, (D) Bar

chart demonstrating signal amplitude increase corresponding to phantom imaging in B with and without amplifier, at

different concentrations of intralipid.

5 Fluence Compensation

The initial PA pressure is proportional to the product of local optical absorption coefficient and

the optical fluence deposited at the same location. However, the optical fluence could vary

significantly in living tissue which necessitates the use of optical fluence compensation for
67
quantitative PA imaging . If the optical parameters of the tissue, e.g., absorption coefficient,

scattering coefficient, and anisotropy factor, are known, the optical fluence can be calculated by

solving the photon diffusion equation using Monte Carlo (MC) or finite element method 68, 69. With

the known optical fluence distribution, a fluence-compensated absorption map can be calculated

via dividing the PA reconstructed image by its corresponding fluence map.

We evaluated the fluence compensation algorithm on a phantom made of a gelatin, intralipid,

and ink mixture (see Figure 11A and B). We used 0.5 mm diameter carbon lead in three layers of

gelatin phantom with thicknesses of 5, 8, and 7 mm, respectively. We added intralipid (Sigma

Aldrich, USA) as scatterers and used ink to represent absorption. The concentration of the

intralipid solutions were 4%, 1%, and 6%, and those of ink were 0.1%, 0.4%, and 0.2%,

respectively; these values were chosen to represent various biological tissues. Using Mie calculator
70
, the scattering coefficient, absorption coefficient, and anisotropy factor of the layers were

calculated at 532 nm. The scattering coefficients were 15.4 cm-1, 3.1 cm-1 and 22.3 cm-1,

respectively. The absorption coefficients were 0.3 cm-1, 0.9 cm-1, and 0.4 cm-1, respectively. The

absorption coefficient of water, i.e., 0.11 cm-1, was added to these numbers. We used another layer

23
of pure gelatin with the thickness of 5 mm. The scattering coefficient of both pure gelatin and US

gel were considered 0.05 cm-1. The absorption coefficient of US gel (3mm layer) was considered

0.11 cm-1. The experimental protocol was as follows: (1) we imaged the phantom using the L7-4

probe and generated an US image (see Figure 11C) and a PA image (see Figure 11D); (2) we then
71
segmented the US image using the segmentation method given in ; (3) the abovementioned

optical properties were assigned to each segment, creating the phantom optical model; (4) fluence

map (see Figure 11E) of the phantom optical model was generated by an MC simulation (using

MCX software 72); (5) finally, the PA image was divided by the MC simulated image. The resultant

image was a fluence-compensated PA image (see Figure 11F). The PA intensity variation between

the peaks (at the location of imaging targets) in the fluence-compensated PA image was less than

5%. In this experiment we assumed that the field of view of each element of the transducer is

cylindrical.

Fig. 11 Fluence Compensation algorithm validation. (A) Schematic of a phantom with 3 layers of gelatin phantom

mixed with intralipid/ink solution and 3 carbon lead imaging targets, (B) photograph of the phantom made in a cube

24
box and magnified to show the layers, (C) US image of the phantom, (D) PA image before fluence compensation, (E)

Monte Carlo simulation of light propagation in the phantom, (F) PA image after fluence compensation.

6 PA Image Reconstruction

After detecting PA signals by the linear array, the channel data are given to an image reconstruction

algorithm and an optical absorption distribution map, i.e., PA image, is reconstructed 73-77. Several
78-81
image reconstruction algorithms have been studied for linear-array PA imaging . The most

basic algorithm is delay-and-sum (DAS). DAS follows the dynamic focusing protocol, in which

the focus is adjusted for each pixel of the imaging target; the detected signals are delayed

proportional to the distance between the focal point and the position of the element in the imaging

array. Finally, the delayed signals are summed up, and an image is formed. As a result of this

summation, the on-axis signals will overlap on each other while off-axis signals will be suppressed.

In the remaining of this section, we used two signal quality metrics to compare the performance

of different image reconstruction algorithms: (i) resolution, which is defined as the full width half

maximum (FWHM) of the lateral profile of the image of a point source, and (ii) sidelobe, defined

as the power of the first lobe next to the main lobe. Although DAS is simple to implement and

fast, it treats all the detected signals the same way which results in low resolution images with high

sidelobes; this is mainly due to the constructive overlap of the off-axis signals 82. Delay-Multiply-

and-Sum (DMAS) can generate images with a finer resolution and lower sidelobes compared to

DAS 83. Double stage DMAS (DS-DMAS) algorithm offers finer resolution, and lower sidelobes
80, 82, 84
than DMAS; DS-DMAS uses two stages of correlation process to suppress off-axis signals.

Minimum variance (MV) significantly improves the image resolution compared to DAS, DMAS

and DS-DMAS, however the produced images have high sidelobes; this problem has been
78, 85
addressed in MV-DMAS . The Eigen-space version of MV-DMAS, i.e., EIMV-DMAS,

25
79
provides similar resolution to MV-DMAS with lower sidelobes . Methods that use MV

algorithm, are sensitive to the quality of the received data; this issue limits the application of MV-

based algorithms for PA image reconstruction, because PA signals are usually weak. Among the

image reconstruction algorithms described above (see Table 6), DS-DMAS is the most suitable

one for linear-array based PA image reconstruction, because it reconstructs images with fine

resolution and low sidelobes, and has a low computational complexity.

Table 6. Performance comparison between image reconstruction algorithms.


Algorithm Resolution Sidelobes Complexity Reference
82
DAS Low High Low
85, 86
MV High High High
83, 87
DMAS Low Medium Medium
80, 82, 84
DS-DMAS Medium Low Medium
78, 85
MV-DMAS High High High
79
EIMV-DMAS High Low High

Four wire phantoms were prepared for this study (Figure 12A). Transducers were securely held

by a clamp which was attached to an x-y translation stage. The phantom container was fixed to the

optical table and filled with distilled water. Each phantom was imaged 50 times. Figures 12(B-E)

demonstrate images of the four wire phantoms taken with the L7-4 probe where (i) is the US image,

(ii) is the PA image reconstructed in the Vantage system, and (iii) is the PA image reconstructed

with DS-DMAS. Figures 12(F-I) demonstrate images of the four wire phantom taken with the L22-

14v probe where (i) is the US image, (ii) is the PA image reconstructed in the Vantage system, and

(iii) is the PA image reconstructed with DS-DMAS.

26
Fig. 12 Comparison of the performance of DS-DMAS to Vantage PA reconstruction method. (A) Wire phantoms for

resolution study, (i) one-wire, (ii) two-wire, (iii) two-wire cross, and (iv) three-wire. (B-I) US and PA images produced

by L7-4 and L22-14v probes where the image reconstruction is: (i) Vantage default US, (ii) Vantage default PA, and

(iii) photoacoustic DS-DMAS. L7-4 probe: (B) single wire cross section, (C) 3-wire cross section, (D) 2-wire cross

section, and (E) 2-wire cross section. L22-14v probe: (F) single wire cross section, (G) 2-wire cross section, (H) 2-

wire cross section, and (I) 3-wire cross section. Images of the targets are shown in white dotted circles or straight

lines.

27
7 Conclusions

Photoacoustic imaging is a complement to the already established US imaging technique and may

significantly increase its scope of application in diagnostic imaging and therapeutic monitoring.

Combining with commercial medical US systems, the development of PAI can be accelerated by

taking advantage of US image reconstruction and processing. Among different configurations,

linear-array PAI is becoming popular, mainly because linear-array US transducers can easily be

manufactured, hence the production cost is lower as compared to custom-made curved or ring

arrays 88, 89. Moreover, these transducers are commonly used in clinical applications, implying the

PAI system built with it will have high clinical translatability.

We discussed technical considerations for US/PA imaging implementation using the Vantage

research system. With the information we presented in the body of this technical note and the four

appendices, we described most of the experimental considerations one should know when working

with the Vantage system for PAI tests. Although the Verasonics Vantage System has many

capabilities and advantages, it has some limitations. The transducer connector limits which

transducers can directly connect to the system. Verasonics offers some connector adapters to

alleviate this problem, but it can make switching between transducers cumbersome. Another

limitation is that the signal pre-amplification occurs far (~1-2 meters) from the probe which

increases the noise level for weak PA signals. Further, the maximum sampling rate of the system

is 62.5MHz which may limit the use of high frequency transducers.

Appendix A: Absorption Spectrum of Black Tape and Carbon Pencil Lead

Measuring the absorption spectrum of the imaging target, i.e., amplitude of the optical energy

versus wavelength, is essential for the analysis of results in PA experiments. Here we explain a

method of how to extract the absorption spectra of two example imaging targets, vinyl black tape
28
and carbon pencil lead. The internal energy readings are not proportional to the energy deposited

on the sample due to dispersion, i.e., variable attenuation of wavelengths in the optical fiber. Using

the PhocusMobil (Opotek, CA, USA) laser, that has a built-in energy meter, we first obtained the

energy ratio by using a separate energy meter (QE25LP-S-MB-QED-D0, Gentec-EO, Quebec

City, Canada) to record the energy after the illumination fiber and dividing it by the internal energy

reading inside the PhocusMobil. This yields the relation between the internal Opotek energy

reading to the energy deposited on the sample. We imaged a 2 mm carbon lead phantom and a

black vinyl tape phantom, and simultaneously recorded the internal laser energies. The PA data

was compensated by dividing the raw PA signals by the product of the energies recorded with each

phantom and the energy ratio. The absorption spectrum of the black tape and carbon pencil lead

after performing a fourth order polynomial fit are shown in Figure 13A and B, respectively.

Fig. 13 Measurement of absorption spectrum of two commonly used imaging targets: (A) black tape, and (B) carbon

pencil lead.

29
Appendix B: MATLAB Codes for Data Acquisition, Processing, Image Reconstruction

MATLAB code for the Vantage system for real-time data acquisition, data processing, image

reconstruction, and display in order of variable appearance in the script is provided as a link in

“Code, Data, and Materials Availability” section 90. Explanation of the code is below:

Start and end depth of imaging are set for US and PA imaging with P(1).startDepth and

P(1).endDepth and P(2).startDepth and P(2).endDepth, respectively. na sets the number of steered

US angles that are transmitted for a single US frame. The default number is 7. oneway, sets whether

or not the system runs in simulation mode: 0 for simulation mode is off or 1 for simulation mode

is on. Simulation mode ignores hardware and mimics transducer elements to create virtual signals

and test if the MATLAB loader program VSX is communicating correctly with RunAcq. RunAcq

communicates receive and transmit commands to the ultrasound hardware. Flash2Qdelay is the

time between trigger input and start of acquisition in microseconds and must equal the time delay

between the flash lamp output and Q-switch output from the laser. We used an oscilloscope to find

the exact delay time and set the time value to the flash2Qdelay variable (see Table 3 for the values

we used for our system). PA_PRF is set to the pulse repetition rate of the laser. The next section

of the code involves setting up system parameters, such as data buffers, and the transducer array

specifications. For Vantage 128 system and a 128-element linear array transducer,

Resource.Parameters.numTransmit and Resource.Parameter.numRcvChannels are set to 128. For

the two linear array transducers we used, Trans.name is set to either ‘L7-4’ or ‘L22-14v’.

ComputeTrans(Trans) populates all the attributes of the specified transducer. PData defines the

pixel grid to be manipulated by the image reconstruction software. PData(n).PDelta defines the

spacing between pixels in all dimensions, PData(n).Size defines rows, columns, and sections of

the data, and PData(n).Origin defines the x,y,z coordinates of the left corner of the reconstructed

30
image, where n is 1 for the US image and n is 2 for the PA image. RcvBuffer, InterBuffer and

ImageBuffer are used to store the US and PA raw data and reconstructed images. Transmit

waveform (TW), transmit (TX), and transmit power controller (TPC) are transmit objects. TW

structure array defines specification of a transmit waveform (type, frequency, duty cycle, duration,

and polarity), where TW(1) is for US and TW(2) for PA. The TX structure array defines the

specification of the beam characteristics of each transmit action, including which transmitters are

active in the aperture (apodization), and the delay time before transmit for each active transmitter.

With 7 US transmits, TX(1:7) is defined for US transmit events. For PA, only one TX structure is

needed, so we define TX(8) for this event. In TX(8),all transmitters are turned off for the receive-

only beamformer. TPC(1) for US and TPC(2) for PA are defined, where TPC sets the transmit

power level for each specific transmit event. Receive objects are defined next and populate all the

characteristics of the receive phase of an acquisition event. The Transmit and Receive periods both

start with the data acquisition. Next, the time-gain-control (TGC) object defines the time-gain-

compensation curve for the receive portion of the acquisition event. To define a TGC waveform,

the user specifies the TGC.CntrlPts array and TGC.rangeMax. TGC.Waveform is then synthesized

and applied to the received data. The next section describes the reconstruction protocol. Recon

structure provides the general attributes of the reconstruction, including the source and destination

buffers to use. Recon.senscutoff is a value from 0.0 to 1.0 that sets the threshold of sensitivity

below which an element is excluded from the reconstruction summation for a given pixel.

Recon.pdatanum(n) specifies the number of the PData structure that defines the pixel locations of

the reconstructed data, where n is 1 for the US image or is 2 for the PA image. Recon.RcvBufFrame

is an optional attribute that when provided, overrides the frame number specified by the ReconInfo

structures. Setting Recon.RcvBufFrame to -1 allows the last acquisition frame transferred into the

31
RcvBuffer to be used in the reconstruction and be displayed for real-time imaging.

Recon.IntbufDest and Recon.ImgBufDest specify the destination buffer and frame that will receive

the reconstructed output, respectively. Recon.RINums is a row vector that specifies the ReconInfo

structure indices associated with the most recent reconstruction. For each Recon, there is an

associated set of ReconInfo objects which contain information on how to perform the

reconstruction. This information includes which data in the data buffer to reconstruct along with

where within PData. Further, ReconInfo chooses which type of reconstruction is performed

between replace, add, or multiply intensity, where each successive reconstructed frame replaces

the previous, is added to the previous, or is multiplied by the previous, respectively. Replace

intensity is chosen for our purpose. Process objects are used to describe the type of processing to

apply to the acquired data. After defining the sequence control objects, event objects, and some

graphical user interface (GUI) controls, the script is complete for data acquisition and displaying

US and PA data.

Appendix C: Optical Absorption of acoustic coupling medium (Water)

In a PA imaging experiment, there are ultrasonic transducers and an acoustic coupling layer

between the transducers and the imaging target 91. Propagation of generated ultrasound waves from

the imaging target is least attenuated when received by the ultrasound sensor through a coupling

agent that exhibits minimal acoustic impedance mismatch with the imaging target. The coupling
92
agent helps minimize degradation and signal loss . Acoustic couplants can be characterized as

liquid, gel, semi-dry, and dry. Liquids and gels generally have lower acoustic impedance than dry

couplant materials. More details about different acoustic couplants can be found in the literature.

Here we describe details on the most widely used acoustic couplant, water. Water has a low

acoustic attenuation and impedance (1.5 MPa.s/m) which makes it suitable as an acoustic coupling

32
material. In terms of water optical properties, it absorbs over a wide range of the electromagnetic

radiation spectrum with rotational transitions and intermolecular vibrations responsible for

absorption in the microwave (≈ 1 mm - 10 cm wavelength) and far-infrared (≈ 10 µm - 1 mm),

intramolecular vibrational transitions responsible for absorption in the infrared (≈ 200nm- 10 µm),
93-95
and electronic transitions responsible for absorption in the ultraviolet region (< 200 nm) . In

regular water (H2O), the first large absorption band occurs at around 980 nm 93, 96, 97, followed by

another band at ~1450 nm. The absorption spectrum of heavy water (D2O) is different from that

of H2O, mainly due to the heavier deuterium nucleus; absorption peaks occur at around 1000 nm,
98, 99
1330 nm, and 1600 nm . The spectral features of water absorption also depend upon the

temperature 93, 98, 100; as temperature decreases, the fraction of hydrogen-bound water molecules is

increased, causing absorption peaks to reduce in intensity, broaden in bandwidth, and shift to lower
96
energy . The optical absorption spectrum of regular water and heavy water at various

temperatures are shown in Figure 14.

Fig. 14 Optical absorption spectrum of regular water (H2O) and heavy water (D2O) at various temperatures in NIR

region. Data for this figure has been extracted from 93, 98 and reproduced.

33
Appendix D: MATLAB Code for Simultaneous Diplay of US and PA Image

The modifications to the original script to simultaneously display US and PA images, in real-time

(more details is described in Section 4.4.), are given in Table 7.

Table 7. Modifications to the original script to simultaneously display US and PA images, in real-time. The
modifications are indicated in red.
Original Modified
Resource.DisplayWindow(1).Title = Resource.DisplayWindow(1).Title =
mfilename; mfilename;
Resource.DisplayWindow(1).pdelta = 0.4; Resource.DisplayWindow(1).pdelta = 0.4;
ScrnSize = get(0,'ScreenSize'); ScrnSize = get(0,'ScreenSize');
DwWidth = DwWidth =
ceil(PData(1).Size(2)*PData(1).PDelta(1)/ ceil(PData(1).Size(2)*PData(1).PDelta(1)
Resource.DisplayWindow(1).pdelta); /Resource.DisplayWindow(1).pdelta);
DwHeight = DwHeight =
ceil(PData(1).Size(1)*PData(1).PDelta(3)/ ceil(PData(1).Size(1)*PData(1).PDelta(3)
Resource.DisplayWindow(1).pdelta); /Resource.DisplayWindow(1).pdelta);
Resource.DisplayWindow(1).Position = Resource.DisplayWindow(1).Position =
[250,(ScrnSize(4)-(DwHeight+150))/2, ... [250,(ScrnSize(4)-(DwHeight+150))/2, ...
DwWidth, DwHeight]; DwWidth, DwHeight];
Resource.DisplayWindow(1).ReferencePt = Resource.DisplayWindow(1).ReferencePt =
[PData(1).Origin(1),0,PData(1).Origin(3)] [PData(1).Origin(1),0,PData(1).Origin(3)
; ];
Resource.DisplayWindow(1).numFrames = 2; Resource.DisplayWindow(1).numFrames = 2;
Resource.DisplayWindow(1).AxesUnits = Resource.DisplayWindow(1).AxesUnits =
'mm'; 'mm';
Resource.DisplayWindow(1).Colormap = Resource.DisplayWindow(1).Colormap =
gray(256); gray(256);
Resource.DisplayWindow(1).splitPalette = Resource.DisplayWindow(1).splitPalette =
1; 1;

Resource.DisplayWindow(2).Title =
mfilename;
Resource.DisplayWindow(2).pdelta = 0.4;
Resource.DisplayWindow(2).Position =
[250,(ScrnSize(4)-(DwHeight+150))/2, ...
DwWidth, DwHeight];
Resource.DisplayWindow(2).ReferencePt =
[PData(1).Origin(1),0,PData(1).Origin(3)
];
Resource.DisplayWindow(2).numFrames = 2;
Resource.DisplayWindow(2).AxesUnits =
'mm';
Resource.DisplayWindow(2).Colormap =
hot;
Resource.DisplayWindow(2).splitPalette =
1;

34
Process(1).classname = 'Image'; Process(1).classname = 'Image';
Process(1).method = 'imageDisplay'; Process(1).method = 'imageDisplay';
Process(1).Parameters = Process(1).Parameters =
{'imgbufnum',1,... {'imgbufnum',1,...
'framenum',-1,... 'framenum',-1,...
'pdatanum',1,... 'pdatanum',1,...
'pgain',1.0,... 'pgain',1.0,...
'reject',10,... 'reject',10,...
'persistMethod','simple',... 'persistMethod','simple',...
'persistLevel',50,... 'persistLevel',50,...
'interpMethod','4pt',... 'interpMethod','4pt',...
'grainRemoval','none',... 'grainRemoval','none',...
'processMethod','none',... 'processMethod','none',...
'averageMethod','none',... 'averageMethod','none',...
'compressMethod','power',... 'compressMethod','power',...
'compressFactor',40,... 'compressFactor',40,...
'mappingMethod','lowerHalf',... 'mappingMethod','full',...
'display',1,... 'display',1,...
'displayWindow',1}; 'displayWindow',1};

Process(2).classname = 'Image'; Process(2).classname = 'Image';


Process(2).method = 'imageDisplay'; Process(2).method = 'imageDisplay';
Process(2).Parameters = Process(2).Parameters =
{'imgbufnum',2,... {'imgbufnum',2,...
'framenum',-1,... 'framenum',-1,...
'pdatanum',2,... 'pdatanum',2,...
'pgain',1.0,... 'pgain',1.0,...
'reject',10,... 'reject',10,...
'persistMethod','dynamic',... 'persistMethod','dynamic',...
'persistLevel',cpers,... 'persistLevel',cpers,...
'interpMethod','4pt',... 'interpMethod','4pt',...
'grainRemoval','none',... 'grainRemoval','none',...
'processMethod','none',... 'processMethod','none',...
'averageMethod','none',... 'averageMethod','none',...
'compressMethod','power',... 'compressMethod','power',...
'compressFactor',40,... 'compressFactor',40,...
'mappingMethod','upperHalf',... 'mappingMethod','full',...
'threshold',cpt,... 'threshold',cpt,...
'display',1,... 'display',1,...
'displayWindow',1}; 'displayWindow',2};

Disclosures

The authors declare no conflict of interest.

Acknowledgments

We would like to thank Verasonics Inc., Kirkland, WA technical support team, especially Mike

Vega, for their suggestions and comments. We are also grateful to the support team from

35
PhotoSound Technologies Inc., especially Dr. Peter Brecht, Houston, TX, for their prompt

response and help to integrate the amplifier with the Vantage system

Code, Data, and Materials Availability

MATLAB code is available at the following link.

https://codeocean.com/capsule/5228184/tree /code/code_appendixB.txt

References

1. Y. Zhou, J. Yao and L. V. Wang, Journal of biomedical optics 21 (6), 061007 (2016).
2. D. Pan, B. Kim, L. V. Wang and G. M. Lanza, Wiley Interdisciplinary Reviews: Nanomedicine
and Nanobiotechnology 5 (6), 517-543 (2013).
3. J. Yao and L. V. Wang, Photoacoustics 2 (2), 87-101 (2014).
4. B. Herzog and F. Sengün, Photochemical & Photobiological Sciences 14 (11), 2054-2063 (2015).
5. V. Dogra, B. Chinni, S. Singh, et al., Journal of biomedical optics 21 (6), 066019 (2016).
6. J. Gong and S. Krishnan, in Dye-Sensitized Solar Cells (Elsevier, pp. 51-81, 2019.
7. A. Fatima, K. Kratkiewicz, R. Manwar, et al., Photoacoustics, 100137 (2019).
8. C. Kim, C. Favazza and L. V. Wang, Chemical reviews 110 (5), 2756-2782 (2010).
9. M. Zafar, K. Kratkiewicz, R. Manwar and M. Avanaki, presented at the Photons Plus Ultrasound:
Imaging and Sensing 2019, 2019 (unpublished).
10. Z. Turani, E. Fatemizadeh, T. Blumetti, et al., Cancer research 79 (8), 2021-2030 (2019).
11. C. Li and L. V. Wang, Physics in medicine and biology 54 (19), R59-R97 (2009).
12. K. Kratkiewicz, R. Manwar, M. Zafar, et al., Applied Sciences 9 (21), 4505 (2019).
13. L. V. Wang, IEEE Journal of Selected Topics in Quantum Electronics 14 (1), 171-179 (2008).
14. A. Hariri, A. Fatima, N. Mohammadian, et al., Journal of biomedical optics 22 (7), 075001
(2017).
15. R. A. Kruger, C. M. Kuzmiak, R. B. Lam, D. R. Reinecke, S. P. Del Rio and D. Steed, Medical
physics 40 (11) (2013).
16. G. Ku and L. V. Wang, Optics letters 30 (5), 507-509 (2005).
17. U. Chitgupi, N. Nyayapathi, J. Kim, et al., Advanced Materials, 1902279 (2019).
18. H. F. Zhang, K. Maslov, G. Stoica and L. V. Wang, Nature biotechnology 24 (7), 848 (2006).
19. S. Mallidi, G. P. Luke and S. Emelianov, Trends in biotechnology 29 (5), 213-221 (2011).
20. S. Yang, D. Xing, Q. Zhou, L. Xiang and Y. Lao, Medical physics 34 (8), 3294-3301 (2007).
21. J. Gamelin, A. Maurudis, A. Aguirre, et al., Optics express 17 (13), 10489-10498 (2009).
22. E. Zhang, J. Laufer, R. Pedley and P. Beard, Physics in medicine and biology 54 (4), 1035
(2009).
23. H.-P. Brecht, R. Su, M. Fronheiser, S. A. Ermilov, A. Conjusteau and A. A. Oraevsky, Journal of
biomedical optics 14 (6), 064007-064007-064008 (2009).
24. L. V. Wang and S. Hu, Science 335 (6075), 1458-1462 (2012).
25. E. Z. Zhang, J. Laufer and P. Beard, presented at the Biomedical Optics (BiOS) 2007, 2007
(unpublished).
26. X. Wang, Y. Pang, G. Ku, X. Xie, G. Stoica and L. V. Wang, Nature biotechnology 21 (7), 803-
806 (2003).

36
27. C. Li, A. Aguirre, J. Gamelin, A. Maurudis, Q. Zhu and L. V. Wang, Journal of biomedical optics
15 (1), 010509-010509-010503 (2010).
28. W. Lu, Q. Huang, G. Ku, et al., Biomaterials 31 (9), 2617-2626 (2010).
29. J. Laufer, P. Johnson, E. Zhang, et al., Journal of biomedical optics 17 (5), 0560161-0560168
(2012).
30. S. Sethuraman, J. H. Amirian, S. H. Litovsky, R. W. Smalling and S. Y. Emelianov, Optics
express 16 (5), 3362-3367 (2008).
31. B. Wang, J. L. Su, A. B. Karpiouk, K. V. Sokolov, R. W. Smalling and S. Y. Emelianov, IEEE
Journal of selected topics in Quantum Electronics 16 (3), 588-599 (2010).
32. S. Sethuraman, S. R. Aglyamov, J. H. Amirian, R. W. Smalling and S. Y. Emelianov, IEEE
transactions on ultrasonics, ferroelectrics, and frequency control 54 (5), 978-986 (2007).
33. S. Hu, K. Maslov, V. Tsytsarev and L. V. Wang, Journal of biomedical optics 14 (4), 040503-
040503-040503 (2009).
34. X. Wang, X. Xie, G. Ku, L. V. Wang and G. Stoica, Journal of biomedical optics 11 (2), 024015
(2006).
35. G. Ku, X. Wang, X. Xie, G. Stoica and L. V. Wang, Applied Optics 44 (5), 770-775 (2005).
36. M. Zafar, K. Kratkiewicz, R. Manwar and M. Avanaki, Applied Sciences 9 (3), 374 (2019).
37. K. Kratkiewicz, R. Manwar, A. Rajabi-Estarabadi, et al., Sensors 19 (12), 2815 (2019).
38. M. Nasiriavanaki, J. Xia, H. Wan, A. Q. Bauer, J. P. Culver and L. V. Wang, Proceedings of the
National Academy of Sciences 111 (1), 21-26 (2014).
39. J. Yao, J. Xia, K. I. Maslov, et al., Neuroimage 64, 257-266 (2013).
40. M. Nasiriavanaki, J. Xia, H. Wan, A. Q. Bauer, J. P. Culver and L. V. Wang, presented at the
Photons Plus Ultrasound: Imaging and Sensing 2014, 2014 (unpublished).
41. A.-R. Mohammadi-Nejad, M. Mahmoudzadeh, M. S. Hassanpour, et al., Photoacoustics (2018).
42. E. Z. Zhang, B. Povazay, J. Laufer, et al., Biomedical optics express 2 (8), 2202-2215 (2011).
43. Z. Chen, E. Rank, K. M. Meiburger, et al., Scientific Reports 7 (1), 17975 (2017).
44. D. Xu, S. Yang, Y. Wang, Y. Gu and D. Xing, Biomedical optics express 7 (6), 2095-2102
(2016).
45. M. Li, J. Oh, X. Xie, et al., Proceedings of the IEEE 96 (3), 481-489 (2008).
46. L. Nie, X. Cai, K. Maslov, A. Garcia-Uribe, M. A. Anastasio and L. V. Wang, Journal of
biomedical optics 17 (11), 110506-110506 (2012).
47. S. A. Ermilov, T. Khamapirad, A. Conjusteau, et al., Journal of biomedical optics 14 (2), 024007-
024007-024014 (2009).
48. J. A. Copland, M. Eghtedari, V. L. Popov, et al., Molecular Imaging & Biology 6 (5), 341-349
(2004).
49. S. Manohar, S. E. Vaartjes, J. C. van Hespen, et al., Optics express 15 (19), 12277-12285 (2007).
50. A. A. Oraevsky, A. A. Karabutov, S. V. Solomatin, et al., presented at the BiOS 2001 The
International Symposium on Biomedical Optics, 2001 (unpublished).
51. S. Hu, B. Rao, K. Maslov and L. V. Wang, Optics letters 35 (1), 1-3 (2010).
52. S. Jiao, M. Jiang, J. Hu, et al., Optics express 18 (4), 3967-3972 (2010).
53. B. Wang, J. L. Su, J. Amirian, S. H. Litovsky, R. Smalling and S. Emelianov, Optics express 18
(5), 4889-4897 (2010).
54. K. Jansen, M. Wu, A. F. W. van der Steen and G. van Soest, Optics Express 21 (18), 21472-
21484 (2013).
55. Y. Deng, N. C. Rouze, M. L. Palmeri and K. R. Nightingale, IEEE transactions on ultrasonics,
ferroelectrics, and frequency control 64 (1), 164-176 (2017).
56. V. Bindal, (2000).
57. H. Pennes, Int. J. Radiation Oncology Biol. Phys. 25, 289-297 (1948).
58. M. Breazeale and O. Leroy, Physical Acoustics: Fundamentals and Applications. (Springer
Science & Business Media, 2012).
59. R. W. Cole, T. Jinadasa and C. M. Brown, Nature protocols 6 (12), 1929 (2011).
37
60. W. H. Blanton, (East Tennessee State University), 2009.
61. J. Xia, (Verasonics Inc., Verasonics Inc.), 2019.
62. S. Mahmoodkalayeh, M. A. Ansari and M. Nasiriavanaki, presented at the Photons Plus
Ultrasound: Imaging and Sensing 2018, 2018 (unpublished).
63. H. Akima, Journal of the ACM (JACM) 17 (4), 589-602 (1970).
64. M. H. Rayyan Manwar, Ali Hariri, Karl Kratkiewicz, Shahryar Noei, Mohammad R. N. Avanaki
Photoacoustic Sensing and Imaging in Biomedicine, MDPI Sensors (2018).
65. M. Zafar, R. Manwar, K. Kratkiewicz, et al., presented at the Photons Plus Ultrasound: Imaging
and Sensing 2019, 2019 (unpublished).
66. V. Ivanov, edited by K. Kratkiewicz (PhotoSound Technologies Inc, PhotoSound Technologies
Inc), pp. 20, 2018.
67. Z. Yuan and H. Jiang, Applied physics letters 88 (23), 231101 (2006).
68. S. R. Arridge, Inverse problems 15 (2), R41 (1999).
69. K. D. Paulsen and H. Jiang, Medical Physics 22 (6), 691-701 (1995).
70. S. Prahl, (OMLC), 2018.
71. L. Zhao, M. Yang, Y. Jiang and C. Li, Journal of Innovative Optical Health Sciences 10 (04),
1740002 (2017).
72. Q. Fang and D. A. Boas, Optics express 17 (22), 20178-20190 (2009).
73. P. Omidi, M. Diop, J. Carson and M. Nasiriavanaki, presented at the Photons Plus Ultrasound:
Imaging and Sensing 2017, 2017 (unpublished).
74. M. Mozaffarzadeh, A. Mahloojifar, M. Nasiriavanaki and M. Orooji, presented at the Photonics
in Dermatology and Plastic Surgery 2018, 2018 (unpublished).
75. P. Omidi, M. Zafar, M. Mozaffarzadeh, et al., Applied Sciences 8 (9), 1570 (2018).
76. M. H. Heidari, M. Mozaffarzadeh, R. Manwar and M. Nasiriavanaki, presented at the Photons
Plus Ultrasound: Imaging and Sensing 2018, 2018 (unpublished).
77. M. Mozaffarzadeh, A. Mahloojifar, M. Nasiriavanaki and M. Orooji, presented at the Photons
Plus Ultrasound: Imaging and Sensing 2018, 2018 (unpublished).
78. M. Mozaffarzadeh, A. Mahloojifar, M. Orooji, K. Kratkiewicz, S. Adabi and M. Nasiriavanaki,
Journal of biomedical optics 23 (2), 026002 (2018).
79. M. Mozaffarzadeh, A. Mahloojifar, V. Periyasamy, M. Pramanik and M. Orooji, IEEE Journal of
Selected Topics in Quantum Electronics 25 (1), 1-8 (2019).
80. M. Mozaffarzadeh, A. Hariri, C. Moore and J. V. Jokerst, Photoacoustics 12, 22-29 (2018).
81. M. Mozaffarzadeh, V. Periyasamy, M. Pramanik and B. Makkiabadi, Journal of biomedical
optics 23 (12), 121604 (2018).
82. M. Mozaffarzadeh, A. Mahloojifar, M. Orooji, S. Adabi and M. Nasiriavanaki, IEEE
Transactions on Biomedical Engineering 65 (1), 31-42 (2018).
83. G. Matrone, A. S. Savoia, G. Caliano and G. Magenes, IEEE transactions on medical imaging 34
(4), 940-949 (2015).
84. M. Mozaffarzadeh, M. Sadeghi, A. Mahloojifar and M. Orooji, Ultrasound in medicine & biology
44 (3), 677-686 (2018).
85. M. Mozaffarzadeh, A. Mahloojifar and M. Orooji, presented at the Digital Optical Technologies
2017, 2017 (unpublished).
86. S. Park, A. B. Karpiouk, S. R. Aglyamov and S. Y. Emelianov, Optics letters 33 (12), 1291-1293
(2008).
87. J. Park, S. Jeon, J. Meng, L. Song, J. S. Lee and C. Kim, Journal of biomedical optics 21 (3),
036010 (2016).
88. T. L. Szabo and P. A. Lewin, Journal of Ultrasound in Medicine 32 (4), 573-582 (2013).
89. B. A. Angelsen, H. Torp, S. Holm, K. Kristoffersen and T. Whittingham, European Journal of
Ultrasound 2 (2), 151-164 (1995).
90. K. K. M. R. Z. M. M. A. Kamran, (Code Ocean Inc., Code Ocean), 2020.
91. N. Meimani, N. Abani, J. Gelovani and M. R. Avanaki, Photoacoustics 7, 27-35 (2017).
38
92. J. Nyholt and G. N. Langlois, (Google Patents, 2013.
93. J. A. Curcio and C. C. Petty, JOSA 41 (5), 302-304 (1951).
94. S. Matcher, M. Cope and D. Delpy, Physics in Medicine & Biology 39 (1), 177 (1994).
95. F. M. Sogandares and E. S. Fry, Applied Optics 36 (33), 8699-8709 (1997).
96. S. Chung, A. Cerussi, S. Merritt, J. Ruth and B. Tromberg, Physics in Medicine & Biology 55
(13), 3753 (2010).
97. M. F. Bakhsheshi and T.-Y. Lee, (Taylor & Francis, 2015.
98. A. Tam and C. Patel, Applied Optics 18 (19), 3348-3358 (1979).
99. S. A. Sullivan, JOSA 53 (8), 962-968 (1963).
100. S. Mahmoodkalayeh, H. Z. Jooya, A. Hariri, et al., Scientific Reports 8 (1), 4873 (2018).

Karl Kratkiewicz is currently a PhD candidate in Biomedical Engineering at Wayne State

University working on photoacoustic computed tomography and transfontanelle photoacoustic

imaging. Karl received his Bachelor's in Biomedical Physics Summa Cum Laude and Master's in

Biomedical Engineering from Wayne State University in 2017 and 2018.

Rayyan Manwar is currently a Post Doctoral Fellow in the Department of Bioengineering

University of Illinois at Chicago, Chicago, IL, USA. He worked as Post Doctoral Research Scholar

at Wayne State University, Detroit MI, USA from 2017-2020. He received his bachelor’s is in

Electrical and Electronic Engineering from Islamic University of Technology (IUT), Bangladesh

in 2011 and PhD from University of Windsor, Windsor, Ontario in 2017. His research interests

include MEMS based design, fabrication and characterization, photoacoustic and ultrasound

imaging.

Yang Zhou received his Master of Engineering (ME) degree in Mechanical Engineering at Wayne

State University, Detroit Michigan and currently enrolled as a PhD student in Biomedical

Engineering department. He received his Bachelor of Engineering degree from Xi’an Jiaotong

University Shaanxi, China.

Moein Mozaffarzadeh received his BSc degree in electrical engineering from Babol Noshirvani

University of Technology, Mazandaran, Iran, in 2015, and his MSc degree in biomedical-

bioelectric engineering from Tarbiat Modares University, Tehran, Iran, in 2017. He is currently a

39
PhD student in the Department of Imaging Physics, Delft University of Technology. His research

interests include photoacoustic/ultrasound image formation, photoacoustic endoscopy and

intravascular photoacoustic imaging.

Kamran Avanaki is an associate professor in the department of Bioengineering and Dermatology

at University of Illinois at Chicago, Chicago, IL, USA. He received a Ph.D. with Outstanding

Achievement in Medical Optical Imaging and Computing from the University of Kent in the

United Kingdom. His bachelor's and master's degrees with honors are in electronics engineering.

In 2014, he completed a three-year postdoctoral fellowship at Washington University in St. Louis,

in the OILab.

Caption List

Fig. 1 Absorption spectrum of (a) endogenous, and (b) exogenous contrast agents. Reproduced

3-5 A =  (  ) Cd
with permission from . According to Beer-Lambert law, absorbance is defined by

where,  (  ) , C , and d are molar absorptivity, concentration, and cross-section thickness of the

contrast agent, respectively. Absorption coefficient is expressed as:  = ( A d ) log10 e 6.

Fig. 2 Principle of PA signal generation, detection, and image reconstruction. Reprinted with

permission from Ref. 7. A high energy and short pulse laser light illuminates the absorber, leading

to a transient temperature rise which results in a thermoelastic expansion of the absorber, and

acoustic (or PA) wave generation. The signals generated from the waves received by an ultrasound

probe are given to a reconstruction algorithm to form a PA image. .

Fig. 3 Vantage system architecture and data acquisition sequencing. (a) Vantage system

architecture for US/PA sequencing and (b) timing diagram of US/PA sequencing in Vantage

system. US: Ultrasound, PA: Photoacoustic, Recon.: Reconstruction. RcvBuffer is the predefined

40
buffer where the US and PA data are stored, Acq: acquisition, TX: transmit, RX: receive, A/D:

analog to digital, VSX: Verasonics script execution, GUI: graphical user interface, RcvBuffer:

receive buffer, CPU: central processing unit

Fig. 4 Photoacoustic signal in (A) time-domain, and (B) frequency-domain, obtained from a 2 mm

diameter carbon lead phantom, imaged with L22-14v transducer at 690 nm wavelength. The red

signal in A shows the overlaid US signal acquired from the same sample.

Fig. 5 US/PA resolution study. (A) Schematic of the experimental setup, including a hair phantom

photograph captured by a 4× objective on a light microscope (SME-F8BH, Amscope, CA, USA).

Resolution study when (B) L7-4 was used, (C) L22-14v was used. (i) US axial resolution versus

depth, (ii) US lateral resolution versus depth, (iii) PA axial resolution versus depth, (iv) PA lateral

resolution versus depth.

Fig. 6 Photoacoustic signal “jitter” comparison between two triggering methods when a 2 mm

carbon lead was imaged. 64th element signals are plotted. (A) (i) Schematic of straightforward

triggering method where laser triggers Vantage system, (ii) schematic of the function generator

driven triggering method where function generator triggers laser flash lamp and Vantage system,

which then triggers laser Q-switch, (B) (i) five sequential PA frames showing significant “jitter”

between frames, (ii) five sequential PA frames showing nearly no “jitter” between frames, and (iii)

averaged frames using methods described in A(i) and A(ii).

Fig. 7 Impact of illumination angle. (A) Experimental setup of illumination angle investigation.

(B) PA signal profile from 64th element of L7-4 probe taken from a 2 mm carbon lead phantom

with different illumination angles. (i) 40 degrees, (ii) 48 degrees.

Fig. 8 PA signal profile from 64th element of L7-4 probe taken from a black tape phantom of 18mm

width with varying sampling rates: (A) 20.8 MHz, (B) 41.6 MHz, (C) 62.4 MHz, and (D)

41
xnor = x ( ymax ) sample no.
interpolated signal using ‘spline’ algorithm. Here, and

y r x = Ratio of y values , where y = positive or negative peaks .

Fig. 9 US image quality improvement. (A) Experimental setup showing an ultrasound probe and

a 3-carbon lead phantom. (L1) 2 mm diameter carbon lead, (L2) 0.9 mm diameter carbon lead,

(L3) 0.2 mm diameter carbon lead, (B-F) US images of the 3 carbon lead phantom for 8, 16, 32,

64, and 128 steered angles, respectively. Green dashed circle encloses object pixels and red dashed

box encloses both object and background pixels, (G) CNR versus number of steered angles.

Fig. 10 PA signal amplification. (A) System setup to test the Photosound AMP128-18 amplifier:

(i) display monitor, (ii) Nd:YAG laser head, (iii) fiber optic bundle, (iv) laser power supply, (v)

laser chiller. (B) Wire imaging results: wire phantom suspended in 0% intralipid imaged (i) without

and (iv) with amplifier, wire phantom suspended in 25% intralipid imaged (ii) without and (v) with

amplifier, and wire phantom suspended in 50% intralipid image (iii) without and (vi) with

amplifier. (C) Raw data from the 64th channel of the linear array when imaging a single wire in

50% intralipid concentration to demonstrate SNR improvement from (i) without amplifier to (ii)

with amplifier, (D) Bar chart demonstrating signal amplitude increase corresponding to phantom

imaging in B with and without amplifier, at different concentrations of intralipid.

Fig. 11 Fluence Compensation algorithm validation. (A) Schematic of a phantom with 3 layers of

gelatin phantom mixed with intralipid/ink solution and 3 carbon lead imaging targets, (B)

photograph of the phantom made in a cube box and magnified to show the layers, (C) US image

of the phantom, (D) PA image before fluence compensation, (E) Monte Carlo simulation of light

propagation in the phantom, (F) PA image after fluence compensation.

Fig. 12 Comparison of the performance of DS-DMAS to Vantage PA reconstruction method. (A)

Wire phantoms for resolution study, (i) one-wire, (ii) two-wire, (iii) two-wire cross, and (iv) three-

42
wire. (B-I) US and PA images produced by L7-4 and L22-14v probes where the image

reconstruction is: (i) Vantage default US, (ii) Vantage default PA, and (iii) photoacoustic DS-

DMAS. L7-4 probe: (B) single wire cross section, (C) 3-wire cross section, (D) 2-wire cross

section, and (E) 2-wire cross section. L22-14v probe: (F) single wire cross section, (G) 2-wire

cross section, (H) 2-wire cross section, and (I) 3-wire cross section. Images of the targets are shown

in white dotted circles or straight lines.

Fig. 13 Measurement of absorption spectrum of two commonly used imaging targets: (A) black

tape, and (B) carbon pencil lead.

Fig. 14 Optical absorption spectrum of regular water (H2O) and heavy water (D2O) at various

temperatures in NIR region. Data for this figure has been extracted from 93, 98 and reproduced.

Table 1 High frequency Vantage 128 system specifications (Verified by Verasonics’ technical

support).

Table 2 Specifications of linear array transducers ATL Philips L7-4, and Verasonics L22-14v.

Table 3 Modifications to original L11-4 script by Verasonics, for ATL Phillips L7-4 and

Verasonics L22-14v probes

Table 4 List of immediately recognizable transducers for the Vantage system.

Table 5 Script modifications to change the sampling rate of L7-4 from 20.8 MHz to 62.4 MHz.

The modifications are indicated in red.

Table 6 Performance comparison between image reconstruction algorithms.

Table 7 Modifications to the original script to simultaneously display US and PA images, in real-

time. The modifications are indicated in red.

43

View publication stats

You might also like