 Research
 Open Access
 Published:
A photoacoustic imaging reconstruction method based on directional total variation with adaptive directivity
BioMedical Engineering OnLine volume 16, Article number: 64 (2017)
Abstract
Background
In photoacoustic tomography (PAT), total variation (TV) based iteration algorithm is reported to have a good performance in PAT image reconstruction. However, classical TV based algorithm fails to preserve the edges and texture details of the image because it is not sensitive to the direction of the image. Therefore, it is of great significance to develop a new PAT reconstruction algorithm to effectively solve the drawback of TV.
Methods
In this paper, a directional total variation with adaptive directivity (DDTV) modelbased PAT image reconstruction algorithm, which weightedly sums the image gradients based on the spatially varying directivity pattern of the image is proposed to overcome the shortcomings of TV. The orientation field of the image is adaptively estimated through a gradientbased approach. The image gradients are weighted at every pixel based on both its anisotropic direction and another parameter, which evaluates the estimated orientation field reliability. An efficient algorithm is derived to solve the iteration problem associated with DDTV and possessing directivity of the image adaptively updated for each iteration step.
Results and conclusion
Several texture images with various directivity patterns are chosen as the phantoms for the numerical simulations. The 180, 90 and 30view circular scans are conducted. Results obtained show that the DDTVbased PAT reconstructed algorithm outperforms the filtered backprojection method (FBP) and TV algorithms in the quality of reconstructed images with the peak signaltonoise rations (PSNR) exceeding those of TV and FBP by about 10 and 18 dB, respectively, for all cases. The Shepp–Logan phantom is studied with further discussion of multimode scanning, convergence speed, robustness and universality aspects. Invitro experiments are performed for both the sparseview circular scanning and linear scanning. The results further prove the effectiveness of the DDTV, which shows better results than that of the TV with sharper image edges and clearer texture details. Both numerical simulation and in vitro experiments confirm that the DDTV provides a significant quality improvement of PAT reconstructed images for various directivity patterns.
Background
Photoacoustic tomography (PAT), also referred to as optoacoustic tomography, is an emerging biomedical imaging modality. It combines a high contrast of the optical imaging with a good resolution of ultrasound one [1] and possesses a noninvasive feature [2], which unique advantages over other traditional imaging techniques help it to find wide applications in many aspects of biomedical fields [3,4,5,6,7], such as small animal imaging [8], tumor detection [9], vessel imaging [10], functional imaging [11], and molecular imaging [12]. In the computedtomographic PAT, which is mainly considered in this study, the laser pulse is usually used to irradiate the biomedical tissue. The tissue absorbs the light and then sends out ultrasound waves. This kind of phenomenon is called photoacoustic effect [1, 2]. A scanning ultrasound transducer or a transducer array is used to detect the generated ultrasound signals around the tissue. The detected signals are then utilized to reconstruct photoacoustic images, which reflect the light absorption of the tissue via a certain algorithm. Therefore, the reconstruction algorithm plays a significant role in the PAT.
Many efforts have been made, in order to find an accurate and efficient photoacoustic image reconstruction method. In 1995, Kruger et al. [6] realized the image reconstruction by utilizing the inverse Radon transform, which is considered a pioneer PAT reconstruction algorithm. After that, they also proposed the inversion of the spherical mean Radon transform method, which was more accurate [13]. The filtered backprojection method (FBP) advanced by Xu et al. [14] was widely applied to PAT, due to its concision and accuracy. Zhang et al. [15, 16] proposed the deconvolution reconstruction algorithm, which had better performances in limitedview sampling and used the heterogeneous speed of sound. There is another type of PAT reconstructed methods called timereversal, which reconstructs images from a forwardpropagation numerical model to generate measured photoacoustic (PA) signals backwards in the time [17,18,19,20]. Xu et al. proposed a timereversalbased reconstructed method for threedimensional broadband diffraction tomography [17] and applied it to PAT [18]. Bradley et al. [19] used the timereversal method to compensate for acoustic absorption in the photoacoustic tomography. Cox et al. [20] found the artifact trapping phenomenon in the timereversal PAT reconstruction and proposed some methods to mitigate these artifacts. Besides the above reconstruction methods, another kind of algorithm, called iterative reconstruction method, has been applied to PAT. This kind of algorithm constructs a forward model, which utilizes the relationship between photoacoustic signals and the light absorption deposition to calculate the reconstructed image iteratively under some optimization conditions [21,22,23]. Thus, this kind of algorithm is also referred to as modelbased algorithm. Paltauf et al. [21] advanced an iteration reconstruction algorithm by minimizing the difference between the detected PA signals and the calculated ones from the image. Ma et al. [22] introduced a filtered meanback projectioniterative reconstruction algorithm to deal with the lineararray detection in practical application. To accelerate the speed of the iterative reconstruction method, DeanBen et al. [24] proposed an angular image discretization modelbased reconstruction method. Rosenthal et al. [25] used wavelet packets to considerably reduce the computational cost for the modelbased algorithm. In order to improve the performance of the modelbased algorithm under the circumstance of sparse sampling, the compressed sensing (CS) theory has been employed [26, 27]. The total variation (TV) is an important sparsity regularizer in the image denoising and CS image reconstruction [28]. It utilizes the sparsity of the natural image gradient to measure the variations in an image [29]. In PAT, Wang et al. [30] presented an adaptive steepestdecentprojection onto convex sets (ASDPOCS) method, which involved the TV in the iteration. Zhang et al. [31] proposed a gradient descent algorithm based on the TV, which provided better results, especially under the condition of sparse sampling. Arridge et al. [32] used the TV regularization enhanced by Bregman iterations to solve the PAT subsampling problems, which achieved both increased acquisition speed and good spatial resolution. However, the TV is only a measure of local changes in images, which is not related to the directions of images. It is reported that the TVbased method tends to produce oversmoothed image edges and texture details [33,34,35]. Thus, the isotropic TVbased PAT reconstructed methods are more applicable to images, which are piecewisesmooth and have no dominant direction, but their operation is deteriorated when applied to images with directional textures. At present, there is no PAT reconstruction algorithm to effectively solve the TV drawback.
In this study, we propose a directivity adaptive direction total variation (DDTV) based PAT image reconstruction algorithm for effective minimization of the TV deficiencies. The spatially varying directivity pattern of the image is firstly estimated. Then the image gradients are weighted by the calculated orientation field of the image. The DDTV is calculated by summing the norm of the weighted gradients. The gradientbased approach [36, 37] is used to efficiently estimate the orientation field of the image. Meanwhile, we also calculate the reliability of the estimated orientation field C _{ k }, which is used to weight the TV in the chosen direction [38]. This makes the DDTV applicable to any kind of images with various directional patterns. Moreover, we also derive an efficient algorithm to solve the iteration problem associated with the DDTV having the directivity of the image adaptively updated for each iteration step.
Finally, the DDTV algorithm is verified through the numerical simulation and in vitro experiments and compared with the FBP and TV. Results obtained show that the DDTV surpasses those two algorithms both in visual quality and quantitative indices. The proposed algorithm demonstrates its superiority, especially for texture images with obvious directivities, where the image edge is preserved better and the texture details are more distinct. In addition, we also compare the peak signaltonoise rations (PSNR), convergence speed, and robustness of the DDTVbased method with those of FBP and TV algorithms.
Arridge et al. [32] also used the TV as the regularizer to solve the PAT reconstruction problem, but their algorithm is quite different from the one proposed in this study, since researchers [32] mainly study the PAT subsampling problems, while our work is mainly focused on the PAT reconstruction problem. Moreover, the TV regularizer used in [32] is the classical TV, which is isotropic and has no relationship with various directivity patterns of images. In this paper, we propose the novel DDTV, which is sensitive to the directions of images. Thus, the edges and texture details of images can be preserved better. Besides, Tick et al. [39] proposed a Bayesian approachbased PAT, which estimates the initial pressure distribution accurately with the uncertainty quantification. In this study, we mainly use a novel DDTV as the sparsity regularizer to solve the PAT reconstruction problem. The Bayesian approach is a probability estimation method, which treats parameters as random variables, and the solution is based on the knowledge of prior information. The DDTVbased algorithm, which is solved iteratively, also needs the directional information of images. However, the directivity of images is estimated and updated for each iteration step with no prior information of the image.
The adaptive directional totalvariation (ADTV) model described in [38, 40] is used in latent fingerprint segmentation with the purpose of decomposing an input image into two layers: cartoon and texture. The DDTV model proposed in this study is applied to the modelbased PAT reconstruction problem as a regularizer in the optimization. Although both DDTV and ADTV models use the same orientation field estimation method and weight the normal TV via spatially varying directions of images, there exist major differences between them. The orientation field estimation of both methods implies calculation of two parameters: orientation θ _{ k }, and reliability of the estimated orientation field C _{ k }. In the ADTV, the orientation vector α _{ k } is computed by multiplying the orientation field (−cosθ _{ k }, sinθ _{ k }) and C _{ k }. The ADTV is obtained by the dot product of the gradient of cartoon layer u and orientation vector α _{ k }. Thus, for the regions with strong orientation patterns, where the value of C _{ k } is large, the textures of this orientation in cartoon layer u are fully depressed, while in the texture layer v they are fully captured. As for the isotropic regions, where C _{ k } approaches zero, α _{ k } becomes a zero vector, so that the ADTV term becomes neglectable, while u depends only on the fidelity term. However, the DDTV in this study is obtained by replacing the unit ball of the L_{2} norm in the TV, which is directionless with spatially varying ellipses having minor axis unit length and major axis length α exceeding unity and oriented along the direction θ _{ k }. The major axis length α _{ k } of the eclipse is calculated via C _{ k }. At regions with strong orientation patterns, where α _{ k } becomes maximum, the TV in that direction is amplified to the largest, while the TV in other directions are weighted depending on the axes of eclipse in these directions. This is different from the ADTV, where the TV in other directions are fully depressed for a strong orientation pattern. As for isotropic regions, where α _{ k } reaches unity, the eclipse turns into a unit circle and the DDTV is reduced to the normal TV. This is also different from the ADTV, wherein the ADTV term is neglectable and the fidelity term becomes dominant. The method proposed in this study is coined DDTV as an abbreviation for “directional total variation with adaptive directivity” to distinguish it from ADTV. Berkels et al. [41] mainly dealt with the cartoon extraction from aerial images, which are mainly characterized by rectangular geometries with varying orientation. They only use the rotation angles to calculate the orthogonal matrix and its dot product with the cartoon part gradients, which is quite different from DDTV that uses both anisotropic directions and the estimated orientation field reliability to weight the image gradients.
The rest of this paper is organized as follows. The first section is “Background”, the second one is the derivation and the framework of the algorithm, “Numerical simulations” describes the numerical simulation results, while the experimental results are presented in “Experimental results”. The last section presents the discussion and conclusion.
Theory and method
Modelbased photoacoustic theory
When irradiated by a laser pulse, the biological tissue absorbs the laser energy and generates ultrasound signals according to the photoacoustic effect. The photoacoustic signals and the laser absorption deposition obey the following relationship [1]:
where p(r, t) is the acoustical pressure measured at the time t and the position r, c is the speed of sound, C _{ p } is the specific heat, β is the isobaric expansion coefficient. H(r, t) is a heating function which can be written as:
where A(r) is the spatial optical absorption distribution of the tissue and I(t) is the temporal laser pulse function.
In this paper, we only consider the twodimension PAT and assume spatially uniform laser irradiation with the laser pulse approximating Dirac’s delta function. A transducer scans the photoacoustic signals at several positions. Then (1) can be solved by a Green’s function and the acoustic pressure detected by the transducer at position r _{0} can be written as [42]:
The analytical reconstruction algorithms are mainly focused on the inverse problem the solutions of (3) to obtain the optical absorption distribution of the tissue A(r). However, the modelbased PAT makes use of (3) to establish a forward model.
Define a new variable g as:
Then, integrating both side of (3), the following equation is derived:
One can see that the right side of (5) is the line integral of A(r) with the path of an arc centered at r _{0} and the radius of ct.
In practical experiments, where the detected signals as well as the photoacoustic image tend to be discretized, g for the lth detection point is discretized to a vector g _{ l } with the length of S, while image A is discretized to a matrix A with the size of N _{ x }×N _{ y }. As it follows from (5), in the discrete form, each element in vector g _{ l } can be expressed via the weighted sum of the elements in matrix A. The size of the weight matrix corresponding to the hth element of g _{ l } is N _{ x }×N _{ y }. Then matrix A and weight matrix are reshaped to the column vector A′ and \({\mathbf{W}}_{l}^{h}\), which have the same length of N_{x}·N_{y}. Thus, the hth element in g _{ l } can be expressed via the dot product of \({\mathbf{W}}_{l}^{h}\) and A′. Arranging these weight vectors \({\mathbf{W}}_{l}^{h}\) in order, the weight matrix W _{ l } corresponding to g _{ l } with the size of N_{x}·N_{y}×S can be constructed. Therefore, for lth detection point, g _{ l } can be expressed as the dot product of W _{ l } and A′ [21]:
where M is the number of the detection points.
The mth element in the weight vector \({\mathbf{W}}_{l}^{h}\) can be written as [31]:
where r _{ l } refers to the position of lth sampling point and r _{ m } refers to the position vector for the mth element in the matrix A, t _{ h } is the time for hth measurement at lth detection point, and Δt is the step size of the discretized time. The weight matrices are determined via the integral path. The elements inside/outside of the integral arc are set to 1 and 0, respectively. However, in practical discretized systems, there are discrete intervals for each discretized variable, so the integral arc may not fully coincide with the discretized points. To reduce this error, we calculate the absolute value \(\left {\frac{{t_{s} }}{\Delta t}  \frac{{\left {{\mathbf{r}}_{l}  {\mathbf{r}}_{m} } \right}}{c\Delta t}} \right\), which is the error between the position of the mth element in the image and the accurate position of the integral arc. When this value exceeds unity, which means that the error exceeds one discrete interval, then this element is not located on the integral arc. When the value is less than one (discrete interval), the weight is from 0 to 1 based on the error calculated by the absolute value. The lager the error, the less the weight, and vice versa. For practical discrete system, g is readily calculated by accumulating values of p. The integral path is determined by the detection points’ and image positions, i.e., the PAT sample pattern. When the latter changes, for example, by shifting from circular scanning to straightline one, the weight matrix W has to be adjusted, according to the sample pattern, making applicable the forward model from A to g in (6). A simple replacement of g by p makes the forward model quite straightforward, transparent, and flexible.
Directivity adaptive directional total variation (DDTV)
The traditional TV, which measures oscillation in an image, is described by the following equation for a discrete image:
where A _{ij} is the pixel value of the image at the coordinate (i,j).
The alternative way to express the TV [43] is:
where B _{2} is the unit ball of the L_{2} norm, and sup is the supremum function. It is obvious that the TV is isotropic, because B _{2} is directionless. Here \(\nabla\) is a linear operator defined as:
where G _{1} and G _{2} are horizontal and vertical gradients of A _{ ij } respectively.
The replacement of B _{2} by an ellipse E _{ α,θ }, which has a unit length minor axis and a major axis length α greater than 1 and is oriented along the direction θ, makes the TV more sensitive to changes at the certain direction, whereas the directivity intensity of that direction is measured by α [43]. Thus, the directional TV (DTV) can be written as:
In [43], only one kind of ellipse is chosen, which has a single direction θ and α. This approach is not universal, since, in most cases, images may have heterogeneous directivity patterns. To overcome this problem, we propose the DDTV with spatially varying θ and α.
The same approach as in [38] is adopted to calculate the spatially different directivity patterns of the image. In order to reduce the calculation efforts, the image is firstly subdivided into blocks of the same size. Then the coarse orientation field for each block is calculated by using the gradientbased approach [36, 37]:
where L is a number of pixels within the block.
To improve the estimation accuracy, O _{k} is further smoothened by the Gaussian smoothing kernel G _{δ} [38]:
Simultaneously, another parameter C _{ k } is calculated, which evaluates the dependability of the estimated direction for each block [38]:
where C _{ k } represents the directivity intensity of kth block within the range of [0,1]. For directionless or isotropic regions, C _{ k } equals to 0, while for strongly oriented regions, C _{ k } approaches 1 [38]. Values θ _{ k }, and C _{ k } are used as the orientation field parameters for all pixels within the block under study.
After completing the calculation, two parameters θ _{ i,j } and C _{ i,j } are obtained for each pixel. Here, the orientation field parameter θ _{ i,j } is used as the spatially varying θ of the ellipse. The major axis length α is defined as:
where α _{ m } is the defined maximum major axis length. For the directionless pattern, C _{ i,j } equals to 0 and α _{ i,j } equals to 1. The ellipse degrades to a circle, which implies that the DTV turns into the TV for that pixel. For the strongest directivity pattern, C _{ i,j } equals to 1 and α _{ i,j } reaches its maximum value α _{ m }.
Thus, DDTV can be described as:
DDTVbased PAT image reconstruction algorithm
Consider the following optimization problem, which has to be solved via the DDTV:
where A* is the reconstructed image. λ is the parameter corresponding to the weight of DDTV value in the optimization.
For each iteration step, the orientation field of the image is firstly estimated by the updated reconstructed image through the method described in “Directivity adaptive directional total variation (DDTV)”. For each pixel (i,j), the rotation matrices R _{ θij } and scaling matrices Λ _{ αij } are controlled by θ _{ i,j } and α _{ i,j } [43]:
After each iteration, the reconstructed image is updated and the orientation field of image is reestimated and updated at the start of the next iteration. Thus, a full DTVregularized inverse problem is solved for each iteration, and then the orientation field is updated.
After that, the problem is solved by minimizing the DDTV via the algorithm proposed in [44, 45]. Define \(\varGamma_{ij} \,=\, \mathop {\text{argmin}}\nolimits_{{\varGamma_{ij} \in B_{2} }} \left\ {Grad({\mathbf{A}}_{{\mathbf{s}}} )_{ij}  \nabla^{T} R_{{\theta_{ij} }} \varLambda_{{\alpha_{ij} }} \varGamma_{ij} } \right\\) and Γ_{ ij } can be solved via the following iteration scheme [44, 45]:
where γ _{ ij } is a coefficient defined as \(\gamma_{ij} = \frac{1}{{8\alpha_{ij}^{2} \lambda^{2} }}\), α _{ ij } is the ellipse major axis length for the pixel (i,j) calculated via (15), H _{ ij } is defined as \({\rm H}_{ij} = \nabla^{T} R_{\theta ij} \varLambda_{\alpha ij}\). Hereinafter A _{ s } is the updated reconstructed image for the sth iteration step, and Grad is the defined gradientdescentbased updating operator, which also are included in the following equations:
As it reported in [44, 45], the final results of this iteration can be obtained from Γ_{ ij }:
Briefly, the iteration steps of the DDTV algorithm can be listed as follows:

1.
Initialization: input A, α _{ m }, λ.

2.
Estimate the orientation fields of A _{ s } via Eqs. (12)–(15). Calculate the rotation matrices R _{ θij } and scaling matrices Λ _{ αij }.

3.
For n = 1 and the defined step number, run the following iteration via Eqs. (1)–(22).

4.
Update the reconstructed image A _{ s+1} via Eq. (23).

5.
If the terminal condition is not satisfied, return to step (2) and continue the iteration. Otherwise, end the iteration.
In the numerical simulations and experiments of this study, we set iteration number 10 as the terminal condition for all cases. Set initial input A to 0 to avoid the input of unnecessary manmade noise caused by the initial guess. The parameters α _{ m } and λ are derived through the experiments.
Numerical simulations
A series of numerical simulations were carried out to validate the proposed DDTVbased algorithm. In order to validate the superiority of the DDTVbased algorithm in the adaptive directional sensitiveness, two kinds of texture images with various directions were chosen as the simulation phantoms. Circular scanning with different sampling points was simulated, and the DDTV results were compared with those of FBP and TV. Then the Shepp–Logan image, which is often adopted to assess the image reconstruction algorithm, was used to verify the effectiveness of the DDTV algorithm quantitatively and qualitatively through the circular, limitedview, and linear scanning options. The PSNR, convergence speed, and robustness of FBP, TV and DDTV algorithms were also analyzed and compared. Finally, several medical images were used to test the universality of the algorithm. The adaptive tunable parameter for lambda. Which was proposed in [31] for the TVbased algorithm, was used in this study. In this case, the initial lambda value was set to 2 for the first iteration and decreased to 0.2, when the iteration number exceeded 10. The iteration time of 10 was set for all cases under study, wherein lambda was relatively large at the beginning of the iteration and decreased as the iterations continued, which provided a good balance of the two parts of the object function. This adaptive tunable parameter proved to be the most effective for the iteration time of 10 [31]. The parameter λ for DDTV was set to maximize the PSNRs of the reconstructed results. In fact, λ determines the weight of DDTV term in the optimization and its large value implies that the DDTVterm is dominant. This would result in a quicker convergence of the algorithm, but too large value of λ will break the balance between the two parts of the objective function. The reconstructed images with a too large λ would much differ from the true ones, due to the data fidelity in the reconstruction being sacrificed to the image regularity. Based on this criterion, a moderate value of λ, which is neither too large nor too small, is preferred.
The simulations were conducted using the Matlab R2013a software installed on a PC with a 2.40 GHz Intel(R) Xeon(R) CPU and 64 GB memory. The speed of sound in the simulation was assumed to be constant and equal to 1500 m/s.
Texture image reconstruction
Two kinds of texture images are selected as the optical absorption distributions of the phantom: the transverse and circular direction patterns, as is shown in Fig. 1a, b. The image dimensions are 128 pixels × 128 pixels, which correspond to the simulation area of 76.8 mm × 76.8 mm, and the scanning radius is 36 mm. The photoacoustic signals are generated by the modified finitedifference timedomain (FDTD) method [46], which are detected circularly with the sampling points of 180, 90, and 30, respectively, while the angular step size is uniform. The modified FDTD method uses photoacoustic Eq. (3) for the numerically produced simulated data. We use the forward projection model described in “Modelbased photoacoustic theory” to reconstruct the image iteratively under DDTV regulation. So the inverse crime can be avoided in our method during the generation of simulated signals. The TV and DDTV iteration numbers are both set to 10, while the parameters λ and α _{ m } are set to 1.2 and 10, respectively. In all following simulations, the size of the block is set to 5×5. The TV and DDTV reconstruction results are compared in Figs. 2 and 3, while the FBP results are also depicted as a reference.
The results obtained strongly indicate a mediocre FBP performance: the contrasts of the image patterns are poor and blurring occurs when the sampling points become sparse. The TV fails to preserve the texture details and the edge sharpness of the image, due its insensitivity to the texture direction. Moreover, there are also some artifacts within the textures, which lead to the contrast reduction. It is clear that, in contrast to TV, the DDTV provides a great improvement of the reconstructed image quality for all directive patterns. The image edges and texture details are well preserved even for the sparseview reconstruction.
Figure 4 shows the orientation field estimation results for two images in the last iteration of 30view DDTV. The direction vectors are multiplied by C _{ k }. Here the arrow length reflects the directivity intensity/strength for that block. As seen from Fig. 4, the arrow directions coincide well with the image directions. The results validate the adaptive direction pattern estimation method in the DDTV.
In order to quantitatively measure the image reconstruction algorithms, we also compare the PSNR of the image, which is defined as:
where N _{ x }, N _{ y } are the image dimensions, and MAXI is the maximum gray value of the image. r _{ i,j } is the gray value of the phantom for pixel (i,j). In this study, images are all normalized to [0,1], so that MAXI equals to 1. Tables 1 and 2 list the PSNR values of three reconstruction algorithms. In all cases under study, the PSNR values provided of the DDTV are much better than those of FBP and TV. In Table 1 corresponding to the transverse texture, the PSNR of the DDTV is by about 16 an 10 dB higher than those of FBP and TV, respectively. As for the circular texture described in Table 2, the PSNR value of the DDTV is about 20 dB and 8.5 dB higher than those of FBP and TV, respectively. Even for the sparseview sampling cases, the DDTV maintains its superiority. For the 30view sampling of the circular texture presented in Table 2, the PSNR value of the DDTV surpasses those of FBP and TV by 20.58 and 12 dB, respectively. The PSNR results of PSNR comply with the visual quality analysis.
To display the details of the reconstructed images, a line of the pixel value of the 30view reconstructed images for the two cases is taken out and compared with that of the original images. Figure 5a, b shows the location of the pixel line in the images, while Fig. 6a, b depicts the comparison curves of the pixel value for two images, respectively. The results obtained show that, as compared to TV, the DDTV profiles are much closer to the standard results, the range abilities of pixel values of the DDTV are much smaller in the homogenous region, and the variation trends coincide better with the true ones.
Shepp–Logan phantom
The Shepp–Logan phantom is also adopted to further evaluate the effectiveness of the algorithm. The image domain dimensions and the scanning radius are the same as those as those in “Texture image reconstruction”. In this case, the angular interval of the sampling points are set to 2°, 4°, 6°, 12°, and the sampling points are 180, 90, 60, and 30, respectively. The parameters λ and α _{ m } are set to 0.01 and 2.5, respectively. The Shepp–Logan phantom is shown in Fig. 7, and the reconstruction results of the DDTV, TV and FBP are shown in Fig. 8. The orientation field estimation results for the last iteration of the DDTV from 30view sampling are displayed in Fig. 9. The arrow directions in Fig. 9 agree with the directivity pattern of the texture edges. In the isotropic parts of the image with weaker directivities, the arrow length is smaller than that in the edge parts with stronger directivities. Table 3 shows the comparison of the PSNR for the three algorithms.
When the number of sampling points is sufficient, all three algorithms provide excellent results, where each part of the phantom in the reconstructed images is distinguishable. However, there is a certain degree of blurs in the reconstructed images of the FBP, while the quality of FBPreconstructed images sharply declines as the sampling points decrease with a lot of artifacts. The DDTV and the TV both have good performances for all sampling cases. However, the DDTVreconstructed images have sharper image edges and clearer texture details, as compared to TV ones. When the number of sampling points is reduced to 30, the DDTV results exhibit no observable changes, while in the TV ones there appear some artifacts. Results of the PSNR listed in Table 3 provide a quantitative proof to the above findings: the PSNR values of DDTV are higher than those of FBP and TV by about 26 and 3.3 dB, respectively. For the sparseview (i.e., 30view) sampling, the PSNR provided by DDTV is by 23.1 dB higher than that of FBP and still exceeds that of TV by 1 dB.
Multimode scanning
To test the validity of the DDTV in terms of multimode scanning, simulations are also conducted under the limitedview and linear scanning conditions. The same simulation environment and parameter settings as in the Shepp–Logan phantom are used. In case of limitedview scanning, the sampling interval is 4° and the number of sample points is equal to 30, which correspond to the sampling angle 120°. In case of linear scanning, the sampling length is 100 mm and the number of sampling points is 20.
The scanning diagram and the reconstruction images of two cases are shown in Fig. 10a–f. Results show that the DDTV can be applied for multimode scanning. But due to the deficiency of the information in some angles, some artifacts appear in both two reconstructed images. In the limitedview scanning results shown in Fig. 10b, there are certain blurs at the image edges, especially in outer contour of the top right corner. For the linearview scanning results shown in Fig. 10d, images oriented in the horizontal direction are relatively good, but there are obvious blurs in the vertical direction of the image due to the less of information in that direction by linearview scanning. The TVreconstructed results for limitedview and linearview scanning conditions are shown in Fig. 10b, e, respectively. Although there are some artifacts due to the lack of some angular information, DDTV exhibits better results, as compared to TV, in the case of limitedview scanning. Further improvements of the limitedview scanning results can be achieved by adding some compensation methods.
Robustness of the algorithm
In practical experiments, the detected signals are readily interfered by the system noises, which are usually the white Gaussian ones, so it is necessary to analyze the noise robustness of the proposed algorithm.
In this part of simulation studies, white Gaussian noises with the signal tonoise ratio (SNR) values of 10, 5, 3, and 0 dB are added to the acquired photoacoustic signals in the case of 30view scanning for texture images depicted in Fig. 1, as well as the Shepp–Logan image shown in Fig. 7. Results obtained via the TV and the DDTV methods are shown in Figs. 11, 12, and 13, while PSNR values are displayed in Tables 4, 5, and 6.
Both DDTV and TV methods succeed to maintain their effectiveness with high SNR. Taking the Shepp–Logan phantom as example, the respective results shown in Fig. 13; Table 6 reveal no strongly manifested influences on the reconstructed images, when SNR = 10 dB. The PSNR values of the DDTV and TV are excellent, amounting to 34.03 and 32.24 dB, respectively. However, the TV performances deteriorate distinctly as the SNR decreases. Especially, when the SNR drops to 0 dB, the reconstructed image of the TV is affected significantly. A lot of background noises can be observed, and the texture details are blurred in the TV outputs, as is shown in Fig. 13d. The PSNR values of the TV decline from 32.24 for 10 dB to 16.96 for 0 dB. On the contrary, the DDTV demonstrates the superiority of noise robustness over the TV. It provides high image quality even under the condition of low SNR. The PSNR of the DDTV for 3 dB SNR is by 5.75 dB higher than that of the TV. For the case of 0 dB SNR, the PSNR of the DDTV is by 9.25 dB higher than that of the TV.
Convergence and calculation
The convergence speeds of two iteration algorithms are compared by the distance between the reconstructed and original images, which is defined as:
where d expresses the degree of difference between the reconstructed and original standard images. The smaller d, the less difference with the original image.
We record d of two methods for each iteration step in cases of 30 and 90view circular texture image reconstructions. The chart of diteration step number is shown in Fig. 14a, b. One can see that the value of d for the DDTV is smaller than that of the TV for each iteration step in both two cases, while the differences of d between two algorithms enlarge as the iteration number increases. For 90view scanning shown in Fig. 14a, the values of d for the DDTV decrease from 0.43 to 0.04 within 10 iteration steps, as compared to variation from 0.46 to 0.12 observed for the TV. For 30view scanning shown in Fig. 14b, the values of d for the DDTV decrease from 0.39 to 0.06 within 10 iteration steps, as compared to the drop from 0.43 to 0.24 of the TV. The results obtained strongly indicate that the DDTV is more accurate and has a faster convergence speed than the TV.
Although TV and DDTV may not converge to the original image, the ultimate goal of using these two algorithms is to reconstruct images, as close as possible to the true ones. The results depicted in Fig. 14 show that as the iteration goes, the distance between the reconstructed and original images decreases within 10 iterations for both algorithms. The distance to iteration curve converges gradually, which validates the convergence of these algorithms. Although the algorithms may only converge to the suboptimal solutions, which approach the optimal one, Eq. (25) still holds for indication of the algorithm convergence property by taking the true images as the gold standard. The distance d in (25) has also been used as an indication of the algorithm convergence in [31] and [47]. Researchers [21,22,23,24,25] also used true images as a standard to calculate the rootmeansquare error for each iteration to study the convergence of the iterative PATreconstructed algorithms.
The computation time is another indicator, which needs to be considered for the evaluation of iterative algorithms. Although TV and DDTV have different optimization algorithms, the proposed DDTV is not a fast algorithm, intended to improve the operation speed. So we compare the computation costs for one iteration step between the TV and the DDTV to show that DDTV has more efficient computation than TV. It is found that the computation times of one iteration step for the TV and the DDTV are roughly the same. For example, the average calculation times of one iteration step for 30view circular scan for a 128 × 128 pixels image are 1.5990 and 1.575 s for the TV and the DDTV, respectively. For 90view, the respective times are 4.257 and 4.366 s for the TV and the DDTV, respectively. Insofar as DDTV has a higher convergence speed than TV, and the computation costs for one iteration are close for the two algorithms, it outperforms the TV according to both factors.
Universality
To test the universality of the algorithm, we also choose three medical images as the original optical absorption distribution, which are two MRI brain images and one angiography image. The reconstruction results of the DDTV and the TV are displayed in Fig. 15. For the MRI brain images, the reconstructed images of the DDTV, as is shown in Fig. 15d–f, i–l, have more distinct image edges and texture details. Moreover, the central paracele and sulci in the first brain image have a higher contrast and the tumor in the second brain image is clearer for the DDTV than that of the TV, as is shown in Fig. 15a–c, g–i. As for the angiography image, one can observe that DDTV images shown in Fig. 15p–r have better performances in terms of image contrast and texture detail preservation. These results confirm the DDTV applicability for various kinds of images.
Experimental results
This Section briefly describes the in vitro experiments conducted to verify the DDTV algorithm. We first made two vessellike phantoms to test the effectiveness of the proposed DDTV algorithm through 90 and 30view circular scanning. The reconstructed results of the DDTV were compared with those of FBP and TV. Linear scanning experiments were also conducted, with the DDTV and TV results being compared and discussed.
The experimental platform is shown in Fig. 16. A laser beam irradiated with a Nd:YAG laser device (Contimuum, Surelite I) is reflected by a mirror and then transferred through a concave lens. The wavelength of the laser in this experiment is 532 nm. The duration and repetition frequency of the laser pulse are 4–6 ns and 10 Hz, respectively, which comply with the experimental requirements. The phantom is homogeneously illuminated from above, and photoacoustic signals are detected from the side. The transducer (V383SU, Panametrics) is unfocused with the center frequency of 3.5 MHz and bandwidth of 1.12 MHz. A stepping motor drives the transducer to scan around the phantom. The sampling frequency of the system is 16.67 MHz.
The experiments on the phantoms (Fig. 17) have been earlier performed by the authors and reported in [31, 47]. These experimental data are referred to in this study, in order to verify whether the proposed algorithm has a better performance than the TVbased one under the same test conditions with the same experimental data. Also, we added a new linear scan experiment in Figs. 20 and 21 to show that our method is capable of accurate image reconstruction under different kinds of scanning modes.
The phantoms used in Fig. 17 are made of gelatin with the tissues to be imaged embedded into the gelatin cylinder. The phantom shown in Fig. 17a uses two rubber bars of 20 and 12 mm in length, respectively. As is shown in Fig. 17b, three wires with the diameter of 1 mm each are embedded into the gelatin. The diameters of both phantoms are 50 mm. Transducer scans around the phantom circularly with the scanning radius of 38 mm. The angular intervals of the sampling points are set to 4° and 12°, which corresponds to 90 and 30view samplings, respectively. The laser energy density is set to meet the ANSI laser radiation safety standards.
The experimental reconstruction results obtained by FBP, TV, and DDTV are shown in Figs. 18 and 19 and confirm that all three algorithms can provide nearly perfect reconstruction results, when the number of sampling points is sufficient. However, for 30view sampling, the image quality of the FBP declines sharply, with appearance of multiple artifacts. Both TV and DDTV maintain their effectiveness in sparseview sampling. As compared to the TV, the DDTV shows its advantages in terms of the reconstruction results. The contrast of the images is enhanced and the edges of the image are more distinct. So the DDTV is preferable to TV for maintaining the image texture details.
We also performed the linear scanning experiments. The phantom is also a gelatin cylinder with the diameter of 50 mm, which is embedded into a rectangular rubber slice acting as the laser beam absorber. The slice dimensions are 9 mm × 14 mm. The phantom is shown in Fig. 20. The photoacoustic signals are sampled uniformly aligning to the longer edge of the rectangle. The sampling interval is 1 mm and the number of the sampling points is 41.
The results of the TV and the DDTV are shown in Fig. 21. Due to the incompleteness of the angular information of the linear scan, platelike artifacts appear in both TV and DDTV results. But the DDTV shows sharper edges of the laser beam absorber. The profile of the rectangle is clearly visible and the pixel values are distributed more uniformly within the rectangle. So the DDTV is more effective than the TV for the linear scan experiments.
Discussion and conclusion
In this study, a novel modelbased photoacoustic image reconstruction algorithm using the directivity adaptive direction total variation (DDTV) is proposed to minimize the deficiencies of the PAT image reconstruction algorithms in terms of image edges and texture details preservation. The classical TV is a sum of L_{2} gradient norms of the image, which measures the variations and penalizes local changes in the image regardless of the image directions. So the TVbased algorithms perceive that the reconstructed images are piecewiseconstant, which may hold for cartoonlike piecewise images, while most natural image reconstructions violate this assumption. The TV models in such applications as image denoising problems minimize the TV along all directions, which leads to a disredard of the important directivity information of images. Therefore, classical TVs are not suitable for images with strong directivity patterns. When applied to PAT image reconstruction methods, the TVbased algorithms tend to produce oversmoothed image edges and texture details, since the TV assumption favors piecewiseconstant solutions. So the problem of TVbased PAT reconstruction is that it fails to obtain adequate results, in terms of texture detail preservation, which is critical for PAT images with spatially varying directivity patterns, such as blood vessel imaging, tumor detection, biological tissue microstructures analysis. The classical TVbased PAT reconstruction methods fail to provide ideal directional information in terms of the reconstructed results. So the application of DDTV to PAT is quite lucrative, since it calculates the TV based on the spatially varying directivity patterns of the image, which makes the TV sensitive to the image directions. The anisotropic directivity pattern of the image is estimated adaptively during the iterations, and the image gradients are weighted by the estimated orientation fields, which makes is applicable to all kinds of images with various directivity patterns. Two kinds of parameters are calculated for the orientation field estimation: direction θ and reliability of the estimated direction C _{ k }, which control the TV directions and the sensitive degree for those directions, respectively. This dualweight method assures the accuracy of the DDTV parameter for the image.
The maximum major axis α _{ m } defines the largest weight of TV in the chosen direction θ _{ k }. When the orientation pattern is strong, the weight of TV in that direction reaches α _{ m }. Small values of α _{ m } may lead to worse performance of DDTV as to preserving the edges and texture details of the image, in comparison to that of TV. In a more extreme case, when α _{ m } equals unity, the ellipse turns into a directionless ball, which makes DDTV equal to TV. Vice versa, too large values of α _{ m } will cause the oversized weight of TV in θ _{ k }, leading to a relatively small TV in that direction. Thus, the reconstructed image will be distorted. Therefore, α _{ m } should be set to a value which is neither too small nor too large. In general, for images with stronger direction patterns α _{ m } should be larger, while for relatively homogeneous ones, it should be smaller. In the simulations, the phantoms of texture images have relatively strong direction patterns, so that α _{ m } = 10 is used in this case to obtain the best performance. The Shepp–Logan phantom is comparatively uniform with a weak directivity, so α _{ m } in this case is set to 2.5. The simulations of this study can be used as a reference for selection of α _{ m }, while in general reconstruction cases α _{ m } can be set between 2–10.
The DDTV was implemented into the modelbased PAT reconstructed algorithm, and the primal–dual based method described in [43] is utilized to solve the optimization problem iteratively. Numerical simulations and in vitro experiments verify the effectiveness of the algorithm. The reconstructed images are studied qualitatively and quantitatively. Simulation results show that the DDTV method has a better performance than FBP and TV, in terms of PSNR, robustness, convergence speed, and visual quality, while its superiority is exhibited even for sparseview and multimode sampling cases. As reported in [31], the TV algorithm converges when the number of iterations is 10, which number is a relatively appropriate choice for TV algorithms. The experimental data show that DDTV also converges about 10th iteration, so the average converging number of 10 was also selected. As shown in “Convergence and calculation”, the line chart of the distance d in Fig. 14 also confirms that the distance between the iteration curves converges when the number of the iterations is 10, which validates the convergence of the algorithms at 10th iteration. From simulations of various images, it can be observed that the dominants of the proposed DDTV algorithm are more prominent for images with a stronger directivity. For the two texture images, chosen in the first part of simulation, the values of PSNR of DDTV are by 16.24 and 9.96 dB higher than those of FBP and TV, respectively, for the transverse texture image. For the circular image, the respective excesses are 19.99 and 8.35 dB over those of FBP and TV. Other texture images with different texture patterns were also tested, with the PSNR values of DDTV exceeding those of FBP and TV by about 18 and 10 dB for all the cases. As for the Shepp–Logan phantom in the second part of simulation, the superiority of DDTV is less pronounced as that of texture images, because the directivity of the Shepp–Logan is not so strong and the major part of the image corresponds to smooth areas with no directions. So, for images a with weak directivity, the performances of DDTV and TV are similar, but the improved quality of the reconstructed images is also confirmed by the Shepp–Logan phantom results. Due to different mechanisms for the reconstructed algorithms, the gray level of the reconstructed images would be different. There may be a certain degree of amplification or narrowing for the gray level of the reconstructed images, as compared to the original ones. Therefore, it is difficult to compare the results for those algorithms under different gray levels. The images are normalized via dividing the gray values of all pixels by the maximum gray value of the images, which has no effect on the quality of images. Also when comparing the PSNR and d for all algorithms, the reconstructed images and the original images should be at the same gray level. Therefore, the images in this study are normalized in the same gray level for comparison. The simulation results also suggest although the DDTV has a higher convergence speed, than that of the TV, it fails to improve the calculation time. The calculation times for an iteration step of the DDTV and the TV are basically the same. Thus, further studies are needed to develop a faster and more efficient method to solve the iterated problem. Moreover, the reconstructed result of the DDTV in the linearview scanning is no better than those obtained via circular and limitedview ones. The information in the vertical direction of the image is not properly reconstructed. So we also need to modify the proposed algorithm to improve the performance for the linear scanning in the future.
Invitro experiments further prove that the proposed DDTV algorithm is able to reconstruct images with a higher quality than FBP and TV algorithms for both circular and linear scanning cases. The contrast of the images is enhanced and the edges of the image are more distinct. The profiles of the optical absorbers are clearer and the texture details information of the images are easier to observe. Thus, the DDTVbased PAT reconstruction algorithm has very promising prospects for biomedical applications, especially those, which need the texture detail information of the tissue. In the experiments, simple gelatinbased phantoms are used. A single, lowfrequency transducer is used in the experiment to detect the PAT signals by scanning around the phantom, which is quite timeconsuming and inconvenient for practical application. In the future, we will improve the experimental system for more complicated biological tissue and in vivo experiments. Consequently, the results of numerical simulations and experiments corroborate that the proposed DDTV is an effective algorithm for the PAT image reconstruction, which has a number of advantages over the available FBP and classical TVbased algorithms.
Abbreviations
 PA:

photoacoustic
 PAT:

photoacoustic imaging
 TV:

total variation
 DDTV:

directional total variation with adaptive directivity
 ADTV:

adaptive directional totalvariation
 PSNR:

peak signaltonoise rations
 FBP:

filtered backprojection method
 CS:

compressed sensing
 ASDPOCS:

adaptive steepestdecentprojection onto convex sets
 DTV:

directional total variation
 FDTD:

finitedifference timedomain
 SNR:

signaltonoise ratio
References
 1.
Xu M, Wang LV. Photoacoustic imaging in biomedicine. Rev Sci Instrum. 2006;77:041101.
 2.
Li C, Wang LV. Photoacoustic tomography and sensing in biomedicine. Phys Med Biol. 2009;54:R59–97.
 3.
Laufer J, Delpy D, Elwell C, Beard P. Quantitative spatially resolved measurement of tissue chromophore concentrations using photoacoustic spectroscopy: application to the measurement of blood oxygenation and haemoglobin concentration. Phys Med Biol. 2007;52:141–68.
 4.
Wang LV. Tutorial on photoacoustic microscopy and computed tomography. IEEE J Sel Top Quantum Electron. 2008;14:171–9.
 5.
Kruger R, Reinecke D, Kruger GA. Thermoacoustic computed tomography–technical considerations. Med Phys. 1999;26:1832–7.
 6.
Kruger RA, Liu P, Appledorn CR. Photoacoustic ultrasound (PAUS)—reconstruction tomography. Med Phys. 1995;22:1605–9.
 7.
Wang X, Pang Y, Ku G, Xie X, Stoica G, Wang LV. Noninvasive laserinduced photoacoustic tomography for structural and functional in vivo imaging of the brain. Nat Biotechnol. 2003;21:803–6.
 8.
Li C, Aguirre A, Gamelin J, Anastasios M, Zhu Q, Wang LV. Realtime photoacoustic tomography of cortical hemodynamics in small animals. J Biomed Optics. 2010;15:0105091–3.
 9.
Zhong J, Wen L, Yang S, Xiang L, Chen Q, Xing D. Imagingguided highefficient photoacoustic tumor therapy with targeting gold nanorods. Nanomed Nanotechnol Biol Med. 2015;11:1499–509.
 10.
Niederhauser JJ, Jaeger M, Lemor R, Weber P, Frenz M. Combined ultrasound and optoacoustic system for realtime highcontrast vascular imaging in vivo. IEEE Trans Med Imaging. 2005;24:436–40.
 11.
Zerda ADL, Zavaleta C, Keren S, Vaithilingam S, Bodapati S, Liu Z, Levi J, Smith BR, TeJen MA, Oralkan O. Carbon nanotubes as photoacoustic molecular imaging agents in living mice. Nat Nanotechnol. 2008;3:557–62.
 12.
Li L, Lungu ZG, Stoica G, Wang LV. Photoacoustic imaging of lacZ gene expression in vivo. J Biomed Optics. 2007;12:020504.
 13.
Kunyansky LA. Explicit inversion formulae for the spherical mean Radon transform. Inverse Prob. 2007;23:373.
 14.
Xu M, Wang LV. Pulsedmicrowaveinduced thermoacoustic tomography: filtered backprojection in a circular measurement configuration. Med Phys. 2002;29:1661–9.
 15.
Zhang C, Wang Y. Deconvolution reconstruction of fullview and limitedview photoacoustic tomography: a simulation study. J Opt Soc Am A. 2008;25:2436–43.
 16.
Zhang C, Li C, Wang LV. Fast and robust deconvolutionbased image reconstruction for photoacoustic tomography in circular geometry: experimental validation. IEEE Photonics J. 2010;2:57–66.
 17.
Xu Y, Wang LV. Time reversal and its application to tomography with diffracting sources. Phys Rev Lett. 2004;92:033902.
 18.
Xu Y, Wang LV. Application of time reversal to thermoacoustic tomography. Proc SPIE. 2004;5302:257–63.
 19.
Treeby BE, Zhang EZ, Cox B. Photoacoustic tomography in absorbing acoustic media using time reversal. Inverse Prob. 2010;26:115003.
 20.
Cox BT, Treeby BE. Artifact trapping during time reversal photoacoustic imaging for acoustically heterogeneous media. IEEE Trans Med Imaging. 2010;29:387–96.
 21.
Paltauf G, Viator JA, Prahl SA, Jacques SL. Iterative reconstruction algorithm for optoacoustic imaging. J Acoust Soc Am. 2002;112:1536–44.
 22.
Ma S, Yang S, Guo H. Limitedview photoacoustic imaging based on lineararray detection and filtered meanbackprojectioniterative reconstruction. J Appl Phys. 2009;106:123104–6.
 23.
Huang C, Wang K, Nie L, Wang LV, Anastasio MA. Fullwave iterative image reconstruction in photoacoustic tomography with acoustically inhomogeneous media. IEEE Trans Med Imaging. 2013;32:1097–110.
 24.
DeanBen XL, Ntziachristos V, Razansky D. Acceleration of optoacoustic modelbased reconstruction using angular image discretization. IEEE Trans Med Imaging. 2012;31:1154–62.
 25.
Rosenthal A, Jetzfellner T, Razansky D, Ntziachristos V. Efficient framework for modelbased tomographic image reconstruction using wavelet packets. Med Imaging IEEE Trans. 2012;31:1346–57.
 26.
Donoho DL. Compressed sensing. IEEE Trans Inf Theory. 2006;52:1289–306.
 27.
Provost J, Lesage F. The application of compressed sensing for photoacoustic tomography. IEEE Trans Med Imaging. 2009;28:585–94.
 28.
Candes EJ, Romberg J, Tao T. Robust uncertainty principles: exact signal reconstruction from highly incomplete frequency information. IEEE Trans Inf Theory. 2006;52:489–509.
 29.
Rudin LI, Osher S, Fatemi E. Nonlinear total variation based noise removal algorithms. Phys D. 1992;60:259–68.
 30.
Wang K, Sidky EY, Anastasio MA, Oraevsky AA, Pan X. Limited data image reconstruction in optoacoustic tomography by constrained total variation minimization. Proc SPIE. 2011;7899:78993–6.
 31.
Zhang Y, Wang Y, Zhang C. Total variation based gradient descent algorithm for sparseview photoacoustic image reconstruction. Ultrasonics. 2012;52:1046–55.
 32.
Arridge S, Beard P, Betcke M, Cox B, Huynh N, Lucka F, Ogunlade O, Zhang E. Accelerated highresolution photoacoustic tomography via compressed sensing. arXiv preprint arXiv. 2016; 1605: 00133.
 33.
Fei X, Wei Z, Xiao L. Iterative directional total variation refinement for compressive sensing image reconstruction. IEEE Signal Process Lett. 2013;20:1070–3.
 34.
Candes EJ, Wakin MB, Boyd SP. Enhancing sparsity by reweighted L1 minimization. J Fourier Anal Appl. 2007;14:877–905.
 35.
Tian Z, Jia X, Yuan K, Pan T, Jiang SB. Lowdose CT reconstruction via edgepreserving total variation regularization. Phys Med Biol. 2011;56:5949–67.
 36.
Hong L, Wan Y, Jain A. Fingerprint image enhancement: algorithm and performance evaluation. IEEE Trans Pattern Anal Mach Intell. 2010;20:777–89.
 37.
Zhu E, Yin J, Hu C, Zhang G. A systematic method for fingerprint ridge orientation estimation and image segmentation. Pattern Recognit. 2006;39:1452–72.
 38.
Zhang J, Lai R, Kuo CCJ. Latent fingerprint detection and segmentation with a directional total variation model. In: 2012 19th IEEE international conference on image processing. New York: IEEE; 2012. p. 1145–1148.
 39.
Tick J, Pulkkinen A, Tarvainen T. Image reconstruction with uncertainty quantification in photoacoustic tomography. J Acoust Soc Am. 2016;139:1951–61.
 40.
Zhang J, Lai R, Kuo CCJ. Adaptive directional totalvariation model for latent fingerprint segmentation. IEEE Trans Inf Forensics Secur. 2013;8:1261–73.
 41.
Berkels B, Burger M, Droske M, Nemitz O, Rumpf M. Cartoon extraction based on anisotropic image classification vision, modeling, and visualization. In: vision, modeling, and visualization 2006: Proceedings, 2006. Aachen: IOS Press; 2006. p. 293.
 42.
Xu M, Wang LV. Timedomain reconstruction for thermoacoustic tomography in a spherical geometry. IEEE Trans Med Imaging. 2002;21:814–22.
 43.
Bayram I, Kamasak ME. Directional total variation. IEEE Signal Process Lett. 2012;19:781–4.
 44.
Chambolle A. An algorithm for total variation minimization and applications. J Math Imaging Vision. 2004;20:89–97.
 45.
Esedo S, Osher SJ. Decomposition of images by the anisotropic RudinOsherFatemi model. Commun Pure Appl Math. 2004;57:1609–26.
 46.
Zhang C, Wang Y. Sound field of thermoacoustic tomography based on a modified finitedifference timedomain method. Chin J Acoust. 2009;3:209–19.
 47.
Zhang C, Wang Y. High total variationbased method for sparseview photoacoustic reconstruction. Chin Optics Lett. 2014;12:81–5.
Authors’ contributions
Study concept and design (JW); drafting of the manuscript (JW); critical revision of the manuscript for important intellectual content (JW,CZ and YW); obtained funding (YW); administrative, technical, and material support (JW and CZ); study supervision (YW). All authors read and approved the final manuscript.
Acknowledgements
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Avaliability of data and materials
The datasets used and/or analyzed during in current study are available from the corresponding author on reasonable requests.
Funding
This work was supported by the National Basic Research Program of China (2015CB755500) and the National Natural Science Foundation of China (No. 11474071).
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Author information
Affiliations
Corresponding author
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.
About this article
Cite this article
Wang, J., Zhang, C. & Wang, Y. A photoacoustic imaging reconstruction method based on directional total variation with adaptive directivity. BioMed Eng OnLine 16, 64 (2017). https://0doiorg.brum.beds.ac.uk/10.1186/s1293801703663
Received:
Accepted:
Published:
DOI: https://0doiorg.brum.beds.ac.uk/10.1186/s1293801703663
Keywords
 Photoacoustic tomography
 Image reconstruction
 Directional total variation
 Directivity adaptive