Europe PMC

This website requires cookies, and the limited processing of your personal data in order to function. By using the site you are agreeing to this as outlined in our privacy notice and cookie policy.

Abstract 


Purpose

Lung tumor tracking during stereotactic radiotherapy with the CyberKnife can misrecognize tumor location under conditions where similar patterns exist in the search area. This study aimed to develop a technique for bone signal suppression during kV-x-ray imaging.

Methods

Paired CT images were created with or without bony structures using a 4D extended cardiac-torso phantom (XCAT phantom) in 56 cases. Subsequently, 3020 2D x-ray images were generated. Images with bone were input into cycle-consistent adversarial network (CycleGAN) and the bone suppressed images on the XCAT phantom (BSIphantom ) were created. They were then compared to images without bone using the structural similarity index measure (SSIM) and peak signal-to-noise ratio (PSNR). Next, 1000 non-simulated treatment images from real cases were input into the training model, and bone-suppressed images of the patient (BSIpatient ) were created. Zero means normalized cross correlation (ZNCC) by template matching between each of the actual treatment images and BSIpatient were calculated.

Results

BSIphantom values were compared to their paired images without bone of the XCAT phantom test data; SSIM and PSNR were 0.90 ± 0.06 and 24.54 ± 4.48, respectively. It was visually confirmed that only bone was selectively suppressed without significantly affecting tumor visualization. The ZNCC values of the actual treatment images and BSIpatient were 0.763 ± 0.136 and 0.773 ± 0.143, respectively. The BSIpatient showed improved recognition accuracy over the actual treatment images.

Conclusions

The proposed bone suppression imaging technique based on CycleGAN improves image recognition, making it possible to achieve highly accurate motion tracking irradiation.

Free full text 


Logo of jacmpLink to Publisher's site
J Appl Clin Med Phys. 2024 Jan; 25(1): e14212.
Published online 2023 Nov 20. https://doi.org/10.1002/acm2.14212
PMCID: PMC10795441
PMID: 37985163

Cycle‐generative adversarial network‐based bone suppression imaging for highly accurate markerless motion tracking of lung tumors for cyberknife irradiation therapy

Abstract

Purpose

Lung tumor tracking during stereotactic radiotherapy with the CyberKnife can misrecognize tumor location under conditions where similar patterns exist in the search area. This study aimed to develop a technique for bone signal suppression during kV‐x‐ray imaging.

Methods

Paired CT images were created with or without bony structures using a 4D extended cardiac‐torso phantom (XCAT phantom) in 56 cases. Subsequently, 3020 2D x‐ray images were generated. Images with bone were input into cycle‐consistent adversarial network (CycleGAN) and the bone suppressed images on the XCAT phantom (BSIphantom) were created. They were then compared to images without bone using the structural similarity index measure (SSIM) and peak signal‐to‐noise ratio (PSNR). Next, 1000 non‐simulated treatment images from real cases were input into the training model, and bone‐suppressed images of the patient (BSIpatient) were created. Zero means normalized cross correlation (ZNCC) by template matching between each of the actual treatment images and BSIpatient were calculated.

Results

BSIphantom values were compared to their paired images without bone of the XCAT phantom test data; SSIM and PSNR were 0.90 ± 0.06 and 24.54 ± 4.48, respectively. It was visually confirmed that only bone was selectively suppressed without significantly affecting tumor visualization. The ZNCC values of the actual treatment images and BSIpatient were 0.763 ± 0.136 and 0.773 ± 0.143, respectively. The BSIpatient showed improved recognition accuracy over the actual treatment images.

Conclusions

The proposed bone suppression imaging technique based on CycleGAN improves image recognition, making it possible to achieve highly accurate motion tracking irradiation.

Keywords: CycleGAN, deep learning, generative adversarial networks (GAN), lung cancer, radiotherapy, stereotactic radiotherapy

1. INTRODUCTION

Stereotactic body radiation therapy (SBRT) has been widely applied as a treatment modality for early stage non‐small cell lung cancer (NSCLC). SBRT combines the techniques of multidirectional irradiation and precise irradiation of lesions to improve local control of tumors confined to the trunk region and reduce adverse events in normal tissue nearby. SBRT is also capable of high efficacy with a single large dose, which cannot be achieved with conventional radiotherapy. 1 The CyberKnife (CK) System (Accuray Inc., Sunnyvale, CA, USA) integrates a robotic‐positioned linear accelerator, an image‐guided system, and respiratory tracking systems. CK has two respiratory tracking systems: the fiducial‐based target tracking system (FTTS) and Xsight Lung Tracking System (XLTS). 2 XLTS is a fiducial‐free real‐time tracking system used to irradiate lung tumors that move with respiration. 3 This facilitates the minimization of radiation exposure to healthy pulmonary tissues, thereby enhancing the efficacy of treatment. 4 , 5 This system, called target locating system (TLS), uses a pair of orthogonal x‐ray imagers, which enable real‐time tracking of moving targets by modeling the correlation between the targets and external surrogate light‐emitting diode markers placed on the patient's chest. 6 With this system, the patient's treatment area is tracked using real‐time imaging, and the error in the irradiated area is less than 1 mm. 7 , 8

Generally, sufficient tumor contrast relative to the surrounding tissue in an x‐ray image is essential for accurate soft‐tissue tracking. XLTS performs template matching to match the image density pattern of the tumor region in digitally reconstructed radiography (DRR) with the most similar region in the live x‐ray image. This may affect visibility, especially in cases with overlapping bone. Because the tumor recognition accuracy is not universal for all lesions, the author implemented an approach for bone suppression imaging using a type of general adversarial network (GAN), called cycle‐consistent adversarial network (CycleGAN). In recent years, GANs have shown state‐of‐the‐art performance in many image‐processing tasks. 9 , 10 In addition, CycleGAN was proposed to learn translation mappings in the absence of aligned paired images. 10 Recently, CycleGAN was used to synthesize CT images from cone beam CT (CBCT) images. 11 However, no research has been conducted to date that incorporates CycleGAN into a CK motion‐tracking algorithm. Therefore, this study aimed to use deep learning to develop a bone suppression imaging technique for kV x‐ray images.

2. MATERIALS AND METHODS

After the images with bone were input to the trained model to create bone suppressed images, the images generated by CycleGAN (BSIphantom) were compared to images without bone. Next, BSIpatient were generated from the images used for actual CK treatment, and the effectiveness of deep learning was compared by template matching for each of the actual treatment images and BSIpatient. The workflow is illustrated in Figure 1.

An external file that holds a picture, illustration, etc.
Object name is ACM2-25-e14212-g005.jpg

Work‐flow of the bone suppression and evaluation framework.

2.1. Data

2.1.1. XCAT phantom (4D extended cardiac‐torso phantom) images

We first created images from XCAT phantom (Duke University, Durham, North Carolina) 12 , 13 to project the images to be used in the deep‐learning datasets. XCAT phantom is a digital anthropomorphic phantom image database that was developed based on the human anatomy database of the National Library of Medicine. Based on patient data and using non‐uniform rational b‐spline surfaces to define the anatomy, the XCAT phantom combines a voxelized approach with a mathematical approach to offer simulated imaging with realistic and detailed organs that remain flexible to allow for anatomical variation and organ deformation. 14 , 15 , 16 In addition, the XCAT phantom is positioned off the center of the image, but patient images in CT images can rotate, shift, or exceed the field of view. Here, we used 56 samples from this phantom, which all differed in age, sex, nationality, height, and weight. The dataset included 56 men and women aged 18−78 years, weight 52−120 kg, height 153−190 cm, and BMI:18‐39 kg/m2. The dxcat2 code of the XCATv2 software package was used by turning off the activity unit of the bone to obtain images with and without bony structures, as shown in Figure 2a. For settings, bone size parameter of 1, pixel width and slice width of 0.03 cm, and array size of 512 × 512 were used.

An external file that holds a picture, illustration, etc.
Object name is ACM2-25-e14212-g003.jpg

Creation of image datasets with and without bone XCAT phantom. (a): With (L)/without (R) bone generated from XCAT Phantom, (b): Pseudo‐projecting the XCAT phantom, Projections with 5 degree difference, (d): The projected images with (L)/without (R) bone as in (b) were divided into 9 segments.

Next, to create 2D CK images based on these images, we used 3D Slicer ver.4.11 software (MIT, Massachusetts, USA) to create 45° images, as shown in Figure 2b. Images at 40° and 50° were created in the same manner in order to increase the number of datasets used, as shown in Figure 2c. The View‐Up vector was set to (x,y,z) = (−1,0,45), Normal vector to (x,y,z) = (any angle,−45,0), and Isocentor position to (x,y,z) = (0,0,0). These were then divided into nine segments, as shown in Figure 2d. This process created a total of 3020 images, of which, data from 340 images were used for testing.

2.1.2. Treatment images

To evaluate the model, data from actual treated patients were required. This was obtained from 50 patients with metastatic lung cancer (tumor size 8–64 mm) and primary lung cancer (tumor size 10–62 mm) undergoing SBRT at our center between February 2020 and January 2022. 40 patients were male and 10 were female. The ages of the patients were between 46 and 93 years (average age, 74.14). All patients were treated with the CK G4 system using 2‐view XLTS (100 kV/200 mA/100 ms).

2.2. Optimization and training methods

The training methods and parameters were based on CycleGAN. Similar to pix2pix, 17 Zhu et al. proposed CycleGAN, a type of image‐generation algorithm using GAN. 10 Figure 3 shows the CycleGAN procedure. The domain of the images was defined, and images were collected for each domain as training data. The main feature of CycleGAN is that there is no need to prepare a pair of training data. The set of images for each domain is denoted as X and Y, and a generator is prepared to perform X‐to‐Y and Y‐to‐X transformations on them. The default generator architecture of CycleGAN is ResNet, while the default discriminator architecture is a PatchGAN classifier. 17 , 18 In addition, two discriminators corresponding to both were prepared. The proposed method enables non‐pairwise image transformations by learning the error (loss) between adversarial loss used in the GAN and the cycle consistency Loss proposed in this study. However, it is shown to be less accurate than pix2pix because of unsupervised learning.

An external file that holds a picture, illustration, etc.
Object name is ACM2-25-e14212-g004.jpg

Procedure of CycleGAN. CycleGAN is a method that achieves image transformation by learning domain relationships such as fields and regions between two images.

In this study, the algorithm was run on a workstation with an NVIDIA® GeForce® RTX 2080 SUPER™ 8GB GPU. Adam 19 was used as the optimizer, with β1 = 0.5 and β2 = 0.999. The learning rate was set to 0.0002 for both the generator and discriminator. Both the generator and discriminator were updated once for each iteration, and the training was terminated when each network was considered to have been generated with a certain degree of accuracy.

2.3. Evaluation methods

The BSIphantom was compared to images without bone from the test data, and the similarity was calculated. We evaluated the models with two different objective image quality metrics, namely, the structural similarity index (SSIM) 20 and peak signal‐to‐noise ratio (PSNR). 21

SSIM is a method to evaluate image quality based on the structural information of an image; it is shown in Equations (1) and (2). Equation (1) was used to obtain the SSIM for each block, and Equation (2) was used to obtain the average SSIM for all blocks.

SSIM(x,y)=(2μxμy+C1)(2σxy+C2)μ2x+μ2y+C1σ2x+σ2y+C2
(1)

MSSIM=1/Mj=1MSSIM(xj,yj)
(2)

where x is one block of the reference image, y is one block of the test image, and μ is the average per block of the Gaussian filtered image. Similarly, σ2 refers to the variance per block of the Gaussian filtered image: C 1 and C 2 are constants, (0.01 × 255)2 and (0.03 × 255)2, respectively. M is the number of blocks.

PSNR is an index obtained by the ratio of the mean squared error (MSE) of the reference image and the test image to the maximum grayscale value (PS = 255). The PSNR is shown in Equation (3), and the MSE is shown in Equation (4).

PSNR=10log10PS2MSE
(3)

MSE=1N1Nxiyi2
(4)

where xi represents the reference image; yi represents the gray scale value of the test image; N represents the total number of images.

Next, template matching 22 , 23 was performed on each of the actual treatment images and BSIpatient to calculate the zero‐mean normalized cross‐correlation (ZNCC). 24 Therefore, in this study, the search area was the entire image and the goal was to capture the tumor contour. The detection method using template matching is the most basic and determines the location of the target in the input image by comparing the input image with the partial image of the target to be tracked, which was prepared in advance. ZNCC is a measure that uses the average luminance values of the template and input image to compensate for the effects of the lighting environment. The location with the highest value is the target location. In addition, ZNCC is a general measure of image similarity and ranges from −1 to 1, with a value of 1 when the two images are exactly matched. After 1000 actual treatment images were run through CycleGAN to BSIpatient, template matching was performed on each of the actual treatment images and BSIpatient to calculate ZNCC.

2.4. Statistical analysis

All statistical analyses were performed using SPSS Statistics for Windows (version 27.0; IBM Corp., Armonk, NY, Version 27.0. Armonk, NY: IBM Corp). Differences between the groups were compared using independent T‐tests. Alpha was set to 0.05.

3. RESULTS

Figure 4 shows examples of BSIphantom compared to the test data without bone, and the similarity was calculated. The bone intensity was generally reduced, although some remained intact. The mean of SSIM and PSNR were 0.90 ± 0.06 and 24.54 ± 4.48, respectively, indicating high similarity. Figure 4g,h show examples of image generation using CycleGAN. The blue line indicates a tumor tracking volume (TTV). The bone intensity was suppressed, and tumor contours remained after image generation.

An external file that holds a picture, illustration, etc.
Object name is ACM2-25-e14212-g002.jpg

Examples between the test data and BSI. (a): the test data with bone, (b): the test data without bone, (c): BSIphantom, (d–f): partially enlarged versions of (a–c), (g): actual treatment images and (h): BSIpatient, the blue line in the patient data shows the TTV.

Next, template matching was performed using the actual treatment images and BSIpatient, and the ZNCC was calculated for each. Figure 5 shows two examples of template matching. The middle figure shows heatmaps for the similarity of each image. ZNCC is vectorized each cell in the heatmap and calculated the similarity based on its angle. The red color is indicated higher similarity; ZNCC was significantly different and higher for images without bone structures. The heatmap visualizes the model based on these differences.

An external file that holds a picture, illustration, etc.
Object name is ACM2-25-e14212-g001.jpg

Template matching for each of actual treatment images and BSIpatient. The upper figure shows (a,c): actual treatment images and (b,d): BSIpatient. The green box in the patient data shows the region of high similarity in template matching, demarcated by rectangular windows. The middle figure shows heatmaps for the similarity of each image. The bottom figure shows the subtraction images for both pairs.

Table 1 shows the average ± standard deviation of ZNCC for the 767 images that could be detected by template matching. The p‐value was less than 0.05, which was considered statistically significant, indicating that the BSIpatient had higher ZNCC than the actual treatment images.

TABLE 1

ZNCC was calculated by template matching of actual treatment images and BSIpatient, respectively.

Image typesTotal number of imagesNumber of images identified correctly (%)ZNCC
Actual treatment images1000767 (76.7)0.763 ± 0.136 *
BSIpatient 1000800 (80.0)0.773 ± 0.143 *
* p < 0.05 significantly different compared with the reference.

4. DISCUSSION

This is the first report that tried to incorporate CycleGAN into a CK motion‐tracking algorithm, and the promising result is expected to gain worldwide recognition. We propose the use of CycleGAN to learn to suppress the intensity of bony structures on TLS radiographs. Experimental evaluations on the XCAT phantom datasets validated the effectiveness of the framework developed to do this.

Using existing methods, tumors are difficult to detect because the visibility of the tumor in the live image is reduced when the tumor overlaps bony structures, and the image quality is degraded. In this study, deep learning and evaluation were performed using a dataset created from the XCAT phantom to generate bone‐suppression images for the XLTS. This method holds the promise of achieving a high level of accuracy through the training of the model using images with and without bone. However, in this study, actual treatment images were run through CycleGAN to generate images without bone, and there were cases in which bone intensity reduction was difficult. This may be due to the possibility that the images were not recognized as bone, in which case, improved datasets may be required. As a measure to create data sets that suppress bone intensity, image processing techniques such as dual energy subtraction have been used recently in order to improve the detection of pulmonary nodules. 25 In a study examining the detection of thoracic masses, the use of soft‐tissue images significantly improved the detection of masses. 26 In addition, a model was constructed to generate soft tissue chest images using GAN, and comparison with test data showed that the generated images had a high degree of similarity, indicating their clinical usefulness. 27

In addition, there was a discrepancy between the location of the tumor and that indicated by template matching. Because template matching searches for the location where the similarity between the input image and template is maximized, it is difficult to accommodate changes in the appearance of the target, such as rotation. To solve this problem, variable template matching, in which the template is updated when tracking is successful, has been proposed, but there is an accumulation of errors due to background contamination in the template and difficulty in determining the success or failure itself. 28 , 29 , 30

As a limitation, XLTS is used uncertainty values to evaluate detection uncertainty, but the algorithm varies from version to version and the details are not publicly available. Therefore, this was addressed in this study by calculating the similarity using ZNCC.

For images in which tumors could not be recognized using this method, the reason for the failure to recognize the tumors was considered to be poor background removal processing. If the tumor to be recognized is complex and contains considerable noise, noise removal cannot be performed well. Furthermore, recognition becomes impossible owing to the inclusion of noise and the tumor.

Therefore, our future studies will investigate a new recognition method that utilizes new features such as noise reduction and contrast.

5. CONCLUSION

In this study, we generated and evaluated CycleGAN images using a dataset created with the XCAT phantom to generate bone suppression images for the XLTS. The proposed bone suppression imaging technique based on CycleGAN improves the image similarity in template matching, making it possible to achieve highly accurate markerless motion tracking irradiation. The possibilities for future studies are included evaluation experiments of the constructed algorithms and their implementation in phantoms on simulated tumors for clinical applications.

AUTHOR CONTRIBUTIONS

The authors confirm contribution to the paper as follows: study conception and design: Zennosuke Mochizuki, MS, Masahide Saito, PhD; data collection: Zennosuke Mochizuki, MS; analysis and interpretation of results: Zennosuke Mochizuki, MS, Masahide Saito, PhD; draft manuscript preparation: Zennosuke Mochizuki, MS, Masahide Saito, PhD. All authors reviewed the results and approved the final version of the manuscript.

CONFLICT OF INTEREST STATEMENT

The authors declare no potential conflicts of interest.

ACKNOWLEDGEMENTS

I would like to thank my sincere gratitude to the people described here for their expertise and support in all aspects of the research and for their help in the preparation of the manuscript.

Notes

Mochizuki Z, Saito M, Suzuki T, et al. Cycle‐generative adversarial network‐based bone suppression imaging for highly accurate markerless motion tracking of lung tumors for cyberknife irradiation therapy. J Appl Clin Med Phys. 2024;25:e14212. 10.1002/acm2.14212 [Europe PMC free article] [Abstract] [CrossRef] [Google Scholar]

REFERENCES

1. Grills IS, Hugo G, Kestin LL, et al. Image‐guided radiotherapy via daily online cone‐beam CT substantially reduces margin requirements for stereotactic lung radiotherapy. Int J Radiat Oncol Biol Phys. 2008;70(4):1045‐1056. [Abstract] [Google Scholar]
2. Nuyttens JJ, van de Pol M. The CyberKnife radiosurgery system for lung cancer. Expert Rev Med Devices. 2012;9(5):465‐475. 10.1586/erd.12.35 [Abstract] [CrossRef] [Google Scholar]
3. Dufournaud Y, Schmid C, Horaud R. Image matching with scale adjustment image matching with scale adjustment. Comput Vis Image Underst. 2004;93(2):175‐194. [Google Scholar]
4. Kilby W, Dooley J, Kuduvalli G, Sayeh S, Maurer Jr C. The CyberKnife® robotic radiosurgery system in 2010. Technol Cancer Res Treat. 2010;9(5):433‐452. [Abstract] [Google Scholar]
5. Casutt A, Kinj R, Ozsahin E‐M, von Garnier C, Lovis A. Fiducial markers for stereotactic lung radiation therapy: review of the transthoracic, endovascular and endobronchial approaches. Eur Respir Rev. 2022;31(163):210149. [Europe PMC free article] [Abstract] [Google Scholar]
6. Akino Y, Shiomi H, Sumida I, et al. Impacts of respiratory phase shifts on motion‐tracking accuracy of the CyberKnife Synchrony™ Respiratory Tracking System. Med Phys. 2019;46(9):3757‐3766. [Abstract] [Google Scholar]
7. Adler Jr JR, Murphy MJ, Chang SD, Hancock SL. Image‐guided robotic radiosurgery. Neurosurgery. 1999;44(6):1299‐1306. [Abstract] [Google Scholar]
8. Chang SD, Main W, Martin DP, Gibbs IC, Heilbrun MP. An analysis of the accuracy of the CyberKnife: a robotic frameless stereotactic radiosurgical system. Neurosurgery. 2003;52(1):140‐147. [Abstract] [Google Scholar]
9. Tang H, Liu H, Xu D, Torr PH, Sebe N. Attentiongan: unpaired image‐to‐image translation using attention‐guided generative adversarial networks. IEEE Trans Neural Netw Learn Syst. 2021;34(4):1972‐1987. [Abstract] [Google Scholar]
10. Zhu J‐Y, Park T, Isola P, Efros AA. Unpaired image‐to‐image translation using cycle‐consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision. 2017.
11. Liang X, Chen L, Nguyen D, et al. Generating synthesized computed tomography (CT) from cone‐beam computed tomography (CBCT) using CycleGAN for adaptive radiation therapy. Phys Med Biol. 2019;64(12):125002. [Abstract] [Google Scholar]
12. Segars WP, Sturgeon G, Mendonca S, Grimes J, Tsui BM. 4D XCAT phantom for multimodality imaging research. Med Phys. 2010;37(9):4902‐4915. [Europe PMC free article] [Abstract] [Google Scholar]
13. Segars W, Bond J, Frush J, et al. Population of anatomically variable 4D XCAT adult phantoms for imaging research and optimization. Med Phys. 2013;40(4):043701. [Europe PMC free article] [Abstract] [Google Scholar]
14. Segars WP. Development and application of the new dynamic NURBS‐based cardiac‐torso (NCAT) phantom . The University of North Carolina at Chapel Hill; 2001. [Google Scholar]
15. Segars W, Mahesh M, Beck T, Frey E, Tsui B. Validation of the 4D NCAT simulation tools for use in high‐resolution x‐ray CT research. SPIE. 2005;5745:828‐834. [Google Scholar]
16. Segars WP, Tsui BM, Frey E, Fishman E. Extension of the 4D NCAT Phantom to Dynamic x‐ray CT Simulation. IEEE; 2003:3195‐3199. [Google Scholar]
17. Isola P, Zhu J‐Y, Zhou T, Efros AA. Image‐to‐image translation with conditional adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2017.
18. He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2016.
19. Kingma DP, Ba JL. Adam: A method for stochastic optimization. arXiv preprint arXiv:14126980. 2014.
20. Wang Z, Bovik AC, Sheikh HR, Simoncelli EP. Image quality assessment: from error visibility to structural similarity. IEEE Trans Image Process. 2004;13(4):600‐612. [Abstract] [Google Scholar]
21. Seshadrinathan K, Soundararajan R, Bovik AC, Cormack LK. Study of subjective and objective quality assessment of video. IEEE Trans Image Process. 2010;19(6):1427‐1441. [Abstract] [Google Scholar]
22. Rosenfeld A, Kak AC. Digital Picture Processing. Elsevier Science; 2014. [Google Scholar]
23. Fieguth P, Terzopoulos D. Color‐based Tracking of Heads and Other Mobile Objects at Video Frame Rates. IEEE; 1997:21‐27. [Google Scholar]
24. Lewis JP. Fast template matching. in Proc. Vis Interface. 1995;120‐123. [Google Scholar]
25. Oda S, Awai K, Funama Y, et al. Detection of small pulmonary nodules on chest radiographs: efficacy of dual‐energy subtraction technique using flat‐panel detector chest radiography. Clin Radiol. 2010;65(8):609‐615. [Abstract] [Google Scholar]
26. Li F, Engelmann R, L Pesce L, Doi K, Metz CE, MacMahon H. Small lung cancers: improved detection by use of bone suppression imaging—comparison with dual‐energy subtraction chest radiography. Radiology. 2011;261(3):937‐949. [Europe PMC free article] [Abstract] [Google Scholar]
27. Zhou B, Lin X, Eck B, Hou J, Wilson D. Generation of Virtual Dual Energy Images from Standard Single‐Shot Radiographs Using Multi‐Scale and Conditional Adversarial Network. Springer; 2018:298‐313. [Google Scholar]
28. Matthews L, Ishikawa T, Baker S. The template update problem. IEEE Trans Pattern Anal Mach Intell. 2004;26(6):810‐815. [Abstract] [Google Scholar]
29. Cahn von Seelen UM, Bajcsy R. Adaptive correlation tracking of targets with changing scale. Technical Reports (CIS). 1996:p. 234. [Google Scholar]
30. Han B, Davis L. On‐line Density‐Based Appearance Modeling for Object Tracking. IEEE; 2005:1492‐1499. [Google Scholar]

Articles from Journal of Applied Clinical Medical Physics are provided here courtesy of Wiley

Citations & impact 


Similar Articles 


To arrive at the top five similar articles we use a word-weighted algorithm to compare words from the Title and Abstract of each citation.