A transformer-based patient-agnostic tumor motion tracking model (TransTracking): A feasibility study in radiotherapy for lung cancer.
OpenAlex 토픽 ·
Advanced Radiotherapy Techniques
Lung Cancer Diagnosis and Treatment
Augmented Reality Applications
[BACKGROUND] Accurate and real-time localization of thoracic tumor targets is essential for effective radiation therapy.
APA
Maidina Abuduxiku, Ningning Chai, et al. (2026). A transformer-based patient-agnostic tumor motion tracking model (TransTracking): A feasibility study in radiotherapy for lung cancer.. Medical physics, 53(5), e70449. https://doi.org/10.1002/mp.70449
MLA
Maidina Abuduxiku, et al.. "A transformer-based patient-agnostic tumor motion tracking model (TransTracking): A feasibility study in radiotherapy for lung cancer.." Medical physics, vol. 53, no. 5, 2026, pp. e70449.
PMID
42027140
DOI
10.1002/mp.70449
Abstract
[BACKGROUND] Accurate and real-time localization of thoracic tumor targets is essential for effective radiation therapy. Recently, Transformer architectures have demonstrated strong global reasoning capabilities across multiple frames by leveraging both self-attention and cross-attention mechanisms. Transformers have therefore been applied to object tracking with great success. By combining Image Guided Radiation Therapy (IGRT) technologies and deep learning-based object tracking architecture, it is possible to deliver radiation doses to the target area with high accuracy.
[PURPOSE] This study develops a transformer-based patient-agnostic tracking model (TransTracking) for surface and markerless internal target tracking in thoracic tumor radiotherapy.
[METHODS] We trained the TransTracking model using the training splits of publicly available object tracking datasets. Subsequently, for internal target tracking, the model is fine-tuned using 10,000 digitally reconstructed radiograph (DRR) images generated from the actual 4DCT datasets of 25 patients. The DRR images are annotated with bounding boxes of the moving tumor. Our method learns to directly predict the target classification and bounding-box regression weights through end-to-end training, enabling accurate target localization in each frame for both surface and internal target tracking sequences. The tracking performance of the trained model was evaluated in 20 volunteers for surface tracking and using DRR images generated from 20 4DCT datasets for internal tumor tracking. To address the limited availability of medical images for training, we conducted the data augmentation procedure to 4DCT datasets and expanded the data scale 40-fold in total.
[RESULTS] For the surface marker tracking, the mean absolute deviation (MAD) ± standard deviation (SD) between the model-predicted and the actual positions for 20 volunteers was 0.07 ± 0.06 mm, 0.12 ± 0.13 mm, and 0.29 ± 0.20 mm in left-right, superior-inferior, and anterior-posterior directions, respectively. In each directional axis, over 85% of frames exhibited a model-predicted target position within 0.5 mm of the corresponding ground-truth position. For the internal tumor tracking, the MAD ± SD between the predicted and annotated center positions of the tumor bounding boxes is 1.49 ± 1.39 mm, with a mean Intersection over Union (IoU) value of 0.83 and an area under curve (AUC) score of 82% for 20 patients. Additionally, our transformer-based model can extract the target position in 81 ms after an image is acquired.
[CONCLUSIONS] This study proposed a novel Transformer-based deep learning method aimed at training a patient-agnostic tumor motion tracking model in radiotherapy. The model enables real-time, high-precision tracking of surface markers using vision cameras, offering a cost-effective and compact solution. Additionally, we demonstrated that our method can accurately locate tumor target areas in DRR images with high precision, without the need for individualized training or the implantation of fiducial markers. This feasibility study demonstrates the strong potential of our strategy as a clinically viable solution for moving tumor IGRT.
[PURPOSE] This study develops a transformer-based patient-agnostic tracking model (TransTracking) for surface and markerless internal target tracking in thoracic tumor radiotherapy.
[METHODS] We trained the TransTracking model using the training splits of publicly available object tracking datasets. Subsequently, for internal target tracking, the model is fine-tuned using 10,000 digitally reconstructed radiograph (DRR) images generated from the actual 4DCT datasets of 25 patients. The DRR images are annotated with bounding boxes of the moving tumor. Our method learns to directly predict the target classification and bounding-box regression weights through end-to-end training, enabling accurate target localization in each frame for both surface and internal target tracking sequences. The tracking performance of the trained model was evaluated in 20 volunteers for surface tracking and using DRR images generated from 20 4DCT datasets for internal tumor tracking. To address the limited availability of medical images for training, we conducted the data augmentation procedure to 4DCT datasets and expanded the data scale 40-fold in total.
[RESULTS] For the surface marker tracking, the mean absolute deviation (MAD) ± standard deviation (SD) between the model-predicted and the actual positions for 20 volunteers was 0.07 ± 0.06 mm, 0.12 ± 0.13 mm, and 0.29 ± 0.20 mm in left-right, superior-inferior, and anterior-posterior directions, respectively. In each directional axis, over 85% of frames exhibited a model-predicted target position within 0.5 mm of the corresponding ground-truth position. For the internal tumor tracking, the MAD ± SD between the predicted and annotated center positions of the tumor bounding boxes is 1.49 ± 1.39 mm, with a mean Intersection over Union (IoU) value of 0.83 and an area under curve (AUC) score of 82% for 20 patients. Additionally, our transformer-based model can extract the target position in 81 ms after an image is acquired.
[CONCLUSIONS] This study proposed a novel Transformer-based deep learning method aimed at training a patient-agnostic tumor motion tracking model in radiotherapy. The model enables real-time, high-precision tracking of surface markers using vision cameras, offering a cost-effective and compact solution. Additionally, we demonstrated that our method can accurately locate tumor target areas in DRR images with high precision, without the need for individualized training or the implantation of fiducial markers. This feasibility study demonstrates the strong potential of our strategy as a clinically viable solution for moving tumor IGRT.
MeSH Terms
Humans; Feasibility Studies; Lung Neoplasms; Radiotherapy, Image-Guided; Movement; Four-Dimensional Computed Tomography; Deep Learning