- Research
- Open access
- Published:
Prior information guided auto-segmentation of clinical target volume of tumor bed in postoperative breast cancer radiotherapy
Radiation Oncology volumeĀ 18, ArticleĀ number:Ā 170 (2023)
Abstract
Background
Accurate delineation of clinical target volume of tumor bed (CTV-TB) is important but it is also challenging due to surgical effects and soft tissue contrast. Recently a few auto-segmentation methods were developed to improve the process. However, those methods had comparatively low segmentation accuracy. In this study the prior information was introduced to aid auto-segmentation of CTV-TB based on a deep-learning model.
Methods
To aid the delineation of CTV-TB, the tumor contour on preoperative CT was transformed onto postoperative CT via deformable image registration. Both original and transformed tumor contours were used for prior information in training an auto-segmentation model. Then, the CTV-TB contour on postoperative CT was predicted by the model. 110 pairs of preoperative and postoperative CT images were used with a 5-fold cross-validation strategy. The predicted contour was compared with the clinically approved contour for accuracy evaluation using dice similarity coefficient (DSC) and Hausdorff distance.
Results
The average DSC of the deep-learning model with prior information was improved than the one without prior information (0.808 vs. 0.734, Pā<ā0.05). The average DSC of the deep-learning model with prior information was higher than that of the traditional method (0.808 vs. 0.622, Pā<ā0.05).
Conclusions
The introduction of prior information in deep-learning model can improve segmentation accuracy of CTV-TB. The proposed method provided an effective way to automatically delineate CTV-TB in postoperative breast cancer radiotherapy.
Introduction
Breast cancer has become the most frequently diagnosed cancer. In 2020, there were 2.26 million women diagnosed with breast cancer and 0.68 million deaths globally [1]. Breast-conserving surgery followed by postoperative radiotherapy has become the established treatment procedure for early-stage breast cancer patients [2]. For postoperative breast cancer radiotherapy, it is important to accurately delineate the tumor bed and its target volume. However, the target volume delineation is susceptible to the number of surgical clips, clarity and size of seroma, inter-observer variability and other factors [3]. And because of asymmetric excision of the tumor during surgery, the uniform expansion of the resection cavity may not be proper to represent clinical target volume of tumor bed (CTV-TB) [4]. In general, the target volume of breast tumor bed is manually delineated by radiation oncologists in current clinical practice. However, as mentioned above, manual delineation is affected by many factors. Overall it is time-consuming and labor intensive. And there exists obvious inter-observer variability [5].
Given the intrinsic characteristics, there are several difficulties in segmenting tumor bed and its target volume for postoperative breast cancer radiotherapy. First, the contrast of soft tissue on CT image is relatively low. And the high-density marker (lead wires and titanium clips) would possibly cause metal artifacts, which compromised the image quality to some extent. Second, the contrast between tumor bed and surrounding normal breast tissue is low. The density within the region of tumor bed is close to soft tissue on the whole. Third, the size, shape and location of tumor bed varied considerably from patient to patient.
Deep learning models were popularly used in automatic segmentation of medical image [6, 7]. For postoperative breast cancer radiotherapy, there have been several researches in auto-segmenting whole breast CTV and organs at risk (OARs) [8,9,10]. However, there are few models developed for segmenting tumor bed and its CTV-TB due to its intrinsic complexity. Dai et al. employed a 3D U-Net to segment tumor bed, whole breast CTV and several organs at risk (OAR) on planning CT and CBCT-generated synthetic CT [11]. The results showed that the tumor bed on synthetic CT was obviously larger than the one manually contoured by physicians. The DSC (0.63āĀ±ā0.08) was lower compared to those achieved in general medical image segmentation tasks. Kazemimoghadam proposed a saliency-based deep learning method for segmenting tumor bed [12]. It incorporated the salient information provided by titanium clip into the deep-learning model. The DSC (0.76āĀ±ā0.03) was slightly better than that of Daiās method but still lower in general.
Motivated by Kazemimoghadamās method which encoded locations of titanium clips and salient regions in the deep-learning model, we proposed a method to incorporate tumor location information into the deep-learning model for segmenting CTV-TB on postoperative CT. The tumor contour on preoperative CT and its transformed contour on postoperative CT both provided prior information in searching for the potential location of CTV-TB. The rest of paper was organized as follows. In methods section, the delineation of CTV-TB and generation of prior information were first introduced. Then, the scheme of model learning and predicting was explained in detail. In results section, the effect of prior information and the model performance were evaluated. Finally, the advantages and disadvantages of the proposed method were discussed, and the future work was prospected.
Methods
Patient dataset
110 left-sided breast cancer patient undergone breast-conserving surgery (BCS) and eligible for whole breast irradiation (WBI) plus boost irradiation were enrolled in this study. The median age of patients was 50Ā years (range, 44ā59Ā years), and the pathological diagnosis was all invasive ductal carcinoma with a stage of T1-T2N0M0. No patient received oncoplastic surgery. All patients underwent a lumpectomy with sentinel lymph node dissection. Tumor-negative margins were ensured during a single operation. Equal or more than 5 surgical clips were used to mark the boundaries of the lumpectomy cavity. All enrolled patients had either no seroma or a seroma clarity score ofāā¤ā3 in the lumpectomy cavity. This study was approved by the Institutional Ethics Committee of Cancer Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College. Consent was waived due to the retrospective nature of the study.
The patient dataset consisted of 110 pairs of preoperative and postoperative CTs, which were acquired in the supine position. The preoperative CT was acquired averagely one week before surgery. They were reconstructed with dimensions of 512āĆā512, slice thickness of 5.0Ā mm, and pixel size of 0.68ā0.94Ā mm. The postoperative CT was acquired averagely 10Ā weeks after surgery and used for the purpose of radiotherapy treatment planning. They were reconstructed with dimensions of 512āĆā512, slice thickness of 5.0Ā mm, and pixel size of 1.18ā1.37Ā mm. All CTs were pre-processed using 3D Slicer (RRID:SCR_005619) [13, 14]. They were first resampled to an isotropic resolution of 1āĆā1āĆā5Ā mm and then cropped to dimensions of 256āĆā256āĆā32 around the breast's centroid [15].
Contour delineation
The distribution of multiple regions of interest (ROIs) on preoperative and postoperative CTs is illustrated in Fig.Ā 1. Before surgery, patient was CT scanned for diagnostic purpose and the location of primary tumor (PT) was manually delineated by radiation oncologist. After surgery, the volume of actually excised tissue (pathological volume, PV) was estimated by its maximum diameter in three dimensions (provided in pathological report). And the excision volume (EV) on preoperative CT image is estimated by adding a given margin to PT as shown in Fig.Ā 1A. Different margin was tested from 1 to 3Ā cm. And it was found that the volume of PT with 2Ā cm margin is closest to the volume of PV. Thus, 2Ā cm margin to PT on preoperative CT image was used to represent excision volume (EV).
In a few weeks after surgery, patient was CT scanned again and proceeded to radiotherapy. As shown in Fig.Ā 1B, clinical target volume of tumor bed (CTV-TB) is generated by adding 1Ā cm margin to the contour of tumor bed (TB). The margin was used to account for the subclinical lesions and potential invaded regions. In practice, the contour of TB was manually delineated by radiation oncologist according to the surgical marks and postoperative changes. Due to the poor clarity of lumpectomy cavity and relatively low soft tissue contrast, TB contouring is difficult and challenging.
Prior information
As the location of TB is at the same place of EV before surgery, the TB contour on postoperative CT would highly correlates with the EV contour on preoperative CT. Accordingly, the TB contour plus 1Ā cm margin (TB1cm), i.e. CTV-TB, on postoperative CT would highly correlate with the EV contour plus 1Ā cm margin (EV1cm) on preoperative CT. Therefore, it would be reasonable to create virtual EV1cm on postoperative CT, and used it as prior location information in searching for CTV-TB contour on postoperative CT. For reaching this goal, the deformable image registration (DIR) between preoperative and postoperative CTs was performed on Elastix (RRID:SCR_009619) [16, 17]. As a result, the deformation vector field (DVF) was obtained and used to generate the transformed EV1cm (T-EV1cm) on postoperative CT from the EV1cm on preoperative CT.
To enhance the effect of tumor contour on CTs, the regions of EV1cm and T-EV1cm were processed via image enhancement tool. In detail, the pixel values within these ROIs were multiplied by an integer number such as 25, while the pixel values outside them was multiplied by a fraction number such as 0.1. The effect of CT images before and after image enhancement is shown in Fig.Ā 2. The preoperative and postoperative CTs before image enhancement are shown in Fig.Ā 2A, B, while the preoperative and postoperative CTs after image enhancement are shown in Fig.Ā 2C, D. Clearly, the intensities of tumor contours on CTs were significantly enlarged comparing with those of the surrounding tissue.
Deep-learning model
A 3D U-Net used to solve many segmentation problems was employed in this study [18,19,20]. The detail of network architecture and setting was described in Additional file 1. In brief, it has an encoder part to analyze the whole image and a decoder part to produce full resolution segmentation. 3D U-Net takes 3D volume as inputs and applies 3D convolution, 3D max-pooling and 3D up-convolutional layers which has an entirely 3D architecture. In this study, there were two 3D input channels (enhanced preoperative and postoperative CT) and one 3D output channel (predicted label) in the deep learning model. A five-fold cross-validation was applied to the 110 patient dataset. One fold (22 patients) was used for testing, and the remaining four folds (88 patients) were used for training.
The weights of convolution layers are initialized by a normal distribution according to the published studies [18, 19]. The Dice similarity coefficient (DSC) was used as the loss function [19]. The Adaptive moment estimation (Adam) with batch size of 4 and weight decay of 3eā5 was used for optimization [21]. The initial learning rate was set as 0.0005, the learning rate drop factor as 0.95, and the validation frequency as 20. The network was implemented with Matlab (version 2020a) (MathWorks, Natick, MA 01760) and trained with maximal 50 epochs. The test was performed on a workstation equipped with one NVIDIA Geforce GTX 1080 TI GPU.
Auto-segmentation of CTV-TB
The overall workflow for segmenting CTV-TB on postoperative CT is shown in Fig.Ā 3 and the main steps are labeled by numbers. (1) Both preoperative and postoperative CTs were registered by DIR. As a result, the DVF was obtained. (2) T-EV1cm on postoperative CT was generated by deforming EV1cm on preoperative CT via the obtained DVF. (3) Both EV1cm and T-EV1cm were processed by image enhancement tool and the resulting 3D images were fed into the deep-learning model. (4) The CTV-TB contour on postoperative CT was predicted by the deep learning model. (5) The similarity between the predicted and clinically approved CTV-TB contours was evaluated.
Evaluations
The DSC and Hausdorff distance (HD) were used to evaluate the similarity between the predicted and clinically approved contours of CTV-TB on postoperative CT. The DSC is defined as follows [22]:
where A is the clinically approved CTV-TB contour manually delineated by the radiation oncologist and B is the predicted CTV-TB contour by the model. Aāā©āB is the volume that A and B have in common. The DSC results in values between 0 and 1, where 0 represents no intersection and 1 reflects perfect overlap. The HD is defined as [23]:
where
and \(\left\| \cdot \right\|\) is some underlying norm on the points of A and B (e.g., the L2 or Euclidean norm). \({\text{h}}\left( {{\text{A}},{\text{B}}} \right)\) identify the point a \(\in\) A that is farthest from any point of B and measures the distance from a to its nearest neighbor in B. The Hausdorff distance \({\text{HD}}\left( {{\text{A}},{\text{B}}} \right)\) is the maximum of \({\text{h}}\left( {{\text{A}},{\text{B}}} \right)\) and \({\text{h}}\left( {{\text{B}},{\text{A}}} \right)\) and measures the largest degree of mismatch between A and B. The overlap between A and B increases with smaller \({\text{HD}}\left( {{\text{A}},{\text{B}}} \right)\).
The performance of the deep-learning model with prior information (using Fig.Ā 2C, D as inputs) was compared with the same 3D U-Net without prior information (using Fig.Ā 2B, C as inputs). This will investigate the effect of prior information on segmentation accuracy of the deep-learning model. Five-fold cross-validation was used to tune the hyperparameters and the testing data were used to evaluate the performance of the final models. In addition, the performance of the traditional gray-level threshold method was also investigated. The gray-level threshold method partitions the gray levels in an image into two classes: those below a user-defined threshold and those above. In our study, CT values above threshold (40 HU) within the breast region were auto-segmented as CTV-TB contour. For statistical analysis, the paired t-test was performed if the data were normally distributed. Otherwise, the Wilcoxon Signed-Rank Test for Paired Samples (non-parametric test) was performed. A level of Pā<ā0.05 was considered statistically significant. All statistical analyses were performed in R Project for Statistical Computing (RRID:SCR_001905) (version 3.6.3).
Results
The training time for 3D U-Net was approximately 30Ā h, while the prediction time was 20Ā s per patient. In this binary segmentation, each pixel is labeled as CTV-TB or non-CTV-TB. The average values of DSC (meanāĀ±āstandard deviation) were 0.808āĀ±ā0.065 and 0.734āĀ±ā0.085 for the deep-learning models with and without prior information. Comparatively, the average value of DSC (meanāĀ±āstandard deviation) was 0.622āĀ±ā0.090 for the traditional gray-level threshold method. For fair comparison, the average values of HD (meanāĀ±āstandard deviation) were 19.254āĀ±ā6.012 and 47.975āĀ±ā22.214 for the deep-learning models with and without prior information. Comparatively, the average value of HD (meanāĀ±āstandard deviation) was 60.512āĀ±ā28.145 for the traditional gray-level threshold method.
The results of ShapiroāWilk normality test confirmed that the data were normally distributed, so the paired t-test was used. The difference of average DSC and HD resulting from the deep-learning models with and without prior information was statistically significant (0.808 vs. 0.734, Pā=ā0.0014ā<ā0.05; 19.254Ā vs. 47.975, Pā=ā0.002ā<ā0.05). Besides, the difference between the deep-learning models with prior information and the traditional gray-level threshold method were statistically significant (0.808Ā vs. 0.622, Pā=ā0.0005ā<ā0.05; 19.254 vs. 60.512, Pā=ā0.001ā<ā0.05).
As shown in Fig.Ā 4, the segmentation results of prior information guided deep-learning model are displayed in three orthogonal views. The predicted labels of CTV-TB were overlaid on postoperative CT images with ground-truth (clinically approved) labels, where the predicted labels in white and the ground truths in black. It showed that the majority of both contours were similar. The predicted contour had smoother boundary than the contour which was manually delineated by radiation oncologist.
Discussion
In this study, a prior information guided deep-learning model was developed to automatically segment CTV-TB from postoperative CT. The results showed that the introduction of prior information succeeded in identifying low-contrast CTV-TB from surrounding normal tissue on postoperative CT. This improvement would be attributed to the introduction of EV1cm on preoperative CT and T-EV1cm on postoperative CT, which bring prior information about the approximate CTV-TB contour and set a good starting point for the deep learning model. Besides, the DSC and HD of the transformed T-EV1cm contour and clinically approved CTV-TB contour were 0.551āĀ±ā0.110 and 49.875āĀ±ā23.514. It is worth noting that intensity-based DIR method is challenged due to the large deformations and non-correspondence caused by tumor resection and clip insertion. Thus the introduction of T-EV1cm on postoperative CT just provide the approximate location of CTV-TB contour.
3D U-Net was previously used in segmenting tumor bed on CBCT-generated synthetic CT and the DSC was lower [11]. Later, Kazemimoghadam incorporated the salient information provided by titanium clip into the U-Net model for tumor bed segmentation. The DSC was improved but limited [12]. In our study the DSC was further improved. The improved DSC would be attributed to the introduction of prior information, which limited the searching range for potential CTV-TB contour on postoperative CT. It should be noted that both Kazemimoghadamās and our methods incorporated prior information in segmentation model to aid the searching of final target.
There are certain limitations of this study. First, the training set is small, which requires massive cross-validation to ensure the stability of learning model. More data will be collected in the future to make the model more robust. Second, the intensity-based DIR was used to generate DVF for transforming tumor contour onto postoperative CT. More advanced DIR methods would be investigated and adopted for future study. Third, only CT image was used for the input of deep learning model. It would be more interesting to include other image modalities such as CBCT, Ultrasound and MRI. With these inputs the segmentation accuracy of the deep learning model would be further improved. Fourth, manual contouring on preoperative CT image was still needed. In the future, auto-segmentation method would be investigated to further improve efficiency.
Conclusions
Incorporating prior information of tumor location into deep learning model improved the segmentation accuracy of CTV-TB contour on postoperative CT. The tumor contours on both preoperative and postoperative CT provided the approximate CTV-TB contour, which facilitated the subsequent searching by the deep learning model. The proposed method demonstrated an effective way in auto-segmentation of CTV-TB in postoperative breast cancer radiotherapy.
Availability of data and materials
The datasets used and/or analysed during the current study are available from the corresponding author on reasonable request.
Abbreviations
- CTV-TB:
-
Clinical target volume of tumor bed
- TB:
-
Tumor bed
- CT:
-
Computed tomography
- DIR:
-
Deformable image registration
- DSC:
-
Dice similarity coefficient
- HD:
-
Hausdorff distance
- OARs:
-
Organs at risk
- BCS:
-
Breast-conserving surgery
- WBI:
-
Whole breast irradiation
- ROIs:
-
Regions of interest
- PT:
-
Primary tumor
- PV:
-
Pathological volume
- EV:
-
Excision volume
- DVF:
-
Deformation vector field
- Adam:
-
Adaptive moment estimation
References
Sung H, Ferlay J, Siegel RL, Laversanne M, Soerjomataram I, Jemal A, et al. Global cancer statistics 2020: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA A Cancer J Clin. 2021;71(3):209ā49.
Litiere S, Werutsky G, Fentiman IS, Rutgers E, Christiaens MR, Van Limbergen E, et al. Breast conserving therapy versus mastectomy for stage I-II breast cancer: 20 year follow-up of the EORTC 10801 phase 3 randomised trial. Lancet Oncol. 2012;13(4):412ā9.
van Mourik AM, Elkhuizen PH, Minkema D, Duppen JC, van Vliet-Vroegindeweij C. Multiinstitutional study on target volume delineation variation in breast radiotherapy in the presence of guidelines. Radiother Oncol J Eur Soc Therap Radiol Oncol. 2010;94(3):286ā91.
Major T, Gutierrez C, Guix B, van Limbergen E, Strnad V, Polgar C. Recommendations from GEC ESTRO breast cancer working group (II): target definition and target delineation for accelerated or boost partial breast irradiation using multicatheter interstitial brachytherapy after breast conserving open cavity surgery. Radiother Oncol J Eur Soc Therap Radiol Oncol. 2016;118(1):199ā204.
Landis DM, Luo W, Song J, Bellon JR, Punglia RS, Wong JS, et al. Variability among breast radiation oncologists in delineation of the postsurgical lumpectomy cavity. Int J Radiat Oncol Biol Phys. 2007;67(5):1299ā308.
Shelhamer E, Long J, Darrell T. Fully convolutional networks for semantic segmentation. IEEE Trans Pattern Anal Mach Intell. 2017;39(4):640ā51.
Badrinarayanan V, Kendall A, Cipolla R. SegNet: a deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans Pattern Anal Mach Intell. 2017;39(12):2481ā95.
Men K, Zhang T, Chen X, Chen B, Tang Y, Wang S, et al. Fully automatic and robust segmentation of the clinical target volume for radiotherapy of breast cancer using big data and deep learning. Phys Med PM Int J Devot Appl Phys Med Biol Offic J Italian Assoc Biomed Phys. 2018;50:13ā9.
Choi MS, Choi BS, Chung SY, Kim N, Chun J, Kim YB, et al. Clinical evaluation of atlas- and deep learning-based automatic segmentation of multiple organs and clinical target volumes for breast cancer. Radiotherapy Oncol J Eur Soc Therap Radiol Oncol. 2020;153:139ā45.
Chung SY, Chang JS, Choi MS, Chang Y, Choi BS, Chun J, et al. Clinical feasibility of deep learning-based auto-segmentation of target volumes and organs-at-risk in breast cancer patients after breast-conserving surgery. Radiat Oncol. 2021;16(1):44.
Dai Z, Zhang Y, Zhu L, Tan J, Yang G, Zhang B, et al. Geometric and dosimetric evaluation of deep learning-based automatic delineation on CBCT-synthesized CT and planning CT for breast cancer adaptive radiotherapy: a multi-institutional study. Front Oncol. 2021;11:725507.
Kazemimoghadam M, Chi W, Rahimi A, Kim N, Alluri P, Nwachukwu C, et al. Saliency-guided deep learning network for automatic tumor bed volume delineation in post-operative breast irradiation. Phys Med Biol. 2021;66(17):175019.
Fedorov A, Beichel R, Kalpathy-Cramer J, Finet J, Fillion-Robin JC, Pujol S, et al. 3D Slicer as an image computing platform for the Quantitative Imaging Network. Magn Reson Imaging. 2012;30(9):1323ā41.
Pinter C, Lasso A, Wang A, Jaffray D, Fichtinger G. SlicerRT: radiation therapy research toolkit for 3D Slicer. Med Phys. 2012;39(10):6332ā8.
Xie X, Song Y, Ye F, Yan H, Wang S, Zhao X, et al. Prior information guided auto-contouring of breast gland for deformable image registration in postoperative breast cancer radiotherapy. Quant Imaging Med Surg. 2021;11(12):4721ā30.
Klein S, Staring M, Murphy K, Viergever MA, Pluim JP. elastix: a toolbox for intensity-based medical image registration. IEEE Trans Med Imaging. 2010;29(1):196ā205.
Shamonin DP, Bron EE, Lelieveldt BP, Smits M, Klein S, Staring M. Fast parallel image registration on CPU and GPU for diagnostic classification of Alzheimerās disease. Front Neuroinform. 2013;7:50.
ĆiƧek Ć, Abdulkadir A, Lienkamp S, Brox T, Ronneberger O. 3D U-Net: learning dense volumetric segmentation from sparse annotation. 2016.
Ronneberger O, Fischer P, Brox T. U-Net: convolutional networks for biomedical image segmentation. 2015.
Szegedy C, Vanhoucke V, Ioffe S, Shlens J, Wojna Z. Rethinking the inception architecture for computer Vision2015 December 01, 2015:[arXiv:1512.00567]. Available from: https://ui.adsabs.harvard.edu/abs/2015arXiv151200567S.
Kingma DP, Ba J. Adam. A method for stochastic optimization. CoRR. 2015; https://arxiv.org/abs/1412.6980.
Dice LR. Measures of the amount of ecologic association between species. Ecology. 1945;26:297ā302.
Huttenlocher D, Klanderman G, Rucklidge W. Comparing images using the Hausdorff distance. IEEE Trans Pattern Anal Mach Intell. 1993;15:850ā63.
Acknowledgements
Not applicable.
Funding
This work was partially supported by the National Natural Science Foundation of China (11875320) and Beijing Hope Run Special Fund of Cancer Foundation of China (LC2018A08, LC2020L01).
Author information
Authors and Affiliations
Contributions
(I) Conception and design: XX; (II) Administrative support: XMZ, JRD; (III) Provision of study materials or patients: FY; (IV) Collection and assembly of data: YCS, SLW; (V) Data analysis and interpretation: XX, HY; (VI) Manuscript writing: All authors; (VII) Final approval of manuscript: All authors.
Corresponding authors
Ethics declarations
Ethics approval and consent to participate
This study was approved by the Institutional Ethics Committee of Cancer Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College. Consent was waived due to the retrospective nature of the study.
Consent for publication
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Additional file 1:
The detail of network architecture and setting.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.
About this article
Cite this article
Xie, X., Song, Y., Ye, F. et al. Prior information guided auto-segmentation of clinical target volume of tumor bed in postoperative breast cancer radiotherapy. Radiat Oncol 18, 170 (2023). https://doi.org/10.1186/s13014-023-02355-9
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s13014-023-02355-9