Literature DB >> 35305023

Robust automated radiation therapy treatment planning using scenario-specific dose prediction and robust dose mimicking.

Oskar Eriksson1, Tianfang Zhang1,2.   

Abstract

PURPOSE: We present a framework for robust automated treatment planning using machine learning, comprising scenario-specific dose prediction and robust dose mimicking.
METHODS: The scenario dose prediction pipeline is divided into the prediction of nominal dose from input image and the prediction of scenario dose from nominal dose, each using a deep learning model with U-net architecture. By using a specially developed dose-volume histogram-based loss function, the predicted scenario doses are ensured sufficient target coverage despite the possibility of the training data being non-robust. Deliverable plans may then be created by solving a robust dose mimicking problem with the predictions as scenario-specific reference doses.
RESULTS: Numerical experiments are performed using a data set of 52 intensity-modulated proton therapy plans for prostate patients. We show that the predicted scenario doses resemble their respective ground truth well, in particular while having target coverage comparable to that of the nominal scenario. The deliverable plans produced by the subsequent robust dose mimicking were showed to be robust against the same scenario set considered for prediction.
CONCLUSIONS: We demonstrate the feasibility and merits of the proposed methodology for incorporating robustness into automated treatment planning algorithms.
© 2022 The Authors. Medical Physics published by Wiley Periodicals LLC on behalf of American Association of Physicists in Medicine.

Entities:  

Keywords:  dose mimicking; knowledge-based planning; robust optimization; scenario dose prediction

Mesh:

Year:  2022        PMID: 35305023      PMCID: PMC9310773          DOI: 10.1002/mp.15622

Source DB:  PubMed          Journal:  Med Phys        ISSN: 0094-2405            Impact factor:   4.506


INTRODUCTION

Radiation therapy treatment planning is a time‐consuming process that typically requires multiple iterations between a dosimetrist and an oncologist. In recent years, automated treatment planning methods have been developed to speed up the process while ensuring a consistent quality of treatment plans, using historically delivered treatment plans to aid in the process of creating plans for new patients. A common approach is using a machine learning model to predict a reference dose distribution for each new patient, , , which is then used in an optimization problem to find a configuration for a treatment machine that would deliver a similar dose. These parts are commonly referred to as dose prediction and dose mimicking, respectively. However, in many cases, especially in proton therapy, it is important to take into account uncertainties in the treatment delivery such as patient setup and density calculations. While robust optimization is the current state‐of‐the‐art method of handling such uncertainties, optimizing on, for example, the near‐worst‐case over a set of sampled scenarios, current methods for automated planning in the literature have yet to be able to incorporate robustness. The contribution of this paper is a framework that unifies ideas from automated treatment planning and robust optimization by predicting a set of scenario reference doses, each corresponding to a specific scenario, to be used in a robust dose mimicking problem. Typically, for the non‐robust case, machine learning–based automated treatment planning is divided into two steps: predicting the achievable values of certain dose‐related quantities from the patient geometry, and finding machine parameters of a plan to reconstruct the same values. For example, the dose‐related quantities may be the spatial dose distribution, dose–volume histograms (DVHs), or a combination thereof. A common approach to spatial dose prediction is using a neural network, often with a U‐net architecture, to predict a dose value for each voxel in the discretized dose grid. , , , , , Likewise, for DVH prediction, one may either use the DVHs evaluated on a spatially predicted dose distribution or employ separate models for the purpose. , , , , , Using the reference dose and the reference DVHs, a dose mimicking optimization problem is then constructed, where the goal is to find a set of machine parameters for a specific treatment machine such that the delivered dose is as similar as possible to the reference dose and DVHs. , , Thus, given that the plans used for training the models are clinically acceptable and follow the desired protocol, the idea is that the automated planning pipeline should produce a high‐quality plan for each new patient. In robust optimization, reference dose levels are usually set for different regions of interest (ROIs), and the goal of the optimization is to find a resulting treatment plan that minimizes some cost functional that depends on the outcome in a number of fixed scenarios. These scenarios may be seen as samples from some probability distribution specifying the geometric uncertainties in the treatment delivery that we want to account for. Typical examples of such uncertainties include setup uncertainties, related to inaccuracies in the setup of the patient or the inaccuracy of the treatment machine, as well as range uncertainties, related to uncertainties in the CT imaging or the conversion from CT values to density values. For particle modalities such as protons, since the dose delivered is relatively sensitive to the density of the material through which they pass, using margins around the clinical target volume (CTV) is often insufficient to account for density uncertainty effects. Hence, robust optimization is especially important in such cases. In robust dose mimicking, one would use a reference dose corresponding to each scenario, with the goal of finding a robust treatment plan that is good across most or all of these scenarios. However, in a robust prediction–mimicking pipeline, how to choose these scenario reference doses is a matter previously unaddressed in the literature. In this paper, we present a method of performing robust automated treatment planning, combining spatial dose prediction and robust optimization through scenario dose prediction and robust dose mimicking. We propose to predict scenario doses using a two‐step pipeline: by first predicting the nominal dose using a U‐net model and subsequently deforming, using a second U‐net model, the nominal dose to scenario doses corresponding to a given set of scenarios. Specifically, starting from a non‐robustly planned data set, we propose to use a DVH‐based loss function when training the scenario model to ensure that the predicted scenario doses have comparable target coverage to that of the nominal dose. The predicted scenario doses are then used as scenario‐specific reference doses in a robust dose mimicking problem, creating a robust deliverable plan. Numerical experiments, designed for a proof‐of‐concept study, are performed on a data set of prostate cancer patients treated with intensity‐modulated proton therapy. We show that the proposed scenario dose prediction pipeline fulfills its purpose satisfactorily, with predictions mostly following the non‐robust ground truth but with increased target coverage, and that the resulting deliverable plans are robust with respect to the considered scenario set. Compared to manually generated benchmark plans, the automatically generated plans are similar in terms of DVH, dose statistic spread, and spatial dose. In particular, we demonstrate the feasibility of a data‐driven, robust automated treatment planning framework.

MATERIALS AND METHODS

Let and be spaces of patient geometries and dose distributions, respectively, and a given set of scenarios representing realizations of systematic setup or range uncertainties. For each , the patient geometry is defined as its CT image with ROI delineations along with its spatial location under s. A dose distribution that has been delivered to this patient previously is referred to as the scenario dose, denoted by . The nominal scenario, corresponding to no setup or range errors, is denoted by s 0. Given a data set of pairs of nominal patient geometries and doses, our proposed pipeline follows the classical prediction–mimicking division, with a scenario‐specific spatial dose prediction model followed by a robust dose mimicking optimization.

Pipeline

We propose to extend the typical nominal spatial dose prediction and dose mimicking procedure with a scenario‐specific dose prediction component, as illustrated in Figure 1. First, a nominal model predicts the nominal dose from the nominal geometry . To predict the scenario dose for any scenario , is deformed in accordance with the change from the nominal geometry to the scenario geometry using a second scenario model. One such scenario dose is predicted for each , yielding a set of predicted scenario doses.
FIGURE 1

An illustration of the proposed pipeline. A nominal dose is predicted from a nominal geometry using a nominal model. Furthermore, for each scenario , is deformed into a scenario geometry . The scenario model is then used to predict a scenario dose from these for each scenario. The scenario doses are finally used in robust dose mimicking yielding a set of machine parameters η

An illustration of the proposed pipeline. A nominal dose is predicted from a nominal geometry using a nominal model. Furthermore, for each scenario , is deformed into a scenario geometry . The scenario model is then used to predict a scenario dose from these for each scenario. The scenario doses are finally used in robust dose mimicking yielding a set of machine parameters η To train such a scenario model, ideally, the training data would consist of robust treatment plans—in particular, a set of geometry–dose pairs where all doses are planned robustly with respect to the scenario set . However, to remove the need of having access to a complete data set of robust plans, which is a relatively strict requirement, we propose an alternative method of training the scenario model. In this framework, each plan in a data set of previously delivered non‐robust treatment plans is deformed in accordance with a number of scenarios. As we want the resulting robust treatment plans to have sufficient target coverage, we want our scenario reference doses to have sufficient target coverage as well. Therefore, we propose to train the scenario model using a loss function that both penalizes deviation from the non‐robust ground truth scenario dose, as well as deviation from the target coverage of the nominal dose. The training pipeline is illustrated in Figure 2.
FIGURE 2

An illustration of training the proposed pipeline. For each scenario s, the ground truth dose is calculated using an accurate dose calculation algorithm from the corresponding scenario geometry . The scenario model is trained using a loss function consisting of a spatial loss that depends on the predicted scenario dose and , as well as a DVH loss that depends on and

An illustration of training the proposed pipeline. For each scenario s, the ground truth dose is calculated using an accurate dose calculation algorithm from the corresponding scenario geometry . The scenario model is trained using a loss function consisting of a spatial loss that depends on the predicted scenario dose and , as well as a DVH loss that depends on and

Algorithm

To maintain the target coverage equivalent to that of the nominal dose, but still predicting a realistic scenario dose, we propose to train the scenario model using a loss function that combines a voxel‐level spatial loss with a DVH loss defined on the targets. In particular, the spatial loss is given by the weighted mean‐squared error where the are nonnegative voxel weights such that . For the DVH loss, let be the set of all ROIs, each represented as index sets of voxels, and let be the local dose vector for each R. Let, furthermore be the set of all target ROIs which are to be covered robustly, typically chosen to comprise all CTVs. Denoting by the dose‐at‐volume at volume level , which is given for each R by the DVH loss may be written as The idea of utilizing a DVH loss for training dose prediction models has previously been explored by Nguyen et al and Zhang et al, but for other purposes. Upon predicting in a scenario with non‐robust scenario ground truth and nominal ground truth , the loss contribution is then given by weighting the DVH loss by a factor α. For both the nominal and scenario models, we propose to use an architecture based on the 3D U‐net, with dimensions adapted for the present dose prediction purpose. The architecture used in our experiments is illustrated in Figure 3. This architecture consists of a contracting path, consisting of a series of convolution, activation, batch normalization, and max pooling operations, serving as a form of feature extraction, and a symmetric expanding path that instead uses deconvolution operations, with skip connections between the two paths. Similar architectures have previously been used successfully for the task of dose prediction. , , , , ,
FIGURE 3

The proposed 3D U‐net architecture. The size of each row is denoted at the far left and the number of channels in each layer is denoted above the layer

The proposed 3D U‐net architecture. The size of each row is denoted at the far left and the number of channels in each layer is denoted above the layer

Robust dose mimicking

Having trained the nominal and scenario models to obtain for a new patient a set of predictions, we may use the same predictions as scenario‐specific reference doses in a dose mimicking to create a robust deliverable plan. Let denote the machine parameters, the physical limitations of which are articulated by constraining η to lie in the feasible set , and let be the scenario doses resulting from the plan defined by η. Partitioning into non‐robust and robust ROIs, we can write a general robust dose mimicking problem as where is a cost functional expressing the conservativeness of the robust optimization and each is on the form Here, the cost functional is commonly chosen as a weighted‐power‐mean function with exponent , where corresponds to stochastic programming and to minimax optimization. As for and , common choices are one‐sided quadratic functions penalizing deviations in spatial dose and DVH, respectively.

Computational study

We applied our framework to a cohort of 52 retrospective treatment plans for prostate cancer patients, divided into a training set of 37 patients, a test set of 10 patients, and a validation set of 5 patients. The patients originally received volumetric modulated arc therapy, so we utilized the framework proposed by Kierkels et al to convert them to proton pencil beam scanning (PBS) plans. We used two beams with isocenters in the middle of the CTV, that is, the prostate, aimed at 90° and 270°, going through the left and right femurs, respectively. The organs at risk (OARs) considered were the rectum, bladder, anal canal, and left and right femurs. The dose prescribed to all the patients was a median dose of () to the CTV, delivered over 35 fractions. For the uncertainties, similar to Fredriksson et al, we used a density uncertainty of , and a slightly smaller setup uncertainty of . For the setup uncertainty, the isocenter was shifted in the unit directions as well as diagonally. The identity shift was also included for both the density and setup uncertainties, resulting in a total of scenarios with corresponding geometries for each patient. For both the nominal and scenario models, we used the architecture in Figure 3, with sizes and channels adjusted for our purposes. The voxel size used was , and the input and output resolution of the models were voxels. For the nominal model, the input channels consisted of , that is, the binary masks for the CTV and each OAR as well as a CT image of the patient. Recall that the purpose of the scenario model is to deform the nominal dose in accordance with a specific scenario s—therefore, the scenario model had input channels consisting of both the nominal geometry and a scenario geometry , as well as the nominal dose . Both models were trained using the total loss described in Section 2.2. For the scenario model, we used the training data set of size resulting from including all scenarios for each patient. In both the nominal and scenario models, data augmentation was applied in the form of rotations around the transversal axis, drawn uniformly at random between and 10° at each forward pass through the network. The DVH loss weighting factor α was chosen using grid search over the values in . The models were trained until the loss on the validation set converged, which was around 500 epochs for the nominal model and 50 epochs for the scenario model. To select the weighting factor α, we inspected the dose washes and DVH curves. For both models, depending on the value of α, we saw a trade‐off between the ability to predict a dose that gives a good DVH for the CTV, and a realistic dose outside of the CTV. For the nominal and the scenario model, we used and , respectively. We observed that these weighting factors gave a satisfactory CTV coverage, while still giving a realistic dose outside of the CTV. The spatial loss in (1) chosen for our experiments is a voxel‐level mean squared error loss. As voxels closer to the CTV are generally more important, we weighted the contribution of each voxel depending on its distance from the CTV using the weighting , where is the Euclidean distance from the voxel i to the CTV, is a constant and is a minimum weight threshold. For our experiments, we used and . The robust dose mimicking was performed using a research version of the treatment planning system RayStation 11A (RaySearch Laboratories) with sequential quadratic programming–based optimization, creating deliverable proton PBS plans. The in‐house dose mimicking algorithm was used, with one‐sided quadratic loss functions and , voxel‐level weights determined partly depending on the corresponding isodose level on the nominal predicted dose, and a weighted‐power‐mean cost functional with exponent 8 approximating minimax optimization. The ROIs considered in the dose mimicking problem were the CTV, bladder, rectum, and ring ROIs with distances and from the CTV, all included in the robust subset . The mimicking optimization was divided into three runs of 60, 60, and 8 iterations, respectively. In particular, approximate doses during optimization were computed using a Monte Carlo algorithm using 104 ions per spot, and final doses using the same algorithm with a statistical uncertainty of 0.5%. Finally, the plans produced by the proposed automated algorithms were benchmarked to manually generated robust treatment plans for the same patients. The in‐house inverse planning algorithm in RayStation 11A was employed to create the comparison plans using the optimization functions displayed in Table 1, of which all were robust and those on the CTV were formulated as constraints. In particular, the optimization was run using the same weighted‐power‐mean exponent, number of iterations, and dose computation settings as for the automatically generated plans. While additional fine‐tuning of the manual plans are likely to be needed before reaching fully clinical quality, they serve the purpose of a comprehensive and transparent benchmark for the proposed automated planning algorithm.
TABLE 1

The optimization formulation used for creating the manual comparison plans

ROIOptimization functionRobustConstraintWeight
CTVAt least 7400cGy at 98% volumeYesYes
CTVAt most 7900cGy at 2% volumeYesYes
BladderAt most 6500cGy at 10% volumeYesNo1
RectumAt most 6500cGy at 10% volumeYesNo1
Ring 0--1cm At most 7000cGy mean doseYesNo1
Ring 0--1cm At most 4000cGy mean doseYesNo1
The optimization formulation used for creating the manual comparison plans

RESULTS

To verify the feasibility of the proposed pipeline, we perform a qualitative analysis for one test patient and a quantitative analysis based on the entire test data set. In Figure 4, two scenarios for a test patient are visualized. The non‐robust ground truth dose fails to give a sufficient target coverage—in particular, in the transversal view, one can see that large parts of the CTV receive less than 95% of the prescribed dose. However, the scenario dose prediction model has been trained to predict doses with a better target coverage, and we see indeed that the predicted dose successfully covers the CTV in these scenarios. Finally, the automatically generated robust deliverable dose is generated by performing a robust dose mimicking using all the predicted scenario doses as reference doses and is expected to ensure coverage of the CTV in most or all of the scenarios. For the two scenarios displayed, the CTVs are well‐covered, whereas the dose decays slower beyond the CTV outline than the corresponding predicted scenario doses—this is due to the robust plan needing to account for the outcome in all scenarios. The manually generated benchmark has similar target coverage and decay around the CTV compared to the automatically generated dose, but slightly more dose spillage in the area beyond 2 cm from the CTV.
FIGURE 4

Two different scenarios with the non‐robust (first row), the predicted (second row), the robust (third row), and the manual (fourth row) doses. The left column shows a coronal slice of scenario A where the patient has been translated down with respect to the image. The right column shows a transversal slice of scenario B where the patient has been translated down with respect to the image and a density shift of has been applied

Two different scenarios with the non‐robust (first row), the predicted (second row), the robust (third row), and the manual (fourth row) doses. The left column shows a coronal slice of scenario A where the patient has been translated down with respect to the image. The right column shows a transversal slice of scenario B where the patient has been translated down with respect to the image and a density shift of has been applied Furthermore, the DVHs corresponding to the scenarios in Figure 4 are displayed in Figure 5. The DVHs of the predicted scenario doses are relatively similar to those of the non‐robust dose in all ROIs except for the CTV, where the target coverage is instead more similar to that of the prescribed dose. This is what we wanted to achieve with our scenario dose prediction since the spatial component of the loss function used to train the scenario model is expected to make the predicted doses similar to the ground truth scenario doses, while the DVH component is aimed at maintaining the target coverage of the nominal dose. We can also see that the automatically generated robust dose has a similar target coverage as the predicted dose in both scenarios, meaning that the target coverage of the predicted doses successfully propagates to the robust dose—however, naturally, this comes with the cost of a slight increase in dosage to the rectum, bladder, and the target surroundings. The automatically generated robust dose has similar DVHs compared to those of the manual benchmark.
FIGURE 5

The DVHs for scenario A (left) and scenario B (right)

The DVHs for scenario A (left) and scenario B (right) Moreover, in Figure 6 and Table 2, a number of dose statistics aggregated across all test patients and scenarios are presented. For CTV and , we see that the predicted and automatically generated robust doses have a lower spread and are more focused around the prescribed dose than the non‐robust dose, whereas they have a slightly higher spread compared to the manual plans. This indicates that the automatically generated dose is in fact robust with respect to the CTV given the specified uncertainty parameters. For bladder and rectum , we see that the automatically generated dose gives a higher dosage in general to the OARs than the non‐robust and predicted doses, which is an expected effect of delivering more dose to the area around the CTV due to the overlap between these OARs and the CTV in the different scenarios—however, the dosage is similar to that of the manual plans. Finally, we include two ring ROIs, representing a border of around the CTV and representing a border of around the aforementioned ring ROI, with the purpose of illustrating the decay of dose beyond the target. We can see here, as well as in Figures 4 and 5, that the decrease is slower for the automatically generated dose than for the non‐robust and predicted doses, which is an expected effect of delivering more of the prescribed dose. However, the decrease is very similar to that of the manually generated dose. In summary, the proposed pipeline is able to generate doses that are robust with respect to the scenarios, while we see certain expected effects from delivering more dose than in the non‐robust case.
FIGURE 6

Boxplot of dose statistics for the different dose types evaluated across the 45 scenarios for each test patient

TABLE 2

The minimum, maximum, mean, and standard deviation of the dose statistics for the different dose types evaluated across the 45 scenarios for each test patient

GoalTypeMin (cGy)Max (cGy)Mean (cGy)Std (cGy)
Non‐robust469775236488542
CTV, D98% Predicted71197512740173
Robust697575747386111
Manual73427511742931
Non‐robust783686888117159
CTV, D2% Predicted78878000793925
Robust78197987789431
Manual78848044794828
Non‐robust404755538851793
Bladder, D10% Predicted70737535001869
Robust1093762254811648
Manual868762054951726
Non‐robust651702933381586
Rectum, D10% Predicted371667529471524
Robust1620766253621462
Manual1974755753491316
Non‐robust479564405742369
Ring 0--1cm, Dmean Predicted521658625565136
Robust632171816867157
Manual651572026892136
Non‐robust160631012297350
Ring 1--2cm, Dmean Predicted190223152104104
Robust295242943755306
Manual315441023746244
The minimum, maximum, mean, and standard deviation of the dose statistics for the different dose types evaluated across the 45 scenarios for each test patient Boxplot of dose statistics for the different dose types evaluated across the 45 scenarios for each test patient

DISCUSSION

In this work, we have presented a data‐driven approach to robust automated radiation therapy treatment planning. Using a data set of non‐robust proton plans and a two‐step scenario dose prediction model with a DVH‐based loss function term, we were able to predict relatively realistic scenario doses with target coverage comparable to that of the nominal ground truth dose. By using them as scenario‐specific reference doses in a robust dose mimicking problem, we were able to create robust deliverable plans consistently delivering sufficient target coverage at the cost of an expectedly higher dosage to surrounding tissue than the predicted and non‐robust doses. Compared to manually generated benchmark plans, the produced plans were similar in terms of DVH, dose statistic spread, and spatial dose. While additional postprocessing of the automatically generated plans may be needed in order to be considered clinical, as is the case with automated planning algorithms in general, the results serve to showcase the feasibility of our type of workflow—that is, the combination of scenario dose prediction and robust dose mimicking. Among the advantages of our method are the nonrequirement of a robustly planned data set for training, the generality and flexibility associated with separating the task of scenario dose prediction into a nominal and scenario model, and the more rigorous handling of setup and range uncertainties through robust optimization rather than through, for example, margins. Although experiments were performed on proton plans, for photons, especially, one may want to have the choice of whether or not to use robust planning. In such cases, access to a completely robust training data set may be too high a requirement. By separating the dose prediction pipeline into the image‐to‐nominal and nominal‐to‐scenario parts, we may use instead use data obtained from robust evaluation to learn the physical deformations associated with each scenario. Along with the DVH loss aimed at controlling target coverage of the scenario model outputs, the result is a highly general framework for handling robustness in automated treatment planning. However, one disadvantage of the method is that there is a lack of theoretical rigor for predicting realistic scenario doses with better target coverage. With our loss functions, the optimal output of the scenario model is a dose that is identical to the non‐robust ground truth outside the CTV and identical to the nominal dose within the CTV, which is a discontinuous dose. In practice, smoothness is introduced by the convolutional design of the neural network, but this notion is hard to control exactly. Insofar as the predicted scenario doses are similar to the non‐robust scenario ground truth far from the target, similar to the nominal ground truth in the target and some smoothened mixture in between, they can be understood to represent a theoretically ideal scenario dose given a fixed nominal dose. They should, however, be strictly more realistic than the nominal dose in each nonnominal scenario—indeed, using the nominal dose as reference dose in all scenarios is guaranteed not to be achievable. Hence, even though the predicted scenario doses may not always be physically realizable, the introduction of an additional scenario‐specific dose prediction shrinks the gap between reference dose and physically realizable dose in the dose mimicking phase. Apart from addressing the smoothness issue, future research may include evaluating the proposed methodology on data sets with different treatment modalities, delivery techniques, and robustness parameters. For protons and heavy ions, one may also generalize the current scenario dose prediction to a version in which beam doses are predicted separately, with the addition of beam‐specific objective functions in the dose mimicking optimization problem. The scenario dose prediction may also be used in other contexts than automated planning, for example, for quality assurance purposes. Moreover, one may try to combine the current framework with a semiautomatic multicriteria optimization methodology such as in Zhang et al. All in all, the incorporation of robustness into machine learning–automated treatment planning enables ample new opportunities to be explored.

CONCLUSIONS

We have presented a new data‐driven approach to robust automated treatment planning, combining prediction of spatial scenario doses with robust dose mimicking. By dividing the former part into a nominal and a scenario dose prediction model, and using a DVH loss during training, we are able to predict for each new patient scenario doses with a robustly covered target using a non‐robust training data set. Subsequently, through robust dose mimicking, we obtain plans robust against the same scenario set. The numerical results serve to demonstrate the feasibility of the proposed methodology, which has the potential of facilitating the incorporation of robustness into automated planning.

CONFLICT OF INTEREST

The authors have no conflicts to disclose.
  24 in total

1.  Dosimetric features-driven machine learning model for DVH prediction in VMAT treatment planning.

Authors:  Ming Ma; Nataliya Kovalchuk; Mark K Buyyounouski; Lei Xing; Yong Yang
Journal:  Med Phys       Date:  2019-01-02       Impact factor: 4.071

2.  Minimax optimization for handling range and setup uncertainties in proton therapy.

Authors:  Albin Fredriksson; Anders Forsgren; Björn Hårdemark
Journal:  Med Phys       Date:  2011-03       Impact factor: 4.071

Review 3.  Artificial intelligence in radiotherapy.

Authors:  Sarkar Siddique; James C L Chow
Journal:  Rep Pract Oncol Radiother       Date:  2020-05-06

4.  Automated improvement of radiation therapy treatment plans by optimization under reference dose constraints.

Authors:  Albin Fredriksson
Journal:  Phys Med Biol       Date:  2012-11-06       Impact factor: 3.609

5.  Data-driven approach to generating achievable dose-volume histogram objectives in intensity-modulated radiotherapy planning.

Authors:  Binbin Wu; Francesco Ricchetti; Giuseppe Sanguineti; Michael Kazhdan; Patricio Simari; Robert Jacques; Russell Taylor; Todd McNutt
Journal:  Int J Radiat Oncol Biol Phys       Date:  2010-08-26       Impact factor: 7.038

6.  3D radiotherapy dose prediction on head and neck cancer patients with a hierarchically densely connected U-net deep learning architecture.

Authors:  Dan Nguyen; Xun Jia; David Sher; Mu-Han Lin; Zohaib Iqbal; Hui Liu; Steve Jiang
Journal:  Phys Med Biol       Date:  2019-03-18       Impact factor: 3.609

7.  Automatic treatment planning based on three-dimensional dose distribution predicted from deep learning technique.

Authors:  Jiawei Fan; Jiazhou Wang; Zhi Chen; Chaosu Hu; Zhen Zhang; Weigang Hu
Journal:  Med Phys       Date:  2018-11-28       Impact factor: 4.071

8.  Probabilistic Pareto plan generation for semiautomated multicriteria radiation therapy treatment planning.

Authors:  Tianfang Zhang; Rasmus Bokrantz; Jimmy Olsson
Journal:  Phys Med Biol       Date:  2022-02-04       Impact factor: 3.609

9.  A convolutional neural network approach for IMRT dose distribution prediction in prostate cancer patients.

Authors:  Tomohiro Kajikawa; Noriyuki Kadoya; Kengo Ito; Yoshiki Takayama; Takahito Chiba; Seiji Tomori; Hikaru Nemoto; Suguru Dobashi; Ken Takeda; Keiichi Jingu
Journal:  J Radiat Res       Date:  2019-10-23       Impact factor: 2.724

10.  Robust automated radiation therapy treatment planning using scenario-specific dose prediction and robust dose mimicking.

Authors:  Oskar Eriksson; Tianfang Zhang
Journal:  Med Phys       Date:  2022-03-30       Impact factor: 4.506

View more
  1 in total

1.  Robust automated radiation therapy treatment planning using scenario-specific dose prediction and robust dose mimicking.

Authors:  Oskar Eriksson; Tianfang Zhang
Journal:  Med Phys       Date:  2022-03-30       Impact factor: 4.506

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.