Literature DB >> 32543097

Predictive quality assurance of a linear accelerator based on the machine performance check application using statistical process control and ARIMA forecast modeling.

Wayo Puyati1,2, Amnach Khawne1, Michael Barnes3,4, Benjamin Zwan4,5, Peter Greer3,4, Todsaporn Fuangrod6.   

Abstract

PURPOSE: A predictive linac quality assurance system based on the output of the Machine Performance Check (MPC) application was developed using statistical process control and autoregressive integrated moving average forecast modeling. The aim of this study is to demonstrate the feasibility of predictive quality assurance based on MPC tests that allow proactive preventative maintenance procedures to be carried out to better ensure optimal linac performance and minimize downtime. METHOD AND MATERIALS: Daily MPC data were acquired for a total of 490 measurements. The initial 85% of data were used in prediction model learning with the autoregressive integrated moving average technique and in calculating upper and lower control limits for statistical process control analysis. The remaining 15% of data were used in testing the accuracy of the predictions of the proposed system. Two types of prediction were studied, namely, one-step-ahead values for predicting the next day's quality assurance results and six-step-ahead values for predicting up to a week ahead. Results that fall within the upper and lower control limits indicate a normal stage of machine performance, while the tolerance, determined from AAPM TG-142, is the clinically required performance. The gap between the control limits and the clinical tolerances (as the warning stage) provides a window of opportunity for rectifying linac performance issues before they become clinically significant. The accuracy of the predictive model was tested using the root-mean-square error, absolute error, and average accuracy rate for all MPC test parameters.
RESULTS: The accuracy of the predictive model is considered high (average root-mean-square error and absolute error for all parameters of less than 0.05). The average accuracy rate for indicating the normal/warning stages was higher than 85.00%.
CONCLUSION: Predictive quality assurance with the MPC will allow preventative maintenance, which could lead to improved linac performance and a reduction in unscheduled linac downtime.
© 2020 The Authors. Journal of Applied Clinical Medical Physics published by Wiley Periodicals, Inc. on behalf of American Association of Physicists in Medicine.

Entities:  

Keywords:  autoregressive integrated moving average forecast modeling; machine performance check; predictive quality assurance; statistical process control

Mesh:

Year:  2020        PMID: 32543097      PMCID: PMC7484849          DOI: 10.1002/acm2.12917

Source DB:  PubMed          Journal:  J Appl Clin Med Phys        ISSN: 1526-9914            Impact factor:   2.102


INTRODUCTION

Daily quality assurance (QA) testing of linear accelerators (linacs) is standard practice for ensuring the safe and accurate delivery of radiotherapy treatments. The American Association of Physicists in Medicine (AAPM) has published recommendations on daily linac QA procedures with tolerances. , , However, the test equipment and exact methodology are non‐standardized and remedial action is often only reactively performed once tolerances have been breached. With the TrueBeam 2.0 platform, Varian (Varian Medical Systems, Palo Alto, CA, USA) released the Machine Performance Check (MPC) application, which is now also available on the Varian Halcyon linac. MPC is a fully integrated image‐based tool for assessing the performance of linac critical functions. MPC tests are divided into two categories. First, beam constancy checks use a single megavolt image per beam energy without a phantom in place to assess the output, beam center, and uniformity constancy against a user‐defined baseline. Second, geometric tests use a series of kilovoltage and 6 MV beam images of the IsoCal phantom situated in a specific bracket on the couch top to assess the radiation isocenter size, coincidence of megavolt and kilovolt isocenters, accuracy of collimator and gantry angles, accuracy of jaw and multileaf collimator (MLC) leaf positions, and accuracy of the couch positioning including pitch and roll. All measurements are highly automated, and the user is simply required to set up the IsoCal phantom and bracket on the treatment couch at position H2 and to beam‐on for each required energy. For the geometric tests, the system makes all required motions automatically and beams‐on when all is in position. Images are automatically analyzed at the linac console. Beam constancy check results are presented relative to the baseline, whereas geometric tests are self‐referenced. Functionality for presenting trends in results is also embedded in the MPC module and data can be exported in.csv format. The MPC application has now been evaluated by multiple authors as a daily linac QA tool. , , , , Statistical process control (SPC) is a statistical method for detecting the defects of a process and was first presented by Shewhart. SPC has become a standard method of quality control. In SPC, continual observations are used to calculate a control chart, which includes a maximum control limit and minimum control limit that define a quality level. The control chart is often used in monitoring a process and detecting failure states at a point of measurement. Binny et al. applied SPC to analyze the QA output variation in helical and static output for periods of up to 4 yr in an effort to improve helical tomotherapy QA. Meanwhile, López‐Tarjuelo et al. adopted SPC in the daily quality control of linac electron beams. Fuangrod et al. applied SPC in constructing a clinically significant threshold of a real‐time treatment verification system. Recently, SPC has been used with MPC data in another study by Binny et al. In that study, SPC analysis was conducted for MPC data across six TrueBeam linacs for 12 months in an attempt to determine MPC tolerances. However, the study did not attempt to forecast MPC results to be used for predictive QA. The concept of predictive QA that allows preemptive maintenance based on SPC analysis has been studied for radiotherapy linacs. , , , Predictive QA testing would allow radiotherapy departments to be proactive in their maintenance by remedying faults before clinical tolerances are breached. In theory, this should provide for improved linac performance consistency and reduced unscheduled linac downtime that can be disruptive to departmental workflow. Previous predictive QA studies have either been based on readouts from the linac itself , , and hence lack independence or have been based on film measurements, which are impractical on a daily basis. The present study presents a framework of predictive linac QA based on MPC data. Control limits were determined from the SPC analysis of MPC data over the long term and used to determine the standard linac performance. When this performance is within clinical tolerances, there is a window of opportunity to rectify the problem before it becomes clinically significant. Such remedial action can be scheduled out of standard treatment hours to avert disruption to the clinical workflow. The forecasting of QA results provides a measure of how long this window of opportunity is.

MATERIALS AND METHODS

System overview

Figure 1 presents the proposed predictive QA system, which can be divided into the four steps of prediction, display, evaluation, and relearning. Daily MPC test data are prepared and loaded into the system. The system predicts both one‐step‐ahead and six‐step‐ahead MPC test results using a predictive model, which has been learnt from historical MPC data. In the display step, the predictive MPC data for each parameter are displayed on the constructed control chart which shows the upper control limit (UCL), lower control limit (LCL), central line (CL), and tolerance levels. The tolerance levels are defined by AAPM TG‐142 recommendations. However, some MPC tests are not required by TG‐142 daily QA, the standard MPC thresholds were used instead. The trend of predictive data is also plotted from the predictive six‐step‐ahead MPC test result for each parameter. The system was developed and implemented using in‐house software development using MATLAB version 2019a (Mathworks, Natick, MA, USA).
Fig. 1

System overview.

System overview. In the evaluation step, the predictive MPC test results are compared with the UCL and LCL and the system generates a warning flag if the control limits are breached. This flag escalates from warning to fail if the result exceeds the TG‐142 tolerance. Finally, the MPC test results are added to the historical MPC data for relearning and recalculating the UCL and LCL. The system therefore self‐adjusts to reflect the current linac uncertainty and behavior.

MPC tests and data collection

There are 29 MPC tests that can be grouped into four categories, namely, isocenter, collimation, gantry, and couch tests. In collimation, there are four MLC tests — denoted mean offset leaf‐bank A and B and max offset leaf‐bank A and B — and the system also provides results for individual MLC leaves from leaf number 2 to 59 for both banks A and B. There are thus a total of 141 MPC test parameters (25 MPC tests + 116 individual MLC leaves). For the predictive model and SPC development, historical MPC data were taken from daily MPC tests from a single Varian TrueBeam Stx linac with HD MLC, 6MV energy, and As 1200 EPID over 490 days. The data were divided into two data sets. The first 85% of data (416 days) were used in constructing the SPC control chart and calculating the predictive model for each MPC test. The remaining 15% of data (74 days) were compared to the predictive model results to validate its accuracy and evaluate the system performance.

Control chart construction

The control chart and UCL and LCL are constructed according towhere is the average of data observations, is an average of the moving range , and is 2.66. UCL, LCL, and CL are, respectively, the UCL, LCL, and CL.

Predictive model calculation: autoregressive integrated moving average

The autoregressive integrated moving average (ARIMA) model is a statistical predictive model whose fitting allows future prediction of the time series. The ARIMA methodology includes an autoregressive term (AR), which is the weighted sum of recent differenced values, moving average term (MA), which is the weighted sum of the forecasting error, and integrated term, which is the degree of differencing for nonstationary elimination, if necessary. The general form of ARIMA (p,d,q) is as follows: where is the predictive value at , denotes differenced values in the time series at , is a constant value, denotes the weights of differenced values of AR ( ), denotes the orders of predicted error, denotes the weights of predicted error ( ), is the order of AR, is the order of the MA, and is the order of differencing in nonstationary elimination. Maximum likelihood estimation is adopted to estimate the parameter and error terms of the ARIMA model. The Akaike information criterion is used to select the optimal model, which is the model that best fits the recorded time‐series data. The ARIMA model is also used to predict the n‐step‐ahead value, which can be used to represent the future trend of the time‐series data.

Predictive model and system evaluation

The accuracy of the predictive model was evaluated by comparing the ARIMA predicted results with the actual results in the final 15% of the MPC data. Model accuracy was assessed using the root‐mean‐square error (RMSE) and absolute error (MAE). when is the predicted value, is the actual value of state , and is the number of observations. RMSE and MSE are considered as the standard error measurements of model in predicting quantitative data. In addition, the model can also be evaluated by comparing the predicted results against the control chart (pass, fail, or warning status). The accuracy of the model in this sense can be calculated using the average accuracy: The average accuracy rate represents the average overall effectiveness of prediction. Let be the number of true positives of state , which is the number of correct predictions of state . Let be number of the true negatives of state , which is the number of correct predictions of nonstate . Let be the number of the false positives of state , which is the number of incorrect predictions of state . Let be the number of false negatives of state , which is the number of incorrect predictions of nonstate .

RESULTS

SPC analysis

Figures 2 and 3 present the UCL and LCL for each MPC test. In the figures, a white area represents the normal stage level (ranging from the UCL to LCL) and a grey area is a warning stage level (ranging from the UCL/LCL to the tolerance).
Fig. 2

Results of the derived LCL and UCL of the MPC (isocenter, collimation, gantry, and couch) against the TG 142 tolerance.

Fig. 3

Results of the derived LCL and UCL of the MPC (beam, MV imager, and kV imager) against the TG‐142 tolerance.

Results of the derived LCL and UCL of the MPC (isocenter, collimation, gantry, and couch) against the TG 142 tolerance. Results of the derived LCL and UCL of the MPC (beam, MV imager, and kV imager) against the TG‐142 tolerance.

Predictive QA system

Figure 4 shows an example of predicted QA data. Figure 4(a) and 4(b) present the MPC test for the beam output change. The one‐step‐ahead predictive value is calculated as well as its weekly trend line (from six‐step‐ahead predictive values). All predictive values are within the SPC control limits and the trend is that the beam output slightly drops over the week but remains at normal‐stage levels. Figure 4(c) presents prediction results of individual MLC leaf tests, showing that the one‐step‐ahead result remains located at normal‐stage levels for all leaves. Figure 4(d) presents the six‐step‐ahead prediction for MLC leaf #30. The prediction indicates that the leaf remains between the UCL and LCL. Figure 5 presents the comparison between predicted and actual of beam center shift, beam output change, and beam uniformity change for all tested data. It found that the beam output change had the systematic offset in predicted value. Figure 6 demonstrated how the system predicts that the data will exceed the warning level (UCL). Figures are not presented for all MPC parameters for readability purposes.
Fig. 4

(a) Historical overview of the beam output change (%); (b) sample of one‐step‐ahead predictive data, six‐step‐ahead predictive values, and trend of the beam output change; (c) overview of the one‐step‐ahead prediction of the MLC against XmR; and (d) a sample comparison of the current check and one‐step‐ahead prediction of the MLC with XmR and the trend line for leaf No. 30.

Fig. 5

Comparison of predicted and actual beam QA data including center shift, output change, and uniformity change.

Fig. 6

Example of trend line to detect the output change exceeded the UCL that demonstrates the system is able to flag the warning stage before it occurs.

(a) Historical overview of the beam output change (%); (b) sample of one‐step‐ahead predictive data, six‐step‐ahead predictive values, and trend of the beam output change; (c) overview of the one‐step‐ahead prediction of the MLC against XmR; and (d) a sample comparison of the current check and one‐step‐ahead prediction of the MLC with XmR and the trend line for leaf No. 30. Comparison of predicted and actual beam QA data including center shift, output change, and uniformity change. Example of trend line to detect the output change exceeded the UCL that demonstrates the system is able to flag the warning stage before it occurs.

Predictive model performance

The performance of the ARIMA model in predicting daily MPC results is presented in Table 1. The majority of tests make accurate predictions with an RMSE less than or equal to 0.05. Exceptions are the jaw collimation, gantry relative, kilovolt imager tangential, beam output change, and beam uniformity change, with the worst case being the output measured at 0.14 RMSE. A lower RMSE corresponds to the predictive model better correlating with the actual MPC test results. A similar pattern of evaluation results was observed with the MAE metric in that most of the results are less than or equal to 0.05. Figure 5 presents the results for individual MLC leaves. The results show that the high numbered leaves are predicted better than the low numbered leaves for Bank A, while the prediction is more consistent across the bank for bank B. The reason for these results is unclear.
Table 1

Results of the predictive model evaluation and accuracy of system output.

MPC testRMSEMAE

Average accuracy rate

(%)

Categories Test
IsocenterSize (mm)0.020.0197.80
MV imager projection offset (mm)0.030.0279.12
kV imager projection offset (mm)0.040.0361.54
CollimationMaximal offset leaves A (mm)0.020.02100.00
Maximal offset leaves B (mm)0.030.0398.97
Mean offset leaves A (mm)0.020.0298.97
Mean offset leaves B (mm)0.030.0298.97
Individual MLC leaf A (mm)0.030.0298.56
Individual MLC leaf B (mm)0.030.0298.13
Jaw X1 (mm)0.100.0647.42
Jaw X2 (mm)0.130.0595.88
Jaw Y1 (mm)0.090.0764.21
Jaw Y2 (mm)0.060.0579.17
Rotation offset (°)0.020.02100.00
GantryAbsolute (°)0.020.0192.86
Relative (°)0.060.06100.00
CouchLateral (mm)0.040.04100.00
Longitudinal (mm)0.030.0264.86
Pitch (°)0.010.0198.63
Roll (°)0.010.0095.38
Rotation (°)0.010.0148.10
Vertical (mm)0.050.0491.78
Rotation‐induced couch shift (mm)0.020.02100.00
Kilovolt imagerIn‐plane rotation (°)0.000.00100.00
Source axial0.040.0395.60
Tangential0.120.1098.90
Megavolt imagerIn‐plane rotation (°)0.000.00100.00
Source Axial0.010.0160.44
Tangential0.010.01100.00
BeamCenter shift (mm)0.050.0489.52
Beam output change (%)0.140.1294.29
Uniformity change (%)0.100.0793.33
Results of the predictive model evaluation and accuracy of system output. Average accuracy rate (%) The average accuracy rate was applied to assess the ability of the models to accurately flag the warning stage. The fourth column in Table I shows the results of the average accuracy rate of the one‐step‐ahead prediction. It is seen that most of the average accuracy rates for all tests were higher than 85.00%. Exceptions are the tests denoted megavolt and kilovolt imager projection offset, Jaw X1, Jaw Y1 & Y2, couch longitudinal and rotation, and megavolt imager source axial. It can be noted that QA test with minor variability or drift will result in tight UCL and LCL so that sensitivity is maintained. Figures 7 and 8 present the results of the average accuracy for individual MLC leaves. The average accuracy is more than 90.00% for all leaves of both banks with the exception of leaf number 19 of bank B.
Fig. 7

Performance of the predictive model for individual leaves evaluated using the MAE and RMSE.

Fig. 8

Evaluation of the warning stage prediction using the average accuracy for an individual leaf of the MLC.

Performance of the predictive model for individual leaves evaluated using the MAE and RMSE. Evaluation of the warning stage prediction using the average accuracy for an individual leaf of the MLC.

DISCUSSION

MPC is advantageous for providing predictive QA data as MPC: provides a measurement that is independent of the linac parameters being tested. is quick and easy to perform and thus suitable for daily measurements, producing large, high‐frequency datasets for predictive analysis. is largely free of user variability. assesses a large number of required linac parameters, meaning that other test devices are not required. generates digital data that are integrated into the linac system, making it simple for Varian to transmit results to a central database via its Smart Connect functionality. provides a standardized method for daily linac QA, meaning that the performances of all linacs running the MPC can be compared. This has the potential for individual linacs with outlying performance to be identified and to provide feedback data to Varian for improvement of the linac design and construction. In this study, the ARIMA forecasting model was applied to predict both the next‐day (one‐step‐ahead) and next‐week (six‐step‐ahead) MPC QA results. SPC control charts were constructed to determine the upper and lower bounds of standard machine performance. For all parameters tested, these bounds were found to be within clinical tolerances (defined as TG‐142 tolerances in this study), providing a window of opportunity in which a performance issue could be rectified before it became clinically significant. For the MLC results of [Fig. 4(c)], individual models were generated for the leave banks and individual leaves. Since the leaf bank results are maximum and mean of all of the individual leaves, then it is expected that the model for each individual leaf would be more precise than for the whole bank. This is demonstrated in the data. The accuracy required for prediction in the clinical setting is likely not high. Clinical physicists would most likely simply need a rough indication of how long they have before a failed test is expected so that they can assign urgency to the problem being addressed and organize remediation. For a TG‐142 tolerance fail, the decision may well be to remove the linac from clinical use until the failure has been investigated, but for results within the warning stage between control limits and clinical tolerance, the investigation would likely be delayed until outside of normal treatment hours to avoid disruption to the clinical schedule. In these cases, it is likely that investigation will be made at either the end of the clinical day or on the weekend. In these cases, one‐step‐ahead and six‐step‐ahead prediction is appropriate. The testing in the present study found that the majority of the one‐step‐ahead predictions were accurate (RMSE and MAE << 0.10) and the majority of average accuracy results were greater than 85.00%. The worst predictive performance was for the beam output change. The beam output change has high variability over time and is recorded as a nonstationary time‐series dataset, which affects the accuracy of the predictive model. Resolution of this problem may require a different predictive model. Examples of possible models include the EWMA model, recurrent neural networks, and artificial neural networks, which could be investigated in future work. An additional weakness of the ARIMA model is that the data require cleaning to eliminate poor‐quality data, such as user errors. The ARIMA model is sensitive to all values in the learning process and learning from poor‐quality data will lead to poor prediction. Moreover, the learning should continually progress by taking new measurements into account. A possible source of error in the prediction model could be caused by EPID detector drift distorting the MPC results. To mitigate this source of uncertainty, throughout the study the EPID dark‐field calibration was updated monthly and pixel defect map annually or as dead pixels were identified. It is noted that MPC is independent of the flood field calibration. EPID response constancy has also been extensively studied in the literature. The short‐ and long‐term dose–response reproducibility of amorphous Silicon EPIDs has been found to be consistently within 1.0 % and 0.5 %, respectively, for all models once linac output variation from nominal had been accounted. , , , , A study by King et al pertained specifically to the long‐term pixel stability of Varian aS500 EPID panels. In King's study it was found that over a 3‐year period mean pixel variations were between 0.29 and 0.6 % and that more than 99% of all pixels showed variations less than 1 %. Also the MPC studies of provided comparison of MPC tests against independent measurements over the medium term with good agreement generally observed. An exception to this was for the output parameter for which Barnes et al detected a divergence of MPC output vs. independent measures. Barnes et al suggest that this can be mitigated via regular intercomparison of MPC output with ion‐chamber readings and rebaselining MPC if a 1 % difference is detected. Such a measure, along with regular routine EPID calibration and QA, is recommended if the forecasting methodology of this study was implanted clinically. Maintenance and recalibration events are also a potential source of sudden systematic change in MPC results as demonstrated by Barnes et al. These events were not accounted for in this study. Such systematic changes will influence the prediction model and hence at such events the prediction model input data should be reassessed and model training data collection potentially restarted. Due to the nature of prediction models being based on large learning‐phase datasets the prediction models are not designed to detect large sudden one‐off jumps in data such as might be expected with a linac component failure. Linac interlocks are still required to mitigate treatment delivery errors from such events as well as routine retrospective QA. Predictive QA is more suited to detecting and forecasting gradual drifts and failures that repeat at regular intervals. The results of the present study suggest that the approach of predictive QA based on MPC data is feasible, but additional data on more linacs are required and the method needs to be tested further in terms of sensitivity for the system to be clinically useable. Such study is proposed as future work.

CONCLUSION

The concept of linac predictive QA with the MPC using SPC and the ARIMA forecast model was demonstrated and its accuracy and performance evaluated. A window of opportunity between SPC control limits and clinical tolerances based on TG‐142 was demonstrated, suggesting that the MPC is an appropriate tool for predictive QA. The concept can be developed further with a greater number of linacs, sensitivity testing, and the evaluation of other predictive model techniques. Such testing thus has the potential to reduce the linac unscheduled downtime and allow linac performance parameters to be controlled within tolerances tighter than those typically applied in the clinic.

CONFLICT OF INTEREST

The authors have no conflict of interest to disclose.
  23 in total

1.  Dosimetric properties of an amorphous silicon electronic portal imaging device for verification of dynamic intensity modulated radiation therapy.

Authors:  Peter B Greer; Carmen C Popescu
Journal:  Med Phys       Date:  2003-07       Impact factor: 4.071

2.  Quality assurance for image-guided radiation therapy utilizing CT-based technologies: a report of the AAPM TG-179.

Authors:  Jean-Pierre Bissonnette; Peter A Balter; Lei Dong; Katja M Langen; D Michael Lovelock; Moyed Miften; Douglas J Moseley; Jean Pouliot; Jan-Jakob Sonke; Sua Yoo
Journal:  Med Phys       Date:  2012-04       Impact factor: 4.071

3.  Dose-response characteristics of an amorphous silicon EPID.

Authors:  Peter Winkler; Alfred Hefner; Dietmar Georg
Journal:  Med Phys       Date:  2005-10       Impact factor: 4.071

4.  Task Group 142 report: quality assurance of medical accelerators.

Authors:  Eric E Klein; Joseph Hanley; John Bayouth; Fang-Fang Yin; William Simon; Sean Dresser; Christopher Serago; Francisco Aguirre; Lijun Ma; Bijan Arjomandy; Chihray Liu; Carlos Sandin; Todd Holmes
Journal:  Med Phys       Date:  2009-09       Impact factor: 4.071

5.  SU-E-T-207: Flatness and Symmetry Threshold Detection Using Statistical Process Control.

Authors:  C Able; C Hampton; A Baydush
Journal:  Med Phys       Date:  2012-06       Impact factor: 4.071

6.  The long-term stability of amorphous silicon flat panel imaging devices for dosimetry purposes.

Authors:  R J W Louwe; L N McDermott; J J Sonke; R Tielenburg; M Wendling; M B van Herk; B J Mijnheer
Journal:  Med Phys       Date:  2004-11       Impact factor: 4.071

7.  Initial investigation using statistical process control for quality control of accelerator beam steering.

Authors:  Charles M Able; Carnell J Hampton; Alan H Baydush; Michael T Munley
Journal:  Radiat Oncol       Date:  2011-12-28       Impact factor: 3.481

8.  AAPM Medical Physics Practice Guideline 8.a.: Linear accelerator performance tests.

Authors:  Koren Smith; Peter Balter; John Duhon; Gerald A White; David L Vassy; Robin A Miller; Christopher F Serago; Lynne A Fairobent
Journal:  J Appl Clin Med Phys       Date:  2017-05-26       Impact factor: 2.102

9.  Evaluation of the truebeam machine performance check (MPC) geometric checks for daily IGRT geometric accuracy quality assurance.

Authors:  Michael P Barnes; Peter B Greer
Journal:  J Appl Clin Med Phys       Date:  2017-03-22       Impact factor: 2.102

10.  Evaluation of the truebeam machine performance check (MPC): mechanical and collimation checks.

Authors:  Michael P Barnes; Peter B Greer
Journal:  J Appl Clin Med Phys       Date:  2017-04-17       Impact factor: 2.102

View more
  1 in total

1.  Predicting machine's performance record using the stacked long short-term memory (LSTM) neural networks.

Authors:  Min Ma; Chenbin Liu; Ran Wei; Bin Liang; Jianrong Dai
Journal:  J Appl Clin Med Phys       Date:  2022-02-16       Impact factor: 2.102

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.