Literature DB >> 26103498

AutoLock: a semiautomated system for radiotherapy treatment plan quality control.

Joseph M Dewhurst1, Matthew Lowe, Mark J Hardy, Christopher J Boylan, Philip Whitehurst, Carl G Rowbottom.   

Abstract

A semiautomated system for radiotherapy treatment plan quality control (QC), named AutoLock, is presented. AutoLock is designed to augment treatment plan QC by automatically checking aspects of treatment plans that are well suited to computational evaluation, whilst summarizing more subjective aspects in the form of a checklist. The treatment plan must pass all automated checks and all checklist items must be acknowledged by the planner as correct before the plan is finalized. Thus AutoLock uniquely integrates automated treatment plan QC, an electronic checklist, and plan finalization. In addition to reducing the potential for the propagation of errors, the integration of AutoLock into the plan finalization workflow has improved efficiency at our center. Detailed audit data are presented, demonstrating that the treatment plan QC rejection rate fell by around a third following the clinical introduction of AutoLock.

Entities:  

Mesh:

Year:  2015        PMID: 26103498      PMCID: PMC5690112          DOI: 10.1120/jacmp.v16i3.5396

Source DB:  PubMed          Journal:  J Appl Clin Med Phys        ISSN: 1526-9914            Impact factor:   2.102


INTRODUCTION

In modern radiotherapy there is continuing emphasis on quality management to improve patient safety and reduce errors. , , A key component of radiotherapy quality assurance (QA) is independent checking of treatment plans. This quality control (QC) typically involves independent evaluation of plan quality and verification of technical and safety critical plan details, and has been found to be one of the most effective tools for error prevention. As radiotherapy moves towards more complex treatment delivery techniques, such as intensity‐modulated radiotherapy (IMRT) and volumetric‐modulated arc therapy (VMAT), the scope and complexity of treatment plan QC is increasing. The appropriate application of automation may be a useful tool to deal with some of these additional pressures while also reducing the potential for errors. Indeed, many elements of treatment plan QC are well suited to automation because they are time‐consuming and difficult for a human to perform but can be checked in a straightforward manner and quickly by computational procedures. Efforts to automate some aspects of treatment plan QC have developed in parallel with efforts to automate treatment planning itself. For example, Siochi et al. developed a system to automatically compare treatment planning system (TPS) data with record and verify system (R&V) data. Furhang et al. described a system that checks automatically the agreement between diagnosis, prescription, plan, and R&V values (intraplan review) and compares a treatment plan with other similar plans (interplan review). Yang and Moore developed a comprehensive system that checks automatically the physical and dosimetric integrity of treatment plans. Yang et al. described a system that automatically performs patient chart checking, and Halabi and Lu described a versatile system with the ability to parse TPS output PDF reports. While many elements of treatment plan QC are objective and well suited to automation, some are more subjective or difficult to fully automate and better suited to human evaluation. In such cases, a checklist can be a useful tool. Checklists are an established tool in error management, and are increasingly being applied in the health‐care setting, following UK and international guidelines. , , For example, the clinical implementation of electronic checklists in radiotherapy has been described by Albuquerque et al. and Greenwalt et al. to improve patient safety and reduce errors. Breen and Zhang described the clinical implementation of an automated checklist for head‐and‐neck IMRT treatment planning. However, its effectiveness was limited by low compliance. The ideal treatment plan QC system would accommodate both objective and subjective checks, apply comprehensively to the full breadth of treatment planning, and be fully integrated into the treatment planning workflow. In this paper, we describe the development and clinical implementation of a semiautomated system designed for this purpose, named AutoLock. AutoLock is intended to augment treatment plan QC by automatically checking aspects of treatment plans that are well suited to computational evaluation, whilst summarizing more subjective aspects in the form of a checklist. The principle is to automate as much as possible, while also acting as a filter to condense the large amount of information requiring human evaluation into a manageable checklist. The treatment plan must pass all automated checks, and all checklist items must be actively acknowledged by the planner as correct before the plan is finalized. Thus AutoLock uniquely integrates automated treatment plan QC, an automated checklist, and plan finalization. In addition to reducing the potential for the propagation of errors, the integration of AutoLock into the plan finalization workflow is intended to improve efficiency at our center. By performing automated checks as part of treatment plan finalization, any errors detected can be corrected at a point in the workflow where this is straightforward to do so — such as before the plan is finalized by the planner — thereby reducing the downstream workload. With this in mind, we present detailed audit data demonstrating a reduction in the rate of treatment plan QC rejections following the clinical introduction of AutoLock.

MATERIALS AND METHODS

Treatment planning workflow

Figure 1 shows a simplified version of the AutoLock‐assisted treatment planning workflow and review process at our center. AutoLock is used by the planner to review, approve and finalize the treatment plan before it is independently checked and reviewed by the clinician. AutoLock is positioned as far upstream in the workflow as possible to help avoid the propagation of errors.
Figure 1

Simplified treatment planning workflow at our center and the role of AutoLock within it.

Simplified treatment planning workflow at our center and the role of AutoLock within it.

Program design

AutoLock is written in Java and integrated into the Philips Pinnacle3 TPS using Pinnacle scripting (Philips Healthcare, Andover, MA). (See Yang and Moore for a detailed discussion of a similar technique.) Pinnacle scripting is a powerful feature of the TPS, which allows sequences of user interactions to be automated. In addition, external UNIX commands can be executed, allowing Pinnacle scripts to be combined with a high‐level programming language such as Java. The general principle behind AutoLock is that the scripting functionality of Pinnacle is used to store plan information in a text file; this plan information file is then processed independently of the TPS using logic statements written in Java. Check tolerances and settings are stored in a separate configuration file. In principle, this technique could be applied to any TPS with suitable scripting capabilities. AutoLock is split into two separate modules: Review and LockPlan. The Review module generates a plan information file, performs checks using this file in conjunction with the configuration file, and displays the check results to the user (see Fig. 2). If the results are then accepted by the user, the details of the review process are stored in a temporary HTML file called the “Plan Lock Sheet”.
Figure 2

AutoLock Review module workflow.

AutoLock Review module workflow. Once a plan has been reviewed and accepted by the planner, the LockPlan module is used to finalize the plan. The LockPlan module first checks that the plan has not changed since successful review, by regenerating the plan information file and comparing a checksum generated using this file with a checksum generated using the original plan information file. If changes are detected, the user is prompted to rerun the Review module. If no changes are detected, the plan and HTML Plan Lock Sheet are finalized (see Fig. 3), or “locked”. The Plan Lock Sheet acts as a permanent record of the planner review and finalization process and can be retrieved as necessary — for example, during independent checking of the plan.
Figure 3

AutoLock LockPlan module workflow.

AutoLock LockPlan module workflow.

Check result taxonomy

Plan check results can be “Pass”, “Fail”, “For review”, “Not checked”, or “Error” (see Table 1). Each result is accompanied by a textual description, which explains the details of the result to the user.
Table 1

AutoLock check result taxonomy and meanings.

Result Meaning
PassThe check has passed
FailA problem has been detected with the treatment plan
For reviewNo specific problems have been detected but the result must be reviewed and acknowledged as correct by the user
Not checkedThe check is not required for the current treatment plan
ErrorA programming error has occurred
AutoLock check result taxonomy and meanings. “Pass” and “Fail” results occur only where a check can be fully automated, and subjectivity is handled with “For review”. “For review” results occur where the check cannot be fully automated or human evaluation is required; a summary of such results is automatically added to a checklist for the planner to review. To allow full flexibility in the system, “Fail” results can also be acknowledged and thereby overridden by the planner. The details of an AutoLock check can vary depending on the type of plan, and in some cases a check may not apply, in which case a “Not checked” result occurs. “Pass” and “Not checked” results do not require user acknowledgment. As stated above, the general principle applied in AutoLock is to automate as much as possible (i.e., maximize the number of “Pass” and “Fail” results and minimize the use of “For review”). It should be noted that a “For review” result does not necessarily imply that no automation has occurred; rather, as much automation as possible has occurred. In this way, AutoLock effectively acts as a filter to condense the large amount of information requiring human evaluation into a manageable checklist. For completeness, the possibility of an “Error” result is included and indicates an unexpected programming event. As with “Fail” results, “Error” results can be overridden by the planner to allow flexibility but would, in practice, be thoroughly investigated before proceeding any further. All check results and acknowledgements appear on the Plan Lock Sheet, which can be reviewed during independent checking.

User interface

Figure 4 shows a screenshot of the AutoLock graphical user interface (GUI), which appears to the planner upon running the AutoLock Review module. Results are split into tabs depending on the check result; Fig. 4 shows the “For review” tab. Any result requiring acknowledgement is acknowledged by completing a tick box. Comments can be added to each item as required by the planner and appear on the Plan Lock Sheet.
Figure 4

Screenshot of the “For review” tab in the AutoLock graphical user interface; each checklist item must be acknowledged by the user by completing a tick box.

Screenshot of the “For review” tab in the AutoLock graphical user interface; each checklist item must be acknowledged by the user by completing a tick box. Only once all items requiring acknowledgement have been acknowledged can the review be accepted in the “Accept” tab. This data completeness check means an active response, in the form of completing tick boxes, is required before the review can be accepted. Once the review has been accepted, the plan can be finalized using the LockPlan module. Figure 5 shows an example of the Plan Lock Sheet; as with the GUI, results are organized depending on the check result.
Figure 5

Extract from the Plan Lock Sheet, which shows all results of the plan review, split by result type.

Extract from the Plan Lock Sheet, which shows all results of the plan review, split by result type.

Checks

Plan checks have been split up into discrete checks, which are given a unique number and assigned to one of four categories: 1) General plan checks; 2) Prescription checks; 3) Beam checks; 4) Control point checks. There are currently 31 checks implemented in AutoLock, which includes checks of prescription, dose calculation parameters, density overrides, beam labeling, and positions (see 2, 3, 4, 5). The implemented checks were chosen based on a review of practice, an audit of treatment plan QC rejection reasons, a review of the literature, , , , , , and a survey of checkers. It should be noted that many of the checks are tailored specifically to working practice at our center with the Pinnacle TPS. However, AutoLock is designed as an expandable model and, therefore, more or different checks can be added in a straightforward manner.
Table 2

AutoLock general plan checks.

ID Name Description
1.1Single trial named finalPlan has only one trial and is labeled correctly
1.2Couch removedCouch is removed (does not check couch removal position)
1.3Dose grid voxel size 0.3 cm or lessDose grid voxel size is set appropriately
1.4Air threshold is 0.81g/cm3 or lessPatient–air threshold is set appropriately
1.5Origin and isocenter POI checkOrigin POI matches localizer, origin and isocenter POIs are correctly formatted
1.6Overrides checkOverrides are set appropriately; a list of overrides present is displayed for the user to review
1.7Plan name lengthPlan name length is less than required for DICOM export
1.8ROIs inside dose gridAll ROIs are covered entirely by dose grid (exceptions can be set in the configuration file)
1.9Has the MU checker run?For 3D conformal plans, prompts the planner to check the plan runs through the independent monitor unit checking software
1.10Plan comment contains finalPlan is labeled correctly
1.11Correct version of PinnacleCurrent clinical version of Pinnacle is in use
Table 3

AutoLock prescription checks.

ID Name Description
2.1Prescription dosePrescription options are set correctly; the prescription dose and number of fractions are displayed for the user to review
2.2Prescription methodCorrectly prescribed to point dose or ROI mean (and not monitor units); the prescription point or ROI is displayed for the user to review
2.3Weights proportional toBeam weights are correctly set proportional to monitor units or point dose
2.4IsodosesIsodose lines set appropriately; a list of isodoses present is displayed for the user to review
2.5Max point dose less than or equal to 110%Flag for potential hotspot issues
Table 4

AutoLock beam checks.

ID Name Description
3.1Consistent machine and timestampAll beams are set to the same beam model and timestamp matches the quality system; the beam model in use is displayed for the user to review
3.2Consistent beam typesAll beams are set to the same type (e.g., conventional, IMRT or VMAT)
3.3All beams set to AC or CCCAll beams are set to either adaptive convolve or collapsed cone convolution
3.4Consistent heterogeneityAll beams are set to heterogeneous or homogeneous and appropriate CT‐density table is in use
3.5All beams set to consistent isocenterAll beams set to the same isocenter
3.6All beams set to consistent prescriptionAll beams set to the same prescription unless multiple prescriptions are present, in which case the user must review
3.7All beams have a field ID enteredAll beams have a field ID entered to enable unique identification in the R&V system, regardless of beam name
3.8Beam labels checkBeam labels are correct
3.9In vivo points check In vivo dosimetry POIs have been generated correctly (name, at depth dmax, and position within field)
Table 5

AutoLock control point checks.

ID Name Description
4.1No control points with less than 3 MUsNo control points have less than the minimum 3 monitor units
4.2Y jaw backupY jaws are backed up correctly
4.3X jaw backupX jaws are backed up correctly
4.4Elekta breast IMRT back edge checkBack edge jaw position is consistent for Elekta breast IMRT plans
4.5No blocks present for plans not using MLCNo blocks are present for plan not using the MLC
4.6Varian CAX matchlinesNo matchlines present near the central axis for Varian step‐and‐shoot IMRT plans
AutoLock general plan checks. AutoLock prescription checks. AutoLock beam checks. AutoLock control point checks.

Audit

AutoLock log file audit

AutoLock log files were collected for plans produced in a six‐month period since clinical introduction at our center. Changes in check results between the first time AutoLock was run and last time AutoLock was run were counted. Any check result changing from “Fail” to “Pass”, “Fail” to “For review”, or “For review” to “Pass” was considered to indicate a positive AutoLock prompted action. The percentage action rate was calculated over all plans, on a month‐by‐month basis, and for each check individually.

Treatment plan QC rejection audit

To assess the broader impact of AutoLock on the workflow, treatment plan QC rejections were compared before and after clinical introduction. At our center, each time a treatment plan is rejected at the independent checking stage, it is assigned a code indicating the reason for audit purposes (see Table 6). These rejection codes were compared before and after the clinical introduction of AutoLock. The first month of audit data following the introduction of AutoLock was excluded from the analysis to avoid the initial ramp‐up phase. The remaining five months of data were compared with the five months immediately prior to the introduction of AutoLock.
Table 6

Thirteen‐category treatment plan QC rejection coding system.

ID Name Description
1ContouringROI delineation and expansion
2Density correctionDensity overrides
3DocumentationMissing or incorrect information in plan printout
4Dose calculationBeam model, dose calculation, voxel grid size, couch removal
5InVivo points In vivo dosimetry POIs
6Jaw/MLCJaw backup, control point constraints
7Modification of treatment by clinicianPatient setup, ill‐fitting shell, changes to patient, PTV, etc.
8OptimizationCoverage, OAR doses, beam energy, shielding
9OriginLocalization
10PrescriptionPrescription and method
11Wedge angleWedge angle not permitted in quality system
12Wrong linacBeam model does not match linac booking
13OtherReason not described in system; recurrent reasons can be reviewed for inclusion via an additional category
Thirteen‐category treatment plan QC rejection coding system.

RESULTS

AutoLock was commissioned and introduced into routine clinical use at our center in January 2014. The commissioning process involved independent testing using a suite of test data containing deliberate errors, independent code verification, independent beta, and usability testing and training. Over the period of our audit, AutoLock was used to finalize 2,384 Pinnacle treatment plans, out of a total of 2,487 (i.e., 96%). Of the 2,384 plans finalized with AutoLock, 1,585 were 3D conformal or forward‐planned IMRT and 799 were inverse‐planned IMRT or VMAT. The majority of plans finalized without AutoLock were produced during January and February, during the initial ramp‐up phase. Compliance from March onwards was close to 100%, as demonstrated by the dashed line in Fig. 6.
Figure 6

AutoLock compliance rate and action rate as a function of time, showing usage increased to 100% over the audit period and an upwards trend in action rate.

AutoLock compliance rate and action rate as a function of time, showing usage increased to 100% over the audit period and an upwards trend in action rate. The overall action rate for all plans finalized with AutoLock was 28%. The action rate was 27% for 3D conformal and forward‐planned IMRT and 30% for inverse‐planned IMRT or VMAT. The solid line in Fig. 6 shows the action rate over all plans finalized with AutoLock on a month‐by‐month basis. Figure 7 shows the action rates for each individual check over all plans finalized with AutoLock.
Figure 7

AutoLock action rates for individual checks.

AutoLock action rates for individual checks. Figure 8 shows rejections code rates before and after the introduction of AutoLock, as well as rejections not assigned a rejection code. The overall treatment plan QC rejection rate was 32% lower after the introduction of AutoLock.
Figure 8

Treatment plan QC rejection codes before and after the introduction of AutoLock.

Treatment plan QC rejection codes before and after the introduction of AutoLock.

DISCUSSION

The results in 6, 7, 8 represent a fascinating insight into the impact of AutoLock, as well as the relative frequency of issues encountered during treatment plan QC. Figure 6 shows that the compliance rate rose to 100% over the period of our audit, considerably higher than achieved in the study by Breen and Zhang. This may be explained by a number of factors. Firstly, AutoLock is fully integrated into the treatment planning workflow and acts as the finalization process. This integration has helped AutoLock to quickly become established as routine treatment planning practice; indeed, since performing this audit, finalization with AutoLock has become mandated in our quality management system. Secondly, anecdotal evidence suggests that AutoLock has become a popular tool amongst planners as it is highly user‐friendly and provides instant, impersonal feedback. User friendliness was a key consideration in the design of AutoLock, as it is vital to the safe and effective use of software. Figure 6 also demonstrates a general upwards trend in the action rate, which may be explained by two factors. Firstly, AutoLock may be run at any point during the planning process and, therefore, the first time AutoLock is run, may not correspond to when a planner is ready to finalize a plan. The rising action rate suggests that planners may be running AutoLock earlier in the planning process. A second possibility is that planners familiar with AutoLock may have started to use it to catch issues that they would have checked manually had AutoLock been unavailable. In either case, this is suggestive of a workflow improvement, as planners may well be spending less time manually checking treatment plan details and instead utilizing the automation provided by AutoLock. Quantification of time savings for planners and checkers will be tackled with future work. Figure 7 shows that action rates vary considerably between different plan checks. It should be noted that these action rates are likely to represent significant overestimates of the planning error rates that would have occurred had AutoLock not been available, for the same two reasons described in the previous paragraph. Nevertheless, these results suggest that AutoLock has had a positive impact at our center by prompting planners to correct issues before plan finalization. This is supported by the observed 32% reduction in the rate of treatment plan QC rejections following the introduction of AutoLock; a similar reduction was found in by Breen and Zhang. The impact of AutoLock would be further appreciated by showing a correlation between the results in Fig. 7 and Fig. 8 (i.e., between AutoLock action rates and QC rejection reasons). However, this is complicated by a number of issues. Firstly, a significant fraction of QC rejections have not been assigned a rejection code. Secondly, some AutoLock checks apply to multiple rejection codes and vice‐versa, and in some cases the correspondence is only partial. For example, check number 1.2 checks that the couch has been removed but does not check the position of the couch removal, which is left for human evaluation. The effects of this check would be observed in rejection category 4 (dose calculation). However, this category also encompasses a number of other AutoLock checks including 1.3, 1.4, 1.8, 3.1, 3.2, and 3.3 and there is no apparent reduction in the rate of rejections for this category. However, there is a more direct correspondence with other categories. For example, in category 2 (density correction) a rejection rate reduction of around 70% was observed. This category closely corresponds to check number 1.6 which, amongst other things, displays the overrides present in the plan for the user to review. A rejection rate reduction of around 70% was also observed for category 9 (origin), which closely corresponds to check number 1.5, which checks coincidence of the origin POI and localizer, while positioning of the localizer is left for human evaluation. Smaller reductions in rejection rates occur in all other categories where AutoLock would be expected to have an impact: 3 (documentation), 5 (InVivo points), 6 (jaw/MLC), 10 (prescription), and 12 (wrong linac). These results will direct the future development of AutoLock checks and the rejection coding system as part of the continuous cycle of quality improvement at our center. Aside from these results, many of the benefits of AutoLock are difficult to measure. For example, automated systems like AutoLock are particularly suited to catching infrequent errors that a human checker may miss on account of their infrequency. Indeed, a number of AutoLock checks showed an action rate of or very close to 0%. Nevertheless, there is still value in including such checks, particularly as the only cost to the department is in the initial development and commissioning process; every subsequent automated check is effectively “free” to the department. Many of the checks implemented in AutoLock are designed specifically to improve workflow at our center. For example, the check with the highest action rate — checking that all beams have a “field ID” entered — is a relatively minor issue, but from a workflow point of view avoids potential downstream issues when the plan is imported into the R&V system. Smoothing the workflow is a key benefit of AutoLock because many small plan details can be checked automatically before plan finalization, thus increasing the consistency of treatment plan generation. Furthermore, one would expect that the reduction of variability will lead to greater prospects for quality improvement. Indeed, the intention behind AutoLock is to automate as much treatment plan QC as possible, and thereby allow the planner and checker to focus their attention on overall treatment plan quality and those safety critical details that are difficult or impossible to fully automate. More work is required to investigate whether a similar approach can be beneficial in areas of the complex radiotherapy process other than treatment planning.

CONCLUSIONS

We have developed and clinically implemented a semiautomated system for treatment plan QC, named AutoLock. AutoLock is designed to augment treatment plan QC by automatically checking aspects of treatment plans that are well suited to computational evaluation, whilst summarizing more subjective aspects in the form of a checklist. The treatment plan must pass all automated checks and all checklist items must be acknowledged by the planner as correct before the plan is finalized. Thus AutoLock uniquely integrates automated treatment plan QC, an electronic checklist, and plan finalization. AutoLock provides instant, impersonal feedback to the planner and acts as a filter to condense the large amount of information requiring human assessment into a manageable checklist. In addition to reducing the potential for propagation of errors, the integration of AutoLock into the plan finalization workflow is intended to improve efficiency at our center. AutoLock has been in clinical use at our center since January 2014 and the treatment plan QC rejection rate fell by around a third following its introduction.
  10 in total

1.  Implementation of electronic checklists in an oncology medical record: initial clinical experience.

Authors:  Kevin V Albuquerque; Alexis A Miller; John C Roeske
Journal:  J Oncol Pract       Date:  2011-06-17       Impact factor: 3.840

2.  Quality control quantification (QCQ): a tool to measure the value of quality control checks in radiation oncology.

Authors:  Eric C Ford; Stephanie Terezakis; Annette Souranis; Kendra Harris; Hiram Gay; Sasa Mutic
Journal:  Int J Radiat Oncol Biol Phys       Date:  2012-06-09       Impact factor: 7.038

3.  Audit of an automated checklist for quality control of radiotherapy treatment plans.

Authors:  Stephen L Breen; Beibei Zhang
Journal:  Radiother Oncol       Date:  2010-10-09       Impact factor: 6.280

4.  Automated radiotherapy treatment plan integrity verification.

Authors:  Deshan Yang; Kevin L Moore
Journal:  Med Phys       Date:  2012-03       Impact factor: 4.071

5.  Vision 20/20: Automation and advanced computing in clinical radiation oncology.

Authors:  Kevin L Moore; George C Kagadis; Todd R McNutt; Vitali Moiseenko; Sasa Mutic
Journal:  Med Phys       Date:  2014-01       Impact factor: 4.071

6.  Technical note: electronic chart checks in a paperless radiation therapy clinic.

Authors:  Deshan Yang; Yu Wu; Ryan S Brame; Sridhar Yaddanapudi; Dharanipathy Rangaraj; H Harold Li; S Murty Goddu; Sasa Mutic
Journal:  Med Phys       Date:  2012-08       Impact factor: 4.071

7.  Recommendations for safer radiotherapy: what's the message?

Authors:  Peter Dunscombe
Journal:  Front Oncol       Date:  2012-09-28       Impact factor: 6.244

8.  Radiation therapy plan checks in a paperless clinic.

Authors:  R Alfredo Siochi; Edward C Pennington; Timothy J Waldron; John E Bayouth
Journal:  J Appl Clin Med Phys       Date:  2009-01-27       Impact factor: 2.102

Review 9.  Automating the initial physics chart checking process.

Authors:  Eli E Furhang; James Dolan; Jussi K Sillanpaa; Louis B Harrison
Journal:  J Appl Clin Med Phys       Date:  2009-02-11       Impact factor: 2.102

10.  Automating checks of plan check automation.

Authors:  Tarek Halabi; Hsiao-Ming Lu
Journal:  J Appl Clin Med Phys       Date:  2014-07-08       Impact factor: 2.102

  10 in total
  9 in total

Review 1.  Automated Plan Checking Software Demonstrates Continuous and Sustained Improvements in Safety and Quality: A 3-year Longitudinal Analysis.

Authors:  Delaney Stuhr; Ying Zhou; Hai Pham; Jian-Ping Xiong; Shi Liu; James G Mechalakos; Sean L Berry
Journal:  Pract Radiat Oncol       Date:  2021-10-17

2.  Strategies for effective physics plan and chart review in radiation therapy: Report of AAPM Task Group 275.

Authors:  Eric Ford; Leigh Conroy; Lei Dong; Luis Fong de Los Santos; Anne Greener; Grace Gwe-Ya Kim; Jennifer Johnson; Perry Johnson; James G Mechalakos; Brian Napolitano; Stephanie Parker; Deborah Schofield; Koren Smith; Ellen Yorke; Michelle Wells
Journal:  Med Phys       Date:  2020-04-15       Impact factor: 4.071

3.  Improving treatment plan evaluation with automation.

Authors:  Elizabeth L Covington; Xiaoping Chen; Kelly C Younge; Choonik Lee; Martha M Matuszak; Marc L Kessler; Wayne Keranen; Eduardo Acosta; Ashley M Dougherty; Stephanie E Filpansick; Jean M Moran
Journal:  J Appl Clin Med Phys       Date:  2016-11-08       Impact factor: 2.102

4.  Using failure mode and effects analysis (FMEA) to generate an initial plan check checklist for improved safety in radiation treatment.

Authors:  Prema Rassiah; Fan-Chi Frances Su; Y Jessica Huang; Dan Spitznagel; Vikren Sarkar; Martin W Szegedi; Hui Zhao; Adam B Paxton; Geoff Nelson; Bill J Salter
Journal:  J Appl Clin Med Phys       Date:  2020-06-25       Impact factor: 2.102

5.  Early detection of potential errors during patient treatment planning.

Authors:  Danielle Lack; Jian Liang; Lisa Benedetti; Cory Knill; Di Yan
Journal:  J Appl Clin Med Phys       Date:  2018-07-05       Impact factor: 2.102

6.  Toward automation of initial chart check for photon/electron EBRT: the clinical implementation of new AAPM task group reports and automation techniques.

Authors:  Huijun Xu; Baoshe Zhang; Mariana Guerrero; Sung-Woo Lee; Narottam Lamichhane; Shifeng Chen; Byongyong Yi
Journal:  J Appl Clin Med Phys       Date:  2021-03-11       Impact factor: 2.102

7.  Automation of DVH Constraint Checks and Physics Quality Control Review Improves Patient Safety in Radiotherapy.

Authors:  Isak Wahlstedt; Nikolaj Jensen
Journal:  J Med Phys       Date:  2021-12-31

8.  Combining automatic plan integrity check (APIC) with standard plan document and checklist method to reduce errors in treatment planning.

Authors:  Ping Xia; Danielle LaHurd; Peng Qi; Anthony Mastroianni; Daesung Lee; Anthony Magnelli; Eric Murray; Matt Kolar; Bingqi Guo; Tim Meier; Samual T Chao; John H Suh; Naichang Yu
Journal:  J Appl Clin Med Phys       Date:  2020-07-17       Impact factor: 2.102

9.  Efficiency and safety increases after the implementation of a multi-institutional automated plan check tool at our institution.

Authors:  Sean L Berry; Ying Zhou; Hai Pham; Sharif Elguindi; James G Mechalakos; Margie Hunt
Journal:  J Appl Clin Med Phys       Date:  2020-03-20       Impact factor: 2.102

  9 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.