Literature DB >> 33870293

Development of a Focused Cardiac Ultrasound Image Acquisition Assessment Tool.

Rosemary Adamson1,2, Amy E Morris1, Jessica Sun Woan1,2, Irene W Y Ma3, Daniel Schnobrich4, Nilam J Soni5,6.   

Abstract

Background: Focused cardiac ultrasound (FCU) is widely used by healthcare providers to answer specific questions about cardiac structure and function at the bedside. Currently, no widely accepted FCU image acquisition checklist exists to assess learners with varying skill levels from different specialties. Objective: The primary objective of this project was to develop a consensus-based FCU image acquisition checklist using a multispecialty group of point-of-care ultrasound (POCUS) experts.
Methods: The essential components of an FCU examination were identified on the basis of published recommendations from echocardiography and international ultrasound societies. A checklist of the essential components of an FCU examination was drafted. A panel of POCUS experts from different medical specialties in the United States and Canada was convened to vote on each checklist item by answering two questions: 1) Is this item important to include in a checklist of essential FCU skills applicable to any medical specialty? and 2) Should the learner be required to successfully complete this item to be considered competent? A modified Delphi approach was used to assess the level of agreement for each checklist item during four rounds of voting. Checklist items that achieved an agreement of 80% or greater were included in the final checklist.
Results: Thirty-one POCUS experts from seven different medical specialties voted on sixty-five items to be included in the FCU image acquisition assessment tool. The majority of POCUS experts (61%) completed all four rounds of voting. During the first round of voting, 59 items reached consensus, and after revision and revoting, an additional 3 items achieved 80% or greater consensus. A total of 62 items were included in the final checklist, and 57 items reached consensus as a requirement for demonstration of competency.
Conclusion: We have developed a multispecialty, consensus-based FCU image acquisition checklist that may be used to assess the skills of learners from different specialties. Future steps include studies to develop additional validity evidence for the use of the FCU assessment tool and to evaluate its utility for the translation of skills into clinical practice.
Copyright © 2020 by the American Thoracic Society.

Entities:  

Keywords:  bedside ultrasound; competency assessment; echocardiography; focused cardiac ultrasound; point-of-care ultrasound

Year:  2020        PMID: 33870293      PMCID: PMC8043312          DOI: 10.34197/ats-scholar.2020-0002OC

Source DB:  PubMed          Journal:  ATS Sch        ISSN: 2690-7097


Focused cardiac ultrasound (FCU) is a bedside examination performed to rapidly detect potentially significant valvular, hemodynamic, or structural abnormalities to guide appropriate early triage and management (1, 2). A broad range of healthcare providers with varying levels of clinical training, ranging from medical students to experienced subspecialty attending physicians, have begun to use FCU in recent years (3–8). Noncardiologists with basic FCU training can reliably obtain cardiac views in the assessment of acutely ill patients to identify clinically significant findings such as severe left ventricular systolic dysfunction or a large pericardial effusion (9–21). Though not a replacement for traditional transthoracic echocardiography, FCU can identify certain cardiac abnormalities with higher accuracy than physical examination alone (22–27). Furthermore, FCU is readily available in settings with limited resources or situations in which rapid diagnosis is crucial to guide immediate decision making, such as hemodynamically unstable patients or those in respiratory distress (28–33). As more providers have begun to use FCU, questions have been raised about training and assessment of competence of providers incorporating FCU into clinical care (1, 34–38). For cardiologists, competence in echocardiography is most often determined by board certification by the National Board of Echocardiography, which entails performing a specified number of echocardiograms, completing formal training in cardiovascular diseases, and passing a knowledge assessment (39). Although certification in medical procedures has traditionally been determined by performing a specified number of examinations, recent literature in different specialties has suggested that competence may be unrelated to numerical thresholds (40–42). Development of structured assessment tools reflects a broad movement in medical training toward the use of objective assessments to determine competence (43). Nielsen and colleagues developed one such tool for cardiologists to assess competence in echocardiography (44). Structured assessment tools to objectively assess competence in noncardiac applications of point-of-care ultrasound (POCUS) have been published (45–47). For FCU specifically, a few structured tools to assess skills in cardiac image acquisition have been published, but they have important limitations. Most important, currently available assessment tools represent the consensus of individual specialties or institutions, which limits their generalizability. Furthermore, current cardiac ultrasound skill assessment tools broadly evaluate image acquisition (i.e., do not define individual components of an adequate cardiac image acquisition), which limits their ability to assess performance of specific skills (46, 48–51). Given these limitations, currently available FCU image acquisition assessment tools have not been widely adopted across institutions for assessing the competence of noncardiologists. Development of a standardized FCU image acquisition assessment tool has important implications for an increasing number of specialties, including critical care medicine, anesthesiology, and emergency medicine, that require FCU training during residency or fellowship per the American College of Graduate Medical Education, as well as for an increasing number of medical schools that are teaching students how to perform FCU examinations (3, 7). The purpose of this study was to develop a multispecialty, consensus-based FCU image acquisition checklist that includes the specific skills needed to acquire the five common FCU views. A structured and objective FCU image acquisition checklist is a critical step toward developing a standardized competency assessment tool for learners with varying skills within different specialties.

Methods

We conducted a literature review to identify the key components of an FCU examination. Sixty-four separate components of an FCU examination were identified by three of the authors (R.A., A.E.M., and J.S.W.), drawn from international ultrasound and echocardiography society guidelines, previously published checklists and curricula, and frequently cited textbooks and educational websites (1, 28, 44, 46, 52). We were intentionally inclusive in our approach, and these items were reviewed by all the authors to ensure that the list was complete. Through a series of conference calls and e-mail correspondences, the FCU examination components were organized and incorporated into the initial checklist with the following sections: 1) ultrasound knowledge and ultrasound machine functions; 2) scanning techniques; and 3) key anatomic structures and features of an optimal image for the parasternal long axis (PLAX), parasternal short axis (PSAX), apical four-chamber (A4C), subcostal four-chamber (S4C), and subcostal inferior vena cava (IVC) views. Between August 2016 and February 2017, a multispecialty group of POCUS experts was invited to participate on the expert panel. Experts were defined as individuals who teach POCUS courses at the local, national, or international level and/or have been published on the subject and who regularly use FCU in their clinical practice. A target of at least 20 panel members from a diverse range of clinical specialties that use POCUS was sought. All expert panel members were required to disclose any conflicts of interest before inclusion on the panel, and any conflicts of interest were adjudicated by the project leaders. This project was reviewed by the University of Minnesota Institutional Review Board, which determined the study to be exempt from full institutional review board review. The checklist items were entered in an internet-based electronic data collection tool (Research Electronic Data Capture [REDCap]) hosted on the server of the University of Texas Health Science Center in San Antonio, Texas (see Appendix E1 in the data supplement) (53). The checklist was circulated to all expert panel members for voting. For each item, panel members were asked two yes-or-no questions with skip logic incorporated to ease the burden of answering questions: 1) Is the item important enough to include in a checklist of essential FCU skills? and 2) Should the item be required for a learner to be considered competent? Panel members were encouraged to provide any relevant comments while voting on each item. Four rounds of voting were conducted between April 2017 and January 2018. A modified Delphi approach was used to determine consensus on each item to be included in the FCU image acquisition checklist. If at least 80% of respondents voted to accept or reject an item, the panel’s voting was accepted as consensus, and the item was either accepted or rejected from inclusion, respectively, on the checklist. If the panel’s agreement failed to reach 80%, the item was revised on the basis of the panel’s comments and reconsidered during up to two additional rounds of voting. After the first round of voting, respondents were informed of the level of consensus in the prior round of voting for each item included. In the final round of voting, respondents could select from among a prespecified list of reasons why they did not agree with the item, or they could offer their own explanation (Figure 1).
Figure 1.

Modified Delphi consensus method.

Modified Delphi consensus method. The authors held a conference call after the completion of each round of voting to review the results. The voting results, including all free-text comments, were shared by e-mail with preliminary discussion of the results before the conference calls. During the conference calls, all free-text comments were reviewed, and the authors discussed whether and how to modify checklist items on the basis of the comments. Any changes to checklist items were shared by e-mail after each conference call to confirm consensus on the new wording before the next round of voting.

Results

A total of 33 POCUS experts were invited to participate on the expert panel, and 31 participated in at least one round of voting. The expert panel consisted of physicians specializing in anesthesiology, cardiology, critical care, emergency medicine, internal medicine, pediatrics, and pulmonology. More than half of the experts had completed 1 year of fellowship training in ultrasound, had been practicing FCU for at least 8 years, and teaching FCU for at least 5 years (Table 1). Response rates of the expert panel were 84% in round 1, 87% in round 2, 84% in round 3, and 74% in round 4. A total of 19 experts (61%) completed all four rounds of voting.
Table 1.

Demographics of the expert panel

Characteristicn (%)
Specialty 
 Internal medicine and/or pediatrics9 (29)
 Emergency medicine6 (19)
 Critical care and pulmonology6 (19)
 Cardiology4 (13)
 Anesthesiology2 (6)
 Critical care and emergency medicine2 (6)
 Critical care and anesthesiology2 (6)
Sex, F13 (42)
U.S. region or Canadian province 
 Northeast (New York, Pennsylvania)6 (19)
 South (Georgia, North Carolina, South Carolina, Texas)6 (19)
 Midwest (Illinois, Kansas, Michigan, Minnesota)4 (13)
 West (California, Oregon, Washington)4 (26)
 Alberta4 (13)
 Nova Scotia1 (3)
 Ontario1 (3)
 Quebec1 (3)
Completed ultrasound fellowship 
 Yes18 (58)
Experience using focused cardiac ultrasound 
 0–4 yr2 (6)
 5–8 yr13 (42)
 >8 yr16 (52)
Experience teaching focused cardiac ultrasound 
 0–4 yr10 (32)
 5–8 yr9 (29)
 >8 yr12 (39)
Years assessing focused cardiac ultrasound skills 
 0–4 yr12 (39)
 5–8 yr9 (29)
 >8 yr10 (32)
Ultrasound-related peer-reviewed publications 
 0–520 (65)
 6–155 (16)
 >156 (19)
Voting 
 Voted in all 4 rounds19 (61)
 Voted in 3 rounds6 (19)
 Voted in 2 rounds2 (6)
 Voted in 1 round only4 (13)
Demographics of the expert panel

Round 1 Voting

A table summarizing all voting results is presented in Appendix E2. The initial FCU image acquisition checklist contained 64 items divided into seven sections: basic ultrasound machine settings, cardiac ultrasound setup, and the five most common cardiac ultrasound views (PLAX, PSAX, A4C, S4C, and subcostal IVC). After the first round of voting, 50 checklist items (78%) achieved 80% or greater agreement as being both important to include in the checklist and required for competency. Of the remaining 14 checklist items, 9 items achieved 80% or greater consensus as being important to include in the checklist but not required for competency. The remaining five checklist items did not achieve consensus for being either important to include (or exclude) on the checklist or required for competency.

Round 2 Voting

The nine checklist items from round 1 that did not reach consensus as a requirement for competency were modified on the basis of expert panel comments for subsequent voting. One checklist item that stated, “measures the largest and smallest diameters of IVC,” was separated into two items: “assesses the largest and smallest diameters of IVC qualitatively using eyeballing” and “measures the largest and smallest diameters of IVC quantitatively using calipers.” Thus, round 2 voting included a total of 10 checklist items. Two additional items reached consensus for inclusion on the checklist (items 30 and 40). Three items achieved consensus for being important but not required for competency (items 6, 55, and 62). Five checklist items on quantitative assessment of IVC diameter failed to reach consensus as being either important or required for competency (items 1 and 7; draft checklist items F5b, F6, and F7; listed in Table 2).
Table 2.

Results of checklist items not achieving consensus during round one voting

Proposed Checklist ItemModified Checklist ItemImportant to Include? (Round)*Required for Competency? (Round)*
1. Enters appropriate identifying information (e.g., provider name, patient name, and medical record number)A1. Enters appropriate identifying information if applicable (e.g., provider name, patient name, and medical record number)Yes (1)No
6. Places patient in supine or left lateral decubitus positionB1. Places patient in optimal position to obtain parasternal long-axis views, typically supine or left lateral decubitus position, when feasibleYes (2)Yes (4)
7. Begins examination with transducer located to left of sternum at approximately the fourth intercostal spaceB2. Begins examination with transducer placed left of sternum, approximately midsternal levelYes (1)No
10. Correctly identifies the right ventricular outflow tractB5. Correctly identifies the right ventricleYes (1)Yes (4)
23. Left ventricle is approximately centered on screen.Yes (1)No
28. Positions patient appropriately: left lateral decubitus position or supine. If unable to obtain adequate views in supine position, moves to left lateral decubitus position.Yes (1)Yes (4)
30. Cardiac apex is centered on screen.D3. Attempts to place cardiac apex approximately in center of screenYes (1)Yes (2)
31. Interventricular septum is approximately vertical.Yes (1)Yes (4)
40. Places patient in the supine position. If unable to obtain adequate views, bends knees to relax abdominal muscles.E1. Places patient in optimal position to obtain subcostal views, typically supine position. If unable to obtain adequate views, repositions patient (e.g., bends knees to relax abdominal muscles).Yes (1)Yes (2)
51. Places transducer subxiphoid in cephalad–caudad orientation with transducer marker toward patient’s headF1. Places transducer subxiphoid in cephalad–caudad orientationYes (1)Yes (4)
55. Measures the largest and smallest diameters of IVCF5a. Assesses the largest and smallest diameters of IVC qualitatively using eyeballingYes (2)No
 F5b. Measures the largest and smallest diameters of IVC quantitatively using calipersNoNo
Draft item F6. When using M-mode, places caliper at approximately 90 degrees to IVCF6. When using M-mode, attempts to place calipers approximately perpendicular to IVCNoNo
Draft item F7. When using M-mode, measures largest and smallest diametersF7. Measures the largest and smallest diameters of IVC using M-mode and calipersNoNo
62. Digitally stores and archives ultrasound images.G3. Digitally stores and archives ultrasound images if applicableYes (2)No

Numbers in parentheses indicate the voting round in which consensus was achieved.

Based on two rounds of voting.

Excluded from checklist because less than 20% of experts voted to include this item in round 3.

Results of checklist items not achieving consensus during round one voting Numbers in parentheses indicate the voting round in which consensus was achieved. Based on two rounds of voting. Excluded from checklist because less than 20% of experts voted to include this item in round 3.

Round 3 Voting

Round 3 included five items. Two items on general machine use had reached consensus as being important but not required for competency: “enters appropriate identifying information” (item 1) and “digitally stores and archives ultrasound images” (item 7). Several experts commented on technical and institutional limitations that can affect storage and archiving capability unrelated to an operator’s skills. Despite clarifying the wording of these two checklist items (Table 2), 77% and 50% of experts voted in favor of requiring items 1 and 7, respectively, for competency in round 3. Thus, items 1 and 7 are included in the final checklist as being important but not required for competency. The other three items in round 3 voting pertained to quantitative assessment of IVC diameter, but none of these items reached consensus as a requirement for competency (“measures IVC diameters with calipers” [draft checklist item F5b], “places calipers perpendicular to IVC” [draft checklist item F6], and “measures IVC diameter in M-mode with calipers” [draft checklist item F7]). All three items were therefore excluded from the final checklist. During data analysis after voting in round 3, it was discovered that some checklist items had duplicate responses from the same expert panel member. After removing duplicate responses, five checklist items that had reached consensus in round 1 were now below the 80% threshold to be required for competency: item 10 (76%), item 23 (72%), item 28 (72%), item 31 (76%), and item 51 (79%). Three checklist items from round 2 voting were similarly affected: item 6 (74%), item 7 (78%), and item 55 (69%). Therefore, a fourth round of voting was conducted to reevaluate and finalize the results for these pending eight items. Five items were found to have been correctly included as required for competency (items 6, 10, 28, 31, and 51), and three items were included as being important but not required for competency (items 7, 23, and 55) in the final checklist. Of note, the removal of duplicate responses did not affect the inclusion or exclusion of any items on the final checklist, and the additional round of voting provided reassurance of accurate representation of the expert panel’s consensus on these items. In summary, the initial checklist included 64 items, and the final checklist included 62 items based on expert panel consensus. Of these 62 items, 57 achieved consensus as a requirement for demonstration of competency in image acquisition for FCU examinations (Table 3). The items that did not achieve consensus and were excluded from the final checklist focused on quantitative measurement of IVC diameter using calipers or M-mode. Of the five items included in the final checklist considered to be important but not required for competency, two items were related to basic machine operation (entering patient identifying information [item 1] and digitally storing and archiving images [item 62]), two items involved nuanced aspects of cardiac image acquisition (beginning the image acquisition of a PLAX view from the left midsternal level [item 7] and centering the left ventricle on the screen in the PSAX view [item 23]), and one item involved assessment of IVC diameter (assessing the largest and smallest diameters of the IVC qualitatively [item 55]).
Table 3.

Focused Cardiac Ultrasound Skills Checklist developed using modified Delphi process

Checklist Item
Required for Competency?
Section A: Appropriate use of ultrasound machine and basic settings
1. Enters appropriate identifying information (e.g., provider name, patient name, and medical record number)No
2. Selects cardiac modeYes
3. Places indicator on screen in location consistent with local or specialty conventionYes
4. Selects phased-array transducerYes
5. Interacts in a professional manner with the patientYes
 
Section B: Parasternal long axis (PLAX)
6. Places patient in optimal position to obtain parasternal long-axis view, typically supine or left lateral decubitus position, when feasibleYes
7. Begins examination with transducer placed left of sternum, approximately midsternal levelNo
8. Positions transducer in correct axis with transducer marker toward right shoulder/left hipYes
9. Displays left ventricular cavity at its fullest diameterYes
10. Correctly identifies right ventricleYes
11. Correctly identifies left ventricular outflow tractYes
12. Correctly identifies aortic valveYes
13. Correctly identifies ascending aortaYes
14. Correctly identifies left ventricleYes
15. Correctly identifies septumYes
16. Correctly identifies left atriumYes
17. Correctly identifies mitral valveYes
18. Correctly identifies anterior leaflet of mitral valveYes
19. Correctly identifies pericardiumYes
20. Correctly identifies descending thoracic aortaYes
 
Section C: Parasternal short axis (PSAX) at midpapillary muscle level
21. Rotates transducer 90 degrees clockwise from parasternal long-axis viewYes
22. Captures a circular left ventricular image in normal patientsYes
23. Left ventricle is approximately centered on screenNo
24. Correctly identifies right ventricleYes
25. Correctly identifies left ventricleYes
26. Correctly identifies both papillary musclesYes
27. Correctly identifies septumYes
 
Section D: Apical 4-chamber (A4C)
28. Positions patient appropriately: left lateral decubitus position or supine. If unable to obtain adequate views in supine position, moves to left lateral decubitus position.Yes
29. Left ventricle and atrium appear on right side of the screen, and right ventricle and atrium are on the left.Yes
30. Attempts to place cardiac apex approximately in center of screenYes
31. Interventricular septum is approximately verticalYes
32. Avoids foreshortening. Heart should appear long and oval rather than short and globular in normal patients.Yes
33. Correctly identifies right ventricleYes
34. Correctly identifies tricuspid valveYes
35. Correctly identifies right atriumYes
36. Correctly identifies left ventricleYes
37. Correctly identifies mitral valveYes
38. Correctly identifies left atriumYes
39. Correctly identifies septumYes
 
Section E: Subcostal long axis (SLAX) or subcostal 4-chamber (S4C)
40. Places patient in optimal position to obtain subcostal views, typically supine position. If unable to obtain adequate views, repositions patient (e.g., bends knees to relax abdominal muscles).Yes
41. Places transducer in subxiphoid positionYes
42. Positions transducer in correct orientation to obtain 4-chamber view of the heartYes
43. Correctly identifies liverYes
44. Correctly identifies right ventricleYes
45. Correctly identifies tricuspid valveYes
46. Correctly identifies right atriumYes
47. Correctly identifies septumYes
48. Correctly identifies left ventricleYes
49. Correctly identifies mitral valveYes
50. Correctly identifies left atriumYes
 
Section F: Subcostal inferior vena cava (SIVC)
51. Places transducer subxiphoid in cephalad–caudad orientationYes
52. Obtains view of inferior vena cava (IVC) entering right atriumYes
53. Acquires view at widest diameter of IVCYes
54. Chooses an appropriate point to assess IVC diameterYes
55. Assesses the largest and smallest diameters of IVC qualitatively using estimation (“eyeballing”)No
56. Correctly identifies IVCYes
57. Correctly identifies liverYes
58. Correctly identifies right atriumYes
59. Correctly identifies hepatic veinYes
 
Section G: Examination conclusion
60. Adjusts depth appropriately in order to capture all key structuresYes
61. Adjusts gain appropriately in order to optimize visualization of all key structuresYes
62. Digitally stores and archives ultrasound images, if applicableNo
Focused Cardiac Ultrasound Skills Checklist developed using modified Delphi process

Discussion

Using a large international panel of multispecialty POCUS experts, we have developed a consensus-based FCU image acquisition checklist. The final checklist contains 62 items, 57 of which are required for demonstration of competency in FCU image acquisition. The checklist covers skills in basic ultrasound machine operation, techniques of cardiac image acquisition, and structures visualized from the five most common cardiac views. At least five FCU assessment tools have been published previously, but none have been widely accepted for assessing competence in FCU image acquisition by noncardiologists. Important limitations of these previously published checklists include lack of inclusion of all standard FCU views, recruitment of a small number of experts, and focus on use by a single specialty (46, 48–51). Patrawalla and colleagues developed a multisystem critical care ultrasound assessment tool that was limited by the number of expert panel members (n = 4) and inclusion of only the parasternal long- and short-axis views (46). Similarly, Rebel and colleagues and Mitchell and colleagues published multisystem checklists for anesthesiology residents, but only the former checklist specified a few required structures to be identified from the parasternal short-axis and apical four-chamber views (50, 51). The tools developed by Gaudet and colleagues and Millington and colleagues included all five standard FCU views, but the number of experts recruited to develop these tools was small and limited to critical care (48, 49). Echocardiography-certified critical care physicians developed the checklist by Gaudet and colleagues, which includes four levels of the parasternal short-axis view and quantitative IVC assessments, which can be considered advanced skills, limiting its generalizability to basic FCU users from other specialties. The FCU skill assessment tool published by Millington and colleagues used a global rating scale approach with a single question to assess image quality of each standard FCU view, rather than articulating the individual components of each view (49). Global rating scales have some advantages over checklists. For example, many checklists include items not essential to the procedure itself, such as cleaning the ultrasound machine, which can cloud judgment regarding technical skill competence. If every checklist item has equal weight, learners who correctly perform most steps of a procedure yet make a critical mistake can theoretically still receive a passing score (54). In such cases, global rating scales may be more effective at discriminating between competent and incompetent performances (55). However, the use of global rating scales requires significant training and expertise among the examiners to clearly define the essential components of a given skill and maintain consistency in rating learners. Given the limited availability of POCUS experts at institutions, it is probable that many FCU skill assessors have limited experience in using global rating scales, and therefore a checklist with specific items was an appropriate tool to develop. We have created an FCU image acquisition checklist that is unique compared with prior tools because it was developed by a multispecialty panel of experts from different institutions. The expert panel represented seven specialties from different regions of the United States and Canada. Our goal in recruiting a broad array of experts was to create a checklist that is useful for all FCU users. Despite this diversity, our panel reached consensus on 81% of the checklist items during the first round of voting. The use of such a broad expert panel provides validity evidence for the content of our checklist. The distinction between “important to include” and “required for competence” of individual checklist items merits further discussion. By asking our expert panel to determine whether each element is not only important but also required for competence, we hoped to establish consensus on the components of a universal FCU exam while identifying critical items that should not be missed if a learner is deemed to be competent. The process of developing this consensus-based checklist revealed some important caveats of assessing FCU image acquisition skills. First, the panel agreed it is important for FCU users to enter patient data and archive images, but the majority believed it was not necessary for learners to demonstrate these skills to be deemed competent. The panel believed the most emphasis should be placed on using FCU to make accurate, rapid clinical decisions in the care of acutely ill patients and less emphasis should be placed on entering patient data and archiving images. Furthermore, the expert panel expressed concerns about the variability in access to image-archiving software and potential challenges in connecting handheld ultrasound devices to institutional servers. Second, there was a common sentiment in favor of focusing on the quality of an acquired image rather than details of the maneuver used to obtain the view. Third, there was controversy about the importance and clinical utility of quantitative measurements of respiratory variation of IVC diameter during an FCU examination. Although a strong majority (85%) agreed that an assessment of IVC diameter should be included in the checklist, none of the items with quantitative IVC measurements reached the 80% threshold for consensus. Experts commented on the minimal advantage of a quantitative versus qualitative assessment and concerns about inexperienced FCU users misinterpreting or applying undue clinical significance to quantitative IVC measurements. These views reflect the ongoing debate about the clinical utility of IVC assessments in certain settings (56, 57). The final checklist therefore contains items that require the user to visualize the IVC but not measure its diameter. Our study has important limitations related to the use of an expert panel, the length of the checklist, and the focus on image acquisition. Though our panel of POCUS experts was based exclusively in North America, our panel only interacted electronically via e-mail and the online surveying tool. Compared with in-person meetings or video-based conference calls, our purely electronic interaction may have precluded discussion of important nuances of FCU image acquisition, especially in areas of disagreement. However, to mitigate loss of these qualitative data, we provided a free-text comment box under each checklist item and reviewed the panel’s comments with each round of voting. In addition, not all experts completed all rounds of voting. When selecting expert panelists for this modified Delphi design, an approach that can be difficult to coordinate because of the sheer number of participants, we invited individuals with known or suspected high levels of engagement in cardiac ultrasound from a broad range of specialties and diverse practice settings. We clearly communicated the survey completion deadline in the invitation e-mail for each round of voting and sent one reminder e-mail before the deadline. Regarding the length of the checklist, although a checklist that includes 62 items may seem long, many of the items can be assessed simultaneously in each view. Typically, checklist items that require identification of structures in a view can be completed relatively quickly given the binary nature (yes or no) of these checklist items. For ease of use, we have developed a one-page version of the checklist that may be preferred (Table E1). Future studies should evaluate the feasibility of using the checklist and assess whether the number of items can be reduced without impacting the discriminatory performance of the checklist. Last, our checklist focuses only on FCU image acquisition skill, which is essential but not sufficient to determine whether a provider is competent to use FCU in patient management. For a complete assessment of FCU competence, a learner’s image interpretation and clinical integration skills must also be assessed. However, acquisition of high-quality FCU images is a critical first step toward accurately interpreting cardiac structure and function, and therefore our checklist addresses this important first step toward integrating FCU in patient care. Furthermore, some ancillary skills, such as appropriate care and cleaning of the ultrasound machine, were not captured by our checklist.

Conclusions

We have developed an FCU image acquisition checklist based on the consensus of a large, multispecialty panel of ultrasound experts. Our checklist contains 62 items that may be used as a standardized tool to provide both summative and formative assessments for a broad range of medical specialties and disciplines that perform FCU examinations. The checklist can help identify healthcare providers who are competent to perform FCU examinations, as well as individuals who require further practice. Future steps include development of a visual anchoring system for each view to aid users, as well as gathering usage data to provide further evidence of validity of the checklist for determining learner competence in FCU performance in patient care.
  55 in total

1.  American College of Chest Physicians/La Société de Réanimation de Langue Française statement on competence in critical care ultrasonography.

Authors:  Paul H Mayo; Yannick Beaulieu; Peter Doelken; David Feller-Kopman; Christopher Harrod; Adolfo Kaplan; John Oropello; Antoine Vieillard-Baron; Olivier Axler; Daniel Lichtenstein; Eric Maury; Michel Slama; Philippe Vignon
Journal:  Chest       Date:  2009-02-02       Impact factor: 9.410

Review 2.  Point-of-care cardiac ultrasound techniques in the physical examination: better at the bedside.

Authors:  Bruce J Kimura
Journal:  Heart       Date:  2017-03-04       Impact factor: 5.994

3.  Development of a Point-Of-Care Cardiovascular Ultrasound Program for Preclinical Medical Students.

Authors:  Anirudh Kumar; Naman Barman; Jacob Lurie; Helen He; Martin Goldman; Stephen A McCullough
Journal:  J Am Soc Echocardiogr       Date:  2018-07-07       Impact factor: 5.251

4.  The Rapid Assessment of Competency in Echocardiography Scale: Validation of a Tool for Point-of-Care Ultrasound.

Authors:  Scott J Millington; Robert T Arntfield; Michael Hewak; Stanley J Hamstra; Yanick Beaulieu; Benjamin Hibbert; Seth Koenig; Pierre Kory; Paul Mayo; Jordan R Schoenherr
Journal:  J Ultrasound Med       Date:  2016-05-31       Impact factor: 2.153

5.  POINT: Should Acute Fluid Resuscitation Be Guided Primarily by Inferior Vena Cava Ultrasound for Patients in Shock? Yes.

Authors:  Gregory A Schmidt
Journal:  Chest       Date:  2016-12-08       Impact factor: 9.410

6.  Cardiopulmonary limited ultrasound examination for "quick-look" bedside application.

Authors:  Bruce J Kimura; Norihiro Yogo; Charles W O'Connell; James N Phan; Brian K Showalter; Tanya Wolfson
Journal:  Am J Cardiol       Date:  2011-06-09       Impact factor: 2.778

7.  Focused echocardiographic evaluation in life support and peri-resuscitation of emergency patients: a prospective trial.

Authors:  Raoul Breitkreutz; Susanna Price; Holger V Steiger; Florian H Seeger; Hendrik Ilper; Hanns Ackermann; Marcus Rudolph; Shahana Uddin; Markus A Weigand; Edgar Müller; Felix Walcher
Journal:  Resuscitation       Date:  2010-11       Impact factor: 5.262

8.  Pocket-size hand-held cardiac ultrasound as an adjunct to clinical examination in the hands of medical students and junior doctors.

Authors:  Vasileios F Panoulas; Anna-Lena Daigeler; Anura S N Malaweera; Amrit S Lota; Dinnish Baskaran; Syed Rahman; Petros Nihoyannopoulos
Journal:  Eur Heart J Cardiovasc Imaging       Date:  2012-07-24       Impact factor: 6.875

Review 9.  Point of care cardiac ultrasound applications in the emergency department and intensive care unit--a review.

Authors:  Robert T Arntfield; Scott J Millington
Journal:  Curr Cardiol Rev       Date:  2012-05

10.  Use of an error-focused checklist to identify incompetence in lumbar puncture performances.

Authors:  Irene W Y Ma; Debra Pugh; Briseida Mema; Mary E Brindle; Lara Cooke; Julie N Stromer
Journal:  Med Educ       Date:  2015-10       Impact factor: 6.251

View more
  1 in total

1.  Development of a multisystem point of care ultrasound skills assessment checklist.

Authors:  Nilam J Soni; Robert Nathanson; Mark Andreae; Rahul Khosla; Karthik Vadamalai; Karthik Kode; Jeremy S Boyd; Charles M LoPresti; Dana Resop; Zahir Basrai; Jason Williams; Brian Bales; Harald Sauthoff; Erin Wetherbee; Elizabeth K Haro; Natalie Smith; Michael J Mader; Jacqueline Pugh; Erin P Finley; Christopher K Schott
Journal:  Ultrasound J       Date:  2022-05-12
  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.