Literature DB >> 28639355

How do the characteristics of breast cancer diagnostic assessment programmes influence service delivery: A mixed methods study.

A R Gagliardi1, G Honein-AbouHaidar1, T Stuart-McEwan1, J Smylie2, A Arnaout2, J Seely2, F C Wright3, M J Dobrow4, M C Brouwers5, K Bukhanov1, D R McCready1.   

Abstract

Diagnostic assessment programmes (DAPs) coordinate multidisciplinary teamwork (MDT), and improve wait times and patient satisfaction. No research has established optimal DAP design. This study explored how DAP characteristics influence service delivery. A mixed methods case study of four breast cancer DAPs was conducted including qualitative interviews with health-care providers and retrospective chart review. Data were integrated using multiple approaches. Twenty-three providers were interviewed; 411 medical records were reviewed. The number of visits and wait times from referral to diagnosis and consultation were lowest at a one-stop model. DAP characteristics (rural-remote region, human resources, referral volume, organisation of services, adherence to service delivery targets and one-stop model) may influence service delivery (number of visits, wait times). MDT, influenced by other DAP characteristics (co-location of staff, patient navigators, team functioning), may also influence service delivery. While the one-stop model may be ideal, all sites experienced similar and unique challenges. Further research is needed to understand how to optimise the organisation and delivery of DAP services. Measures reflecting individual, team and patient-reported outcomes should be used to assess the effectiveness and impact of DAPs in addition to more traditional measures such as wait times.
© 2017 The Authors. European Journal of Cancer Care Published by John Wiley & Sons Ltd.

Entities:  

Keywords:  breast cancer; diagnostic techniques and procedures; interprofessional relations; patient care team; systems integration

Mesh:

Year:  2017        PMID: 28639355      PMCID: PMC5900983          DOI: 10.1111/ecc.12727

Source DB:  PubMed          Journal:  Eur J Cancer Care (Engl)        ISSN: 0961-5423            Impact factor:   2.520


INTRODUCTION

Cancer management requires coordinated delivery of services by different professionals, in different settings and at different time points (Brar, Look Hong, & Wright, 2014). Multidisciplinary teamwork (MDT) improves clinical and patient‐reported outcomes for cancer by improving treatment decisions, and their implementation and documentation; attendance and professional diversity at joint meetings; role clarity among team members; team effectiveness and staff satisfaction; and guideline‐adherent care delivery (Fleissig, Jenkins, Catt, & Fallowfield, 2006; Lamb et al., 2011; Lemieux‐Charles & McGuire, 2006). MDT is frequently operationalised through multidisciplinary cancer conferences (MCCs) which improve cancer management and associated outcomes (Hayward et al., 2003; Hong, Wright, Gagliardi, & Paszat, 2010; Taylor, Shewbridge, Harris, & Green, 2013; Wright, De Vito, Langer, & Hunter, 2007). Timely diagnosis of cancer improves access to MCCs, leading to earlier treatment and a potentially better prognosis (Brar et al., 2014). Clinicians whom we surveyed suggested the need to improve MDT earlier in the cancer trajectory given numerous barriers of access to, and coordination of diagnosis (Gagliardi, Wright, Davis, Urbach, & McLeod, 2008). Similar challenges at the interface between primary and oncology specialty care have been reported elsewhere (Kekhlyudov & Latosinsky, 2010; Sussman & Baldwin, 2010). We reviewed 22 studies that evaluated MDT for cancer patients and found that none examined MDT for diagnosis (Gagliardi, Dobrow, & Wright, 2011). Diagnostic centres or programmes can bridge the primary–specialty care interface and deliver timely, coordinated diagnostic services (Kekhlyudov & Latosinsky, 2010). In our previous research and that of others, diagnostic assessment programmes (DAPs) reduced time from referral to specialist visit and first treatment, and improved patient satisfaction with services and personal care received (Brouwers et al., 2009; Castellanos et al., 2008; Gagliardi, Grunfeld, & Evans, 2004). Recommendations issued in our jurisdiction and elsewhere to guide DAP structure and function were largely consensus based (Brouwers et al., 2009; Wilson et al., 2013). Further comparative research is needed to identify the ideal characteristics of DAPs that promote MDT and enhance the delivery of diagnostic services. This may reveal one or more optimal models for DAP design that could be broadly adopted. As a first step to prepare for future comparative research, the purpose of this study was to explore whether and how DAP characteristics influenced MDT and diagnostic service delivery. This knowledge serves as a baseline assessment of the participating centres, and could provide guidance to others for planning, evaluating or improving DAP services.

METHODS

Design

A mixed methods multiple case study was conducted involving four breast cancer diagnostic assessment programmes (DAPs), chosen because they shared the goal of coordinating diagnostic assessment for patients with suspected cancer, but varied by health region and by regional characteristics (urban, rural, remote and size of population served), factors that may have influenced DAP design (Fetters, Curry, & Creswell, 2013; Yin, 1999). A convergent mixed methods approach was used where the priority of qualitative and quantitative methods was equal; qualitative and quantitative data collection and analysis were concurrent; and qualitative and quantitative data were integrated and interpreted following analysis. Findings are reported based on Good Reporting of A Mixed Methods Study (GRAMMS) criteria (O'Cathain, Murphy, & Nicholl, 2008). Ethical review boards at participating sites approved the study.

Qualitative analysis of DAP characteristics

A study representative at each site was interviewed to learn about DAP characteristics according to those recommended in our jurisdiction (Brouwers et al., 2009), and the type, sequence and target (goal to be achieved) timing of diagnostic services. They also provided the names and contact information of other DAP staff for additional interviews. Basic qualitative description was employed (Sandelowski, 2000). Rigour was optimised using qualitative research and reporting standards (Barbour, 2001; Clark, 2003). Purposive sampling was used to recruit participants who varied by professional role. Individuals were invited by email, and asked to sign and return a consent form prior to being interviewed. Telephone interviews were conducted by a trained research assistant. Participants were asked to describe examples of MDT, associated outcomes, facilitators and challenges, and recommendations to enhance MDT. Interviews were held from January 29 to October 15, 2013, audio‐recorded and transcribed. An initial goal of five individuals from each site (one nurse, one physician, one referring physician, one other health professional and one administrative staff) was set for a minimum of 20 participants. Sampling proceeded to thematic saturation. Themes were identified using constant comparative technique (Auerbach & Silverstein, 2003). ARG, GH and the research assistant independently read transcripts to identify, define and organise themes. Data (quotes labelled by theme) were tabulated by theme and participating site.

Quantitative analysis of diagnostic services

Eligible patients were aged 18 and older who were referred to participating DAPs for assessment of suspected primary breast cancer from January 1, 2012 to December 31, 2012. Sampling was based on 2011 referral volumes which varied across sites. From site B, C and D, 80 patients (15% of patients at site with lowest 2011 referral volume) were randomly sampled. From site A, 200 patients were randomly sampled to accommodate another study. From the initial sample of 440, patients were excluded if they were referred for a second opinion (3) or consultation only (1) rather than undergoing diagnostic assessment, had metastasis from another primary cancer (4) or recurrent breast cancer (19), or had no recorded referral date (2), leaving 411 eligible for analysis. Reporting complied with standards for observational studies (von Elm et al., 2008). A data abstraction form was developed to collect data on the type and timing of diagnostic procedures performed after referral (Hulvat, Hansen, & Jeruss, 2009; Pruthi et al., 2007). Data included patient demographic characteristics (date of birth, gender); type of procedure that confirmed the diagnostic result (imaging only—one or more of mammography, ultrasound and/or magnetic resonance imaging; biopsy following one or more imaging procedures—fine‐needle aspiration, core or open) and results (positive for cancer, negative for cancer, still suspicious requiring follow‐up). Recorded dates included: referral (date when referral form received by DAP), confirmatory procedure (date when confirmatory diagnostic procedure performed), diagnosis (date when finding was recorded in patient record) and consultation (date of meeting to discuss treatment or follow‐up plan). Four trained abstractors collected data from medical records at participating sites. Dates for all procedures were identified except for two patients (one at site B and one at site C) for whom the date of consultation was not recorded in the medical record. Data were collected between June 2013 and August 2014. Summary statistics were used to assess the proportion of patients whose confirmatory procedure was imaging or biopsy, and median number of DAP visits and wait time in days (plus interquartile range) from referral date to confirmatory procedures, diagnosis and consultation. If referral and confirmatory procedure dates coincided they were counted as one visit. ANOVA was used to compare continuous variables and the Chi‐square test to compare proportions by site. The number of visits and wait times were not normally distributed, therefore these measures were compared by site using the Kruskal–Wallis non‐parametric test, and we reported the Dunn's adjusted p values based on multiple comparisons between groups. Analyses were performed with IBM SPSS (version 21, SPSS Statistics/IBM Corp, Chicago IL, USA).

Integration of findings

Data were integrated by translating coded qualitative data into counts (transformation approach); weaving the qualitative findings through the description of quantitative findings (narrative approach); and visually depicting potential associations between qualitative and quantitative findings (joint display) (Fetters et al., 2013). This enabled the assessment of coherence between qualitative and quantitative findings (confirmation, expansion and/or discordance). Integration of the findings was independently assessed by two investigators (ARG, GH) who met to discuss the findings and achieve consensus. This was refined according to review and feedback from participants and the study team.

RESULTS

Organisational characteristics

Table 1 compares organisational characteristics across DAPs. Sites were similar in terms of providing a single point of access for regional referrals, maintaining protected scheduling times for patients referred to the DAP and other operational features such as a dedicated steering or oversight committee, protocols or pathways to guide service delivery, and the collection and reporting of performance data. Apart from sampling criteria (health region, urban versus rural/remote, size of population served), sites differed in total volume of patients referred in 2012 (A 836; B 7,773; C 513; D 670), service delivery model (site A offered single‐visit diagnosis), days per week of service (site C operated fewer than 5 days per week) and human resources (site A and B featured more full‐time human resources compared with C and D). Sites differed in triage criteria, time to schedule first visit from referral, time to first visit, whether an additional visit was needed for biopsy, time to biopsy and whether consultation with patients to discuss the results took place in the DAP or with the referring physician. As a result target time from referral to diagnosis and to consult, and target number of total visits varied across DAPs.
Table 1

Characteristics of participating DAPs

CharacteristicsParticipating site
ABCD
Regional characteristics
Health regionUrbanUrban–ruralUrban–ruralRural–remote
Population1.2 million1.2 million775,000236,000
DAP launch date2006199720072007
Total patients referred in 20128367,773513670
Diagnostic service delivery model
Scope of care diagnostic onlyYesYesYesYes
Single locationYesYesYesYes
Single‐visit diagnosisYesNoNoNo
Patient risk level servedAllAllAllAll
Regional access
Single point of entryYesYesYesYes
Accepts referral from all sourcesYesYesYesYes
Operational features
Days per week552 to 35
Referral and triage criteriaYesYesYesYes
Protected booking slotsYesYesYesYes
Dedicated governance structureYesYesYesYes
Guidelines/service frameworkYesYesYesYes
Performance reportingYesYesYesYes
Human resources
Medical directorPFPP
Clinical directorPP
Clinical managerFFF
Reception/clerical/bookingFFPP
Social workerPFPP
Other supportive careFPPP
Patient navigatorFFFF
Nurse practitioner/advanced practice nurseF
Registered nurseFFPP
Surgical oncologistFFPP
Medical oncologistFPPP
General physicianFF
RadiologistPFPP
Imaging technologistPFPP
PathologistFFPP
Pathology technologistFPPP
Plastic surgeonPFPP
Total full‐time staff101222
Target time to diagnosis* Within 5 to 15 days for all patientsNo biopsy—within 2 to 10 days; biopsy—2 to 12 daysNo biopsy—within 10 days; biopsy—16 daysNo biopsy—within 14 days; biopsy—35 to 49 days
Target time to consult* Coincides with first visit for all patientsNo biopsy—first visit; biopsy—9 to 22 daysNo biopsy—within 10 days; biopsy—21 to 26 daysNo biopsy—first visit
Target number of total visits* 1 to 22 to 32 to 3 (2 if consult with referring physician)2 (all consults with referring physician)

F, full‐time; P, part‐time; *Target—refers to intended/planned according to goals/internal protocols.

Characteristics of participating DAPs F, full‐time; P, part‐time; *Target—refers to intended/planned according to goals/internal protocols.

Multidisciplinary teamwork

Twenty‐three individuals were interviewed (Table 2). Themes related to number and type of MDT examples, facilitators and challenges, and perceived benefits were largely similar across sites (Online Resource 1). Scheduling given staffing shortages was particularly problematic for site D (rural–remote region, staffing). Unintended consequences and suggestions to enable or enhance MDT were largely expressed by those at site A (one‐stop model) and B (large referral volume).
Table 2

Interview participants

Professional roleParticipating siteTotal
ABCD
Medical director019 (surgical oncologist)1
Clinical director0070252
Clinical manager004 (NP)034 (RN)011 (mammography technologist)4
002 (NP)
Patient navigator004 (NP)012 (RN)037 (Radiation technologist)2
002 (NP)
Surgical oncologist006 (plastic surgeon)017 (surgical oncologist)2
NP/APN/RN030 (RN)1
Medical Oncologist0241
Radiologist0400273
035
Pathologist0231
Referring Primary Care physician0330362
Social worker0161
Administrator or clerk0151
General practitioner Oncologist0091
Technologist (mammography, ultrasound or MRI)0050412
Total785424

NP, nurse practitioner; APN, advanced practice nurse; RN, registered nurse.

Interview participants NP, nurse practitioner; APN, advanced practice nurse; RN, registered nurse.

Patient characteristics, procedures and findings

A total of 411 medical records were reviewed (Table 3). The mean age was 56 years, and patients at site D were significantly older. More patients at site D had imaging and fewer had biopsy as the confirmatory procedure (p < .001). The number of patients diagnosed with cancer differed across sites, ranging from 1 (1.3%) at site D to 72 (39.3%) at site A (p < .01). The site D coordinator confirmed a high rate of “inappropriate referrals” that were found to be negative for cancer based on confirmatory imaging (organisation of services).
Table 3

Patient characteristics, and confirmatory diagnostic procedures and findings

Patient characteristics and diagnostic findingsParticipating site n (%)Total
ABCD
Number of patients183806880411
Age group (years)
23–4026 (14.2)10 (12.5)7 (10.3)43 (10.5)
41–5045 (24.6)22 (27.5)15 (22.1)82 (20.0)
51–6056 (30.6)23 (28.8)16 (23.5)47 (58.8)a 142 (34.5)
61–7035 (19.1)14 (17.5)22 (32.4)30 (37.5)a 101 (24.6)
>7121 (11.5)11 (13.8)8 (11.8)3 (3.8)43 (10.5)
Mean age (years)54.454.956.360.2a 56.0
Gender
Female180 (98.4)80 (100.0)68 (100.0)80 (100.0)408 (99.3)
Male3 (1.5)0 (0.0)0 (0.0)0 (0.0)3 (0.7)
Confirmatory procedure
Imaging62 (33.9)48 (60.0)26 (38.2)70 (87.5)a 206 (50.1)
Biopsy121 (66.1)32 (40.0)42 (61.8)10 (12.5)a 205 (49.9)
Findings
Positive72 (39.3)b 16 (20.0)24 (35.3)1 (1.3)a 113 (27.5)
Negative56 (30.6)49 (61.3)c 25 (36.8)26 (32.5)156 (38.0)
Follow‐up55 (30.1)15 (18.8)19 (27.9)53 (66.3)d 142 (34.5)

More patients at site D were aged 51 to 60, or 61 to 70, and mean age was higher compared with other sites; more patients at site D had imaging and fewer had biopsy as the confirmatory diagnosis; fewer patients at site D were diagnosed with cancer compared with other sites, p < .05.

More patients at site A were positive for cancer compared with site B, p < .05.

More patients at site B were negative for cancer compared with other sites, p < .05.

More patients at site D required follow‐up compared with other sites, p < .05.

Patient characteristics, and confirmatory diagnostic procedures and findings More patients at site D were aged 51 to 60, or 61 to 70, and mean age was higher compared with other sites; more patients at site D had imaging and fewer had biopsy as the confirmatory diagnosis; fewer patients at site D were diagnosed with cancer compared with other sites, p < .05. More patients at site A were positive for cancer compared with site B, p < .05. More patients at site B were negative for cancer compared with other sites, p < .05. More patients at site D required follow‐up compared with other sites, p < .05.

Number of visits

For patients with an image‐confirmed diagnosis (206, 50.1%), the median number of visits from referral to diagnosis was similar across all sites (1.0, interquartile range 1.0 to 1.0) (Table 4).
Table 4

Number of visits from referral to diagnosis and consultation

End‐pointParticipating site (n patients, median number of visits from referral to end‐point in days, interquartile range)Total
ABCD
Diagnosis (image confirmed)62482670206
1.01.01.01.01.0
1.0 to 1.01.0 to 1.01.0 to 1.01.0 to 1.01.0 to 1.0
Diagnosis (biopsy confirmed)121324210205
1.01.01.02.0a 1.0
1.0 to 2.01.0 to 2.01.0 to 1.02.0 to 2.81.0 to 1.0
Consultation15817231199
2.03.0b 2.03.02.0
2.0 to 2.02.0 to 3.02.0 to 2.03.0 to 3.02.0 to 2.0
Target number of total visits from referral to consultation (Table 2)1 to 22 to 32 to 3 (2 days if consult with referring physician)2 (consult with referring physician)

Diagnosis—date when result of confirmatory diagnostic procedure recorded in patient medical record.

Consultation—date of meeting with patient to discuss treatment or follow‐up plan.

Patients referred to site D had significantly more visits compared with other sites, p < .05.

Patients at site B had significantly more visits compared with sites A and C, p < .05.

Number of visits from referral to diagnosis and consultation Diagnosis—date when result of confirmatory diagnostic procedure recorded in patient medical record. Consultation—date of meeting with patient to discuss treatment or follow‐up plan. Patients referred to site D had significantly more visits compared with other sites, p < .05. Patients at site B had significantly more visits compared with sites A and C, p < .05. For patients with a biopsy‐confirmed diagnosis (205, 49.9%), the median number of visits from referral to diagnosis was highest at site D (2.0, p < .01). At site D scheduling had to accommodate radiologists from elsewhere were periodically hired on a weekly basis to compensate for the lack of a local full‐time radiologist (staffing), and the flight schedules of women who had to fly from remote communities (rural–remote region). The median number of visits from referral to consultation was higher at site B (3.0, p < .01) compared with sites A and C. Apart from standard mammography and ultrasound, the 17 patients at site B with a median of 3.0 visits underwent additional procedures (one or more of repeat mammography, repeat ultrasound, MRI, CT of the chest or abdomen, bone scan or biopsy) on one or more visits (organisation of services, service delivery targets).

Wait times

Table 5 reports wait times from referral to confirmatory procedure, diagnosis and consultation. The median wait time from referral to confirmatory imaging was similar across all sites (15.0 days, interquartile range 8.0–23.0).
Table 5

Wait time from referral to confirmatory procedure, diagnosis and consultation

End‐pointParticipating site (n patients, median wait time from referral to end‐point in business days, interquartile range)Total
ABCD
Confirmatory imaging procedure62482670206
14.514.514.515.015.0
8.0 to 27.05.0 to 22.59.0 to 22.09.0 to 21.08.0 to 23.0
Confirmatory biopsy procedure121324210205
10.0a 19.0a 14.5a 38.513.0
6.0 to 17.07.5 to 28.09.0 to 25.029.0 to 48.07.0 to 23.0
Diagnosis (image confirmed)62482670206
15.015.014.521.017.0
8.0 to 27.06.0 to 23.010.0 to 22.012.0 to 28.09.5 to 26.0
Diagnosis (biopsy confirmed)121324210205
13.0b 32.017.044.516.0
9.0 to 20.019.0 to 37.013.0 to 28.031.0 to 53.010.0 to 27.0
Consultation15817231199
16.0c 40.023.084.018.0
10.0 to 26.024.0 to 54.020.0 to 40.084.0 to 84.011.0 to 29.0
Target wait time from referral to diagnosis (Table 2)6 to 16 (all patients)2 to 10 (image) 7 to 21 (biopsy)10 (image) 21 to 26 (biopsy)14 (image) 35 to 49 (biopsy)
Target wait time from referral to consult (Table 2)First visit for all patientsFirst visit (image); 9 to 22 days (biopsy)10 days (image); 21 to 26 days (biopsy)First visit (image); referring physician (biopsy)

Referral—date when referral form received by the DAP.

Confirmatory procedure—type of procedure used to confirm diagnosis.

Diagnosis—date when result of confirmatory diagnostic procedure recorded in patient medical record.

Consultation—date of meeting with patient to discuss treatment or follow‐up plan.

Significantly lower for site A compared with all other sites; and for sites B and C compared with site D, p < .05.

Significantly lower for site A compared with sites B and D, p < .05.

Significantly lower for site A compared with sites B and C, p < .05.

Wait time from referral to confirmatory procedure, diagnosis and consultation Referral—date when referral form received by the DAP. Confirmatory procedure—type of procedure used to confirm diagnosis. Diagnosis—date when result of confirmatory diagnostic procedure recorded in patient medical record. Consultation—date of meeting with patient to discuss treatment or follow‐up plan. Significantly lower for site A compared with all other sites; and for sites B and C compared with site D, p < .05. Significantly lower for site A compared with sites B and D, p < .05. Significantly lower for site A compared with sites B and C, p < .05. The median wait time s from referral to confirmatory biopsy (10.0 days, interquartile range 6.0–17.0), referral to biopsy‐confirmed diagnosis (13.0 days, interquartile range 9.0–20.0) and referral to consultation (16.0 days, interquartile range 10.0–26.0) were significantly lower at site A (p < .01) compared with all other sites (one‐stop model). At site A only, the time from referral to biopsy‐confirmed diagnosis was lower than the time from referral to image‐confirmed diagnosis, reflecting triage prioritisation criteria for higher‐risk cases (service delivery targets). The median wait time of 38.5 days from referral to confirmatory biopsy for 10 patients at site D was likely influenced by the mutual availability of outside radiologists and patients from remote communities (rural–remote region, staffing). Integration of qualitative and quantitative data generated a conceptual framework that visually displays how DAP characteristics may influence MDT and diagnostic service delivery (Figure 1). Integration revealed concordance between qualitative and quantitative findings. Qualitative data revealed that several DAP characteristics influenced MDT including rural–remote population, workload and human resource limitations. Quantitative data, when interpreted based on qualitative findings, found that similar DAP characteristics influenced service delivery: rural–remote region, human resources, referral volume, organisation of services and one‐stop service delivery model could explain differences across sites in number of visits and wait times.
Figure 1

Conceptual framework of factors influencing diagnostic service delivery

Conceptual framework of factors influencing diagnostic service delivery Instances of discordance were also identified. Qualitative data identified that all sites specified service delivery targets based on triage of risk. Quantitative data showed that wait time for biopsy‐confirmed diagnosis (higher‐risk cases) was shorter than image‐confirmed diagnosis at site A only, the one‐stop service delivery model. Other sites were likely unable to adhere to, or achieve service delivery targets due to the noted challenges of rural–remote population, workload/referral volume and human resource limitations. This discordance further supports the potential relationship between DAP characteristics and diagnostic service delivery. Integrated findings contribute to an expansion in the understanding of MDT in the diagnostic context. MDT was said to achieve several beneficial outcomes at the level of individual providers and teams which, in turn, enhanced the efficiency of service delivery and the patient experience by reducing wait times, and the number of visits needed to establish a diagnosis. This study also identified unanticipated consequences at site A and B, likely associated with the pressure of having to achieve one‐stop service and provide services to a high volume of referrals.

DISCUSSION

This study revealed that DAP characteristics (rural–remote region, human resources, referral volume, organisation of services, adherence to service delivery targets and one‐stop service delivery model) may influence the efficiency of service delivery (number of visits, wait times). Other DAP characteristics (co‐location of staff, patient navigators, team functioning) may influence MDT and team effectiveness which were also thought to influence the efficiency of service delivery (number of visits, wait times). Previous research found that DAPs reduced wait times and improved patient satisfaction, but had not investigated the underlying mechanism (Brouwers et al., 2009; Castellanos et al., 2008; Gagliardi et al., 2004). This study's findings are unique in that they provide preliminary insight on DAP features that could be enhanced to improve service delivery including the organisation of services and MDT. This study is further distinguished from other research by the evaluation of multiple sites with differing features, and the use of a mixed methods approach to evaluate DAP design and impact in a holistic manner based on the perspectives of a variety of actors and a range of measures. In contrast, other research evaluated single sites and most commonly reported wait times only (Baliski et al., 2014; Oon et al., 2014; Royle et al., 2014). The findings are similar to those of a systematic review of health‐care team effectiveness literature from 1985 to 2004 which found that the type and diversity of clinical expertise involved in team decision making largely accounted for improvements in patient care and organisational effectiveness, while collaboration, conflict resolution, participation and cohesion were most likely to influence staff satisfaction and perceived team effectiveness (Lemieux‐Charles & McGuire, 2006). This suggests that measures reflecting individual provider, team and patient‐reported outcomes should be used to assess the effectiveness and impact of DAPs in addition to more traditional measures such as wait times. Operations management principles have been used to simulate a demand–supply model for a one‐stop skin cancer clinic and found that by managing triage criteria, resource allocation and capacity planning, the time to treatment of new patients could be reduced by 90% with the same resources (Romero et al., 2013). While modelling may be a useful first step in identifying alternative DAP designs, real‐world studies are needed to pilot the feasibility and impact of various DAP models. In this study the one‐stop model required the fewest visits for diagnosis and, similar to other studies, achieved the lowest wait times to diagnosis (Brouwers et al., 2009; Gagliardi et al., 2004). Although these findings may not be surprising, there are several implications to consider. The one‐stop site included in this study experienced challenges similar to those at other sites, and additional challenges unique to the one‐stop model. Similarly, in a study of a rapid access prostate cancer clinic, the diagnosis of cancer increased resulting in a considerable increase in workload for surgeons (Oon et al., 2014). Moreover, the one‐stop model may not be possible to implement in all settings given resource limitations or regional characteristics. Further research is needed to understand how to optimise efficiency in DAPs that are not able to offer one‐stop service. Participants suggested that one way to improve DAP services was to optimise scope of practice. Current research on nurse navigation in the cancer care continuum largely focuses on supportive care or survivor follow‐up; little research has studied the navigation content or services that should be offered (Post et al., 2015; Shockney, 2015; Wells et al., 2016). Further research is also needed to examine how DAPs can enhance patient‐reported outcomes, a concept that has evolved from patient satisfaction to person‐centred care (Harper, De Costa, Garrett‐Mayer, & Sterba, 2015; Rathert, Wyrwich, & Boren, 2013). Several study limitations must be noted. We may not have identified and evaluated all DAP characteristics relevant to diagnostic service delivery. Few individuals representing each profession were interviewed at each site, however, we did achieve thematic saturation within and across sites. Only four sites that diagnosed one type of cancer participated, and the sample of patients was small, thus findings may not be transferrable. Further research may confirm whether these findings are true of DAPs in other settings or for the diagnosis of different types of cancer. In conclusion, DAP characteristics (rural–remote region, human resources, referral volume, organisation of services, adherence to service delivery targets and one‐stop service delivery model) may influence service delivery (number of visits, wait times). MDT, influenced by other DAP characteristics (co‐location of staff, patient navigators, team functioning), may also influence the number of visits and wait times. Insights generated by this research, captured as a conceptual framework of the factors that influence diagnostic service delivery, could be used by other to plan, evaluate and improve diagnostic services for cancer patients. While the one‐stop model achieved fewer visits and a shorter wait time compared with other sites, all sites experienced similar and unique challenges. Further research is needed to understand how to optimise the organisation and delivery of DAP services. Click here for additional data file. Click here for additional data file.
  34 in total

1.  The Strengthening the Reporting of Observational Studies in Epidemiology (STROBE) statement: guidelines for reporting observational studies.

Authors:  Erik von Elm; Douglas G Altman; Matthias Egger; Stuart J Pocock; Peter C Gøtzsche; Jan P Vandenbroucke
Journal:  J Clin Epidemiol       Date:  2008-04       Impact factor: 6.437

2.  Achieving integration in mixed methods designs-principles and practices.

Authors:  Michael D Fetters; Leslie A Curry; John W Creswell
Journal:  Health Serv Res       Date:  2013-10-23       Impact factor: 3.402

3.  Effect of patient navigation on satisfaction with cancer-related care.

Authors:  Kristen J Wells; Paul C Winters; Pascal Jean-Pierre; Victoria Warren-Mears; Douglas Post; Mary Ann S Van Duyn; Kevin Fiscella; Julie Darnell; Karen M Freund
Journal:  Support Care Cancer       Date:  2015-10-05       Impact factor: 3.603

Review 4.  How can we improve cancer care? A review of interprofessional collaboration models and their use in clinical management.

Authors:  Anna R Gagliardi; Mark J Dobrow; Frances C Wright
Journal:  Surg Oncol       Date:  2011-07-16       Impact factor: 3.279

Review 5.  Patient-centered care and outcomes: a systematic review of the literature.

Authors:  Cheryl Rathert; Mary D Wyrwich; Suzanne Austin Boren
Journal:  Med Care Res Rev       Date:  2012-11-20       Impact factor: 3.929

Review 6.  Multidisciplinary cancer conferences: a systematic review and development of practice standards.

Authors:  F C Wright; C De Vito; B Langer; A Hunter
Journal:  Eur J Cancer       Date:  2007-02-27       Impact factor: 9.162

7.  Effects of Patient Navigation on Patient Satisfaction Outcomes.

Authors:  Douglas M Post; Ann Scheck McAlearney; Gregory S Young; Jessica L Krok-Schoen; Jesse J Plascak; Electra D Paskett
Journal:  J Cancer Educ       Date:  2015-12       Impact factor: 2.037

Review 8.  Multidisciplinary care for patients with breast cancer.

Authors:  Melissa C Hulvat; Nora M Hansen; Jacqueline S Jeruss
Journal:  Surg Clin North Am       Date:  2009-02       Impact factor: 2.741

9.  The requirements of a specialist Breast Centre.

Authors:  A R M Wilson; L Marotti; S Bianchi; L Biganzoli; S Claassen; T Decker; A Frigerio; A Goldhirsch; E G Gustafsson; R E Mansel; R Orecchia; A Ponti; P Poortmans; P Regitnig; M Rosselli Del Turco; E J Th Rutgers; C van Asperen; C A Wells; Y Wengström; L Cataliotti
Journal:  Eur J Cancer       Date:  2013-08-19       Impact factor: 9.162

10.  Breast cancer teams: the impact of constitution, new cancer workload, and methods of operation on their effectiveness.

Authors:  R Haward; Z Amir; C Borrill; J Dawson; J Scully; M West; R Sainsbury
Journal:  Br J Cancer       Date:  2003-07-07       Impact factor: 7.640

View more
  4 in total

1.  Cost analysis of breast cancer diagnostic assessment programs.

Authors:  G N Honein-AbouHaidar; J S Hoch; M J Dobrow; T Stuart-McEwan; D R McCready; A R Gagliardi
Journal:  Curr Oncol       Date:  2017-10-25       Impact factor: 3.677

2.  How do the characteristics of breast cancer diagnostic assessment programmes influence service delivery: A mixed methods study.

Authors:  A R Gagliardi; G Honein-AbouHaidar; T Stuart-McEwan; J Smylie; A Arnaout; J Seely; F C Wright; M J Dobrow; M C Brouwers; K Bukhanov; D R McCready
Journal:  Eur J Cancer Care (Engl)       Date:  2017-06-21       Impact factor: 2.520

3.  Recommendations to improve patient-centred care for ductal carcinoma in situ: Qualitative focus groups with women.

Authors:  Bryanna B Nyhof; Frances C Wright; Nicole J Look Hong; Gary Groot; Lucy Helyer; Pamela Meiers; May Lynn Quan; Nancy N Baxter; Robin Urquhart; Rebecca Warburton; Anna R Gagliardi
Journal:  Health Expect       Date:  2019-09-18       Impact factor: 3.377

4.  Implementation of the Canadian National Standards in Breast Cancer Surgical Care: Gaps, Barriers, Enablers and Opportunities.

Authors:  Angel Arnaout; Anubha Prashad; Nadine Dunk; Jess Rogers; Christian Finley; Annemarie Edwards; Mary Argent-Katwala
Journal:  Curr Oncol       Date:  2021-01-19       Impact factor: 3.677

  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.