Literature DB >> 34079603

Collaborative Development of an Older Adult Experience Survey for Specialized Geriatric Services.

Ronaye Gilsenan1, Rhonda Schwartz2, Iris A Gutmanis3, Adam M B Day4, David P Ryan5, Rosemary R A Brander6, Kelly Milne1, Frank Molnar7.   

Abstract

BACKGROUND: While generic, site, and disease-specific patient experience surveys exist, such surveys have limited relevance to frail, medically complex older adults attending appointment-based specialized geriatric services (SGS). The study objective was to develop and evaluate a patient experience survey specific to this population.
METHODS: Using established survey research methods, this study was conducted collaboratively with older adults (patients and family members/friends) at three Ontario sites offering SGS. The study was done in three phases: Phase One-literature review, evidence alignment, and operationalization of core survey items; Phase Two-cognitive interviews and refinement; and Phase Three-pilot testing, survey item analysis, and refinement.
RESULTS: Based on an evidence-informed framework, the "Older Adult Experience Survey" includes 12 core items, two global rating items, two open-ended questions, and two demographic questions. The summed 12 core items demonstrated acceptable internal consistency (Cronbach's alpha: 0.83), and the correlation between the summed score and a global question was 0.59, providing evidence of construct validity. The survey also demonstrated face and content validity.
CONCLUSION: This open access, collaboratively developed, psychometrically sound patient experience survey can be used to assess, then improve, the clinical experience and quality of care of older adults attending appointment-based SGS clinics/programs.
© 2021 Author(s). Published by the Canadian Geriatrics Society.

Entities:  

Keywords:  appointment-based; frail older adults; patient experience; specialized geriatric services; survey

Year:  2021        PMID: 34079603      PMCID: PMC8137462          DOI: 10.5770/cgj.24.487

Source DB:  PubMed          Journal:  Can Geriatr J        ISSN: 1925-8348


INTRODUCTION

Ongoing assessment of “patient experience”( is key to improving health-care quality and reducing costs.( While generic ,( site(, and disease-specific( patient experience surveys have been developed and implemented, item wording is not specific to appointment-based specialized geriatric services (SGS) and may not have included dimensions relevant to older frail adults.( Further, existing SGS patient satisfaction/experience surveys include various wording and rating scales that impede provincial reporting. In Ontario, a collaborative of 11 regional programs provide SGS to the ever-increasing number of older adults( living with, or at risk for, frailty( whose health, dignity, and independence are challenged due to multiple complex medical, functional, and psychosocial issues. This SGS collaborative works with primary care physicians, community professionals, and others, by offering a spectrum of hospital and community-based services to older adults. The objective of this study was to develop and evaluate a minimum set of core survey items for measuring the experience of older adults in appointment-based SGS settings.

METHODS

Using established survey research methods,( this study was conducted in three phases (see Table 1) as approved by the Health Sciences North Research Ethics Board, the Ottawa Health Science Network Research Ethics Board, and The Scarborough Hospital Research Ethics Board.
TABLE 1

Methods used to develop and test the Older Adult Experience Survey

PhaseActionsDetails
Phase 1: Framework Identification and Item SelectionReview of existing surveys used by RGPs to identify quality improvement initiativesPerformance Measurement Committee (PMC) establishedReview of existing surveys; Identification of common items; examination of formatting, layout, and overall design
Literature review:Conceptualizing patient experienceReview of existing measures/frameworksSelection of a patient experience frameworkCollection/sharing of existing patient experience surveys and relevant literature
Framework dimension/subdimension selection and refinementItem developmentGroup consensus based methodology (see appendix A) used by PMC to identify dimensions and subdimensions of the selected primary care framework applicable to SGSItem refinement guided by principles of clarity, simplicity, conceptual specificity, and contextual relevanceSurvey face validity evaluated by the PMC

Phase Two: Survey RefinementReview of draft survey items by target populationREB submission and approvalDevelopment of semi-structured interview guideSite selection and recruitment of patients/familyCognitive interviews (see appendix B) with consenting patients/family members: face validation; review of formatting, layout and design; item-by-item review regarding utility, relevance, and wording of each item; assessment of comprehensibility and acceptability
Pilot survey finalizedItems revised by PMC based on de-identified notes taken during the cognitive interviews

Phase Three: Pilot TestingPilot testing of core items with target populationSite selection and patient recruitmentStaff training in survey administrationDistribution and collection of surveys
Assessment of pilot survey psychometric propertiesData entered into SPSS v. 24 databaseDetermination of item missingness, item frequency distributions, tests of normalityAssessment of internal consistency: Cronbach’s alphaAssessment of construct convergent validity: Spearman rho correlation between “Overall, I felt that the care and services I experienced were [0=poor, 10=excellent] and the12-item summed scoreAssessment of construct divergent validity: item by item Mann-Whitney U or Kruskal Wallis tests by site and time of yearScores summed by domain: inter-domain Spearman rho correlations determinedResponses to “What worked well?” and “What could be improved?” coded using framework’s subdimensionsSurvey endorsement by RGPs of Ontario Executive group
Survey reviewReview of item wording, scoring and formatting by PMCPost discharge, consenting patients randomly given both the pilot survey and the post pilot version followed by cognitive interviewsScore deviations examined
DocumentationImplementation guide developedSurvey ( https://rgps.on.ca/resources/rgps-of-ontario-older-adult-experience-survey)/ and implementation guide (https://rgps.on.ca/resources/rgps-of-ontario-older-adult-experience-survey-implementation-guide/) posted

Phase One: Literature Review, Evidence Alignment, and Operationalization of Core Survey Items

The literature review was built off work done as part of the Canadian primary health-care system renewal,( as both SGS and primary care are largely appointment-based services. The work conducted to develop the Ontario Primary Care Performance Measurement Framework,( along with Wong and Haggerty’s scoping review( and other publicly available articles describing patient experience frameworks and surveys, formed the basis of this non-exhaustive literature review. Conventional search methods included a search of electronic databases (e.g., Google for grey literature and hand searches of key articles). Rigid inclusion/exclusion criteria were not applied, but articles that focused on older adults and appointment-based services were of prime interest. The search was limited to articles written in English between 2002 and 2016. An evidence-informed patient experience framework was selected based on its relevancy to SGS settings, the SGS population, and alignment with other Canadian work in this area. Following this, a group consensus approach based on the Delphi methodology( was used to identify SGS applicable dimensions and sub-dimensions, and to draft item wording for an SGS patient experience survey (see Appendix A).

Phase Two: Cognitive Interviews and Refinement

As informed by Willis and Artino,( semi-structured cognitive interviews (see Appendix B & C) were conducted at three SGS sites (Ottawa, Scarborough, Sudbury). A convenience sample of 5–15 older adults( was required. Older adults who attended SGS appointment-based services during the study period and who were able to speak and understand English were asked to provide insights into the utility, relevance, and wording of each draft item. Based on their feedback, a pilot version of the survey was finalized using the methodology described in Appendix B.

Phase Three: Pilot Testing, Survey Item Analysis and Refinement

The pilot version of the survey was tested at two sites (Ottawa, Scarborough) with another convenience sample of cognitively intact older adults. It was determined that 73 patients per site were needed, assuming a 10% margin of error and a 95% confidence interval around a sample proportion of 50%. The statistical analysis was done using SPSS statistical software (SPSS 24, IBM, Armonk, NY, USA). Item-by-item frequency distributions were generated, and Cronbach’s alpha was calculated for the summed core items. The distribution specific correlation between the global item (“Overall, I felt that the care and services I experienced were…”), and the summed core-item score was determined to provide evidence of construct validity. Correlations among the framework dimensions were then examined. Finally, framework subdimensions were used to code responses to the open-ended questions (What worked well? What could be improved?). Subsequently, the pilot survey was revised and re-evaluated by a convenience sample of older patients who received SGS services at the Scarborough site. These patients were given both the pilot survey and the revised pilot survey in random order and then participated in cognitive interviews (see Appendix D). Item-by-item frequency distributions were compared using distribution appropriate statistics and qualitative responses were coded.

RESULTS

Wong and Haggerty’s( primary care framework was selected and used to guide the development of the SGS patient experience survey. All six framework dimensions and 12/17 subdimensions were deemed applicable to SGS (see Table 2). This Phase One draft survey included 16 core items and two global items.
TABLE 2

Wong and Haggerty( dimensions and subdimensions retained for Older Adult Experience Survey

SubdimensionsPHASE 1 Group Consensus ApproachRounds 1–4: modified W&H dimensions and subdimensions to better fit SGSRationale for excluding specific primary care patient experience sub-dimensionsPHASE 1 Group Consensus ApproachRounds 5–10: developed core item wording (draft version for cognitive interview)PHASE 2 Group Consensus ApproachRound 11: refinements made after cognitive interviews (pilot version)PHASE 3 Group Consensus ApproachRound 12: refinements made after pilot testing (final version)
1. Access

First contact accessibilityThe ability to obtain patient or client initiated needed care (including advice and support) from the provider of choice within a time frame appropriate to the urgency of the problemCurrently in SGS most referrals are initiated by a patient’s primary care practitioner, and therefore, care is not initiated by the client.

AccommodationRelationship between how resources are organized to accept patients or clients (including appointment systems, hours of operation, walk-in facilities, telephone services) & the patient’s or clients’ ability to accommodate factors to realize access1. The time I had to wait for my first appointment was reasonable
2. Someone was available to talk to me if I needed it

Economic accessibilityThe extent to which direct or indirect costs related to care impeded decisions to access needed care or continue recommended careLow priority rating. Concern that only actionable core items are included in the survey and that SGS does not have the ability to impact these costs (e.g., parking).

2. Interpersonal Communication

General communicationAbility of the provider to elicit and understand patient or client concerns and to explain health and health care issuesMy concerns were listened to3. My concerns were addressed
Explanations were given in a way I could understand4. Information was given in a way I could understand

RespectfulnessAbility of practitioners to provide care that meets expectations of users about how people should be treated, such as regard for dignity & provision of adequate privacy5. I was treated with respect
I was comfortable sharing my story(Item removed. Deemed not relevant by patients)

Shared decision-makingExtent to which patients or clients are involved in making decisions about their treatmentI was able to contribute to decisions about my care6. I was included in making decisions about my care, as much as I wanted to be

Whole person careExtent to which providers address the physical, emotional and social aspects of a patient’s or client’s health & consider the community context in their care7. Time was taken to learn about me as a person

3. Continuity and Coordination

Relational continuityA therapeutic relationship between the patient or client and one or more identified providers that spans separate health care episodes & delivers care that is consistent with client’s bio-psycho-social needs.Low priority rating. SGS are typically of a short duration and tend not to span separate health care episodes.

Information continuityExtent to which information is used to make current care appropriate to the patient or clientI was confident that results of my visit were shared with others as needed12. It was clear who would receive information about my care

CoordinationProvision and organization of a combination of health services and information with which to meet a patient’s or client’s health needs, including services available from other community health service providersI was connected to services that I neededI was referred to other programs and/or services that I needed11. I was referred to other programs/services that I needed

Team functioningAbility of SGS providers to work effectively as an inter-professional team to manage and deliver quality patient or client care(Initially included and then excluded after cognitive interviews)Patients and caregivers accessing SGS programs don’t always see the interactions between providers. And, if they do see them, they have difficulty differentiating SGS providers from other types of care providers, especially those that do home visits.The advice I received was consistent(Item removed. Deemed not relevant by patients)
The people I saw worked together well(Item removed. Deemed not relevant by patients)

4. Comprehensiveness of Services

Comprehensive-ness of services providedThe provision, either directly or indirectly, of a full range of services to meet patient’s or clients’ health care needs and caregiver information and support needs after being seen by a specialized geriatric service. This includes health promotion, prevention, diagnosis and treatment of common conditions, referral to other clinicians, management of chronic conditions, rehabilitation, palliative care and in some models, social servicesThe people I saw thought about all of my needs10. The (add program/clinic name) met my needs

Health promotion and secondary preventionHealth promotion is the process of enabling people to increase control over, and to improve, their health. Secondary prevention aims to reduce the impact of a disorder by detecting & treating it as soon as possible to halt or slow its progressHigh priority item initially, but later became low priority. It was seen to be captured in “comprehensiveness of services provided”

5. Trust

TrustExpectation that other person will behave in a way that is beneficial and that allows for risks to be taken based on this expectation. For example, patient or client trust in the SGS team provides the basis for taking the risk of sharing personal infoI had confidence in the care I received8. I had confidence in the people I saw

6. Patient Reported Impacts of Care

Patient activationPatient’s or client’s ability or readiness to engage in health behaviours that will maintain or improve their health statusI was able to follow the advice I was givenI could achieve the goals that were agreed to9. I will be able to use the advice I was given

Patient safetyPatient’s or client’s report of medication errors (given or taken the wrong drug or dose) or incorrect medical or laboratory reports and communication with their provider about not taking their prescribed medication or medication side effects(Initially included, then excluded after cognitive interviews)Patients did not see safety as a relevant concernThe care I received was safe(Item removed. Deemed not relevant by patients)

Confidence in the primary health care systemThe perception that allows patients or clients of health care to make decisions since they assume (and expect relative certainty about providers delivering safe & technically competent careConfidence in the primary health care system drew a low priority rating by members. Not considered very applicable to SGS.Total: 16 core itemsTotal: 12 core items (4 items removed) 8 items rewordedTotal: 12 core items

7. Overall Ratings

Overall, I felt that the care and services I received were:13. Overall, at the (add program/clinic name)I had a:
14. I would recommend this program to my family or friends if they needed it15. What could be improved?
16. What worked well?

8. Demographics
My Age:
My Gender
Interviewees (n=19) indicated that the draft survey items measured all key aspects of their patient experience, thereby providing some evidence of both face and content validity. Feedback led to the rewording of eight items and the deletion of four items pertaining to three subdimensions. Based on these findings, a pilot survey was generated that included 12 items scored on a 5-point Likert scale, one global question scored on an 11-point Likert scale, a willingness to recommend item rated on a 4-point Likert scale, and two open-ended questions (see Appendix B). Of the estimated 257 patients who met the phase three study inclusion criteria, 145 were recruited [Ottawa: n=75/114 (65.8%); Scarborough: n=70/123 (56.9%)]. Due to missing values, summed scores for the core 12 items were generated for 131 patients. Summed scores ranged from 43 to 60. The mean of the summed core items was 56.9 (SD: 3.9) and the median was 59 (interquartile range (IQR): 6). Cronbach’s alpha was 0.83, demonstrating acceptable internal consistency.( As the frequency distribution associated with summed scores deviated significantly from a normal distribution (Shapiro-Wilk test: 0.80, p < .001; skewness: −1.180), non-parametric tests (Spearman rho correlations, Mann-Whitney U or Kruskal Wallis tests) were used to assess statistical associations and group differences. The Spearman rho correlation between the 12-item summed score and the global experience rating was 0.59, providing evidence of construct convergent validity. Although inter-domain Spearman rho correlations varied from 0.19 (trust and access) to 0.58 (comprehensiveness of services and continuity and coordination), all correlations were statistically significant (p < .05). More than 60% of all patients selected the top category for any one of the 12 core survey items and 54 patients (41.2%) selected the top response for all items (see Table 3). Despite relatively little dispersion, differences were detected by time of year and/or by site/program for 6 of the 12 survey items (Mann-Whitney U or Kruskal Wallis test, p < .5), providing some evidence of construct divergent validity. Qualitative survey responses were mapped to nine of the ten framework subdimension and provided further information on perceived strengths and areas for improvement.
TABLE 3

Pilot survey item-by-item analysis

ItemI felt that …Missing% Picking Highest CategoryDiffered By Site (S)/By Month (M)ABItem-Total Correlation
… the time I had to wait for my first appointment was reasonable1 (0.7%)70.80%S, M0.55
…someone was available to talk to me if I needed it078.60%M0.38
…my concerns were addressed2 (1.4%)79.00%0.48
…information was given in a way I could understand1 (0.7%)81.90%S0.48
…I was treated with respect1 (0.7%)92.40%S0.40
…I was included in making decisions about my care, as much as I wanted to be1 (0.7%)77.80%0.47
…time was taken to learn about me as a person073.10%0.43
…I had confidence in the people I saw083.40%0.43
…I could achieve the goals that were agreed to4 (2.8%)61.70%S0.53
…the program met my needs4 (2.8%)78.70%M0.70
…I was referred to other programs and/or services that I needed3 (2.1%)76.80%0.52
…it was clear who would receive information about my care2 (1.4%)81.80%0.50

Significant difference (p < .05), Mann-Whitney U or Kruskal Wallis test.

Difference by site (S): Ottawa or Scarborough site or by month (M): Jan–March, April–May, June–July.

Item-total Correlation = correlation between the item score and the summed framework-based items excluding that item.

The pilot survey was then revised. Two of the core items were reworded, two demographic items were added, instructions were shortened, anchors for the 12 core survey items and one global assessment question were changed, the survey name was modified, and minor changes were made to formatting. Testing of the revised pilot survey was conducted with additional patients (n=5). No one expressed concerns regarding the above revisions. Four patients found the revised survey easier to complete. Response selections were identical for four of the 12 core items and only once did a response switch by two points. Further, there were no significant differences in mean summed core-item scores (pilot version vs. revised pilot version: 55.4 [SD: 4.8] and 55.2 [SD: 4.9], respectively) and median scores did not differ significantly by version (56 [IQR: 8.5] vs. 57 [13], respectively; Mann-Whitney U: p = 1.00).

DISCUSSION

A framework-based patient experience survey specific to frail, medically complex older adults attending appointment-based SGS was developed and tested by incorporating input from older adults, their family members/friends, and experts in geriatrics and research. When taken together, the 12 core items of the Older Adult Experience Survey demonstrated acceptable internal consistency (Cronbach’s alpha: 0.83). Slightly more than 40% of patients had the maximum score, perhaps accurately reflecting perceived patient experience or perhaps indicating a ceiling effect. Despite this finding, the survey was able to identify meaningful group differences. Users are encouraged to use a mixed methods approach to triangulate qualitative and quantitative information.( Testing was conducted on an English language, paper-based version of the survey at three sites providing appointment-based SGS. Further examination of interrater and test/retest reliability, structure, discriminant validity and response rates is warranted. Psychometric properties will need to be re-evaluated if the survey is translated to another language or to an electronic version. Future studies may provide evidence of the survey’s clinical utility and ability to identify areas for quality improvement that will lead to improved quality of patient care. Findings may also provide insights for system planners at the local, regional, and provincial levels.

CONCLUSIONS

Based on an evidence-informed framework, the collaboratively developed Older Adult Experience Survey demonstrates acceptable internal consistency, as well as face, content, construct convergent and construct divergent validity.
DetailsDelphi: A group-based consensus research method with structured voting where a panel of ‘experts’ complete a series of questionnaires interspersed with controlled feedback and their voting is anonymous (McKenna, 1994).
Group Consensus Approach: Participants attended “real time” teleconference calls and had in-depth discussions regarding potential survey wording prior to voting.
Recruitment: 8 participants were recruited who had expertise in geriatrics and research.
Documentation: GCA worksheets were sent to participants prior to each teleconference. The worksheet included proposed survey wording to be voted on, a space to vote “Yes” or “No” and space to present concerns and rewording suggestions. A summary of the group discussion and the voting on survey wording revisions were forwarded to members after each round.
Facilitation: The proposed wording (e.g., framework wording or core item wording) were presented by the facilitator, who then facilitated a group discussion (regarding the principles for inclusion).
Item Evaluation for Inclusion: Each item was evaluated based on cognitive interview feedback and discussion amongst GCA participants regarding key principles for inclusion:

Clarity– clear definitions of each dimension and subdimension that apply to SGS

Simplicity – short and common words, short sentences and simple concepts

Contextual specificity – applicable wording for SGS programs in different settings and by different providers

Contextual relevance – relevant questions that solicit feedback to support the implementation of QI initiatives within SGS

Voting: Voting was conducted during structured meetings. If participants were unable to attend, they could submit their vote in advance via the GCA worksheet.
Level of agreement: 100% level of agreement was needed. If this was not reached, a discussion regarding rationale for non-agreement occurred, additional wording options were generated, and participants voted on the revised wording options.

Implementation Phase 1GCA Rounds 1–4 (framework development)
Reviewed and minimally modified the dimensions and subdimensions of the Wong and Haggerty 2013 primary care patient experience framework to better fit specialized geriatric services
6/6 dimensions included. 12/17 subdimensions included in SGS version
Participants: 8
GCA Rounds 5–10 (core item development)
16 core items developed that aligned with 12 subdimensions
Participants: 8

Phase 2GCA Round 11 (core survey item development POST cognitive interviews)
4/16 core items removed and resulted in 2/12 sub-dimensions being removed.
8/16 core items reworded
a 5-point Likert scale was selected for the 12 core items (1=strongly disagree, 3=neutral, 5=strongly agree), an 11-point Likert scale (0=poor, 10=excellent) was selected for the one global question, a 4-point Likert scale was chosen for the willingness to recommend item (1=definitely no; 4=definitely yes), and two open-ended questions were added to gather details about what worked well and what could be improved.
Participants: 8

Phase 3GCA Round 12 (survey development, core items and formatting POST pilot testing)
2/12 core items re-worded
2 demographic items added
Anchors for 12 core survey items and one global assessment question were changed to 1= no, definitely not; 5= yes definitely and 0=poor experience, 10=excellent experience, respectively
Survey name was changed to The Older Adult Experience Survey (OAES) and minor changes to survey formatting (1 legal size page to two letter size pages)
SMOG Readability Test conducted (Flesch-Kincaid grade level: 6.2)
Participants: 6
DetailsDefinition: Cognitive interviewing is an evidenced based, qualitative method designed to investigate participant thought processes while answering survey questions.(17)
Format: Involves (1) “think-aloud interviewing where respondents actively verbalize their thoughts as they attempt to answer the survey questions” and (2) “verbal probing where interviewer administers a series of probe questions specifically designed to elicit detailed information” (p. 354(17)).
Semi-structured: Potential probing questions are developed in advance of the interview, while additional probing questions may be developed during the interview based on the participants’ responses.
# to be recruited: 5–15 participants(18)

Phase 2Implementation
12 cognitive interviews conducted with 19 participants at three SGS sites (The Ottawa Hospital Geriatric Day Hospital, The Scarborough Hospital GAIN Clinic and the North East Specialized Geriatric Centre)
Survey probes were developed in advance (e.g., I noticed that you were hesitating, tell me what you were thinking?). Specific survey wording was identified for additional clarification (e.g., what does the term “reasonable” mean to you?). Participants were asked if each survey item was useful and relevant to their experience of SGS care and whether any items should be reworded/added/removed
See Appendix C for the Phase 2 cognitive interview guide
Analysis
One analyst amalgamated and analyzed the data from all sites
Quantitative analysis: Percent agreement with the relevance of each core item
Qualitative analysis: Core survey item feedback was themed according to the conceptual framework (e.g., dimensions, subdimensions). Feedback from participants regarding specific item wording was grouped (e.g., “2 mentioned…”) and feedback detailed as to why wording was confusing, or how it could be improved

Phase 3Implementation
5 cognitive interviews with 5 patients at one site (The Scarborough Hospital GAIN Clinic)
Participants reviewed the pre-pilot and post-pilot versions of the survey.
Three patients completed the post-pilot survey first while two patients completed the pilot-tested version first.
Specific changes to the format of the survey were provided to participants along with the rationale for those changes and participants were asked “do you have any concerns with this change?” and “if yes, please tell us about your concerns”.
See Appendix D for the Phase 3 cognitive interview guide
Analysis
One analyst analyzed the data from the one site
Quantitative analysis: percent who (1) noticed differences between the two surveys, (2) had concerns about specific format changes, (3) stated which version was easier to complete, (4) stated that changes impacted how they responded to the question
Qualitative analysis: aligned thematic analysis with quantitative categories to provide specific examples about their experience completing the different versions of the survey
Survey Core Items being TestedProbe Question
The time I had to wait for my first appointment was reasonableACCOMMODATION: Relationship between how resources are organized to accept patients“first appointment”“reasonable”
Someone was available to talk to me if I needed itACCOMMODATION: Relationship between how resources are organized to accept patients & the patient’s ability to accommodate factors to realize access“available”
My concerns were listened toCOMMUNIATION: Ability of provider to elicit and understand patient concerns“concerns”
Explanations were given in a way I could understandCOMMUNICATION: Ability of the provider to explain health and health care issues“explanations”
I was comfortable sharing my storyRESPECTFULNESS: Ability of practitioners to provide care that meets expectations about how people should be treated, such as regard for dignity & provision of adequate privacy“sharing my story”
I was treated with respectRESPECTFULNESS: Ability of practitioners to provide care that meets expectations about how people should be treated, such as regard for dignity & provision of adequate privacy“respect”
I was able to contribute to decisions about my careSHARED DECISION MAKING: Extent patients are involved in making decisions about their treatment“contribute”
Time was taken to learn about me as a personWHOLE PERSON CARE: Extent providers address the physical, emotional and social aspects of a patient’s health & consider the community context in their careWhat does this statement mean to you?Could you paraphrase this?
I had confidence in the care I receivedTRUST: Expectation that other person will behave in a way that is beneficial and that allows for risks to be taken based on this expectation“confidence”“care”
The care I received was safeSAFE: Patient’s report of medication errors or incorrect medical or laboratory reports and communication with their provider about not taking their prescribed medication or medication side effects“safe”
The advice I received was consistentTEAM FUNCTIONING: Ability of SGS providers to work effectively as an inter-professional team to manage and deliver quality patient or client care“advice”“consistent”
I could follow the advice I was givenPATIENT ACTIVATION: Patient’s ability or readiness to engage in health behaviours that will maintain or improve their health status“advice”
The people I saw worked together wellTEAM FUNCTIONING: Ability of SGS providers to work effectively as an inter-professional team to manage and deliver quality patient or client careWhat does this statement mean to you? Could you paraphrase this?
The people I saw thought about all of my needsCOMPREHENSIVENESS OF SERVICES PROVIDED: The provision, either directly or indirectly, of a full range of services to meet patient’s health care needs and caregiver information and support needs after being seen by a specialized geriatric service“people I saw”“all”
I was connected to other services that I neededCOORDINATION: Provision and organization of a combination of health services and information with which to meet a patient’s health needs, including services available from other community health service providers“connected”
16. I was confident that results of my visit were shared with others as neededINFORMATION CONTINUITY: Extent to which information is used to make current care appropriate to the patient or client“results”“with others as needed”
Overall, I felt that the care and services I received were:“care and services”
I would recommend this program to family or friends if they needed it“if they needed it”
  12 in total

1.  What Do Our Respondents Think We're Asking? Using Cognitive Interviewing to Improve Medical Education Surveys.

Authors:  Gordon B Willis; Anthony R Artino
Journal:  J Grad Med Educ       Date:  2013-09

Review 2.  The Delphi technique: a worthwhile research approach for nursing?

Authors:  H P McKenna
Journal:  J Adv Nurs       Date:  1994-06       Impact factor: 3.187

3.  Frailty in older adults: evidence for a phenotype.

Authors:  L P Fried; C M Tangen; J Walston; A B Newman; C Hirsch; J Gottdiener; T Seeman; R Tracy; W J Kop; G Burke; M A McBurnie
Journal:  J Gerontol A Biol Sci Med Sci       Date:  2001-03       Impact factor: 6.053

4.  The Warwick Patient Experiences Framework: patient-based evidence in clinical guidelines.

Authors:  Sophie Staniszewska; Felicity Boardman; Lee Gunn; Julie Roberts; Diane Clay; Kate Seers; Jo Brett; Liz Avital; Ian Bullock; Norma O' Flynn
Journal:  Int J Qual Health Care       Date:  2014-02-20       Impact factor: 2.038

5.  Measuring the patient experience in primary care: Comparing e-mail and waiting room survey delivery in a family health team.

Authors:  Morgan Slater; Tara Kiran
Journal:  Can Fam Physician       Date:  2016-12       Impact factor: 3.275

6.  Development of the Patient Activation Measure (PAM): conceptualizing and measuring activation in patients and consumers.

Authors:  Judith H Hibbard; Jean Stockard; Eldon R Mahoney; Martin Tusler
Journal:  Health Serv Res       Date:  2004-08       Impact factor: 3.402

7.  What explains worse patient experience in London? Evidence from secondary analysis of the Cancer Patient Experience Survey.

Authors:  Catherine L Saunders; Gary A Abel; Georgios Lyratzopoulos
Journal:  BMJ Open       Date:  2014-01-03       Impact factor: 2.692

8.  A short generic patient experience questionnaire: howRwe development and validation.

Authors:  Tim Benson; Henry W W Potts
Journal:  BMC Health Serv Res       Date:  2014-10-22       Impact factor: 2.655

9.  Making sense of Cronbach's alpha.

Authors:  Mohsen Tavakol; Reg Dennick
Journal:  Int J Med Educ       Date:  2011-06-27

10.  Establishing a Primary Care Performance Measurement Framework for Ontario.

Authors:  Wissam Haj-Ali; Brian Hutchison
Journal:  Healthc Policy       Date:  2017-02
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.