| Literature DB >> 26358965 |
Z Belshaw1, L Asher2, N D Harvey2, R S Dean2.
Abstract
Assessment of quality of life (QoL) is an important, increasingly popular outcome measure in veterinary research and practice, particularly in dogs. In humans, QoL is commonly assessed by self-reporting and since this is not possible for animals, it is crucial that instruments designed to measure QoL are tested for reliability and validity. Using a systematic, replicable literature search strategy, the aim of this study was to find published, peer-reviewed instruments for QoL assessment in dogs and to assess the quality of these. CAB Abstracts and PubMed were searched in July 2013 using terms relevant to dogs, wellbeing and QoL. Inclusion and exclusion criteria were applied. When instruments were not published in full, authors were contacted to obtain them. Criteria were applied to assess the quality, validity and reliability of the 52 instruments obtained. Twenty-seven additional instruments used in peer-reviewed publications were not included because they had not been fully described in the publication or were not provided by authors upon request. Most of the instruments reviewed (48/52) were disease-specific rather than generic. Only four publications provided a definition of QoL or wellbeing. Only 11/52 instruments demonstrated evidence of assessing reliability or validity, and the quality of these instruments was variable. Many novel, unvalidated instruments have been generated and applied as clinical outcomes before it was known whether they measured QoL. This rapid review can be used to identify currently available and validated canine QoL instruments, and to assess the validity and quality of new or existing instruments.Entities:
Keywords: Dog; Quality of life; Review; Validation; Welfare
Mesh:
Year: 2015 PMID: 26358965 PMCID: PMC4641869 DOI: 10.1016/j.tvjl.2015.07.016
Source DB: PubMed Journal: Vet J ISSN: 1090-0233 Impact factor: 2.688
Summary of information extracted from the publications reviewed.
| Information extracted | Potential answers |
|---|---|
| Was the instrument designed for use in dogs with a specific disease type? | Oncology; gastroenterology; hepatology; cardiorespiratory; dermatology |
| Was the keyword defined? | Free text |
| What was length of the recall period? | At the time of instrument administration; time (in units given in publication or description of information given); not stated |
| Was the instrument reproduced, described or referenced? | Free text |
| Was there any evidence that the authors had attempted to validate the instrument prior to its use to collect clinical data? | Yes/no |
| Brief description of the instrument in the format available for review | Free text |
| What was the function of the instrument as stated by the authors? | Free text |
| Which publications cite this instrument as found by the Scopus search? | Free text |
| Information relevant to reliability, validity and quality | |
| Who is completing the questions in the instrument? | One or more of: owner; veterinarian; clinical investigator; unclear |
| Was the method behind construction of the instrument described? | Yes; yes some questions adapted from elsewhere; yes, references an unvalidated instrument not found elsewhere in this search; no |
| Was a scoring or weighting applied to the results? | Yes (explanation of method used); no |
Validated instruments only.
Unvalidated instruments only.
All validated instruments were for completion by owners.
Assessment criteria for reliability (adapted from Taylor and Mills, 2006).
| Test | Aim of test | Criteria | Legend for |
|---|---|---|---|
| Intra-rater reliability | To assess reliability in scoring when one person repeat-scores the animal | Does the same person repeatedly score the same animal under the same conditions within a short time period? | 1 |
| Is that time period clearly stated? | 2 | ||
| Is the gap between repeat scores a minimum of 1 week, ideally a minimum of 2 weeks? | 3 | ||
| Is the consistency of scoring between first and subsequent assessments compared? (Tests for comparison are typically correlation coefficients such as the intra-class correlation coefficient, Kappa coefficient, Pearson's, Spearman's Rank or Kendall's tau-b.) | 4 | ||
| Have reliability statistics been assessed against a stated threshold? | 5 | ||
| Inter-rater reliability | To assess reliability in scoring when scorers simultaneously score the same animal | Do multiple people simultaneously score the same animal? | 6 |
| Does the methodology describe a circumstance which ensures that the scores of each rater are independent and unbiased by each other? | 7 | ||
| Is the consistency of scores between raters compared? (Tests for comparison are typically correlation coefficients such as the intra-class correlation coefficient, Kappa coefficient, Pearson's, Spearman's Rank or Kendall's tau-b.) | 8 | ||
| Have reliability statistics been assessed against a stated threshold? | 9 | ||
| Test–retest reliability | To assess consistency in scoring when a long period of time has elapsed | Does the same person score the same animal under the same conditions after a considerable time interval? (Length of interval might be constrained by the health condition; criterion 14) | 10 |
| Is the time period clearly stated? | 11 | ||
| Is the gap between repeat scores a minimum of 2 weeks? (Longer time periods were preferred.) | 12 | ||
| Is the consistency of scores compared? (Tests for comparison are typically correlation coefficients based upon rank order consistency such as the intra-class correlation coefficient, Spearman's Rank or Kendall's tau-b.) | 13 | ||
| Where relevant, is it acknowledged that for rapidly changing health conditions, this assessment is not always possible, or that time intervals might need to be shorter? | 14 | ||
| Have reliability statistics been assessed against a stated threshold? | 15 | ||
| Internal consistency | To assess whether, if questions are grouped in any form, there is a correlation between questions within the groups | Has an attempt been made to determine whether correlations exist between questions which are grouped together? | 16 |
| Is the method of grouping the questions stated? | 17 | ||
| Is the method of grouping appropriate? (Methods include factor analysis and principal component analysis; each has their own criteria for appropriate use.) | 18 | ||
| Has an analysis been performed to look for correlations between questions within groups, factors or components? | 19 | ||
| Is the method of analysis appropriate? (Methods of analysing within group correlations include Cronbach's Alpha and intra-class correlation coefficients.) | 20 |
Assessment criteria for validity (adapted from Taylor and Mills, 2006).
| Test | Aim of test | Criteria | Legend for |
|---|---|---|---|
| Content validity (face validity is a form of this) | To assess whether individual questions really ask what they are meant to be asking | Has an attempt been made to ensure that the questions in the instrument truly ask what they should? (e.g. do questions in the area of comfort truly ask about comfort?) | 21 |
| Is the method by which this has been performed described? | 22 | ||
| Is the method appropriate? (Methods include consultation with a panel of experts which in this context might be veterinarians, dog owners, canine behaviour experts etc.) | 23 | ||
| Construct validity | Whether questions, or groups of questions, ask what they are meant to be asking. This is assessed by comparing constructs which are hypothesised to be related. A construct is something which cannot be proved or objectively measured, e.g. quality of life, happiness. | Has an attempt been made to statistically check whether questions truly assess the broad area which they were designed to assess by comparing the relationships between questions/groups, or between questions/groups and other observable responses? (e.g. questions about comfort should be negatively associated with questions about pain level; and comfort scores should be negatively associated with sleep quality, while pain scores should be positively associated with reduced movement) | 24 |
| Have hypotheses about expected positive (convergent) and/or negative (divergent) associations between tested measures been clearly stated before analysis? This is critical to the assessment of construct validity. | 25 | ||
| Is the method by which the assessment has been made described? | 26 | ||
| Is this method appropriate? (Potential methods are numerous but include comparing the distribution of scores to other observable measures, or comparisons between scores within the instrument.) | 27 | ||
| Criterion (concurrent) validity | How this instrument compares to an independent reference standard measure. A criterion is something which can be objectively and definitively measured, e.g. age, a hip score. A measurement of a construct should not be used as the comparator in criterion testing. | Has the instrument been compared to a different instrument/measurement (criterion measure, standard reference test, reference standard) which measures the same thing? | 28 |
| Do the authors state that the criterion method used has been validated, or provide a reference? | 29 | ||
| Have hypotheses about expected associations between the instrument and the comparison measure been clearly stated, including the directionality of the expected correlation, before being tested? This is critical to the assessment of criterion validity. | 30 | ||
| Has the time when the criterion measurement was performed been clearly stated (typically at the same time as concurrent validity)? | 31 | ||
| Did the instrument produce results comparable to a reference standard? | 32 | ||
| Criterion (known groups) validity | Whether the instrument can distinguish between groups of veterinary patients, e.g. dogs with different severities of heart disease, or dogs given placebo vs. treatment | Has the instrument been assessed for its ability to distinguish clinically relevant differences between known groups? | 33 |
| Have hypotheses about expected associations between the instrument and the comparison measure been clearly stated, including the directionality of the expected correlation, before being tested? This is critical to the assessment of criterion validity. | 34 | ||
| Is the time when the assessment of the known group was performed clearly stated? | 35 | ||
| Has the instrument been shown to distinguish between different populations or groups? | 36 |
Assessment criteria for, and results of, the quality appraisal of the 11 validated instruments (based on questions by Gill, Feinstein, 1994, Guyatt et al, 1997, Locker, Allen, 2007).
| Are keywords (e.g. QoL) defined within the instrument? | Are the domains of QoL to be measured stated in the publication or instrument? | Do the investigators state why they used this instrument rather than any other? | During a pilot, were owners asked to suggest additional questions which could be included? | Were the questions informed by discussion or qualitative interviews with those who will complete the instrument? | If the authors aimed to measure QoL rather than HRQoL, is the instrument doing so? | Is a single-question overall QoL rating included? | Are multiple items aggregated into a single score? | Are owners asked to indicate which items were personally important to them? | If so is this incorporated into a weighted score? | |
|---|---|---|---|---|---|---|---|---|---|---|
| N | Y | Y | Y | NS | Y | Y | N | N | NA | |
| N | NA | Y | NA | NA | Y | Y | NA | Y | Y | |
| D | N | N | N | NS | NA | N | N | N | NA | |
| N | N | Y | N | N | NA | N | Y | N | NA | |
| D | N | Y | N | N | Y | Y | N | N | NA | |
| Y | Y | Y | N | Y | Y | Y | N | N | NA | |
| Y | Y | Y | N | N | Y | N | N | Y | N | |
| Y | N | Y | Y | NS | N | N | Y | N | NA | |
| D | Y | Y | N | N | Y | N | N | N | NA | |
| D | Y | N | NS | NS | Y | N | Y | N | NA | |
| D | Y | Y | Y | NS | Y | N | N | N | NA |
QOL, Quality of life; HRQoL, Health related quality of life; N, no; Y, yes; NA, not applicable; NS, not stated; D, the authors discuss definitions but do not clearly state which definition they have used.
Fig. 1Summary of the systematic application of inclusion and exclusion criteria.
Summary of information extracted from the 11 validated instruments.
| Publication (name of instrument if stated) | Function of instrument as stated by the authors | Brief description of instrument in the format available for review | Was the instrument validated in dogs with a specific disease type? | What was the recall period? | Was the instrument reproduced, described or referenced? | Publications which cite this publication found in Scopus search |
|---|---|---|---|---|---|---|
| Owners' perceptions of the severity and impact of chronic pain on their dogs with osteoarthritis | Two page, 11 question instrument. Four questions on pain, six on function (both numeric scales) and one scale for QoL (Likert-type). | Chronic pain | Previous 7 days | No. Later publications refer to website for download | ||
| Owner-perceived, weighted quality of life assessments for dogs with spinal cord injuries | Owners asked to choose five areas of life/life activity important to their dog, then to weight these using a laminated disc. Separate visual analogue scales for QoL and owner ability to cope with spinal cord injury. | Spinal cord disease | At the time of completion | Adequately described | ||
| Impact of atopic dermatitis on health-related quality of life of affected dogs and their owners | One page, 14 question instrument proposed for future use. Thirteen questions regarding QoL in the dog related to its skin disease and one about the QoL of the owners. Likert-type scale. | Skin disease | Since last visit to veterinarian | Reproduced | ||
| Health-related quality of life in dogs with cardiac disease | Two page, 18 question instrument. All questions relate to how the dog's heart disease has impacted on its comfort or sociability in the preceding 7 days. Likert-type scale. | Cardiac disease | Previous 7 days | Adequately described and available from author | ||
| Quality of life survey for use in a canine cancer chemotherapy setting | Four page, 30 question instrument. Three sections: 14 questions about how the dog was 6 months previously; 13 questions about the dog's QoL now; three questions about how the owners coped during the chemotherapy. Mixed scale types. | Cancer treated by chemotherapy | At the time of completion and 6 months previously | No. Available from the author | None | |
| Health-related quality of life in canine and feline cancer patients | One page, 24 question instrument. Eight sections, each with three questions. Sections on happiness, mental status, pain, appetite, hygiene, hydration, mobility and general health. Likert-type scales apart from the final global QoL question which is a visual analogue scale. | Cancer | At the time of completion | Reproduced | ||
| To raise awareness of welfare considerations of pet dogs visiting | Four page, 39 question instrument. Seven sections: three questions on comfort; three on exercise; three on diet; three on mental stimulation; four on companionship; 16 across two sections on behaviour. Mix of Likert-type and visual analogue scales and one open question. | No | At the time of completion and ‘at their best’ | Adequately described and available from author | None | |
| Quality of life of dogs with skin diseases and their owners | One page, 15 question instrument. No subdivision into sections, all disease related. Likert-type scale. | Skin disease | Previous seven days | Reproduced | ||
| Multidimensional assessments regarding QoL and the human–animal bond of companion dogs | Four page, 88 question instrument. Four sections: physical (27 questions), psychological (30 questions), social (15 questions) and environment (16 questions). All Likert-type scale. | No | At the time of completion | No. Available from the author | None | |
| Health-related quality-of-life scale for dogs with pain secondary to cancer | One page, 12 question instrument. No subdivision into sections, Likert-type scale. | Cancer | At the time of completion | Reproduced | ||
| A participatory tool in order to encourage discussions and decisions about dogs' quality of life | One page, five question instrument. Questions asking owners how well they provide for five ‘needs’. Visual analogue scale. | No | At the time of completion | Adequately described | None |
The instrument provided by the author contains 88 questions; the publication describes a 91 question instrument.
The illustration of the instrument provided in the publication is different to its description.
Results of reliability and validity assessment performed on all publications that cited the instrument.a
| Reliability and consistency | Validity | |||||||||||||||||||||||||||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Intra-rater | Inter-rater | Test–retest | Internal | Content | Construct | Criterion | Known group | |||||||||||||||||||||||||||||
| Criteria | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | 11 | 12 | 13 | 14 | 15 | 16 | 17 | 18 | 19 | 20 | 21 | 22 | 23 | 24 | 25 | 26 | 27 | 28 | 29 | 30 | 31 | 32 | 33 | 34 | 35 | 36 |
| A | A | A | A | A | A | A | A | A | P | P | A | P | P | P | P | P | P | P | P | P | P | P | P | P | P | P | P | P | P | P | P | P | P | P | P | |
| A | A | A | A | A | N | N | N | N | A | A | A | A | A | A | A | A | A | A | A | P | P | P | P | A | P | P | P | P | A | P | P | P | A | P | P | |
| A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | P | P | P | P | A | P | P | A | A | A | A | A | A | A | A | A | |
| A | A | A | A | A | A | A | A | A | P | A | A | P | A | P | P | A | A | A | P | P | P | P | P | A | P | P | A | A | A | A | A | P | P | P | P | |
| A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | P | A | P | P | A | A | A | A | A | P | A | P | P | |
| A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | P | P | P | A | A | A | A | A | A | A | A | A | A | A | A | A | |
| P | P | A | P | A | A | A | A | A | A | A | A | A | P | A | A | A | A | A | A | P | P | P | P | P | P | P | A | A | A | A | A | A | A | A | A | |
| A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | P | P | P | P | P | P | P | P | P | A | P | P | A | A | A | A | A | P | P | P | P | |
| A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | P | P | P | P | P | A | A | A | P | A | P | P | A | A | A | A | A | P | P | P | P | |
| A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | P | A | P | P | |
| A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | A | P | P | P | A | A | A | A | A | A | A | A | A | A | A | A | A | |
A, absent in any of the publications assessed which contain this instrument; P, present in any of the publications assessed which contain this instrument; N, not applicable for this instrument.
Assessment criteria in Table 2, Table 3.
Heart failure as assessed by the International Small Animal Cardiac Health (ISACH) score was considered to be a construct, not a criterion measure, therefore criterion validity was not performed.
Criterion 27 was likely to have been achieved, but the description of the methodology was inadequate to allow for replication.
The criterion measures used did not fulfil the accepted definitions used in this review.
The context in which this instrument was used means that construct and criterion validity were not relevant, but they could in theory be assessed.