Literature DB >> 32977111

Measuring bonding or attachment in the parent-infant-relationship: A systematic review of parent-report assessment measures, their psychometric properties and clinical utility.

A Wittkowski1, S Vatter2, A Muhinyi2, C Garrett2, M Henderson3.   

Abstract

BACKGROUND: Meaningful, valid and reliable self-report measures can facilitate the identification of important parent-infant-relationship factors, relevant intervention development and subsequent evaluation in community and clinical contexts. We aimed at identifying all available parent-report measures of the parent-infant-relationship or bond and to appraise their psychometric and clinimetric properties.
METHOD: A systematic review (PROSPERO: CRD42017078512) was conducted using the, 2018 COSMIN criteria. Eight electronic databases were searched. Papers describing the development of self-report measures of the parent-infant-bond, attachment or relationship from pregnancy until two years postpartum or the assessment of their psychometric properties were included.
RESULTS: Sixty-five articles evaluating 17 original measures and 13 modified versions were identified and reviewed. The studies' methodological quality (risk of bias) varied between 'very good' and 'inadequate' depending on the measurement property assessed; however, scale development studies were mostly of 'inadequate' quality. Although most measures had good clinical utility, the psychometric evaluation of their properties was largely poor. The original or modified versions of the Postpartum Bonding Questionnaire collectively received the strongest psychometric evaluation ratings with high quality of evidence.
CONCLUSIONS: This novel review revealed that only a few antenatal and postnatal measures demonstrated adequate psychometric properties. Further studies are needed to determine the most robust perinatal measures for researchers and clinicians.
Copyright © 2020. Published by Elsevier Ltd.

Entities:  

Keywords:  Fathers; Measurement; Mothers; Quality assessment; Reliability; Validity

Mesh:

Year:  2020        PMID: 32977111      PMCID: PMC7695805          DOI: 10.1016/j.cpr.2020.101906

Source DB:  PubMed          Journal:  Clin Psychol Rev        ISSN: 0272-7358


Introduction

A large body of research now confirms that the early stages of the parent and infant relationship exert an important influence over a child's future development, psychological wellbeing and life chances (Ainsworth, 1979; Bowlby, 1982) and infancy is considered to be the most cost-effective time to intervene (Doyle, Harmon, Heckman, & Tremblay, 2009). Consequently, various organizations and government reports have been advocating the need to address the early stages of parenting with the intent of strengthening the early parent-infant-relationship (e.g., Allen, 2011; Ellyatt, 2017; Moullin, Waldfogel, & Washbrook, 2014; NHS England, NHS Improvement, and National Collaborating Centre for Mental Health, 2018; National Institute for Health and Care Excellence (NICE), 2013, National Institute for Health and Care Excellence (NICE), 2014; Public Health England, 2019; World Health Organization (WHO) and Calouste Gulbenkian Foundation, 2014; Wright et al., 2015). A good early parent-infant-relationship, in which the parents are sensitive and responsive to their infant's physical and emotional needs, lays the foundation for a child's future self-esteem and resilience, their ability to regulate their emotions and their capacity to form close relationships (Bowlby, 1979, Bowlby, 1988; Thompson, 2000; Winston & Chicot, 2016; Wright et al., 2015). Conversely, poor early relationships place children at increased risk of poor cognitive, social and emotional outcomes (Leclére et al., 2014; van Ijzendoorn, Schuengel, & Bakermans-Kranenburg, 1999; Winston & Chicot, 2016; Wright et al., 2015). Given the importance of the early parent-child-relationship and emotional bond, it is paramount to identify how to support parents in strengthening or improving this relationship effectively when there are any difficulties. An important step in doing so is to be able to identify parents who may be struggling to bond with their developing fetus and/or baby in order to offer them an increased level of support (Royal College of Psychiatrists, 2018). However, the prevalence of difficulties in the early parent-child-relationship can vary depending on how and what is being measured, with some researchers examining bonding via questionnaires (Condon & Corkindale, 1998; van Bussel, Spitz, & Demyttenaere, 2010; Wittkowski, Wieck, & Mann, 2007) and this emotional bond or the reciprocal and interactive relationship between parent and infant often referred to as attachment via observation (Ainsworth, Blehar, Waters, & Wall, 1978; Crittenden, 2001; Bicking Kinsey, & Hupcey, 2013; Lotzin et al., 2015; Noorlander, Bergink, & van den Berg, 2008). Hereby, it is imperative to define the terms ‘bonding’ and ‘attachment’ because they are seen as different concepts but often used synonymously (e.g., Benoit, 2004; Bicking Kinsey & Hupcey, 2013; Redshaw & Martin, 2013). Bonding is described as the tie from the parent to the infant (Bicking Kinsey & Hupcey, 2013; Kennell & McGrath, 2005); it generally consists of feelings and emotions that parents experience towards their infant (Bicking Kinsey & Hupcey, 2013). Attachment is seen as the interplay and reciprocity between the parent and the child (Bicking Kinsey & Hupcey, 2013; Kennell & McGrath, 2005), which usually develops during pregnancy between the parent and the fetus (Condon & Corkindale, 1997). Attachment is part of the parent-child-relationship whereby the parent's role is to ensure the safety, security and protection of the child (Bowlby, 1982). Since the concepts of ‘bonding’ and ‘attachment’ are closely related and have been widely researched, we have opted to include both within the term ‘parent-infant-relationship’. The ‘gold standard’ for the assessment of parent-child-attachment, and as such the reciprocal aspect of the parent-child-relationship, is via the use of behavioral, observational measures used with parents or other caregivers and their children over 1 year old, such as the Strange Situation task (Ainsworth et al., 1978) and the Attachment Q-Set (Waters & Deane, 1985). Several observational assessment tools exist to evaluate attachment and interaction behaviours between parent and child (up to 30 months old) (e.g., for reviews see Gridley et al., 2019; Lotzin et al., 2015; Mesman & Emmen, 2013; Tryphonopoulos, Letourneau, & Ditommaso, 2014). However, these measures have two key limitations. Firstly, they are time- and resource-intensive and require extensive training to administer and interpret. This limits their use by practitioners, for example, in obstetric, pediatric or primary-care services which mostly lack the time, facilities and training to administer these assessments (van Bussel et al., 2010). Secondly, it has been argued that it is impossible to gain a complete understanding of attachment without also assessing the subjective experience of the parent (Condon, 2012; Scopesi et al., 2004). Whilst there may always be challenges in enabling parents to disclose any difficulties in bonding or forming emotional ties with their developing fetus or infant to healthcare professionals during routine appointments because parents fear being stigmatised (Morsbach & Prinz, 2006) and/or the involvement of social services and the potential loss of custody (NICE, 2014), developing reliable, valid and sensitive measures may be useful in assisting with the assessment of the early parent-child-relationship and the quest to endorse emotional experiences and beliefs in facilitating parental disclosure. Furthermore, self-report measures allow us to gain insights into the factors parents perceive to influence their relationship with their child. Given the fact that attachment or bonding in the antenatal period is largely one-sided, consisting mainly of the subjective experiences reported by the parent with little observable behavior (relative to the postnatal period) shown by the fetus, antenatal measures are usually self-reported. Although self-report measures are subject to social desirability bias which can skew interpretation (Arnold & Feldman, 1981; van de Mortel, 2008), they are less costly and labour-intensive to administer (Streiner, Norman, & Cairney, 2015). In addition, they allow us to gain an understanding into the parent's subjective experience of their relationship with their child, which can be meaningful clinically and valuable for research (Condon & Corkindale, 1998; Scopesi et al., 2004). In clinical settings, valid and reliable measures, which are quick and easy to administer, can facilitate screening for difficulties in the parent-infant-relationship and they can also be used to assess change (Brockington et al., 2001). Moreover, the relative ease of administration means that these instruments can be more readily incorporated into large-scale studies and surveys, including those with multiple follow-ups, thereby facilitating research in this area (Pallant, Haines, Hildingsson, Cross, & Rubertsson, 2014). In order to have clinical and research utility, self-report measures must meet criteria for validity and reliability (Crandall, 1976; Streiner et al., 2015) and ideally convergence or concurrent validity with a ‘gold standard’ observational measure. However, when choosing a measure, clinicians or researchers also need to know which measure is suitable for their population and which one accurately assesses change (Streiner et al., 2015), as evidence-based assessment is considered intrinsic to professional practice (e.g., Hunsley & Mash, 2008). Several parent-report measures of the early parent-infant-relationship have been developed, which differ in terms of their focus, format, content, length, theoretical underpinnings, the purpose for which they were developed, and the extent to which information exists regarding their validity and reliability. Whilst recent reviews have explored the associations between pre-and postnatal bonding (McNamara, Townsend, & Herbert, 2019; Tichelman et al., 2019), only three reviews have explicitly assessed self-report measures of the parent-child-relationship and examined their psychometric properties (Perrelli et al., 2014; Pritchett et al., 2011; Van den Bergh & Simons, 2009). These three reviews differed in their focus. Van den Bergh and Simons (2009) critically evaluated information of the construction and psychometric properties of three maternal-fetal attachment measures only: the Prenatal Attachment Inventory (PAI, Müller, 1993), the Maternal-Fetal Attachment Scale (MFAS, Cranley, 1981) and the Maternal Antenatal Attachment Scale (MAAS, Condon, 1993). Although the PAI and the MFAS appeared to have some robust psychometric properties, all three measures had weaknesses and required further psychometric validation. Pritchett et al. (2011) described the validity and reliability of measures of family functioning. However, the inclusion of 107 measures meant that no measures were reviewed in specific detail. Finally, Perrelli, Zambaldi, Cantilino, and Sougey (2014) undertook an integrative review of measures that could be used in pregnancy and in the first year postpartum. Their review identified 23 articles published after 2002 relating to 13 measures, of which ten were measures completed by parents. Whilst this review identified many of the important and widely used parent-report measures of the early parent-infant-relationship, only a relatively small number of research studies and measures were identified. A further limitation of these reviews is that they did not use a formal quality appraisal tool which would have allowed for a detailed assessment of the papers' methodological quality and an easier comparison between measures (Terwee et al., 2012). However, a standardized, evidence-based approach to reporting the psychometric properties is essential in order to ensure that the quality of measures used in clinical practice and service improvements is appropriately high (e.g., Kilbourne et al., 2018). Consequently, a review that can be considered a relevant and comprehensive guide for researchers and clinicians is required, especially given the focus on the expansion of perinatal mental health services (NHS England, NHS Improvement, and National Collaborating Centre for Mental Health, 2018). Taking into consideration the aforementioned gaps in the perinatal field, the main aim of the current systematic review is to provide an overview and evaluation of existing parent- measures of the parent-infant-relationship to assist researchers and clinicians in identifying the most suitable measure to use in their research, practice or service. Specifically, the current review addresses the limitations of previous reviews by bridging the gap between the depth and breadth of the included measures within the systematic review. We aimed to achieve this by a) appraising only measures that assess the parent-infant-relationship in terms of perceived bond or parent-reported attachment rather than broader or related concepts (e.g., maternal self-efficacy, maternal attitudes) in studies that specifically aimed to develop a measure or test its psychometric properties; b) utilising a systematic search strategy to increase confidence that the review included a comprehensive list of measures; c) applying the COnsensus-based Standards for the selection of health Measurement INstruments (COSMIN; Mokkink et al., 2018; Prinsen et al., 2018; Terwee et al., 2018), a comprehensive and systematic tool for appraising and comparing the quality of individual measures in terms of their psychometric properties and clinical utility, and d) identifying relevant administrative properties of the identified measures (Bot et al., 2004).

Methods

This systematic review was conducted in line with the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) guidelines (Moher et al., 2009) and registered with the PROSPERO database (www.crd.york.ac.uk/prospero; registration number CRD42017078512).

Search strategy and paper inclusion

In keeping with the aims of the review, we conducted a systematic literature search in eight electronic databases using four steps. In step 1 we designed a search strategy to retrieve peer-reviewed papers relevant to the development, validation and implementation of self-report measures of the early parent-infant-relationship, which was piloted by one reviewer (CG). The aim of this pilot search was to increase specificity and sensitivity to capture the highest possible proportion of relevant articles. This pilot search resulted in low specificity with too many irrelevant articles being initially retrieved; thus, following further consultation with a university librarian, we refined our search strategy by adding the ‘adjacency’ operator (abbreviated as ‘ADJn’ whereby n refers to a number of words from each other in any order) to our search terms. This strategy led to step 2 whereby another reviewer (SV) conducted the search in three electronic platforms (Ovid, Clarivate and EBSCO) and their eight bibliographic databases from their inception to the end of April 2019: MEDLINE, Embase, PsycINFO, PsycARTICLES, Maternity and Infant Care, Health and Psychosocial Instruments, Web of Science and CINAHL. The search was updated in August 2019. Limitations for language or year of publication were not set because the exclusion of non-English studies could introduce risks of bias and, therefore, each non-English study should be evaluated case-by-case to maintain internal validity (Higgins & Green, 2011; Neimann Rasmussen & Montgomery, 2018). In step 2, we searched the following terms in the title, abstract or keywords in those eight databases (see Appendix A for a sample search strategy): 1) (parent*or maternal or paternal or mother* or father*) adj7 (child or infant or newborn or foet* or fetus or fetal or baby or neonate); 2) (antenat* or prenat* or puerper* or postnat* or postpart* or peripartum or pregnan* or perinat*); 3) (measur* or scale$ or questionnaire$ or construct$ or tool$ or inventor* or instrument$ or test*) adj7 (attachment or relation* or bond* or orientation or synchrony or synchronicity or “emotional availability” or attitude* or belief* or responsiv* or feel* or interact*). Papers retrieved from this search were then screened for measures relevant to the aims of the review. In step 3, further searches with the names of identified measures were conducted in a ninth database (i.e., PubMed) to identify the original development/validation paper(s) for that measure as well as papers reporting further validation work undertaken with any identified and included measures. In the final and fourth step, the reference lists of included articles were checked for additional relevant studies. When the initial development/validation work for a measure was unpublished, further information was sought from study authors. When this was not possible, we extracted relevant development and validation process information about this measure from papers by the original authors. To verify inter-rater reliability of the screening, an independent research assistant (CS) independently double-screened 1% of all identified articles during the screening stage and 20% of potentially eligible articles to determine their inclusion or exclusion. The percentage of inter-rater agreement and Cohen's kappa were calculated on both types of screening to ensure the validity of the screening process.

Inclusion criteria of papers and article selection

Papers were included if they described the initial development and validation of a relevant measure. Papers were also included if they described an attempt to validate and/or to test the psychometric properties of an included measure, and this was the clearly stated aim of the paper. Decisions about the inclusion/exclusion of measures and papers were based on the initial judgment of two reviewers (AW and CG). Their decisions were verified by two other reviewers (SV and AM), and any disagreements were resolved through consultation with the fifth reviewer (MH).

Inclusion and exclusion criteria of measures

Measures were included if they were completed by the parent and assessed the parent's perception of the parent-infant-relationship or bond during the antenatal period or the postnatal period up until an infant age of two years. Measures were excluded if they were not assessing the parent-infant-relationship per se but instead assessed a related concept (e.g., ‘parenting style’ or ‘attitudes to pregnancy’) or if they only assessed the parent-infant-relationship as part of a subscale in a longer inventory (e.g., the MAMA, Kumar, Robson, & Smith, 1984). As the content of measures assessing related constructs (e.g., maternal self-efficacy, maternal attitudes, etc.) can be very similar to those of measures explicitly described by authors as measures of bonding or attachment, we based inclusion decisions on item content rather than author description (e.g., the How I Feel About My Baby Now Scale, FAB, Leifer, 1997; the Mothers' Object Relations Scales Short Form, MORS-SF, Oates, Gervai, Danis, Lakatos, & Davies, 2018). On several occasions, original measure authors or other researchers proposed shortened or alternative versions of measures which had already been identified and included in the current review. For example, we included the Short Postpartum Bonding Questionnaire (SPBQ; Kinsey, Baptiste-Roberts, Zhu, & Kjerulff, 2014), which was based on the original 25-item Postpartum Bonding Questionnaire (PBQ; Brockington et al., 2001) but shortened to address the need for a briefer instrument to measure parent-infant-bonding as part of a large scale telephone interview survey. Similarly, in several cases, researchers (but not the original authors) conducted psychometric testing on slightly different versions of measures (e.g., containing fewer items or having fewer Likert response categories). These alternative versions were also included in the current review and treated as separate independent measures.

Assessing the psychometric properties of included measures

We evaluated the measurement properties of the included studies and measures using: 1) the COSMIN criteria for evaluating the quality of the measure development studies and content validity studies (Terwee et al., 2018), 2) the COSMIN Risk of Bias checklist (Mokkink et al., 2018) to assess the methodological quality of the studies, 3) the COSMIN checklist to examine eight psychometric results, including structural validity, internal consistency, reliability, hypothesis testing for construct validity, cross-cultural validity/measurement invariance, measurement error, criterion validity and responsiveness (Mokkink et al., 2018; Prinsen et al., 2018), and 4) the modified Grading of Recommendations Assessment, Development, and Evaluation (GRADE) approach to examine the quality of the evidence (Mokkink et al., 2018). All materials are available at www.cosmin.nl/index.html.

Step 1: Quality assessment of included studies

The first step in the process to assess the methodological quality of included studies is achieved via the application of the ‘COSMIN Risk of Bias checklist’ (Mokkink et al., 2018). This checklist consists of categories for appraising the quality of the outcome measure development studies as well as the quality of various psychometric measurement properties which are outlined above (see Table 1 for definitions of measurement properties, Mokkink et al., 2018). Content validity was assessed in terms of relevance, comprehensiveness and comprehensibility of the measure's items (Terwee et al., 2018).
Table 1

Definitions and criteria for good measurement properties⁎.

Definitions and criteria for good measurement properties⁎. During the pilot stage we identified that many researchers did not explicitly describe what they explored or evaluated in terms of content validity in their studies. Consequently, we expanded the COSMIN's definitions of ‘relevance’ and ‘comprehensibility’: studies were considered to evidence ‘relevance’ when they evaluated the relevance, appropriateness, suitability and/or acceptability of each item in the target population. In terms of ‘comprehensibility’, studies were rated if they evaluated the understanding, coherence, clarity, meaning and/or ambiguity of the items and whether the response options, instructions and/or the recall period were clear and comprehensible. ‘Comprehensiveness’ was evaluated in accordance with the COSMIN guidelines whereby participants should have been explicitly asked about whether the items comprehensively covered the construct that the outcome measure (or the sub-scale) intended to measure or if the included domains together comprehensively covered the wider construct measured by the total score of the outcome measure (Terwee et al., 2018). Each measurement property (including content validity) was rated across several items assessing different aspects of quality, using a four-point COSMIN Risk of Bias scale (i.e., 4 = ‘very good’, 3 = ‘adequate’, 2 = ‘doubtful’, 1 = ‘inadequate’). An overall score for the methodological quality of a study was determined for each measurement property separately by taking the lowest rating of any of the items in a given category. When the developers of the original version of a measure omitted to provide detailed information on one or more psychometric properties, but there was sufficient information to assume that the study was conducted adequately, we deviated from the stricter COSMIN guidance and opted to give an ‘adequate’ or ‘doubtful’ rating rather than an ‘inadequate’ rating. For example, if in the measure development study it was unknown whether the qualitative data, collected for the purposes of cognitive interview or pilot testing, were coded by one or two researchers independently, we rated it as ‘doubtful’ rather than ‘inadequate’ due to lack of information. Interpretability or the degree to which one can assign qualitative meaning to quantitative scores (Mokkink et al., 2010, Mokkink et al., 2009) is not considered a measurement property but an important characteristic of a measurement instrument (Mokkink et al., 2018). This means that investigators should provide information about clinically meaningful differences in scores between subgroups, floor and ceiling effects, and the minimal (clinically) important change (Mokkink et al., 2009). However, since a limited number of studies reported aspects of interpretability, we could not present this in our review.

Step 2: assessment of study outcomes

The second step involved assessing the study results for each of the included measures, according to the updated 2018 measurement for good measurement properties (Mokkink et al., 2018; Prinsen et al., 2018). These criteria cover eight measurement properties, for each of which the rater is required to assign ‘+’, ‘?’ or ‘–’. A ‘+’ is assigned when the study findings provide good evidence of a measure exhibiting this property (i.e., ‘sufficient’ rating); a ‘?’ is assigned when results are equivocal or appropriate tests have not been performed (i.e., ‘indeterminate’ rating) and a ‘–’ is assigned when appropriate tests have been performed and the result suggests that the measure does not exhibit this property as defined by the checklist criterion (i.e., ‘insufficient’ rating). This checklist and quality criteria are presented in Table 1. The content validity of an outcome measure was evaluated according to the quality and results of the available studies and the outcome measure itself (Terwee et al., 2018). Although the COSMIN guidelines suggest not to rate a study if the quality of the study (according to the risk of bias assessment) was ‘inadequate’, we decided to rate all studies, including those with an ‘inadequate’ quality rating, in order to gain a comprehensive overview of a particular outcome measure. Content validity of each outcome measure was rated according to the development studies (scored as ‘+’, ‘?’ or ‘–’ for ‘sufficient’, ‘indeterminate’ and ‘insufficient’ ratings, respectively), available content validity studies (also scored as ‘+’, ‘?’ or ‘–’) and ratings given by two reviewers (AW and SV) (scored as ‘+’, ‘±’ or ‘–’ for ‘sufficient’, ‘inconsistent’ and ‘insufficient’ ratings, respectively). When no content validity studies were available or only content validity studies of inadequate quality were available, the overall ratings for content validity were determined according to the reviewers' ratings as per COSMIN criteria. In order to rate the structural validity of measures, we had to adapt the criteria as the current 2018 COSMIN criteria for good measurement properties do not include guidance for rating the results of Exploratory Factor Analysis (EFA). Consequently, EFAs were rated as ‘sufficient’ if ≥50% of the variance was explained (as in previous versions of the COSMIN criteria; see Terwee et al., 2012). Such evidence was downgraded for methodological quality based on the risk of bias checklist (i.e., studies using EFA can only be rated as ‘adequate’ rather than ‘very good’). When the % of variance accounted for (in the case of EFA) or model fit statistics (in the case of CFA) were not reported, an ‘indeterminate’ rating was assigned. Finally, when higher quality evidence (e.g., CFA) was available for a given measure, lower quality evidence (e.g., EFA) was ignored. In terms of hypothesis testing for construct validity, the decision was made to include any published measure as a comparator instruments that measured a similar construct (e.g., other attachment measures included in the current review or a subscale from a measure not included in the review, such as the attitudes towards pregnancy and the baby subscale of the MAMA scale, Kumar et al., 1984). To receive a ‘sufficient’ rating, 75% of the correlations tested had to meet the cut-off of r ≥ 0.50 against a comparator instrument measuring a similar construct. Given the lack of an established self-report measure for the construct under study, caution is needed in interpreting the results for this measurement property. We also adapted the criteria for rating reliability due to ambiguity in the COSMIN guidelines in a way that the studies that reported a correlation coefficient (i.e., Pearson's or Spearman's) but did not report the intraclass correlation coefficient (ICC) for (test-retest) reliability would still receive a ‘sufficient’ or ‘insufficient’ rating which would normally receive an ‘indeterminate’ rating if the ICC was not applied. Instead, we decided to reflect this in the quality of evidence (described in Section 2.4.3) whereby studies that did not use the more robust method (i.e., the ICC) would get a lower rating even if the study received a ‘sufficient’ rating on reliability.

Step 3: summary and quality grading of the evidence

As per COSMIN criteria, the psychometric findings reported in each of the included studies were summarized and graded for each measure. This process resulted in each measure being assigned two ratings: 1) an overall rating of ‘sufficient’ (‘+’), ‘insufficient’ (‘–’) or ‘indeterminate’ (‘?’) for the eight psychometric properties (except content validity), or an overall rating of ‘sufficient’ (‘+’), ‘insufficient’ (‘–’) or ‘inconsistent’ (‘±’) for content validity, and 2) an overall rating of methodological quality for each measurement property (‘high’, ‘moderate’, ‘low’ or ‘very low’). The latter rating is achieved through following the modified GRADE approach, which involves consideration of several factors in rating methodological quality of the pooled results (e.g., the evidence for risk of bias, inconsistency of results and imprecision through small sample sizes). Importantly, this approach also takes into account the number of available studies and the methodological quality of each individual study. More detailed information on how GRADE was conducted can be found in the COSMIN manual (Mokkink et al., 2018; Prinsen et al., 2018; Terwee et al., 2018). Definitions for each of the GRADE quality level ratings are shown in Table 2.
Table 2

Definitions of quality levels using the GRADE approach.

Notes: Estimate of the measurement property refers to the pooled or summarized result of the measurement property of a given measure. These definitions were adapted from the GRADE approach.

Definitions of quality levels using the GRADE approach. Notes: Estimate of the measurement property refers to the pooled or summarized result of the measurement property of a given measure. These definitions were adapted from the GRADE approach.

Step 4: assessment of practical administrative properties

Practical, administrative or clinimetric properties that would affect the ease with which each of the measures could be employed in a clinical or research context, but are not covered in the COSMIN or Terwee et al. (2018) checklists, were also assessed. The following properties were assessed: Time to administer: As measure authors did not routinely report this property, this was assessed independently by two reviewers (SV and AM) who completed each measure and timed themselves. As per Bot et al.'s (2004) clinimetric checklist, a positive rating was given when the questionnaires could be completed within 10 min. Ease of scoring refers to the extent to which the measure can be scored by a trained investigator or expert. In accordance with Bot et al.'s (2004) checklist, the scoring method was rated as easy when the items were simply summed, moderate when a visual analogue scale or simple formula was used and difficult when either a visual analogue scale in combination with a formula or a complex formula was used. Readability and comprehension: The Flesch Reading Ease (FRE; Flesch, 1948) method was used to assess readability and comprehension. The text is rated on a 100-point-scale in which 100 represents the easiest text and 0 the hardest. Measures scoring ≥90 using the FRE were considered excellent for this property; measures scoring between 80 and 89 were considered good; measures scoring between 70 and 79 were considered fair and measures with scores below <69 were considered poor. Availability and conditions of use refers to the ease with which researchers/clinicians can obtain the questionnaire and whether it is free to use. If the measure was easily accessible on the internet or through e-mailing the first author, and it was also free to use, availability was classed as excellent. If a measure was difficult to obtain but was free of cost, the measure was classified as good. If a measure was easy to obtain but had a cost, the measure was classed as fair. Finally, if a measure was difficult to obtain and there was a cost for accessing or utilising the instrument, the measure was classified as ‘poor’.

Inter-rater reliability

Extraction of data and the assessment of the methodological quality (i.e., risk of bias) was performed by reviewers independently (AM, SV and CG). The assessment of all psychometric properties (except content validity) was performed by one reviewer (AM). The reliability of ratings was confirmed by having another reviewer (SV), who independently rated 20% of the papers. For the measure development studies and content validity studies, another reviewer (SV) completed risk of bias and rated content validity; 20% of those papers were independently rated by an independent research assistant (CS). Inter-rater reliability was met if Cohen's kappa between the two reviewers was above 0.61, indicating ‘substantial’ agreement (McHugh, 2012), on all psychometric ratings. When this was not achieved, the disagreements were discussed and resolved through consultation with another reviewer (AW).

Results

Review process

The original search identified 15,924 papers. After removing duplicates, the titles and abstracts of 12,081 papers were screened. The titles, abstracts and/or full texts of 220 papers were examined against inclusion and exclusion criteria. In August 2019 the search was repeated which resulted in 600 hits between January and August 2019 which were fully screened. Only two studies were identified: a study describing the development of the MAAS-13 and PAAS-13 (Göbel et al., 2019) which was included in the review and a non-English study describing a Slovenian version of the PAI (presented in Appendix B alongside other non-English papers). The agreement for the screening of 1% of all identified articles was 94.7% (kappa = 0.90) and for the 20% of potentially eligible articles the agreement was 75% (kappa = 0.58). Any discrepancies in the exclusion and inclusion of studies were resolved among all reviewers through discussion. After a detailed assessment, 65 papers evaluating 17 original measures in associated development studies and 13 modified versions, derived from only four of the identified 17 measures, were included in the review (for the references of the included papers, please see Appendix C). In total, 14 antenatal measures (eight original and six modified versions) and 18 postnatal measures (ten original measures plus eight modified version), of which one measure (the Prenatal and Postnatal Bonding Scale, PPBS, Cuijlits et al., 2016) could be used antenatally and postnatally, were reviewed. The majority of these measures were maternal, but we also identified four paternal measures (three antenatal and one postnatal version). The search process and outcome are illustrated in Fig. 1.
Fig. 1

Flowchart of paper selection based on PRISMA guidance.

Flowchart of paper selection based on PRISMA guidance.

Study characteristics and information on measure development

The publication dates of the studies describing the original 17 measures ranged from 1977 to 2018. The validation work undertaken for the original 17 measures included studies conducted in eight different countries, such as the USA (n = 5), Australia (n = 4), the UK (n = 4), the Netherlands (n = 1), Hungary (n = 1), Korea (n = 1), Sweden (n = 1) and India (n = 1); however, the sample in the MORS-SF (Oates et al., 2018) scale development comprised British and Hungarian mothers (see Table 3 for details). Study sample sizes reported by measure authors in their development studies ranged from 19 to 1050 women and 100 to 461 men. The majority of studies included non-clinical samples (n = 15), with only two studies using a clinical sample of women with mental illness (Brockington et al., 2001; Hackney, Braithwaite, & Radcliff, 1996).
Table 3

Overview of the included measures and summary of their administrative and clinimetric properties.

Overview of the included measures and summary of their administrative and clinimetric properties.

Description of identified measures

A description of each of the 17 original measures is presented in Table 3. The majority of measures focused on the assessment of parent reported perceptions of bonding and attachment. However, one measure also focussed on the psychodynamic concept of object relations (i.e., the MORS-SF, Oates et al., 2018) and another measure also focussed on the assessment of maternal role attainment (i.e., the Mother-to-Infant Relations and Feelings Scale, MIRFS, Thorstensson et al., 2012). Although the authors did not set out to test the MIRFS, we included it in this review because Ekström and Nissen (2006) described the MIRFS' development and evaluated its content validity. Eight scales were measures of the parent-fetus-relationship, administered to expectant women and men in the antenatal period of pregnancy: the How I Feel About My Baby Now Scale (FAB, Leifer, 1997), the Maternal-Fetal Attachment Scale (MFAS, Cranley, 1981), the Prenatal Attachment Inventory (PAI, Müller, 1993), the Maternal Antenatal Attachment Scale (MAAS, Condon, 1993), the Pre- and Postnatal Bonding Scale (PPBS, Cuijlits et al., 2016). the Paternal-Fetal Attachment Scale (PFAS; Weaver & Cranley, 1983), the Paternal Antenatal Attachment Scale (PAAS; Condon, 1993) and the Korean Paternal-Fetal Attachment Scale (K-PAFAS; Noh & Yeom, 2017. As per COSMIN criteria, any modified versions of measures were reviewed separately even if they differed from the original scale by only one item. Three modified versions were identified for the original 24-item-MFAS offering different item totals: the MFAS-23 (Müller, 1993; Müller & Ferketich, 1993), the MFAS-20 (Busonera, Cataudella, Lampis, Tommasi, & Zavattini, 2016), and the MFAS-17 (Seimyr, Sjögren, Welles-Nystrom, & Nissen, 2009; Sjögren, Edman, Widstrom, Mathieson, & Uvnas-Moberg, 2004). The original 19-item-MAAS had also been shortened in modified versions referred to as the MAAS-13 (Göbel et al., 2019) and MAAS-12 (Navarro-Aresti, Iraurgi, Iriarte, & Martinez-Pampliega, 2016). Five original and five modified versions of these were measures of the mother-fetus-relationship: the FAB, the MFAS-24, the MFAS-23, the MFAS-20, the MFAS-17, the PAI-21, the MAAS-19, the MAAS-12, the MAAS-12 and the PPBS. Only three measures assessed the father-fetus-relationship, namely the PFAS, the PAAS and the K-PAFAS. The PAAS has also been revised and shortened to the 13-item-PAAS (Göbel et al., 2019). Ten original scales and eight modified versions were measures of the parent-infant-relationship. As can be seen in Table 3, the original scales were the Maternal Attachment Inventory (MAI-26; Müller, 1994), the Mother Infant Attachment Scale (MIAS; Bhakoo, Pershad, Mahajan, & Gambhir, 1994), the Mother and Baby Interaction Scale (MABISC; Hackney et al., 1996), the Maternal Postnatal Attachment Scale (MPAS; Condon & Corkindale, 1998), the Postpartum Bonding Questionnaire (PBQ-25; Brockington et al., 2001), the Mother-to-Infant Bonding Scale (MIBS-8; Taylor, Atkins, Kumar, Adams, & Glover, 2005), the Mother-to-Infant Relations and Feelings Scale (MIRFS; Thorstensson et al., 2012), the PPBS (Cuijlits et al., 2016), the Mothers' Object Relations Scales Short Form (MORS-SF; Oates et al., 2018) and the Paternal Postnatal Attachment Scale (PPAS; Condon, Corkindale, & Boyce, 2008). Only the PPAS was designed to assess the father-infant-relationship. Of these measures, only the PPBS could be used both antenatally and postnatally. The MFAS (Cranley, 1981) and the PFAS (Weaver & Cranley, 1983) as well as the MAAS (Condon, 1993) and the PAAS (Condon, 1993) are maternal and paternal versions of the same measures, respectively, and can be completed by mothers and fathers in the same family. The MAI (Müller, 1994) is the postnatal version of the PAI (Müller, 1993). Condon and colleagues (Condon, 1993; Condon et al., 2008; Condon and Corkindale, 1997, Condon and Corkindale, 1998; Condon, Corkindale, Boyce, & Gamble, 2013) have produced measures based on Condon's (1993) model of human attachment with antenatal and postnatal measures for both mothers and fathers (i.e., MAAS, PAAS, MPAS and PPAS). Of the postnatal measures, the PBQ has received the most attention by other researchers who have produced shorter versions, including the PBQ-22 (Wittkowski, Williams, & Wieck, 2010), the PBQ-19 (Vengadavaradan, Bharadwaj, Sathynarayanan, Durairaj, & Rajaa, 2019), the PBQ-16 (Reck et al., 2006), the PBQ-16-J (Kaneko & Honjo, 2014), the PBQ-14 (Suetsugu, Honjo, Ikeda, & Kamibeppu, 2015) and the Short PBQ with 10 items (Kinsey et al., 2014). Although the PBQ has also been used with fathers at 2 months postpartum in a Swedish study (Edhborg, Matthiesen, Lundh, & Widström, 2005), the PBQ was not specifically developed to be used with fathers. As Edhborg et al. (2005) did not evaluate the psychometric properties of the PBQ in their male sample, this study was not rated in our review. The only other postnatal measure with modified versions is the 8-item-MIBS, which had been reduced to seven items in the MIBS-J-7 (Ohara et al., 2016) and extended to 10 items in the MIBS-J-10 (Yoshida, Conroy, Marks, & Kumar, 2012).

Additional information on the measures' items and target population

The majority of the measures comprise items that are worded as statements on a Likert-scale that typically enquire how the mother is feeling towards the developing fetus or the newborn. For example, the PAI includes items, such as “I stroke the baby through my tummy” or “I enjoy feeling the baby move”. The PBQ includes items, such as “I feel close to my baby” or “My baby winds me up”, and the MPAS includes items, such as “When I am not with the baby, I find myself thinking about the baby: …” or “Taking care of this baby is a heavy burden of responsibility. I believe this is: …”. Only one measure, the MIRFS, was a two-part measure in which seven items (worded as statements) evaluated the mothers' perception about the relationship between the mother and her baby (e.g., “I talk a lot with my baby” and “I do not talk at all with my baby”) and seven items (worded as adjectives) explored the mothers' current feelings towards the baby (e.g., “Difficult” and “Easy”). Although most studies described the population with whom the study was conducted, the majority of the studies did not specify the target population of parents by providing information about the gestation age or the infant's age. Consequently, it was impossible to determine the measures' applicability to parents of infants at different developmental ages and we had to assume that they targeted parents of children younger than two years old. The number of items used in the original 17 measures ranged from five (e.g., the PPBS) to 26 items (e.g., the MAI). Of the original 17 measures, seven measures were unidimensional (FAB, PAI, MAI, PPBS, MABISC, MIBS-8 and MIAS), whereas ten measures included multiple sub-scales (MFAS, PFAS, MAAS, PAAS, K-PAFAS, MPAS, PPAS, PBQ, MORS-SF and MIRFS), which ranged from two (e.g., MAAS, PAAS and MORS-SF) to five subscales (e.g., MFAS). Most measures were designed for the assessment of parents within a non-clinical population who were asked to reflect on their feelings or thoughts in the present moment (e.g., the FAB), the past two weeks (e.g., the PAAS) or during the past month (e.g., the PPBS, the PAI-21, the MABISC). However, six measures did not state a specific recall time and three measures accepted a variable timeframe. As our search did not exclude studies not written in English initially, our review identified that most of the 17 original and 13 modified measures are available in a total of 17 languages including English. Other language versions included measures in Chinese, Korean, Japanese, German, Italian, French, Portuguese, Spanish, Dutch, Swedish, Norwegian, Hungarian, Turkish, Persian, Tamil and Hindi. Four measures were only available in one language: the FAB in English only, the MIAS in Hindi only, the K-PAFAS in Korean only and there appears to be only a Swedish version of the MIRFS. In addition, the MORS-SF was validated on a mixed sample of British and Hungarian women. Although it was our original intention to include measures not written in English, we were unable to a) apply the COSMIN criteria consistently across these studies ourselves and b) identify professional translators trained in the application of the COSMIN criteria for the foreign language papers we identified. For comprehensiveness, the foreign language papers are presented in Appendix B.

Measurement properties assessed

Sixty-five studies pertaining to the 17 measures and their 13 modified versions rated aspects of validity and reliability (see Table 4 and Table 5). Several of the included studies (e.g., Bienfait et al., 2011; Brockington et al., 2006, Brockington et al., 2001) tested for diagnostic accuracy (i.e., sensitivity and specificity of the measure in detecting bonding difficulties), but this property does not fall within the COSMIN taxonomy and consequently it was not rated.
Table 4

Quality of the measure development (n = 17 measures) and content validity (n = 16 measures) (Baldisserotto et al., 2018; Chen et al., 2013; Della Vedova and Burrp, 2017; Golbasi et al., 2015; Linacre, 2002; Lingeswaran and Bindu, 2012; Maydeu-Olivares, 2013; Riera-Martin et al., 2018; Schermelleh-Engel et al., 2003; Shin and Kim, 2007; Siu et al., 2010; Stochl et al., 2012; Yoshida et al., 2012).

Table 5

Synthesis of psychometric properties and quality of evidence (using GRADE)⁎.

Quality of the measure development (n = 17 measures) and content validity (n = 16 measures) (Baldisserotto et al., 2018; Chen et al., 2013; Della Vedova and Burrp, 2017; Golbasi et al., 2015; Linacre, 2002; Lingeswaran and Bindu, 2012; Maydeu-Olivares, 2013; Riera-Martin et al., 2018; Schermelleh-Engel et al., 2003; Shin and Kim, 2007; Siu et al., 2010; Stochl et al., 2012; Yoshida et al., 2012). Synthesis of psychometric properties and quality of evidence (using GRADE)⁎.

Assessment of validity

Content validity

According to the COSMIN criteria for assessing content validity (Terwee et al., 2018), the relevance, comprehensiveness and comprehensibility of the 17 measures and their 13 modified versions were rated separately in a multi-step process. Firstly, the overall quality of the development (i.e., risk of bias checklist) of the 17 original outcome measures were evaluated: three original measure development studies (MABISC, MPAS and PPAS) had a ‘doubtful’ rating and 14 original measure development studies were rated as ‘inadequate’ (see detailed ratings in Table 4). In this step, the quality of the content validity studies was also evaluated according to whether participants in a content validity study had been asked about the relevance, comprehensiveness and comprehensibility of the measure items and whether professionals had been asked about the relevance and comprehensiveness of the measure items. A total of 16 out of 65 studies (Table 4) evaluated content validity either among a participant group (i.e., mothers or fathers) or among a professional group (i.e., midwives, psychologists, psychiatrists, researchers, etc.). ‘Relevance’ and ‘comprehensibility’ was evaluated among participants in six and nine studies, respectively, and among professionals in 10 and five studies, respectively. However, none of the content validity studies evaluated ‘comprehensiveness’ among participants. From all content validity studies evaluating relevance, comprehensiveness and comprehensibility among participants and/or professionals, only two studies received an ‘adequate’ rating: Busonera et al. (2016a) for ‘relevance’ among participants using the MFAS and Riera-Martin et al. (2018) for ‘relevance’ among professionals on the MPAS and the PPAS. The remaining studies received a ‘doubtful’ rating for ‘relevance’, ‘comprehensibility’ and/or ‘comprehensiveness’ when these properties were assessed in a particular study (see Table 4 for detailed ratings). Once the quality of the development studies and content validity studies were rated for each measure, the content validity of an outcome measure was evaluated. With regards to the 17 original measures, the overall content validity was rated as ‘sufficient’ (‘+’) for the following measures: the FAB, the PAI and the MABISC. However, the remaining 14 measures were rated to have ‘inconsistent’ (‘±’) evidence for their overall psychometric properties (see Appendix D for detailed ratings). The reasons for ‘inconsistent’ ratings were explored and in the majority of the cases, the information on relevance, comprehensiveness and comprehensibility presented by the measure authors in the papers was poor. However, the independent evaluation by two reviewers of the scale alone indicated that the information given was ‘sufficient’ and met the COSMIN criterion of ‘+’. In the third step, the quality of the evidence was rated using the GRADE approach. As the development study of the measure received ‘inadequate’ quality ratings, according to COSMIN criteria these studies have to receive a lowered rating in terms of the GRADE. Therefore, three scales (MABISC, MPAS and PPAS) were rated as ‘low’ and 14 measures were rated as ‘very low’ for relevance, comprehensiveness and comprehensibility according to the GRADE approach (Table 5).

Structural validity (part of construct validity)

In order to rate structural validity, we adapted the 2018 COSMIN criteria for good measurement properties as previously outlined, because many of the included studies conducted Exploratory Factor Analysis (EFA); however, the COSMIN criteria does not provide guidance for rating the results of the EFA. Structural validity was tested for the majority of the included measures (27 out of 29; 93%), but not for the FAB or the PBQ-16. Of the prenatal measures, only the MFAS-17 and the PPBS were assigned a ‘sufficient’ rating. The remaining prenatal measures were assigned ‘insufficient’ ratings. Of the postnatal measures, the MABISC, the PBQ-22, and the MIBS-J-7 were assigned ‘sufficient’ ratings, whereas the MPAS, the PBQ-25, the MIBS-J-10, the MORS-SF and the PPAS were assigned ‘insufficient’ ratings. The MAI-26 was assigned an ‘inconsistent’ rating because there was evidence for structural validity but for its two different factor structures. The remaining postnatal measures were assigned ‘indeterminate’ ratings (see Appendix E for detailed ratings). The quality of the evidence was graded from ‘very low’ to ‘high’ for this measurement property.

Hypothesis testing (part of construct validity)

As none of the studies included an observer-rated measure of parent-child attachment, correlations had to meet the agreed cut-off against a self-report instrument only. Twenty-five of 30 measures (83.3%) had studies reporting information for this measurement property. Of the prenatal measures, the MFAS-23 and the PFAS were assigned ‘sufficient’ ratings, but the MFAS-24, MFAS-20, PAI-21, MAAS-19 and PAAS-19 were assigned ‘insufficient’ ratings, and the MFAS-17 and the PPBS were assigned an ‘indeterminate’ rating. Of the postnatal measures, the MABISC and MPAS were assigned ‘sufficient’ ratings, the MAI-26, PBQ-25, PBQ-14, MIBS-8 and PPAS were assigned ‘insufficient’ ratings, and the remaining measures were assigned ‘indeterminate’ ratings (i.e., when the hypothesis was not as defined by our review team). The quality of the evidence for all pre- and postnatal measures was graded ‘high’ for this measurement property.

Cross-cultural validity (part of construct validity)

Although many studies (n = 38) aimed to adapt a given measure to different ethnic and language groups and some included back translation and other necessary procedures, none evaluated cross-cultural validity by comparing multiple groups by factor analysis and testing for differential item functioning (e.g., English- and Dutch-speaking), as stipulated in the COSMIN criteria. For this reason, this property is omitted from Table 5. See Appendix B for on overview of which measures have a version available in a different language.

Measurement invariance (part of construct validity)

Only two measures (6.7%) could be rated for this measurement property (see Table 5). The MPAS and the PPAS were assigned ‘sufficient’ ratings of measurement invariance, demonstrating that these two measures appear to be measuring the same underlying construct (when tested with mothers and fathers, respectively). The quality of evidence for this property was ‘moderate’.

Criterion validity

None of the studies reported on assessment of criterion validity. Hence, this property was omitted from Table 5.

Assessment of reliability

Internal consistency

Twenty-eight of the 30 measures (93.3%) had studies reporting on internal consistency. However, internal consistency could only be rated as ‘indeterminate’ for the majority of these measures because they did not demonstrate at least low evidence of ‘sufficient’ structural validity (as per the COSMIN criteria). Of the prenatal measures, the PPBS was assigned a ‘sufficient’ rating, whereas the MFAS-17 was assigned an ‘insufficient’ rating; the remaining prenatal measures were assigned ‘indeterminate’ ratings. Of the postnatal measures, the PBQ-22 was the only measure to receive a ‘sufficient’ rating. The MIBS-J-7 was assigned an ‘insufficient’ rating, and the remaining measures were assigned ‘indeterminate’ ratings. The quality of the evidence was graded from ‘very low’ to ‘high’ for this measurement property (see Table 5).

Reliability

Eleven of 30 measures (36.7%) had studies that reported test re-test reliability as defined by the COSMIN criteria. Of the maternal measures, the PAI-21, the MPAS, the PBQ-25, the PBQ-14, and the MORS-SF were all assigned ‘sufficient’ ratings, whereas the MAI-31, the MABISC, and the MIBS-J were assigned ‘insufficient’ ratings. Of the paternal measures, the PFAS was assigned a ‘sufficient’ rating whereas the PPAS was assigned an ‘insufficient’ rating. Evidence was graded from ‘very low’ to ‘moderate’ for this measurement property.

Measurement error and responsiveness

None of the included studies measured measurement error and responsiveness as defined by the COSMIN criteria; hence, these could not be rated and were omitted from Table 5. The agreement between the two reviewers was 89.0% (kappa = 0.85) for the risk of bias ratings, 79.7% (kappa = 0.68) for the measurement properties and 84.5% (kappa = 0.73) for quality of evidence (i.e., GRADE).

Clinimetrics/clinical utility

The clinimetrics or clinical utility of the 17 measures, presented in Table 3, were assessed in terms of time of administration, ease of scoring, Flesch Reading Ease (FRE), availability and conditions of use.

Time to administer

Based on Bot et al.'s (2004) suggestion of a desirable completion time of less than 10 min, all 17 measures were assessed independently by two reviewers (SV and AM) and could be completed within 10 min. Most questionnaires (76.5%) took less than 2 min to complete. Four measures (the MAAS, PAAS, MPAS and PPAS) took longer than 2 min to administer because items covered multiple pages.

Ease of scoring

In terms of ease of scoring, most measures (n = 14, 82.4%) received an easy rating due to their use of the Likert-scale scoring system. Response options ranged from four to seven options (see Table 3). Only two prenatal measures (the MPAS and the PPAS) received a moderate rating due to a combination of Likert-scale and simple formula scoring. The MIRFS is a two-part-scale, administered postnatally, in which the first sub-scale is rated as a Likert-scale and the second as a semantic differential scale; thus, it was judged to be difficult to score. As indicated in Table 3, in the majority of measures (n = 11, 64.7%) higher scores indicated stronger bonding or attachment, but in four (25%) postnatal measures higher scores were indicative of greater difficulties in the parent-reported bond with their infant. One scale (the MORS-SF) consisted of two sub-scales, whereby higher scores indicated higher maternal perceived levels of warmth as well as invasiveness. The ease of scores for one measure (the MIRFS) could not be reported because the scale development authors did not specify this in their paper (Thorstensson et al., 2012).

Readability and comprehensiveness

Seventeen measures were assessed using the Flesch Reading Ease (FRE) test. As can be seen in Table 3, two measures (the K-PAFAS and the PPBS) received a poor rating. An excellent rating in terms of readability was given to one scale only (the MIAS). A fair rating was given to seven measures: FAB, PAI-21, MAAS-19, PAAS, MAI, MPAS, and PPAS, and seven measures, namely the MFAS, PFAS, MABISC, PBQ-25, MIBS-8, MORS-SF, and MIRFS, received a good rating. Of those measures, the MFAS and PFAS were the only antenatal measures.

Availability and conditions of use

The majority of scales (n = 14, 82.4%) were easily accessible on the internet and free of charge; thus, receiving an excellent rating. Three scales, namely the FAB, the PFAS and MIAS, were difficult to obtain but free of charge; hence, they received a good rating.

Discussion

In this review we systematically examined the literature to identify, describe and evaluate the psychometric and clinimetric properties of self-report questionnaires for measuring the mother's or father's perception of their bond, attachment or relationship with their child. Seventeen original measures and their 13 modified versions, described in 65 articles from seven countries, were included and their methodological quality was carefully evaluated. Of these, a few measures were antenatal and postnatal measures for mothers (i.e., MAAS, MFAS, MPAS) or fathers (PAAS, PFAS, PPAS) only. The findings indicate that the evidence base for the robustness of self-report questionnaires measuring the parent-infant-relationship or bond is rather limited; consequently, we can only advise that these measures are used with some caution.

Considerations in relation to the COSMIN guidelines

The current 2018 COSMIN criteria appear to be the most stringent and complex to apply due to the multi-step process whereby firstly the quality of the measure development studies and the content validity studies were evaluated, secondly the methodological quality (risk of bias) of all studies was rated, thirdly the psychometric measurement properties were assessed and finally the quality of the evidence was graded. In other reviews of measures, reviewers either did not choose to apply the COSMIN criteria and opted to use other guidelines for rating each psychometric property (e.g., Lotzin et al., 2015; Perrelli et al., 2014), or they used an older COSMIN version (Terwee et al., 2007, in Wittkowski, Garrett, Calam, & Weisberg, 2017; Mokkink et al., 2010, in Bentley, Hartley, & Bucci, 2019, or De Vet, Terwee, Mokkink, & Knol, 2011, in Jewell et al., 2019). Despite using the older COSMIN criteria, reviewers such as Jewell et al. (2019) have highlighted the arbitrary nature of cut-off scores which determine if a measurement property is ‘adequate’ or ‘inadequate’ because in some cases the statistical values indicative of a negative rating were very close to values suggesting a positive one. Furthermore, Jewell et al. (2019) critiqued the use of the ‘worst case counts’ rule because a single flaw in the study would result in only a ‘fair’ or even a negative rating which means that the adequacy and sufficiency of measurement properties and the methodological quality of any evidence are not necessarily a true reflection and most likely an underestimation. This criticism also fits with our observations when applying the COSMIN 2018 guidance. In the application of the latest COSMIN guidance, we also became aware of how much practice and reporting standards have changed over the course of the last few decades; for example, the oldest measure our review identified was published in 1977 (e.g., the FAB). It was frustrating to note that authors reported some relevant information but did so without methodological consistency or rigor. For example, authors did not always report model fit statistics for confirmatory factor analyses so that they can be rated appropriately. Moreover, often authors only reported on correlation coefficients instead of reporting intraclass correlation coefficient (ICC) or kappa scores for (test-retest) reliability. This information is vital because the accurate assessment of a scale's structural validity depends on it. When following the COSMIN 2018 criteria, we evaluated any modified versions of measures separately even if the total item count differed by only one item. However, in our findings we observed that psychometric testing was conducted less rigorously on refined or revised versions of measures compared to the originally developed measures. Thus, the risk of bias ratings, which show the methodological quality of each measure, might have been downgraded in line with the strict rules of COSMIN. This downgrading could be considered unfair given that some development work (e.g., pilot assessment or cognitive interviewing) might have been undertaken with the original scale. Additionally, several of the included studies (e.g., Bienfait et al., 2011; Brockington et al., 2006, Brockington et al., 2001) tested for diagnostic accuracy (i.e., sensitivity and specificity of the measure in detecting bonding difficulties) and discriminant (or divergent) validity but these properties do not fall within the COSMIN taxonomy and consequently were not rated by us although we believe these to be important aspects of psychometric testing. Furthermore, only a few studies assessed aspects of interpretability (e.g., subgroup analyses, minimal important change, floor and ceiling effects, etc.) and thus, we could not report on these properties in our review.

Considerations relating to content validity

A measure's relevance, comprehensiveness and comprehensibility play a major deciding factor in why a measure may be chosen for clinical or research purposes; consequently, content validity may arguably be the most important psychometric property (Mokkink et al., 2018; Terwee et al., 2018). Based on COSMIN criteria, the methodological quality for the content validity of the original development measures identified in this review was ‘doubtful’ for the MABISC, MPAS and PPAS and ‘inadequate’ for the remaining 14 original measures. However, only 15 of the 65 included studies evaluated content validity at all, with the MFAS and PBQ having received the most attention, which was also noted in Tichelman et al.'s review (2019). Despite some studies evaluating ‘relevance’ and/or ‘comprehensibility’ among participants and/or professionals, none of the studies evaluated a measure's ‘comprehensiveness', highlighting the need for further research. Following psychometric evaluation of all 17 measures, only three measures (FAB, PAI-21, and MABISC) were rated as ‘sufficient’ for overall content validity with the remaining measures receiving ‘inconsistent’ ratings. Nonetheless, the quality of the evidence regarding the FAB and the PAI-21 was ‘very low’ and for the MABISC ‘low’ which indicates some uncertainty regarding the trustworthiness of the overall ratings. Despite the increased trend towards paying more attention to evaluating content validity in terms of relevance, comprehensiveness and comprehensibility, particularly since 2010, there was a high percentage of studies being rated as ‘inadequate’ or ‘inconsistent’ with a ‘very low’ quality evidence. To mitigate against the very strict COSMIN criteria, we applied a more flexible approach by including studies that described other relevant aspects of content validity, such as appropriateness, suitability, acceptability, understandability, coherence, ambiguity and clarity of the items or overall measure. We consider these to be important aspects of content validity and they are potentially worthy of consideration as an expansion of content validity in the COSMIN guidelines. Although study authors described their evaluation of content validity, it often remained unclear what they had actually explored and how they had conducted the evaluation, because these studies were not developed or conducted in accordance with the high reporting standards of the COSMIN criteria. Hence, applying the COSMIN 2018 criteria resulted in most studies being rated as ‘doubtful’, ‘inadequate’, ‘inconsistent’ and/or ‘indeterminate’. Furthermore, rating content validity according to the COSMIN criteria is a complex and multi-step process whereby the overall rating depends on the ratings of the development study, content validity study (or studies) and reviewers' ratings. On occasions when there were no content validity studies and the development studies received an ‘indeterminate’ rating, the overall ratings of the study were determined according to the reviewers' ratings which lead to increased subjectivity and to a higher likelihood of giving positive or ‘sufficient’ ratings. When content validation studies had been conducted, the measures tended to receive a lower overall score due to a lower quality of the content validity study. Thus, to minimise the bias and ambiguity, we encourage the reader to refer to individual ratings of the development study, content validity study (or studies) and reviewers' ratings which would give a more accurate overview of the measure's content validity.

Considerations relating to structural validity

The risk of bias for most studies assessing antenatal and postnatal measures was rated to be ‘adequate’ or ‘very good’. However, only two measures, namely the PBQ-22 and the MIBS-J-7 which were adapted versions of original measures, showed ‘sufficient’ evidence for structural validity and at least ‘moderate’ quality of evidence. The fact that many widely used measures had ‘insufficient’ evidence for structural validity is problematic and this issue needs to be explored further in future studies. In addition, most papers identified did not report the CFA estimation method used (e.g., Maximum Likelihood or Weighted Least Squares Mean and Variance), the appropriateness of which depends on several factors (see Rhemtulla, Brosseau-Liard, & Savalei, 2012 for recommendations). We suggest that future studies report this information to increase transparency and facilitate quality ratings. Unlike Chiarotto, Ostelo, Boers, and Terwee (2018), who followed the COSMIN guidance strictly by only reporting the content validity and structural validity of the studies, we did report on other psychometric properties of the identified measures.

Considerations relating to construct validity

Construct validity comprises hypothesis testing, cross-cultural validity/measurement invariance and criterion validity; however, none of the included studies evaluated criterion validity. The risk of bias for hypothesis testing for antenatal and postnatal measures was mostly ‘very good’ with all studies consistently showing ‘high’ quality of evidence. Nevertheless, despite these promising results, only three antenatal measures (MFAS-23, PFAS and K-PAFAS) and two postnatal measures (MABISC and MPAS) offered the best evidence of any hypothesis testing undertaken. It is also important to note that construct validity was not assessed against any ‘gold standard’ self-report measure, since none has yet been identified. Furthermore, none of the studies included in this review used a ‘gold standard’ observer-rated measure of parent-child attachment to assess their measure's construct validity. This is clearly an area of further investigation. Measurement invariance was rarely assessed in the identified measures. Based on ‘adequate’ methodological quality (i.e., risk of bias ratings), the MPAS and the paternal equivalent, the PPAS, were assigned ‘sufficient’ ratings with ‘moderate’ quality of evidence. Due to little evidence of construct validity, more research needs to be undertaken.

Considerations relating to reliability

Except for three antenatal versions (FAB, MAAS-13 and PAAS-13) and three postnatal versions (MABISC, PBQ-16 and MIAS), the quality of evidence for the internal consistency of ten antenatal versions and 13 postnatal versions was considered to be ‘high’. However, most studies did not provide enough information about the internal consistency of the antenatal and postnatal measures assessed and only the PBQ showed ‘sufficient’ evidence for internal consistency. This is because internal consistency could only be rated if the studies demonstrated at least low evidence of ‘sufficient’ structural validity (as per the 2018 COSMIN criteria); if a measure does not demonstrate good structural validity, there is no confidence that those subscales exist. This explains why most of the measures received ‘indeterminate’ ratings. In addition, all of the studies reported internal consistency as Cronbach's Alpha values. Although this approach is the most popular and widely applied statistic of internal consistency (Dunn et al., 2014), it has been criticised for having several flaws. For example, it is considered an inappropriate statistic to estimate a scale's reliability (Peters, 2014) and homogeneity of unidimensionality (Green, Lissitz, & Mulaik, 1977; Schmitt, 1996; Sijtsma, 2009). Thus, researchers evaluating the internal consistency of measures are encouraged to use alternatives, such as McDonald's omega in the future (Dunn et al., 2013; Peters, 2014). The reliability of measures could be assessed for two antenatal measures only because information for the remaining antenatal measures was not available. Whilst the reliability for the PAI-21 was ‘sufficient’ with ‘moderate’ quality of evidence, the reliability for the K-PAFAS was ‘insufficient’ with ‘low’ quality of evidence. Thus, based on the available evidence, the PAI-21 appears to be the most robust antenatal measures to be used with pregnant women. In relation to the 17 versions of postnatal measures, seven versions (i.e., MAI-26, MPAS, PBQ-25, PBQ-14, MIBS-J-10, MORS-SF and PPAS) evaluated the methodological quality (i.e., risk of bias ratings) of reliability in ten studies which varied between ‘adequate’, ‘doubtful’ and ‘inadequate’ ratings. However, conclusive summaries regarding the methodological quality cannot be provided due to these measures presenting with mixed evidence, depending on the country, language and sample size of each conducted study. Of all postnatal measures, only two measures were considered to have ‘sufficient’ evidence for their reliability with ‘moderate’ quality of evidence. Hence, the MPAS and the PBQ-25 appear to be the most reliable postnatal measures. However, it is important to note that none of the identified studies reported measurement error and responsiveness as part of their assessment of a scale's additional reliability. Clearly, more work is needed to fully establish the reliability of these scales.

Considerations relating to clinimetric properties

The assessment of administrative properties of a measure in addition to its psychometric properties has been recommended (e.g., Thornicroft & Slade, 2000; Wittkowski et al., 2017). As it is assumed that their modified and often simplified versions would achieve similar ratings, the clinimetric properties were assessed for the original 17 measures only. Except for three measures (e.g., MPAS, PPAS and MIRFS), the scoring of all measures was rated to be ‘easy’. In addition, none of the measures had an excessive number of items. Hence their completion time was rated as ‘excellent’. With item numbers ranging from seven (e.g., the MIBS-7) to 26 (e.g., the MAI), all measures could be completed under five minutes, which suggests that they are acceptable and feasible measures, suitable for the use in routine outcome assessments. In terms of readability and comprehension, all measures (except for the K-PAFAS) obtained ratings in the ‘fair’ to ‘excellent’ range. Of the antenatal measures, the MFAS obtained the best (i.e., ‘good’) score in readability, whereas six of the nine postnatal measures were rated to be ‘good’ or ‘excellent’ and hence easy to understand (e.g., MABISC, PBQ-25, MIBS-8, MORS-SF, MIRFS and MIAS). Hereby it is noteworthy that the item exploring whether the women or their partners felt that the woman's body was ‘ugly’ (reverse scored) in the MFAS (and the paternal equivalent, the PFAS) is occasionally removed from the scale because it does not refer to maternal feelings (Müller & Ferketich, 1993; Van den Bergh and Simons, 2009). With the Flesch readability scores ranging from 55.6 (K-PAFAS) to 96.0 (MIAS), all of the measures appeared to be acceptable.

Strengths and limitations of the review

A clear strength of this review is its comprehensiveness evidenced by the fact that eight databases were searched and more than 12,000 records were screened, in all languages and publication years. Any measure and study inclusion criteria were determined in advance and registered. In addition, data extraction and evaluation processes were verified by an independent rater and these showed good to excellent inter-rater agreement and reliability. Furthermore, by reporting information pertaining to the clinimetric properties of the identified measures, we assist clinicians and researchers in their assessment of how ‘user-friendly’ a measure is. Furthermore, the strict application of the latest COSMIN criteria (2018) ensured that a rigorous assessment was undertaken of the validity and reliability evidence of the identified measures. Compared to previous reviews in this area (e.g., Perrelli et al., 2014), the use of the most recent and stringent COSMIN criteria adds substantial credibility to this detailed assessment of measures. Finally, although the methodological quality of studies varied, we chose to report those variations rather than exclude those studies. In terms of limitations, it should be acknowledged that the search was restricted to peer-reviewed studies only, which introduces a publication bias (Rothstein, 2014). However, although unpublished measures with good psychometric properties may exist, without being published appropriately their impact in the field may be minimal. For this reason, Lotzin et al. (2015) searched all available literature in their review of observational tools for measuring parent-infant-interaction, but then decided to exclud tools that were only used in one or no peer-reviewed journal articles. Although we included different language versions of identified measures in our evidence synthesis (e.g., K-PAFAS, MIRFS and MIAS), it proved impossible for us to apply the COSMIN criteria to assess the psychometric properties of measures in identified studies if they were not written in English. However, for comprehensiveness we offer further information on those studies, including their sample size and the psychometric properties tested. In addition, when rating the content validity of studies, we chose to deviate from the stricter 2018 COSMIN guidelines on four occasions. Firstly, we opted to rate the psychometric properties of all studies evaluating content validity even if the study's methodological quality (i.e., risk of bias) was ‘inadequate’ as this resulted in a detailed and thorough overview of all included measures. Secondly, we adapted the criteria of rating the risk of bias of the measure development studies slightly in cases where the authors had presented ‘adequate’ evidence regarding the study's conduct; however, this could have resulted in assigning higher risk of bias ratings in some cases. Thirdly, we modified the criteria for structural validity since many studies in this review had undertaken EFA but the COSMIN criteria do not include guidance on how to rate the results of an EFA. Fourthly, we deviated from the guidelines when rating hypothesis testing and the results of this measurement property should be interpreted with caution. Another limitation that should be acknowledged is the fact that we excluded measures in which the parent-infant-relationship was examined but only alongside other and arguably less relevant aspects, such as exploring the woman's body or diet. For example, despite containing seven items that could be said to reflect the mother's attitude towards the developing fetus, the 60-item Maternal Adjustment and Maternal Attitude (MAMA) questionnaire (Kumar et al., 1984) was excluded because it contained many other items relating to the mother's perceptions of her body or items of somatic symptoms, the marital relationship, attitudes to sex and attitudes towards pregnancy. The postnatal version of the MAMA was also excluded, although it was judged to contain slightly more relevant items (n = 9). Only 13 of the 26-item What Being The Parent of a Baby is Like (WBTPBL, Pridham & Chang, 1985) scale related directly to the parent-infant-relationship with the other items asking about the new parent's adaptation to parenthood, relationships with family members and the stress of being a new parent. Finally, although the Maternal Infant Responsiveness Instrument (MIRI; Amankwaa, Younger, Best, & Pickler, 2002) partly met our inclusion criteria, the scale mostly examines parental perceptions of baby responsiveness and was therefore excluded.

Implications for research and practice

The fact that we identified a lack of evidence for robust psychometric properties across a wide variety of antenatal and postnatal parent-report measures is problematic because any conclusions based on these measures will have inherent limitations. Although some of these measures may have been extensively used (e.g., the PBQ) or their use (e.g., the PBQ and the MORS-SF) may have been recommended (Royal College of Psychiatrists, 2018), it is advisable that clinicians and researchers alike scrutinise each measure in order to determine if it fits their purpose. For example, all of these measures were validated using predominantly non-clinical populations (with the exception of the PBQ-25 and the MABISC) and this means that clinicians and researchers need to consider a measure's relevance when applied to their intended population or purpose. Besides, in line with the review's aims, we only included studies specifically evaluating psychometric properties but we are aware that studies may exist that report on a measure's use with clinical populations (e.g., see Wittkowski et al., 2010) and possibly on correlations with observer-rated measures as well. The reader is advised that we did not search for those studies or indeed included them in this review. Given the recent proliferation of measures being adapted for use in other countries and in languages other than English, we believe that there is a need for appropriate and more stringent testing for cross-cultural validity. For example, studies with different cultural or ethnic groups should conduct factor analyses for multiple groups (e.g., in English and in Dutch) and complete measurements of invariance or differential item functioning (DIF) to provide information on whether the measures are comparable when used in differing cultural contexts. This could be one of the future directions when testing psychometric properties of the measures. We also believe that future studies conducting content validity evaluation should describe more explicitly how they evaluated content validity and what aspects they did evaluate and to consult and follow COSMIN criteria when developing the method of a new measure or assessing the method of an already existing measure. This may include conducting a qualitative study (i.e., a focus group or interviews), using appropriate data collection and analysis methods and ideally exploring the relevance, comprehensiveness and comprehensibility of the measure among a sufficient sample of participants and professionals, which would lead to a higher quality and more credible evidence of the measure's content validity.

Conclusion

This is the first systematic review to provide a synthesis of robust validity and reliability evidence for available self-report measures of the parent-infant-relationship. A total of 17 measures and 13 modified versions were identified and evaluated, of which the majority lacked adequate methodological quality despite being widely used and with some being recommended measures. Only the Postpartum Bonding Questionnaire (PBQ), and some of its modified versions, were found to demonstrate sufficient evidence for structural validity, internal consistency and reliability with high quality of evidence. The PBQ was also the most frequently adapted tool which is indicative of its perceived relevance and popularity in this field. However, due to the inadequate methodological quality and insufficient psychometric measurement evaluation of most measures, in addition to the lack of comprehensive psychometric evaluation of many measures, firm conclusions regarding the most valid and reliable parent-infant-relationship measure(s) cannot be drawn. The current review is important and timely given the increasing importance of routine self-report outcome monitoring within a range of perinatal services and within research studies (e.g., NHS England, NHS Improvement, and National Collaborating Centre for Mental Health, 2018). Despite the wealth of antenatal and postnatal measures, the psychometric properties of these tools remain poor and understudied. It is advisable that future researchers developing new or modified measures follow the current COSMIN guidelines and that research into evaluation of psychometric properties would continue in order to bring measures to the industry standard and facilitate the selection of the most robust antenatal and postnatal measures by researchers and clinicians.

Funding

This review was undertaken as part of the THRIVE project funded by the National Institute for Health Research (NIHR) Public Health Research Programme (PHR Project: 11/3002/01), led by MH. As part of this grant, CG, AM and SV were employed as researchers and AW was their supervisor as well as Manchester lead for the THRIVE project. The funder had no direct involvement in the conduct of this review. The views expressed are those of the authors and not necessarily those of the NHS, the NIHR or the Department of Health and Social Care. MH was supported by the MRC/ Quinquennial funding of the Relationships and Health Improvement Programme, which is part of the Social and Public Health Sciences Unit, based at , MC_UU_12017-11 and SPHSU11.

Contributions

AW devised the idea for this review and oversaw the whole review process. CG and AW wrote the initial protocol which was later updated with SV and AM's assistance. SV, CG and AM conducted the literature searches. AM and SV rated the included studies, under AW's supervision. Any disagreements were resolved through consultation with the fifth reviewer, MH. All authors contributed to drafts and approved the final manuscript.

Declaration of Competing Interest

All authors declare that they have no conflict of interest.
Measure(s) evaluatedAuthorsLanguageSample sizePsychometric properties tested
MFASLauriola et al. (2010)Italian254 womenInternal reliability; construct validity
MFASAndrek et al. (2016)Hungarian114 womenInternal reliability
PAIJurgens et al. (2009)French112 womenInternal reliability; construct validity
PAILauriola et al. (2010)Italian254 womenInternal reliability; construct validity
PAIPavše et al. (2019)Slovenian619 womenInternal reliability; structural validity
MAASCamarneiro & Justo (2010)Portuguese212 couplesInternal reliability; structural validity
MAASLauriola et al. (2010)Italian254 womenInternal reliability; construct validity
MAASDenis et al. (2013)French117 womenInternal reliability; construct validity; concurrent validity; divergent validity
MAASNie & Fan (2017)Chinese545 womenInternal reliability; convergent validity
PAASCamarneiro & Justo (2010)Portuguese212 couplesInternal reliability; structural validity
MAIKavlak & Sirin, 2009Turkish165 womenInternal reliability; content validity
PBQYalcin et al. (2014)Turkish189 womenInternal reliability
MIBSFigueiredo et al. (2005)Portuguese456 parentsInternal reliability; test-retest reliability
MIBSYalcin et al. (2014)Turkish189 womenInternal reliability
MORS-SFDanis et al. (2012)Hungarian1164 parentsInternal reliability
RELEVANCE
COMPREHENSIVENESS
COMPREHENSIBILITY
TOTAL CONTENT VALIDITY
Outcome measure development studyContent validity studiesReviewers ratingsRatings of resultsQuality of evidenceOutcome measure development studyContent validity studiesReviewers ratingsRatings of resultsQuality of evidenceOutcome measure development studyContent validity studiesReviewers ratingsRatings of resultsQuality of evidence
FAB[?]None[+; +][+]VERY LOW[?]None[+; +][+]VERY LOW[?]None[+; +][+]VERY LOW[+]
MFAS-24[?][±; ±][+; +][±]VERY LOW[?][−; −][+; +][±]VERY LOW[?][?;?][+; ±][±]VERY LOW[±]
MFAS-20NA[±][+; +][±]VERY LOWNA[−][+; +][±]VERY LOWNA[?][+; ±][±]VERY LOW[±]
PAI[?][+][+; +][+]VERY LOW[?][+][+; +][+]VERY LOW[?][?][+; +][+]VERY LOW[+]
MAAS[+][±][+; +][±]VERY LOW[+][+][+; +][+]VERY LOW[+][?][+; +][+]VERY LOW[±]
PPBS[+]None[+; +][+]VERY LOW[−]None[+; −][±]VERY LOW[−]None[+; +][±]VERY LOW[±]
PFAS[?]None[+; +][+]VERY LOW[+]None[+; +][+]VERY LOW[?]None[+; ±][±]VERY LOW[±]
PAAS[+][−][+; +][±]VERY LOW[+][−][+; +][±]VERY LOW[?][?][+; +][+]VERY LOW[±]
K-PAFAS[+][±][+; ±][±]VERY LOW[+][−][+; +][±]VERY LOW[?][−][+; ±][±]VERY LOW[±]
MAI[?][±; ±][+; +][±]VERY LOW[+][+; +][+; +][+]VERY LOW[?][?;?][+; +][+]VERY LOW[±]
MIAS[?]None[+; ±][±]VERY LOW[+]None[+; +][+]VERY LOW[?]None[+; ±][±]VERY LOW[±]
MABISC[+]None[+; +][+]LOW[+]None[+; +][+]LOW[+]None[+; +][+]LOW[+]
MPAS[+][±][+; +][±]LOW[+][−][+; +][±]LOW[?][−][+; +][±]LOW[±]
PBQ-25[±][+; +][+; +][+]VERY LOW[−][+; −][+; +][±]VERY LOW[?][+; +][+; +][+]VERY LOW[±]
PBQ-19NA[−][+; +][±]VERY LOWNA[−][+; +][±]VERY LOWNA[?][+; +][+]VERY LOW[±]
MIBS[±]None[+; +][+]VERY LOW[−]None[+; ±][±]VERY LOW[?]None[+; +][+]VERY LOW[±]
MIRFS[?][±; ±][+; ±][±]VERY LOW[+][−; −][+; ±][±]VERY LOW[?][?; +][+; ±][±]VERY LOW[±]
MORS-SF[?]None[+; +][+]VERY LOW[−]None[+; +][±]VERY LOW[?]None[+; +][+]VERY LOW[±]
PPAS[+][±][+; +][±]LOW[+][−][+; +][±]LOW[?][−][+; +][±]LOW[±]

Notes. [+] = sufficient. [?] = indeterminate. [±] = inconsistent. [−] = insufficient. NA—not applicable. None = no content validity studies conducted. The multiple ratings per box indicate either multiple content validity studies or multiple reviewers' ratings.

#Type1Type2MeasureLanguage of scalePaper (by first author and year)nStructural validitynInternal consistencynReliabilitynHypothesis testingnMeasurement invariance
1PreMatFABEnglishLeifer, 1977
2PreMatMFAS-24EnglishCranley, 198171Very good [?]71Inadequate [−]
3PreMatTamilLingeswaran, 2012230Inadequate [?]Very good [?]
4PreMatHungarianAndrek, 2016114Inadequate [?]114Very good [?]Inadequate [?]
5PreMatGermanDoster, 2018324Adequate [−]324Very good [?]324Very good [−]
6PreMatMFAS-23EnglishMüller & Ferketich, 1993681Adequate [−]681Very good [?]
7PreMatEnglishMüller, 1993336Adequate [?]336Very good [?]336Very good [+]
8PreMatMFAS-20ItalianBusonera, 2016a482Very good [−]482Very good [?]482Very good [−]
9PreMatMFAS-17SwedishSeimyr, 2009298Adequate [?]298Very good [?]Very good [?]
10PreMatSwedishSjögren, 200476Inadequate [+]76Very good [?]
11PreMatPAI-21EnglishMüller, 19931336Adequate [−]336Very good [?]336Adequate [−]
12PreMatEnglishMüller, 1996196Very good [?]196Very good [−]
13PreMatEnglishGau, 2003349Very good [−]349Very good [?]
14PreMatSwedishPallant, 2014775Very good [−]775Very good [?]
15PreMatItalianBarone, 2014130Adequate [+]130Very good [?]Very good [?]
16PreMatItalianBusonera, 2017a535Very good [−]535Very good [?]535Very good [−]
17PreMatItalianDella Vedova, 2008214Adequate [−]214Very good [?]Very good [?]
18PreMatPersianSamani, 2016322Very good [−]322Very good [?]322Adequate [+]Very good [?]
19PreMatMAAS-19EnglishCondon, 1993150Adequate [−]150Inadequate [?]
20PreMatEnglishCondon, 1997Very good [?]
21PreMatItalianBusonera, 2016b482Very good [−]482Very good [?]482Very good [−]
22PreMatSpanishNavarro-Aresti, 2016525Very good [?]
23PreMatTurkishGolbasi, 2015190Adequate [?]190Very good [?]
24PreMatHungarianMako Deak, 2014237Very good [?]237Very good [+]
25PreMatDutchVan Bussel, 2010b403Very good [?]Inadequate [?]
26PreMatMAAS-13GermanGöbel, 2019263Adequate [−]263Very good [?]
27PreMatMAAS-12SpanishNavarro-Aresti, 2016525Very good [−]525Very good [?]
28PreMatPPBSDutchCuijlits, 20161050Very good [+]529Very good [+]1050Very good [?]
29PrePatPFASEnglishWeaver, 1983100Inadequate [?]100Inadequate [?]
30PrePatSwedishSeimyr, 2009298Adequate [?]298Very good [?]298Very good [?]
31PrePatPAAS-19EnglishCondon, 1993112Doubtful [−]112Inadequate [?]
32PrePatEnglishCondon, 2013204Very good [−]
33PrePatItalianDella Vedova, 201765Doubtful [−]65Very good [?]65Very good [−]
34PrePatPAAS-13GermanGöbel, 2019128Adequate [−]128Very good [?]
35PrePatK-PAFASKoreanNoh, 2017200Very good [−]200Very good [?]200Doubtful [+]200Very good [+]
36PostMatMAI-26EnglishMüller, 19942196Very good [?]148Doubtful [+]196Adequate [−]
37PostMatEnglishMüller, 1996196Very good [?]196Very good [−]
38PostMatKoreanShin, 2007196Adequate [+]196Very good [?]196Very good [−]
39PostMatChineseChen, 2013181Adequate [+]181Very good [?]181Very good [−]
40PostMatMIASHindiBhakoo, 1994100Inadequate [?]100Adequate [?]Very good [?]
41PostMatMABISCEnglishHackney, 1996Inadequate [?]
42PostMatNorwegianHoivik, 201376Doubtful [+]76Doubtful [?]76Very good [+]
43PostMatMPASEnglishCondon, 1998212Adequate [−]212Very good [?]212Adequate [+]212Very good [+]
44PostMatEnglishFeldstein, 200459Very good [?]59Very good [−]
45PostMatItalianScopesi, 2004210Very good
[−]210Very good [?]Inadequate [−]Very good [?]
46PostMatDutchVan Bussel, 2010a263Very good [+]
47PostMatSpanishRiera-Martin, 2016571Very good [−]571Very good [?]Very good [?]571Adequate [+]
48PostMatPBQ-25EnglishBrockington, 2001104Adequate [+]104Inadequate [+]
49PostMatEnglishBrockington, 2006
50PostMatGermanReck, 2006862Adequate [−]862Very good [?]Very good [?]
51PostMatEnglishWittkowski, 200796Very good [?]96Very good [−]
52PostMatEnglishWittkowski, 2010132Adequate [−]
53PostMatChineseSiu, 2010
54PostMatDutchVan Bussel, 2010a263Very good [?]263Very good [−]
55PostMatJapaneseKaneko, 20141786Doubtful [?]1786Very good [?]Very good [?]
56PostMatJapaneseSuetsugu, 2015199Adequate [?]199Very good [?]199Doubtful [+]199Very good [+]
57PostMatSpanishGarcia-Esteve, 2016840Very good [−]
58PostMatPortuguese (Brazil)Baldisserotto, 2018
59PostMatItalianBusonera, 2017b123Adequate [+]123Very good [?]123Very good [−]
60PostMatJapaneseOhashi, 2016364Very good [−]364Very good [?]364Adequate [+]Very good [?]
61PostMatPBQ-16GermanReck, 2006862Inadequate [?]Very good [?]
62PostMatPBQ-22EnglishWittkowski, 2010132Adequate [+]132Very good [+]Very good [?]
63PostMatPBQ-16-J3JapaneseKaneko, 20141786Doubtful [?]1786Very good [?]Very good [?]
64PostMatPBQ-14JapaneseSuetsugu, 2015199Adequate [?]199Very good [?]199Doubtful [+]199Very good [−]
65PostMatPBQ-194TamilVengadavaradan, 2019250Adequate [?]Very good [?]
66PostMatS-PBQEnglishKinsey, 20143005Adequate [?]3005Very good [?]Very good [?]
67PostMatMIBS-8EnglishTaylor, 2005162Adequate [?]162Very good [?]
68PostMatEnglishWittkowski, 200796Very good [?]96Very good [−]
69PostMatDutchVan Bussel, 2010a263Very good [?]263Very good [−]
70PostMatFrenchBienfait, 201178Very good [+]
71PostMatMIBS-J-10JapaneseYoshida, 2012554Very good [−]554Very good [?]554Doubtful [−]Very good [?]
72PostMatMIBS-J-7JapaneseOhara, 2016375Very good [+]751Very good [−]Very good [?]
73PostMatMIRFSSwedishThorstensson, Nissen 2012b395Adequate [?]395Very good [?]
74PostMatSwedishThorstensson, Hertfelt Wahn, 2012a
75PostMatMORS-SF5EnglishOates, 2018100Very good [?]Very good [?]
76PostMatHungarianOates, 2018331Very good [?]Very good [?]
77PostMatEnglishOates, 2019100Inadequate [−]
78PostMatHungarianOates, 2019134Inadequate [−]36Doubtful [+]
79PostMatEnglish and HungarianOates, 2019243Adequate [−]243Very good [?]
80PostPatPPASEnglishCondon, 2008461Adequate [−]461Very good [?]Very good [?]
81PostPatEnglishFeldstein, 200438Very good [?]Very good [?]
82PostPatEnglishCondon, 2013204Inadequate [−]204Very good [−]
83PostPatSpanishRiera-Martin, 2016376Very good [−]376Very good [?]Very good [?]376Adequate [+]

Notes. [+] = sufficient. [−] = insufficient. [?] = indeterminate. [±] = inconsistent. Structural validity ratings were based on the best fitting model presented in the paper (this was not necessarily the factor structure proposed by the original authors). As per the COSMIN criteria, internal consistency could only be rated as sufficient if there was at least low evidence of sufficient structural validity (otherwise an indeterminate rating was assigned). Some papers are listed more than once as the authors tested psychometric properties of more than one measure.

EFA was conducted on 29-item PAI, resulting in a 21-item scale.

EFA was conducted on a 31-item MAI, resulting in a 26-item scale.

PBQ-J-16 contains different items from the PBQ-16.

EFA conducted on full PBQ, resulting in 19-item scale.

MORS-SF EFA conducted on 44-items, resulting in 14-item scale.

  80 in total

Review 1.  Understanding and improving the validity of self-report of parenting.

Authors:  Sarah K Morsbach; Ronald J Prinz
Journal:  Clin Child Fam Psychol Rev       Date:  2006-03

2.  Maternal Attachment Inventory: psychometric evaluation of the Korean version.

Authors:  Hyunjeong Shin; Young Hee Kim
Journal:  J Adv Nurs       Date:  2007-06-21       Impact factor: 3.187

3.  Validity and reliability of the Turkish version of the Maternal Antenatal Attachment Scale.

Authors:  Zehra Golbasi; Tuba Ucar; Nilufer Tugut
Journal:  Jpn J Nurs Sci       Date:  2014-06-24       Impact factor: 1.418

4.  Factor analysis of the Maternal Fetal Attachment Scale.

Authors:  M E Müller; S Ferketich
Journal:  Nurs Res       Date:  1993 May-Jun       Impact factor: 2.381

5.  Development of a self-administered questionnaire to measure maternal adjustment and maternal attitudes during pregnancy and after delivery.

Authors:  R Kumar; K M Robson; A M Smith
Journal:  J Psychosom Res       Date:  1984       Impact factor: 3.006

6.  An exploration of paternal-fetal attachment behavior.

Authors:  R H Weaver; M S Cranley
Journal:  Nurs Res       Date:  1983 Mar-Apr       Impact factor: 2.381

7.  Systematic Review of Self-Report Measures of General Mental Health and Wellbeing in Adolescent Mental Health.

Authors:  Natalie Bentley; Samantha Hartley; Sandra Bucci
Journal:  Clin Child Fam Psychol Rev       Date:  2019-06

8.  Maternal Antenatal Attachment Scale (MAAS): adaptation to Spanish and proposal for a brief version of 12 items.

Authors:  Lucía Navarro-Aresti; Ioseba Iraurgi; Leire Iriarte; Ana Martínez-Pampliega
Journal:  Arch Womens Ment Health       Date:  2015-02-24       Impact factor: 3.633

9.  Mokken scale analysis of mental health and well-being questionnaire item responses: a non-parametric IRT method in empirical research for applied health researchers.

Authors:  Jan Stochl; Peter B Jones; Tim J Croudace
Journal:  BMC Med Res Methodol       Date:  2012-06-11       Impact factor: 4.615

10.  A systemic review of maternal wellbeing and its relationship with maternal fetal attachment and early postpartum bonding.

Authors:  Josephine McNamara; Michelle L Townsend; Jane S Herbert
Journal:  PLoS One       Date:  2019-07-25       Impact factor: 3.240

View more
  8 in total

1.  Mental Health Surveillance Among Children - United States, 2013-2019.

Authors:  Rebecca H Bitsko; Angelika H Claussen; Jesse Lichstein; Lindsey I Black; Sherry Everett Jones; Melissa L Danielson; Jennifer M Hoenig; Shane P Davis Jack; Debra J Brody; Shiromani Gyawali; Matthew J Maenner; Margaret Warner; Kristin M Holland; Ruth Perou; Alex E Crosby; Stephen J Blumberg; Shelli Avenevoli; Jennifer W Kaminski; Reem M Ghandour
Journal:  MMWR Suppl       Date:  2022-02-25

2.  Does Parental Reflective Functioning Mediate the Associations between the Maternal Antenatal and Postnatal Bond with the Child in a Community Sample?

Authors:  Chiara Pazzagli; Livia Buratta; Giulia Cenci; Elena Coletti; Maria Luisa Giuliani; Claudia Mazzeschi
Journal:  Int J Environ Res Public Health       Date:  2022-06-07       Impact factor: 4.614

3.  Measuring PROMIS® Social Relationships in Early Childhood.

Authors:  Courtney K Blackwell; Jin-Shei Lai; Michael Kallen; Katherine B Bevans; Matthew M Davis; Lauren S Wakschlag; David Cella
Journal:  J Pediatr Psychol       Date:  2022-05-13

4.  A Feasibility Randomized Controlled Trial of a Parenting Intervention Offered to Women With Severe Mental Health Problems and Delivered in a Mother and Baby Unit Setting: The IMAgINE Study Outcomes.

Authors:  Anja Wittkowski; Richard Emsley; Penny E Bee; Elizabeth Camacho; Rachel Calam; Kathryn M Abel; Paula Duxbury; Paula Gomez; Kim Cartwright; Holly E Reid
Journal:  Front Psychiatry       Date:  2022-05-16       Impact factor: 5.435

5.  Pre-natal Attachment and Parent-To-Infant Attachment: A Systematic Review.

Authors:  Tommaso Trombetta; Maura Giordano; Fabrizio Santoniccolo; Laura Vismara; Anna Maria Della Vedova; Luca Rollè
Journal:  Front Psychol       Date:  2021-03-17

6.  Identification of Bonding Difficulties in the Peripartum Period Using the Mother-to-Infant Bonding Scale-Japanese Version and Its Tentative Cutoff Points.

Authors:  Koyo Hashijiri; Yuichiro Watanabe; Naoki Fukui; Takaharu Motegi; Maki Ogawa; Jun Egawa; Takayuki Enomoto; Toshiyuki Someya
Journal:  Neuropsychiatr Dis Treat       Date:  2021-11-20       Impact factor: 2.570

7.  The Measurement Properties and Acceptability of a New Parent-Infant Bonding Tool ('Me and My Baby') for Use in United Kingdom Universal Healthcare Settings: A Psychometric, Cross-Sectional Study.

Authors:  Tracey Bywater; Abigail Dunn; Charlotte Endacott; Karen Smith; Paul A Tiffin; Matthew Price; Sarah Blower
Journal:  Front Psychol       Date:  2022-02-14

8.  Factors associated with re-admission in the year after acute postpartum psychiatric treatment.

Authors:  Billie Lever Taylor; Angela Sweeney; Laura C Potts; Kylee Trevillion; Louise M Howard
Journal:  Arch Womens Ment Health       Date:  2022-08-25       Impact factor: 4.405

  8 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.