Literature DB >> 32101577

Repeatability of flatfish reflex impairment assessments based on video recordings.

Sven Sebastian Uhlmann1, Noëlle Yochum2, Bart Ampe1.   

Abstract

Using measures of reflex impairment and injury to quantify an aquatic organism's vitality have gained popularity as survival predictors of discarded non-target fisheries catch. To evaluate the robustness of this method with respect to 'rater' subjectivity, we tested inter- and intra-rater repeatability and the role of 'expectation bias'. From video clips, multiple raters determined impairment levels of four reflexes of beam-trawled common sole (Solea solea) intended for discard. Raters had a range of technical experience, including veterinary students, practicing veterinarians, and fisheries scientists. Expectation bias was evaluated by first assessing a rater's assumption about the effect of air exposure on vitality, then comparing their reflex ratings of the same fish, once when the true air exposure duration was indicated and once when the time was exaggerated (by either 15 or 30 min). Inter-rater repeatability was assessed by having multiple raters evaluate those clips with true air exposure information; and intra- and inter-rater repeatability was determined by having individual raters evaluate a series of duplicated clips, all with true air exposure. Results indicate that inter- and intra-rater repeatability were high (intra-class correlation coefficients of 74% for both), and were not significantly affected by background type nor expectation bias related to assumed impact from prolonged air exposure. This suggests that reflex impairment as a metric for predicting fish survival is robust to involving multiple raters with diverse backgrounds. Bias is potentially more likely to be introduced through subjective reflexes than raters, given that consistency in scoring differed for some reflexes based on rater experience type. This study highlights the need to provide ample training for raters, and that no prior experience is needed to become a reliable rater. Moreover, before implementing reflexes in a vitality study, it is important to evaluate whether the determination of presence/absence is subjective.

Entities:  

Mesh:

Year:  2020        PMID: 32101577      PMCID: PMC7043772          DOI: 10.1371/journal.pone.0229456

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


Introduction

To address concerns over discard practices and animal welfare in commercial fisheries, methods are needed to reliably profile fish condition onboard vessels to describe fishing impacts on both individuals and populations [1-4]. To allow assessments in adverse and remote conditions, responsiveness of a fish to induced stimuli expressed as a binary presence-absence score may be measured as part of the Reflex Action Mortality Predictor Method (RAMP) [5]. This method assumes that physical stress from mechanical interaction with fishing gear during a capture event may trigger internal physiological responses and alters metabolism (i.e., from anaerobic exercise and hypoxia) which then may precipitate in impaired responsiveness, because the pathway of nerve impulses from the receptors to the muscles through the brainstem and/or the spinal cord might be affected [6-7]. To collect such fish condition or vitality information, some methods involve observers (or so-called ‘raters') scoring or rating the extent of external injury and/or responses to stimuli (e.g., reflex impairment), typically using a binary or ordinal scale [7-8]. Depending on the scope of the study, these assessments may be done by different and possibly independent raters with different technical experience and levels of training [4]. When multiple raters are involved in the assessment, rater subjectivity has the potential to affect the accuracy and precision of the vitality assessment [9]. A rater's ‘score' may be influenced by (i) knowing the treatment an animal has received (e.g., expectation bias from a non-blinded experimental design; [10-15]; (ii) their level of experience (scientific background or familiarity with vitality scoring), which may lead to a subjective interpretation of otherwise objective scoring criteria [13, 16–17]; and/or (iii) an assessment criterion or metric that is difficult to discern or that lends itself to subjectivity [14, 17–18]. When using vitality scoring, it is important to evaluate whether the assessment criteria are unbiased. This has been highlighted in studies on fish discard mortality [19] as well as in the medical field [17]. Although (semi-) quantitative condition indicators such as reflex impairment ratings have been collected for fishes (e.g., [4–5, 20–21]) and invertebrates (e.g., [22-25]) around the world (largely as a predictor for fisheries discard survival), their sensitivity towards rater biases has not been thoroughly tested. While [9] showed that three raters were able to similarly score reflexes of European plaice (Pleuronectes platessa), the role of expectation bias, together with experience intra-rater reliability (also termed ‘repeatability'; [11, 26–27] has not been tested for any fish species before. To address uncertainty in inter- and intra-rater repeatability, we evaluated reflex scores of common sole (Solea solea) in the Belgian flatfish beam-trawl fishery. We conducted workshops whereby raters with different backgrounds and levels of scientific experience (i.e., third year veterinary medicine students, practicing veterinarians/food safety inspectors, and fisheries scientists with varying amounts of experience in fish reflex testing) were asked to score fish for reflex impairment (four different reflexes) through video clips using a tagged analogue continuous scale (tVAS; [9]). Through this study, we aimed to (i) quantify the effect of expectation bias in reflex impairment ratings by testing whether falsified air exposure information misled raters (i.e., evaluating whether informing raters that the animal had prolonged air exposure would bias them toward either higher or lower scores) compared to the duplicated clip (false vs. true air exposure information); (ii) evaluate intra-rater repeatability and the influence of experience, by observing if the same rater was able to reproduce a given score from a duplicated clip (only true air exposure information); and (iii) evaluate inter-rater repeatability among rates for the same clip (including clips with only true air exposure information).

Materials and methods

Ethics statement

The handling of animals, including those that were filmed in this research, was approved by the animal ethics commission of the Flanders Research Institute for Agriculture, Fisheries and Food (ILVO, Ref. no. 2016/264). Experiments were performed on-board a commercial Belgian beam-trawler, the R/V Simon Stevin and at a research laboratory in Ostend, Belgium. All research-related handling was designed to minimize any stress cumulative to being captured by beam trawls and sorted on deck. For example, any air exposure during the reflex tests was kept to a minimum and was well within exposure times during conventional, commercial sorting practices. If fish were held captive, housing mimicked natural conditions. The filming did not require any extra handling procedures. Animal ethics approval was granted by the Flanders Research Institute's for Agriculture, Fisheries and Food (ILVO) Animal Care and Ethics Committee (EC2016/264).

Equipment and treatments

To evaluate inter- and intra-rater repeatability, we conducted a series of seven workshops where separately either third-year, veterinary medicine students, practicing veterinarians/food safety inspectors, or fisheries scientists scored four reflexes of common sole from short (<30 s) video sequences (or ‘clips'; Fig 1).
Fig 1

Lecture theatre view to illustrate video projection.

Third-year veterinary medicine students from the University of Ghent independently scored a tail grab reflex response of a common sole (Solea solea) from a video clip projected onto a lecture theatre screen during a workshop session in April 2016. Note: The picture was blurred in parts to guarantee that any person in the audience cannot be identified to comply with the PLOS open-access (CC-BY) license.

Lecture theatre view to illustrate video projection.

Third-year veterinary medicine students from the University of Ghent independently scored a tail grab reflex response of a common sole (Solea solea) from a video clip projected onto a lecture theatre screen during a workshop session in April 2016. Note: The picture was blurred in parts to guarantee that any person in the audience cannot be identified to comply with the PLOS open-access (CC-BY) license. The first scoring session, conducted in April 2015, was attended by third-year, veterinary medicine students from the University of Ghent (N = 120 female and N = 35 male students; N = 2 male non-student experts). The second session, in May 2015, occurred during a lunchtime seminar with fisheries research scientists, with diverse expertise (N = 7 female and N = 11 male). The third session, in December 2015, was during an international workshop of fisheries research scientists with specialist expertise in discard survival studies (N = 5 female, N = 8 male). The fourth session, in January 2016, included seagoing fisheries observers and fisheries technicians (N = 6 female, N = 13 male). The final sessions (5–7) were shown: (5) in April 2016 to third-year veterinary medicine students from the University of Ghent (N = 140 female, N = 39 male students; N = 2 genderless; N = 1 male expert); (6) in December 2016 to practicing veterinarians/food inspectors (N = 12 female, N = 20 male; N = 1 male expert); and (7) in December 2017 to fisheries research scientists with diverse expertise (N = 6 female; N = 7 male; N = 1 genderless). The reflexes that were selected (body flex, righting, head, and tail grab; Table 1) were those used to assess common sole [4] and that were clearly visible in a video clip. Each workshop began with a 15-min lecture with visual aids detailing the utility of the reflex scoring method as an animal welfare indicator and predictor for discard survival (Supporting Video 1, https://doi.org/10.14284/399). Participants were also informed about relevant factors in the catch-and-discarding process that potentially stress fish and result in weaker reflex responses, namely prolonged periods of air exposure on deck, among others. Following the lecture, participants were trained on example video clips showing, for each of the four reflexes, an ‘absent', ‘weak', ‘moderate', or ‘strong' reflex response (Table 1; Supporting Video 1, https://doi.org/10.14284/399). The key criterion associated with each response category was read out loud and given to each participant in the form of a pictogram handout (Fig 2A). These training clips were unique and not used again within the video clips that were scored by the participants.
Table 1

List of scoring criteria for categorical reflex responses (i.e., absent, weak, moderate, and strong) of common sole (Solea solea) in the order tested within 5 s of observation after stimulus (based on [4, 9]).

ReflexStimulusAbsentWeakModerateStrong
Body flexThe fish is held outside the water on the palms of two hands (touching each other) with its bellyfacing up and its head and tail unsupported.No active movement, the body rests limp on the hand.Tail is moving slightly, but not beyond the plain of the hand.Tail is flexing beyond the plain of the hand. Body may move–spasticflexion; or slowly slipping off the hand.The fish is actively trying to move head and tail towards each other; orquickly slipping off the hand.
RightingThe fish is held underwater at the surface on the palms of two hands (touching each other) with its bellyfacing up and then is slowly released.Fish drifts and sinks passively to the bottom of the container.Fish appears stunned, but rights itself very slowly.Fish appears stunned, but starts to turn after a delay. The rotation can be swift.Fish actively and quickly turns underwater.
HeadThe fish’s head is held between thumb and index finger, with either belly or dorsal side facing up.No movement. The body dangles motionless.The fish may move its tail slightly.The fish may exhibit a cramp-like flexion, but noclear curling, nor repeated bending.Fish immediately and repeatedly curls around fingers.
Tail grabThe fish’s tail is held between thumb and index finger.Fish does not struggle free; itremains motionless upon release.Fish does not struggle free; no swimmingmovement, but swims away upon release.Fish does not struggle free, but moves its body as if it attempts to swim away.The fish actively struggles free and swims away.

Intensity of a response increases from absent to strong. The speed of a response for weak and moderate categories may be delayed; for strong it should be immediate.

Fig 2

Example of a pictogram handout (A) and scoresheet (B) to train and score reflexes from video clips. The scoresheet details how to score the body flex reflex (which was also termed ‘bellybend’) response of common sole (Solea solea) on a continuous tagged-analogue visual scale from a short video clip.

Example of a pictogram handout (A) and scoresheet (B) to train and score reflexes from video clips. The scoresheet details how to score the body flex reflex (which was also termed ‘bellybend’) response of common sole (Solea solea) on a continuous tagged-analogue visual scale from a short video clip. Intensity of a response increases from absent to strong. The speed of a response for weak and moderate categories may be delayed; for strong it should be immediate. Video clips were used to test whether the same rater (‘intra-rater repeatability'), or different raters (‘inter-rater repeatability') were able to repeat the same score of the same fish, and whether a rater's score could have been influenced by knowing how much a given fish was exposed to air prior to its reflex test (intra-rater repeatability with testing an expectation bias effect). A total of 36 video clips of the four reflex responses of common sole were picked out of a reference library of video clips, representing a range of impairment states filmed inside a laboratory (N = 5 fish), or on-board a commercial beam trawler (N = 5 fish; Table 2). Overall, clips included reflexes across the categorized spectrum of responses (ranging from absent to strong). Three experienced expert raters who were involved in the development of the reflex scoring methodology scored the 12 unmodified, original clips that were used to create the scoring video used during sessions 2–4 (Fig 3). It showed that the selected clips did not bias reflexes towards either weak or strong responses.
Table 2

Schematic representation of the treatment (inter-rater vs intra-rater repeatability, with or without expectation bias; see shading) assigned to 36 video clips (each <30 s in length) of a given fish’s reflex response per scoring session.

Reflex
FishBody flexRightingHeadTail grab
Session 1
1FFTF
2FTFT
3TTTF
4T, FT, FF, FT, T
5F, FT, FF, FT, T
6T, TF, FT, FT, T
Sessions 2–4
1FFTF
2FTTF
3TTTF
4T, FT, FT, FT, F
5T, FT, FT, FT, F
6T, FT, FT, FT, F
Sessions 5–7
3T, FT, FT, FT, F
4T, FT, FT, F
5T, TT, TT, T
7TTTT
8T, T
9T, TT, TT, TT, T
10T, F
Inter-Rater Repeatability
Inter-and Intra-Rater Repeatability
Inter-and Intra-Rater Repeatability, and Expectation Bias

The notations ‘T,F’ or ‘T,T’ are indicating whether air exposure information was true or falsified on the duplicated video pair, and if not duplicated air exposure was marked as either false or true (‘F’ or ‘T’, respectively). For sessions 1–4, if falsified, 15 minutes were added to the true value, for sessions 5–7, 30 minutes were added.

Fig 3

Frequency distribution of average ‘silver standard’ reflex scores.

Scores by three expert raters who developed the reflex scoring methodology and were experienced raters were considered as the ‘silver standard’. These three raters scored all original clips (N = 12) that went into the making of the scoring video for sessions 2–4.

Frequency distribution of average ‘silver standard’ reflex scores.

Scores by three expert raters who developed the reflex scoring methodology and were experienced raters were considered as the ‘silver standard’. These three raters scored all original clips (N = 12) that went into the making of the scoring video for sessions 2–4. The notations ‘T,F’ or ‘T,T’ are indicating whether air exposure information was true or falsified on the duplicated video pair, and if not duplicated air exposure was marked as either false or true (‘F’ or ‘T’, respectively). For sessions 1–4, if falsified, 15 minutes were added to the true value, for sessions 5–7, 30 minutes were added. To address intra-rater repeatability and bias related to expectation of an effect from prolonged air exposure, between 12 and 16 of the 36 clips were duplicated (Table 2). All duplicates were mirrored or at least slightly modified in Adobe Photoshop by increasing their brightness levels to mislead the viewers in assuming all clips were unique. Onto each clip, the true or falsified number of minutes the fish spent on deck exposed to air prior to the reflex test (‘air exposure') was labelled, together with a date and time stamp. To falsify air exposure times, an arbitrary 15 min or 30 min were added to the true value. These air exposure periods were chosen to i) represent conventional commercial catch sorting times, and ii) to increase the expectation bias potential for each rater. The greater value (i.e., longer air exposure time) was chosen to increase the likelihood of the rater being influenced by this information if the rater had a preconceived idea about the effect of air exposure on vitality.

Data and analyses

To analyse whether a reflex response was biased toward lower or higher tVAS score when an elevated air exposure was falsely indicated on the clip, each rater's scores of duplicate clips were compared with a linear-mixed model (LMMs; lme4 package in R; [28] with as fixed effects: between i) a rater's expectation of an effect of prolonged air exposure on reflex responsiveness, ii) the experience level of each rater, iii) and the reflex type and all possible interactions. Random effects were included for the ID of a given clip, and a rater's ID. The Tukey method was used to compare each reflex for corresponding pairs of duplicate clips shown with either false or true air exposure. These were evaluated by each rater's expectation group (expecting positive, and/or no or negative impact from air exposure) and experience level (no, <100, or ≥100 vertebrate animals previously assessed for reflex responsiveness). A significance level of 0.05 was applied. Inter- and intra-rater repeatability were estimated based on inter- and intra-rater reliability coefficients [29-30] which were implemented using the irr-package [31]. To estimate inter-rater repeatability, all clips with true air exposure information were included in the dataset (scoring sessions 1–7; Table 2), also arbitrarily stratifying all raters by their reflex rating experience, and calculating across all included clips or specifically per reflex type. To estimate both inter- and intra-rater repeatability on the same dataset, only duplicated clips with true air exposure information were included in the dataset (scoring sessions 5–7; Table 2). The intra-class correlation coefficient (ICC) is based on the ratio of the variability among rater's reflex scores over the sum of this variance plus error, thus ranging between 0 and 100. A higher value of ICC reflects a higher agreement among the raters for a given clip or per reflex type. The ICC measure of association was estimated using the psych package in R [32]. In this study, we report the ICC for a single random rater [29].

Results

In total, 436 participants scored video clips during the seven dedicated scoring workshops and produced 13,676 scores, because not all participants were equally able to score each of the 36 clips (Table 3). The majority (N = 401) were unexperienced raters (i.e., never previously scored animals for reflex responsiveness). Of these, the majority were students, but some were scientists, technicians, observers, and practicing veterinarians/food safety inspectors. Fourteen and 21 participants had scored some (<100 animals) or ≥100 fish (i.e.,‘experienced') reflexes before, respectively. One of the raters with some experience had observed behavioural responses among seabirds and seals, but not fish.
Table 3

Number, gender, experience, and expectation of workshop participants per scoring session (1–7), stratified by previous experience in scoring reflex responsiveness of live animals (‘none’: no animals scored; ‘ some’: <100 animals scored; and ‘experienced’: ≥100 animals scored).

SessionNo. participantsMaleFemaleNAExperienceExpectation
PositiveNegativeNA
115735120None546239
20Experienced200
21887None3120
30Experienced120
31321None210
31Some310
33Experienced150
41940None400
53Some710
43Experienced520
5182391402None784162
10Experienced100
6332012None13163
10Experienced100
714461None731
20Some110
10Experienced100
Total4361342963184147105

‘Expectation’ was classified based on the rater’s response to a question on the scoresheet asking whether s/he expected air exposure to impact reflex impairment, either positively or negatively (i.e., the rater believes that prolonged air exposure would exacerbate or reduce reflex impairment, respectively). NA, not all participants revealed their gender or gave a score for the expectation question.

‘Expectation’ was classified based on the rater’s response to a question on the scoresheet asking whether s/he expected air exposure to impact reflex impairment, either positively or negatively (i.e., the rater believes that prolonged air exposure would exacerbate or reduce reflex impairment, respectively). NA, not all participants revealed their gender or gave a score for the expectation question.

Expectation bias

The dataset that included scores of duplicated clips with either true or falsified air exposure information comprised 3,525 scores which were assigned in workshop sessions 2–7 to duplicate clips by those participants who indicated an expectation about the effect of air exposure on reflex responsiveness (Table 2). Scores by participants from session 1 were not included, because not all duplicated clips were paired by true/false air exposure information (Table 2). Based on histogram data indicating a clear distinction at greater and less than 30, a positive expectation (i.e., air exposure would exacerbate reflex impairment) was set at <30, and a negative expectation (i.e., air exposure would reduce reflex impairment) was ≥30 (Fig 4). Of these scores, 70% were scored with a positive expectation by the participant (Fig 4). An expectation of the effect of air exposure on reflex responsiveness did not bias the scoring of reflex clips. The null hypothesis (i.e., no difference in scores due to air exposure information) was not rejected for raters who expected air exposure to positively affect reflex impairment (N = 128; Table 3). Overall, these raters were not more likely to assign a lower score to a duplicated clip that showed falsified air exposure (extra 15 or 30 min) compared to the original, which was stamped with the true air exposure time (Table 4). Generally, where available, the median scores followed what the three expert raters had assigned to each clip (‘silver standard’ score), although for some clips scored by raters with some or experienced raters, their median values were off the mark compared to the silver standard (Fig 5).
Fig 4

Frequency distribution of hypothetical tVAS scores provided by unique raters from scoring sessions 2–7.

N of unique raters is indicated above each bar as it was marked by a rater on their scoresheet in response to a question whether a reflex response would weaken or strengthen when the animal was knowingly exposed to air for a prolonged period (15–30 min).

Table 4

Tukey comparisons of the least-square mean (lsmean) ± SE reflex score of a given reflex type which was scored by a rater with a certain experience and a positive (1) or negative (0) expectation.

ExperienceReflexExpectationAir exposurelsmeanSEl.CLu.CLGroup
NoneBody flex1TRUE50.815.420.681a
1FALSE58.715.428.588.9b
0TRUE55.215.425.085.5ab
0FALSE60.415.430.290.7b
Head1TRUE39.813.413.666a
1FALSE37.013.410.863.2a
0TRUE40.813.414.667.1a
0FALSE38.213.412.064.4a
Righting1TRUE46.313.420.172.5a
1FALSE49.413.423.275.7a
0TRUE48.313.422.174.5a
0FALSE51.313.425.077.5a
Tail grab1TRUE46.013.419.872.2a
1FALSE45.113.419.071.3a
0TRUE47.613.421.473.8a
0FALSE49.013.422.875.2a
SomeBody flex1TRUE60.816.229.092.5ab
1FALSE69.216.237.4100.9b
0TRUE42.917.78.277.6a
0FALSE53.917.719.288.6ab
Head1TRUE42.313.715.469.1a
1FALSE34.113.77.361a
0TRUE39.414.810.368.4a
0FALSE31.914.82.860.9a
Righting1TRUE62.513.735.789.4a
1FALSE51.813.924.679.1a
0TRUE61.914.832.991a
0FALSE50.215.220.380a
Tail grab1TRUE53.013.726.279.9a
1FALSE54.513.727.681.3a
0TRUE51.014.821.980a
0FALSE52.114.823.181.2a
ExperiencedBody flex1TRUE66.116.234.497.9a
1FALSE62.616.230.894.4a
0TRUE74.016.541.7106.3a
0FALSE75.816.543.5108.2a
Head1TRUE41.113.813.968.2a
1FALSE39.413.812.366.5a
0TRUE40.713.813.767.7a
0FALSE38.713.811.765.7a
Righting1TRUE53.113.926.080.3a
1FALSE51.214.023.878.7a
0TRUE57.013.830.084a
0FALSE50.914.023.578.3a
Tail grab1TRUE57.513.830.384.6a
1FALSE51.613.824.578.8a
0TRUE56.413.829.483.4a
0FALSE48.513.821.575.5a

Clips were duplicated within a scoring video and imprinted onto the screened clip with either false (an added 15 or 30 min to the true value) or true air exposure information. A rater’s expectation (scored on a scale of 0 to 100) of the effect of prolonged, onboard air exposure on a fishes’ reflex responsiveness was categorized as to whether it would result in either a weaker (<30; positive expectation; 1) reflex response or no effect (≥30, no or negative/wrong expectation; 0). Our hypothesis was that clips imprinted with false air exposure information would receive a lower score than their duplicate shown with the true value, as the fish would have been weakened from additional air exposure (positive expectation). Groups with the same letter were not significantly different at p = 0.05.

Fig 5

Plot of mean score per video clip doublet (with either ‘true’ or ‘false’ air exposure information) of a given reflex across all workshop participants (a), and then stratified by experience in scoring reflexes of fish: none (b); some (c); and experienced (d). Doublet ID includes a scoring video clip ID (2 or 3) and a running ID number for each doublet, with each clip of a doublet abbreviated by ‘a’ or ‘b’. Treatments include: ‘Intra-rater reliability with expectation bias’ (IOR-exp.), which refers to duplicated clips of the same fish and reflex with either true or false air exposure information; or ‘Intra-rater reliability’ (IOR), which refers to duplicated clips of the same fish and reflex and always true air exposure information. Where available, dots indicate the ‘silver standard’ scores which were averaged across three experienced, expert raters who scored 12 unmodified, original clips.

Frequency distribution of hypothetical tVAS scores provided by unique raters from scoring sessions 2–7.

N of unique raters is indicated above each bar as it was marked by a rater on their scoresheet in response to a question whether a reflex response would weaken or strengthen when the animal was knowingly exposed to air for a prolonged period (15–30 min). Plot of mean score per video clip doublet (with either ‘true’ or ‘false’ air exposure information) of a given reflex across all workshop participants (a), and then stratified by experience in scoring reflexes of fish: none (b); some (c); and experienced (d). Doublet ID includes a scoring video clip ID (2 or 3) and a running ID number for each doublet, with each clip of a doublet abbreviated by ‘a’ or ‘b’. Treatments include: ‘Intra-rater reliability with expectation bias’ (IOR-exp.), which refers to duplicated clips of the same fish and reflex with either true or false air exposure information; or ‘Intra-rater reliability’ (IOR), which refers to duplicated clips of the same fish and reflex and always true air exposure information. Where available, dots indicate the ‘silver standard’ scores which were averaged across three experienced, expert raters who scored 12 unmodified, original clips. Clips were duplicated within a scoring video and imprinted onto the screened clip with either false (an added 15 or 30 min to the true value) or true air exposure information. A rater’s expectation (scored on a scale of 0 to 100) of the effect of prolonged, onboard air exposure on a fishes’ reflex responsiveness was categorized as to whether it would result in either a weaker (<30; positive expectation; 1) reflex response or no effect (≥30, no or negative/wrong expectation; 0). Our hypothesis was that clips imprinted with false air exposure information would receive a lower score than their duplicate shown with the true value, as the fish would have been weakened from additional air exposure (positive expectation). Groups with the same letter were not significantly different at p = 0.05. Nevertheless, for some duplicated clips that were scored by raters with some prior reflex scoring experience, lower scores were assigned to clips as postulated by our null hypothesis (i.e., duplicates with IDs 3_10a & b; Fig 5C). But this difference was not significant (Table 4). In advance of scoring, some raters expected that the reflex would not be affected by air exposure or would even become stronger (N = 85; decreased impairment = negative expectation, Table 3). This aligned with clips of the body flex reflex, for which raters with no reflex assessment experience consistently scored higher for clips with falsified compared with true air exposure (Table 4; Fig 5B). This contrasted our null hypothesis.

Intra- and inter- rater repeatability

When quantifying inter-rater repeatability (dataset included scores of all clips with true air exposure information, some clips were duplicated; 6,664 observations), raters with different experience levels in scoring reflex impairment were able to reproduce the same score for a given clip when scored independently in different scoring sessions with an intra-class correlation coefficients of 76% (68% 84%, lower and upper confidence interval, CI). Participants who had no prior scoring experience produced a lower intra-class correlation coefficient (ICC = 76%, 68% 84% CI) compared with participants who had scored at least some fish throughout their career (ICC = 79%, 71% 87% CI). However, the latter sample size was rather small (N = 29) compared to 396 raters with no experience who were considered in this analysis. A similar pattern resulted when comparing ICC values per reflex type. For example, for the tail grab reflex, raters with at least some experience scored more consistently than raters with no experience (ICC = 86%, 76% 94% upper and lower CI vs ICC = 81%, 69% 92% upper and lower CI, respectively). Similarly, but with a less prominent difference, for the head reflex, raters with at least some experience had an ICC value of 79% (63% 93% lower and upper CI) compared to 78% (61% 93% CI) for raters with no experience. Including only seagoing observers and those experts who developed this methodology, increased the ICC (ICC = 83%, 67% 95% upper and lower CI). However, in contrast, for the righting reflex, the pattern was reversed: raters with no experience scored more consistently than raters with experience (ICC = 71%, 50% 92% upper and lower CI versus ICC = 49%, 27% 83% upper and lower CI, respectively). The least repeatable were the scores for the body flex, regardless of experience (ICC = 15%, 0.07% 46% upper and lower CI versus ICC = <1%, -11% 37% upper and lower CI, respectively for raters with none or at least some experience). The dataset which included duplicated clips with only true air exposure information, to calculate ICC of both intra- and inter-rater reliability comprised 3,664 observations. Across all reflexes, relatively high ICC values of 74% were achieved for inter- and intra-rater reliability, for both. For individual reflexes, highest ICC of both intra- and inter-rater reliability (for both the values were almost the same and differed from beyond the third decimal) were achieved for head (92%), tail grab (78%), righting (45%), and by far the lowest ICC was achieved for body flex (<1%).

Discussion

There is a global effort to determine the limitations and strengths of methods that profile fish condition related to fishing impacts and survival prediction [21, 33]. This study examined whether vitality information is reliable based on the involvement of multiple raters and/or on their experience level, and whether scoring repeatability can be influenced by knowing the treatment a fish has received. Results suggest that vitality assessments using reflex responsiveness are robust. In regard to expectation bias, there was no evidence that the exaggerated air exposure information influenced intra-rater repeatability. Regardless of their experience, raters were not misled to assign lower reflex scores to fish which they believed were exposed to air for a prolonged period of time, even when they expected air exposure to positively impact reflex impairment. This does not mean that other variables cannot invite expectation bias; however, it does suggest that perhaps when focusing on a specific metric over a short time frame, the rater does not subconsciously bias their assessment, especially when the scoring criteria (here between absent and present) are unambiguous. These results are promising if reflexes are to be used in settings with multiple, independent raters and/or with raters who do not have a strong background in reflex assessment. We do however acknowledge that this study was done through video clip analysis rather than having participants handle fish. There is the possibility that tactile experience in fish handling or reflex scoring could result in inter-rater variability among scores. However, [9] found no inter-rater differences when multiple participants scored the same live fish for reflex impairment. While rater experience in conducting reflex assessments did not bias the scoring outcomes (similar to results from [9]), results suggest that bias is potentially more likely to be introduced through subjective reflexes than raters; especially when reflexes were to be presented as <30 sec long video clips. This includes reflexes that are difficult to assess or that elicit responses that are difficult to discern between presence and absence; or reflexes such as body flex and righting which during evaluation were rapidly tested in succession of each other. This supports the need for researchers to scrutinize the selected reflexes that will be used for a vitality study in advance of data collection based on a screening for consistent and unambiguous candidate reflexes among unstressed fish [4], to be deliberate about scoring metrics (i.e., binary vs. continuous scoring), and ideally, establish a concrete physiological link between a stressor and reflex impairment to validate underlying hypotheses that such links exist [20-21]. Ideally, during data collection, each rater should be blinded and unaware about any prior treatments a study animal may have received, likewise an analyst should be unaware of who did the scoring [15]. Attention has to be paid when editing video clips accordingly. Experience may contribute to a subjective interpretation of scoring criteria, when pre-gained routines and self-made ‘rules' may bias an assessment. This study also supports the use of video-taped reflex assessments that can be reviewed at a later time. This has implications for allowing multiple assessments of the same video and to include reviewers who are unable to go to sea for each field trial. It also is beneficial for training purposes to minimize handling of fish. While there is evidence that untrained raters are capable of rating as or even more accurately as experienced raters, for future studies using reflex impairment as a vitality metric, we recommend having a substantial training programme for raters which includes protocols with clear and meaningful definitions, scoring of videos with pictogram-based handouts, repetitive training sessions and continued repeatability checks [34]. In addition, if video assessments are performed, it is helpful to have sheets describing the reflexes in front of the raters, constraining a fixed amount of time to observe each clip, and ensuring only one reflex is shown in a video clip at a time. There is also the potential to have a video shown on a touch screen where the rater could be more in control of viewing; however, time to review should be limited. Blinding and intra- and inter-rater reliability analyses are relevant concepts which should be considered for robust inference within experimental fisheries science, especially where many independent raters are involved. For example, when fish otolith are read for their age (e.g., [35]) or when using vitality indices to evaluate welfare and/or freshness of catches either on-board vessels or at fish auctions. Among domestic farm animals, such assessments are routinely done (e.g., [36]). 21 Jan 2020 PONE-D-19-34394 Repeatability of flatfish reflex impairment assessments based on video recordings PLOS ONE Dear Dr Uhlmann, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Editors comments: The reviewer, while enthusiastic, outlines a number of ways in which the manuscript could be improved. Please pay attention to these in particular, the following three points The intro needs a paragraph to contextualize what exactly is a reflex impairment score and how it relates to physiological status in the fish The discussion needs some work, particularly around comparisons to other work. I also suggest that a table outlining general recommendations for avoiding bias’s in an experimental design so that others can avoid the same pitfalls and a good discussion around this. We would appreciate receiving your revised manuscript by Mar 06 2020 11:59PM. When you are ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols Please include the following items when submitting your revised manuscript: A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). This letter should be uploaded as separate file and labeled 'Response to Reviewers'. A marked-up copy of your manuscript that highlights changes made to the original version. This file should be uploaded as separate file and labeled 'Revised Manuscript with Track Changes'. An unmarked version of your revised paper without tracked changes. This file should be uploaded as separate file and labeled 'Manuscript'. Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out. We look forward to receiving your revised manuscript. Kind regards, Judi Hewitt Academic Editor PLOS ONE Journal Requirements: 1. When submitting your revision, we need you to address these additional requirements. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at http://www.journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and http://www.journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf 2. We note that Figure 1 includes an image of a [patient / participant / in the study]. As per the PLOS ONE policy (http://journals.plos.org/plosone/s/submission-guidelines#loc-human-subjects-research) on papers that include identifying, or potentially identifying, information, the individual(s) or parent(s)/guardian(s) must be informed of the terms of the PLOS open-access (CC-BY) license and provide specific permission for publication of these details under the terms of this license. Please download the Consent Form for Publication in a PLOS Journal (http://journals.plos.org/plosone/s/file?id=8ce6/plos-consent-form-english.pdf). The signed consent form should not be submitted with the manuscript, but should be securely filed in the individual's case notes. Please amend the methods section and ethics statement of the manuscript to explicitly state that the patient/participant has provided consent for publication: “The individual in this manuscript has given written informed consent (as outlined in PLOS consent form) to publish these case details”. If you are unable to obtain consent from the subject of the photograph, you will need to remove the figure and any other textual identifying information or case descriptions for this individual. [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes ********** 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes ********** 3. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes ********** 4. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes ********** 5. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: The work presented by Uhlmann et al. was quite interesting and highly relevant given the increased usage of reflex scores in assessing fish stress status in numerous contexts and settings. Overall, it was quite an enjoyable read and thought the concept and data were very interesting. Although, I have some minor concerns that I would like to see addressed before acceptance of the manuscript. • Weird spacing in the intro. Eg lines 81-83, 62-63 • The intro overall conveys a very good and concise of the literature regarding biases and issues with using reflex impairment scores. However, I think some additional information is need to contextualize what exactly is a reflex impairment score and how it relates to physiological status in the fish. Readers unfamiliar with the topic might be confused as to why this metric is even used. I recommend just including a short paragraph outlining this. • Backing up my point, in the objectives, the authors are indicating using real vs fake air exposure info. If the reader did not understanding the underlying physiological processes associated with air exposure (i.e. hypoxia, metabolic acidosis, substrate depletion, etc.) then making the connection to a reflex score might be difficult. • The video/crowd sourcing data idea is really cool! I commend the authors on thinking outside the box on this one. • Table 7: may wish to include a reference indicating where this information was originally sourced from? In the methods it does indicate that it uses standardized metrics employed by fisheries science with this species. • Line 158, ok but how do the authors know that these reflexes are characterized as such? Was one person quantifying across all of the videos to indicate that this was a representative “strong” or “weak” fish. I would just like some more details pertaining how the representative videos were selected in the first place especially if observer bias is a problem. • Line 167, why were these durations used? It seems unlikely to have a fish air exposed in excess of 30 min. • Table 2: why was gilling, a standard part of the ramp score, no used in the reflex assessments? • Were the veterinarians in the study strictly aquatic vets or was it a mix of students studying general vertebrate anatomy/physiology. I ask as this as this in itself may pose an issue relating to experience working with fish vs non-fish vertebrates. • I realize there is a large volume of data but would it be worth visualizing some of it in the form of a graph. May help the readers get a better sense of the trends in the data as the results section is pretty text heavy which I totally understand is hard to otherwise with a data set like this one. I just think a couple visuals would go a long way in conveying the data. • Overall, I found the discussion quite short and fairly superficial. I recommend incorporating more comparisons to other work especially in the realm of observer bias in animal behaviour for which there are numerous works on. The authors have a lot of data here with some pretty neat trends but don’t really discuss their data to any great length. • It may be worth having either a table or a paragraph outlining general recommendations for avoiding bias’s in an experimental design so that others can avoid the same pitfalls. This is really what this paper is about so I recommend discussing this in some great detail here. • How would these results compare to say taking reflex scores in a field setting (i.e. without camera) where the observer may only have a single opportunity to view a fish’s behaviour? • Outside of sole, how applicable would these results be other species and settings? Presumably, the various body characteristics of the fish may also help enhance/impair scoring especially if the behaviour is relatively subtle. You may wish to highlight context/species specific effects in your discussion as something like a salmon or a shark may not have the same observer bias associated with it. ********** 6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files to be viewed.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org. Please note that Supporting Information files do not need this step. 28 Jan 2020 Journal requirement: naming of files As requested we have paid attention to the file and formatting style and naming requirements as outlined in the provided guides for authors documents. Accordingly, adjustments to the title page were made. Please note that the formatting of the affiliations are inconsistent in the example provided (i.e., there seem to be a space missing between superscript and affiliation address in some instances, check 1 Department, Institution, City, State, Country vs 2). Journal requirement: figure 1 – license to publish Although all persons pictured were facing with their backs to the camera and the purpose of this image is to illustrate the projection of the video clips onto a big screen and not to show the audience or individual participants, we have blurred the audience in photoshop to make sure that no one is identifiable and added this as a comment to its figure caption (L123-L124). Reviewer #1: Comment 1 – Weird spacing in the intro. We have corrected the weird spacing and apologise for it. It was an artefact from sharing a manuscript draft between North American and European Microsoft Office Word packages, which caused compatibility issues with the formatting, and eventually I created a new version by copying and pasting. 2 – Additional information needed in intro. As requested, we included a short paragraph at the beginning of the introduction outlining the concept of reflex scoring and why it is used (L47-L55). Note that the numbering of in-text citations has changed to accommodate for extra referencing of studies. 3 – Link between hypoxia and impairment. See comment 2 above. This paragraph also mentions how the exposure to air from hypoxia could contribute to impaired reflex responsiveness (L53-L55). 4 – Video/crowd sourcing idea. We thank the reviewer for this appreciation of our efforts. It all started off as a ‘pet project’, and soon we realized that we could make a very useful contribution to the study of reflex responsiveness of fish and experimental design in fisheries science in general. 5 – Table 7. We are not sure whether the reviewer refers to Table 1 here, because there is no Table 7? We have added another reference to the caption of Table 1 (L153). 6 – Selection of clips. We have added a sentence to explain how clips were selected and added a histogram of the distribution of averaged ‘silver standard’ scores (Fig 3). These were based on scores by three experts, who were experienced raters who were involved in developing the methodology and who scored 12 unmodified, original clips in a separate video. These 12 clips were then used to create the scoring video by introducing duplicates, and falsifying air exposure information. We also modified Figure 4 (now Figure 5), to indicate how the percentile ranges of pooled raters (by experience) compared to the ‘silver standard’ scores averaged across the three expert raters (L305-L307). 7 – Choice of fictive air exposure periods. These air exposure periods were chosen for two reasons: to represent commercial fishing practices (depending on catch volumes which can be large), it may take at least 15 min to 30 min until all air-exposed catch has been sorted by the crew. Secondly, the longer air exposure period was chosen to increase any expectation potential (L187-L189). 8 – Choice of reflexes. If the reviewer refers to what is called “head complex” or “operculum” in the literature and scores the functioning of the breathing apparatus, then this reflex was not selected, because it was considered that it may be difficult to discern for a rater whether the gills/operculum and/or mouth of a sole would be moving from a video clip. 9 – Veterinarians in the study. In the material and Methods section, we have specified that experience can refer to having assessed any vertebrate animal for reflex behaviours (L212-L213), and mentioned in the Results section how many of the raters with some experience have had experience with scoring behavioural responses among vertebrate animals other than fish (L235-L236). None of the practicing veterinarians had indicated any prior experience with scoring reflexes. 10 – Visualize some of the data. We suggest to include as an online supplement a video which we produced that includes a powerpoint lecture, training video and commentary. Other than that, we think that the structure and experimental design was sufficiently and succinctly summarized in the tables we provided. But we also included another histogram to illustrate the distribution of ‘silver standard’ scores (see comment 6). 11 – Superficial discussion. In places, we have revised the discussion to relate our results in greater depth to recent and relevant literature (L381-L383; L404-L406). 12 – Additional table or paragraph outlining general recommendations. We think that our concluding and several other paragraphs in the discussion list our recommendations, and we do not see the necessity to add another table to keep this article more concise. Submitted filename: reply_to_editor.docx Click here for additional data file. 7 Feb 2020 Repeatability of flatfish reflex impairment assessments based on video recordings PONE-D-19-34394R1 Dear Dr. Uhlmann, We are pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it complies with all outstanding technical requirements. Within one week, you will receive an e-mail containing information on the amendments required prior to publication. When all required modifications have been addressed, you will receive a formal acceptance letter and your manuscript will proceed to our production department and be scheduled for publication. Shortly after the formal acceptance letter is sent, an invoice for payment will follow. To ensure an efficient production and billing process, please log into Editorial Manager at https://www.editorialmanager.com/pone/, click the "Update My Information" link at the top of the page, and update your user information. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org. If your institution or institutions have a press office, please notify them about your upcoming paper to enable them to help maximize its impact. If they will be preparing press materials for this manuscript, you must inform our press team as soon as possible and no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. With kind regards, Judi Hewitt Academic Editor PLOS ONE Additional Editor Comments (optional): Reviewers' comments: 18 Feb 2020 PONE-D-19-34394R1 Repeatability of flatfish reflex impairment assessments based on video recordings Dear Dr. Uhlmann: I am pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department. If your institution or institutions have a press office, please notify them about your upcoming paper at this point, to enable them to help maximize its impact. If they will be preparing press materials for this manuscript, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org. For any other questions or concerns, please email plosone@plos.org. Thank you for submitting your work to PLOS ONE. With kind regards, PLOS ONE Editorial Office Staff on behalf of Dr. Judi Hewitt Academic Editor PLOS ONE
  10 in total

1.  Protocol for assessing brain function in fish and the effectiveness of methods used to stun and kill them.

Authors:  S C Kestin; J W van deVis; D H F Robb
Journal:  Vet Rec       Date:  2002-03-09       Impact factor: 2.695

2.  Considerations in the choice of interobserver reliability estimates.

Authors:  D P Hartmann
Journal:  J Appl Behav Anal       Date:  1977

3.  Artifact, bias, and complexity of assessment: the ABCs of reliability.

Authors:  A E Kazdin
Journal:  J Appl Behav Anal       Date:  1977

Review 4.  Intraclass correlations: uses in assessing rater reliability.

Authors:  P E Shrout; J L Fleiss
Journal:  Psychol Bull       Date:  1979-03       Impact factor: 17.737

5.  Statistical methodology for the concurrent assessment of interrater and intrarater reliability: using goniometric measurements as an example.

Authors:  M Eliasziw; S L Young; M G Woodbury; K Fryday-Field
Journal:  Phys Ther       Date:  1994-08

6.  Inter-observer reliability testing of pig welfare outcome measures proposed for inclusion within farm assurance schemes.

Authors:  Siobhan Mullan; Sandra A Edwards; Andrew Butterworth; Helen R Whay; David C J Main
Journal:  Vet J       Date:  2011-03-04       Impact factor: 2.688

7.  Inter-rater reliability of the Full Outline of UnResponsiveness score and the Glasgow Coma Scale in critically ill patients: a prospective observational study.

Authors:  Michael Fischer; Stephan Rüegg; Adam Czaplinski; Monika Strohmeier; Angelika Lehmann; Franziska Tschan; Patrick R Hunziker; Stephan C Marsch
Journal:  Crit Care       Date:  2010-04-14       Impact factor: 9.097

8.  Validity of the FOUR score coma scale in the medical intensive care unit.

Authors:  Vivek N Iyer; Jayawant N Mandrekar; Richard D Danielson; Alexander Y Zubkov; Jennifer L Elmer; Eelco F M Wijdicks
Journal:  Mayo Clin Proc       Date:  2009-08       Impact factor: 7.616

9.  Inter-rater reliability of categorical versus continuous scoring of fish vitality: Does it affect the utility of the reflex action mortality predictor (RAMP) approach?

Authors:  Pieter Meeremans; Noëlle Yochum; Marc Kochzius; Bart Ampe; Frank A M Tuyttens; Sven Sebastian Uhlmann
Journal:  PLoS One       Date:  2017-07-13       Impact factor: 3.240

10.  Evidence of Experimental Bias in the Life Sciences: Why We Need Blind Data Recording.

Authors:  Luke Holman; Megan L Head; Robert Lanfear; Michael D Jennions
Journal:  PLoS Biol       Date:  2015-07-08       Impact factor: 8.029

  10 in total
  1 in total

1.  Estimating Discard Mortality in Commercial Fisheries without Fish Dying: A 3R Challenge.

Authors:  Niels Madsen; Rasmus Ern; Aage Kristian Olsen Alstrup
Journal:  Animals (Basel)       Date:  2022-03-19       Impact factor: 2.752

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.