Literature DB >> 27047590

Addressing Common Student Technical Errors in Field Data Collection: An Analysis of a Citizen-Science Monitoring Project.

Joanna Philippoff1, Erin Baumgartner2.   

Abstract

The scientific value of citizen-science programs is limited when the data gathered are inconsistent, erroneous, or otherwise unusable. Long-term monitoring studies, such as Our Project In Hawai'i's Intertidal (OPIHI), have clear and consistent procedures and are thus a good model for evaluating the quality of participant data. The purpose of this study was to examine the kinds of errors made by student researchers during OPIHI data collection and factors that increase or decrease the likelihood of these errors. Twenty-four different types of errors were grouped into four broad error categories: missing data, sloppiness, methodological errors, and misidentification errors. "Sloppiness" was the most prevalent error type. Error rates decreased with field trip experience and student age. We suggest strategies to reduce data collection errors applicable to many types of citizen-science projects including emphasizing neat data collection, explicitly addressing and discussing the problems of falsifying data, emphasizing the importance of using standard scientific vocabulary, and giving participants multiple opportunities to practice to build their data collection techniques and skills.

Entities:  

Year:  2016        PMID: 27047590      PMCID: PMC4798815          DOI: 10.1128/jmbe.v17i1.999

Source DB:  PubMed          Journal:  J Microbiol Biol Educ        ISSN: 1935-7877


INTRODUCTION

Citizen science involves engaging cadres of volunteers in scientific research. An increasingly popular research approach, citizen science enables the collection of large-scale datasets across time and space and can play a valuable role in conducting ecological research because it allows the scientific community to address questions that it would otherwise be unable to for logistical or financial reasons (12). Citizen scientists are not professional scientists; they may be volunteers connected to a particular wildlife area or visitors to a website who become intrigued by the opportunity to help collect or examine data. Citizen-science projects provide opportunities for the public to engage in the process of science through authentic scientific experiences, contributing to a more scientifically literate society (10). Citizen science can be a valuable instructional model in K–12 and postsecondary classrooms. However, the educational opportunities afforded by citizen science projects need to be balanced with the importance of collecting scientifically useful data (8). The adoption of citizen-scientist data to examine scientific issues can be hindered by the perception that the data are not reliable (7). When questions arise, citizen-science data are underutilized, which is detrimental to both scientific and educational goals. However, there is mounting evidence supporting the use of citizen-science data from a variety of projects. For example, comparisons between datasets collected by volunteer citizen scientists and those of professional scientists have shown consistent rates of error and bias (3). Volunteers with training, even limited training, are able to collect reliable data when provided with unambiguous standardized protocols (9). Our Project In Hawai’i’s Intertidal (OPIHI) is a citizen science program in which middle and high school students survey rocky intertidal areas in Hawai’i (1, 2). To address reservations scientists may have had about using OPIHI student-generated data, we conducted a validity assessment that demonstrated students’ data quality was robust and similar to that of more experienced professional researchers (5). OPIHI student data resulted in the first description of community-level patterns at multiple intertidal sites across the Hawaiian islands (6). The purpose of this paper is not to determine the ability of students and other citizen scientists to conduct scientific research, which OPIHI and many other projects have already demonstrated, but rather to determine the types of data collection errors made by students involved in a monitoring project. We categorized and analyzed the frequency of occurrence of different types of errors committed during data collection in order to enhance instruction, minimize error, and improve data quality. This focus on prior training does not remove the need for post hoc data filtration procedures in citizen science, such as expert review, but is an important aspect of a robust project improvement cycle. The objectives of this study were to 1) determine the types of data collection errors made by student researchers, 2) determine what factors increase or decrease the likelihood of these errors, and 3) suggest ways to strategize instruction and refine protocols to reduce citizen science data collection errors in the field.

PROCEDURE

We examined the quality of student data sheets from 47 OPIHI field trips. Trips were conducted by nine teachers at eight different schools to 13 different intertidal sites from February 2004 to May 2007. Participating students, from grades 5 to 12, worked in groups with a volunteer chaperone while collecting data. Prior to the field trips, teachers led students in core OPIHI curricula, including lessons on field methodology and species identification (for a full description of OPIHI methods, activities, and protocols see 1 and 5). OPIHI protocols use traditional ecological sampling methods including transects and quadrats. Students completed separate data sheets on waterproof paper for each of up to three different sampling methods along a transect line. We analyzed student data sheets from 233 transect lines. Data sheets included the scientific names of the most abundant intertidal organisms as well as blank spaces for students to write in the names of additional organisms. Because students worked in groups, and different teachers focused on different sampling methods, to avoid pseudoreplication of data one unit of analysis encompassed all of the sampling data collected along a transect. Although each specific error type could occur multiple times on each data sheet, transect groups were considered to have either committed the error (“present”) or not (“absent”). Data collection procedures in citizen-science projects that emphasize recording the presence of organisms, rather than the absence, have been shown to be more reliable than projects that rely on opportunistic data collection (13). OPIHI data sheets followed a standard format, although teachers often modified them to streamline data collection when trips had limited time, when students required differentiation, or to highlight particular classroom content or elements of instruction. Because of these modifications, not all errors were possible on each data sheet. For example, some teachers filled in location information prior to photocopying, thus the error of location omission would not be possible on these data sheets. To correct for different potential numbers of mistakes students on each field trip could make, we calculated the percent error of each specific type of data collection error. The relationship between student error rate and teacher experience, assistant experience, student age and experience, and instructional time was examined using Pearson correlation coefficients and independent-samples t-tests. Experience was defined as the number of OPIHI field trips in which each citizen scientist participated. Information on instructional time was collected in 2007 by surveying teachers to determine how much time they had spent in class on OPIHI curricula.

RESULTS

We identified 24 different types of specific errors that were grouped into four broad categories (Table 1). Of these broad categories, “sloppiness” was the most prevalent error type. “General category,” a specific error in the sloppiness group, had the highest single error type—over 40% of student groups had at least one instance of this error on their data sheets. An example of a “general category” error would be writing “urchin” on the data sheet as opposed to the specific species of urchin, e.g., “Echinometra mathaei.” Using the scientific names for organisms, to a taxonomic level appropriate for field identification and student age, is important in biodiversity monitoring studies to accurately assess trends in ecological patterns.
TABLE 1

Specific student data sheet errors grouped into four broad categories.

Error CategorySpecific ErrorsaN TransectsPercent Error (%)
SloppinessGeneral category utilized (e.g., “urchin”)23341.6
Quadrat cover did not add to the correct number of intercepts or percent (100%) and final tally falsified21135.6
Incorrect substrate and/or incorrect or suspicious substrate size10234.3
Writing so messy not discernable23331.3
Quadrat “ticks” did not add to the correct number of intercepts and never tallied21131.3
Genus but no species name (or species name with no genus)23324.5
Suspicious wind speed12914.7
Quadrat “ticks” never tallied (but do add to correct number of intercepts)2118.1
Missing DataNo start and/or end times3938.5
No (or incomplete) site conditions18227.5
No transect number (metadata geographical information)22620.4
No location11013.6
No date2009.5
No names2138.0
MethodologicalMultiple records at transect intercepts20830.8
Wrong method2337.7
Transient objects recorded as data point (e.g., leaf litter or trash)2336.4
“Water” recorded as data point2332.2
MisidentificationMisidentification or suspicious species identification23318.5
Suspicious amount of space given to small organisms2333.9

Specific errors are in order of frequency, with the most frequent errors in each category listed first.

Specific student data sheet errors grouped into four broad categories. Specific errors are in order of frequency, with the most frequent errors in each category listed first. Over one-third of student groups had quadrat totals that did not add up to the number of intercepts utilized in OPIHI protocols, or 100% cover. Alarmingly, we found these students recorded that their data accurately totaled the correct coverage, indicating they never added their intercepts and falsified the tally. We hypothesize students may have felt too rushed in the field to complete this arithmetic “data check”; further evidence to support this hypothesis is the high prevalence of messy writing errors. Another error with a high rate of occurrence in the sloppiness category was the incorrect or suspicious substrate error (e.g., recording “limestone” for a site known to be a basalt bench). This error, along with high rates of incorrectly recoding multiple species under a single transect point (a methodology error), may indicate inadequate preparation for reporting certain site characteristics and the need for additional instruction and practice using OPIHI methods. Error rates decreased with OPIHI teacher experience (r = −0.20, n = 232, p < 0.01) and volunteer assistant experience; volunteers who had previously facilitated a field trip had significantly fewer errors on their mentee data sheets, t(42) = 3.35, p < 0.01. Older students made fewer errors than younger students (r = −0.34, n = 232, p < 0.01). Interestingly, there was no discernable relationship between OPIHI class instructional time before field trips (range 6.5 to 29 hours) and data sheet error rate. However, instructional time spent on sampling methods (5 to 10 hours) was more consistent than time spent on learning intertidal identifications (0 to 22 hours) prior to OPIHI field trips. Although not significant, overall error rates decreased with the number of field trips students went on. However, the percentage of missing data errors increased. This may be due to students not understanding the importance of research metadata, such as recording start and end times of sampling.

Implications for practice

Based on our systematic analysis of data sheets from a successful monitoring project, we suggest a number of strategies to reduce citizen-science data collection errors. Some of our recommendations will be most relevant to students participating in course-based research, including those at the postsecondary level, as they have implications for implementing active-learning strategies (4). Other recommendations are applicable to any citizen-science program that features a training component. For example, several of the most common errors we identified, particularly those regarding sloppy or falsified data, are likely unique to situations where participation is required as a course component—in contrast to citizen-science projects whose participants are self-selected interested volunteers. Student motivation and enthusiasm for accuracy may be reduced and data collection rushed due to course structure. Other types of errors, such as identification errors, are found across a number of citizen-science projects; our suggestions for decreasing these errors align with those put forth by other researchers (8).

Emphasize neat data collection

In school settings, students can examine the readability of data sheets from previous years or swap data sheets when inputting information; neatness can be included when assessing student performance. Encourage citizen scientists to identify potential sources of error and recommend corrections— implement their suggestions to continually refine and improve data collection procedures. In addition to reducing errors, participant engagement may be increased when they have the opportunity to help guide data collection and entry protocols. Consider the use of technology to help minimize sloppy errors if appropriate to your project and environment.

Explicitly address and discuss the problem of falsifying data

Examining the validity and reliability of volunteer data is good practice for all citizen-science projects. However, these practices often occur after sampling. To facilitate conscientious data collection, discuss the potential scientific and ethical implications of cutting corners as part of project training. Cultivating a classroom or community culture that is collaborative, as opposed to competitive, may decrease these types of errors and allow for instructive discussions should concerns arise.

Practice data collection techniques

In this study, practice appeared to be more effective than direct instruction at building data collection skills. Ecological data collection can be practiced in the field—or even in a classroom or courtyard before a field trip. Other citizen science projects have found error rates decrease with modest training (9); reviews of successful projects show practice is particularly effective for studies requiring taxonomic identification (8). For K–12 and postsecondary students, practice is also a component of enhanced learning. Continued opportunity to engage in data collection protocols not only improves scientific accuracy, it builds scientific content and skills knowledge (11).

Practice identification skills

We hypothesize one of the reasons for the large number of “general category” errors in this study is that both teachers and students are uncomfortable using scientific names. Emphasizing the importance of using common language, in particular scientific terms appropriate to the citizen-science project, helps to standardize data collection and thus can enhance scientific usage.

Tailor data sheets and identification cards to sites

Opportunistic data collection by volunteers can be particularly subject to bias (13). In this study, we found students prioritized species identifications based on the most easily accessible information. To address these errors, students or volunteers can be recruited to refine data sheets or build and refine field guides so information is easily accessible and user-friendly, but the amount is not overwhelming, as it is site-specific.

Recruit experienced volunteers to assist with the data collection process

Depending on your project, volunteers can be adult assistants, scientists, or more experienced peers. Students who have already completed a course can serve as mentors. At the postsecondary level, independent or work study opportunities are ways to recruit a cadre of more experienced volunteers. These approaches to reducing errors in data collection emphasize the importance of the role of an instructor or lead facilitator in enhancing citizen-science data quality. Error rates may be indicative of how careful students or volunteers were during data collection and thus reflective of overall data quality. Being aware of common data collection errors allows educators and trainers to anticipate and address potential pitfalls in data collection.
  4 in total

1.  A new dawn for citizen science.

Authors:  Jonathan Silvertown
Journal:  Trends Ecol Evol       Date:  2009-07-06       Impact factor: 17.712

2.  Expert variability provides perspective on the strengths and weaknesses of citizen-driven intertidal monitoring program.

Authors:  T E Cox; J Philippoff; E Baumgartner; C M Smith
Journal:  Ecol Appl       Date:  2012-06       Impact factor: 4.657

3.  Assessing accuracy in citizen science-based plant phenology monitoring.

Authors:  Kerissa K Fuccillo; Theresa M Crimmins; Catherine E de Rivera; Timothy S Elder
Journal:  Int J Biometeorol       Date:  2014-09-02       Impact factor: 3.787

4.  Process, not product: investigating recommendations for improving citizen science "success".

Authors:  Amy Freitag; Max J Pfeffer
Journal:  PLoS One       Date:  2013-05-15       Impact factor: 3.240

  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.