Literature DB >> 28949323

A global reference database of crowdsourced cropland data collected using the Geo-Wiki platform.

Juan Carlos Laso Bayas1, Myroslava Lesiv1, François Waldner2, Anne Schucknecht3,4, Martina Duerauer1, Linda See1, Steffen Fritz1, Dilek Fraisl1, Inian Moorthy1, Ian McCallum1, Christoph Perger1, Olha Danylo1, Pierre Defourny2, Javier Gallego3, Sven Gilliams5, Ibrar Ul Hassan Akhtar6,7, Swarup Jyoti Baishya8, Mrinal Baruah8, Khangsembou Bungnamei8, Alfredo Campos9,10, Trishna Changkakati8, Anna Cipriani11,12, Krishna Das8, Keemee Das8, Inamani Das8, Kyle Frankel Davis13,14, Purabi Hazarika8, Brian Alan Johnson15, Ziga Malek16, Monia Elisa Molinari17, Kripal Panging8, Chandra Kant Pawe8, Ana Pérez-Hoyos3, Parag Kumar Sahariah18, Dhrubajyoti Sahariah8, Anup Saikia8, Meghna Saikia19, Peter Schlesinger20,21, Elena Seidacaru22, Kuleswar Singha8, John W Wilson23.   

Abstract

A global reference data set on cropland was collected through a crowdsourcing campaign using the Geo-Wiki crowdsourcing tool. The campaign lasted three weeks, with over 80 participants from around the world reviewing almost 36,000 sample units, focussing on cropland identification. For quality assessment purposes, two additional data sets are provided. The first is a control set of 1,793 sample locations validated by students trained in satellite image interpretation. This data set was used to assess the quality of the crowd as the campaign progressed. The second data set contains 60 expert validations for additional evaluation of the quality of the contributions. All data sets are split into two parts: the first part shows all areas classified as cropland and the second part shows cropland average per location and user. After further processing, the data presented here might be suitable to validate and compare medium and high resolution cropland maps generated using remote sensing. These could also be used to train classification algorithms for developing new maps of land cover and cropland extent.

Entities:  

Year:  2017        PMID: 28949323      PMCID: PMC5613736          DOI: 10.1038/sdata.2017.136

Source DB:  PubMed          Journal:  Sci Data        ISSN: 2052-4463            Impact factor:   6.444


Background & Summary

The spatial extent of cropland has been mapped from remote sensing via many different initiatives as part of global land cover mapping activities, e.g., GLC-2000[1], MODIS land cover[2], GlobeLand30[3] and the Climate Change Initiative (CCI) of the European Space Agency[4]. Despite the availability of these and many other products, large spatial disagreement on the location and distribution of cropland still exists[5-7]. As such, quality-assured reference data are needed to undertake robust quantitative assessments and detailed comparisons of global products regarding their representation of cropland extent. Reference data sets can be collected in-situ, e.g., the Land Use Cover Area frame Sample (LUCAS) across EU member states[8], but due to the high costs involved in field surveys, they are more often gathered through interpretation of high or very high resolution satellite imagery. Some of the reference data used to validate different global land cover products are now being made openly available, e.g., through the GOFC-GOLD validation portal[9]. Because these data sets are not specifically tailored to cropland validation, sample sizes are insufficient, making their efficacy in quality assessments questionable, especially given the lack of sensitivity of accuracy indices[7]. To collect reference samples specifically designed for cropland map validation, we conducted a three-week cropland identification campaign during September, 2016. The campaign was implemented using the Geo-Wiki (http://www.geo-wiki.org/) crowdsourcing tool. A schematic showing the design and implementation of the campaign is illustrated in Fig. 1. A secondary motivation of the campaign was to gain a better understanding of crowdsourced data quality as well as the reasons why volunteers participate in crowdsourcing campaigns.
Figure 1

Schematic representation of the design and implementation of the crowdsourcing campaign to collect reference samples designed for cropland map validation, implemented using the Geo-Wiki (http://www.geo-wiki.org/) crowdsourcing tool.

This campaign builds on previous crowdsourcing campaigns using Geo-Wiki[10], e.g., to validate a map of land availability for biofuels[11] and to map wilderness globally[12], while the results from several campaigns were used to produce a global hybrid cropland map[13], among others. The scope of the early campaigns was generally directed towards improving global land cover and land use reference data[10], whereas the campaign described here focuses specifically on cropland data. In addition to validation, the data presented here also represent a valuable training tool that can be used to develop new land cover or cropland extent maps as well as to train algorithms to produce remote sensing-based products[14,15].

Methods

To develop the cropland validation campaign, cropland per se had to be defined, and a sample of systematically selected areas was generated. At the same time, the Geo-Wiki platform was modified to implement the campaign, the incentive scheme was developed, and the control data for quality assurance were collected. This section describes the main components of the campaign as outlined in Fig. 1.

Cropland definition

In order to distinguish cropland from other classes, the definition used for the campaign follows that of GEOGLAM/JECAM[16,17] in which ‘The annual cropland from a remote sensing perspective is a piece of land of a minimum of 0.25 ha (minimum width of 30 m) that is sowed/planted and harvestable at least once within the 12 months after the sowing/planting date. The annual cropland produces an herbaceous cover and is sometimes combined with some tree or woody vegetation’. According to this GEOGLAM/JECAM definition, perennial crops, agroforestry plantations, palm oil, coffee, tree crops and fallows are not included in the cropland class. The following exceptions to this definition were made: Sugarcane plantations and cassava crops are included in the cropland class, although they have a longer vegetation cycle and are not planted yearly. Taken individually, small plots, such as legumes, do not meet the minimum size criterion of the cropland definition. However, when considered as a continuous heterogeneous field, they are included in cropland. Moreover, greenhouse crops cannot be monitored by remote sensing and are thus excluded from the definition. Note that the use of this definition may lead to underestimation of cropland in the situation where legumes or other crops are planted among tree crops such as fruit and nut trees or where fields were fallow for 1 or more years but still cultivated. This would not be picked up in the visual interpretation of the imagery using Geo-Wiki although the use of Google Earth historical imagery and the Normalized Difference Vegetation Index (NDVI) profiling tool may have helped to identify cropland in the latter situation.

Sampling design

A stratified systematic sampling procedure was applied to generate the sample locations where the validation would take place in frames/cells of 1°×1° (geographic coordinate system with latitude and longitude) across the globe. A given replicate corresponds to a relative location in each frame. The scheme was designed to correct the distortion of the non-equal area projection. These cells serve as an instrument for defining a first-phase sample. The strata used were derived from the IIASA cropland probability map[13] with the aim of sampling areas of lower or higher probability of misclassification with different rates. Areas with a cropland probability between 25 and 75% were assumed to be more difficult to classify and were therefore sampled with a higher rate, while areas with very low or very high probability of cropland were sampled at a lower rate as they are easier to classify. Table 1 summarizes the strata and distribution of samples in each stratum. The size for each stratum as well as the calculated weights that should be used for accuracy metrics are also shown.
Table 1

Strata, sample distribution and strata sizes in the cropland validation campaign.

Stratum (% cropland probability)Number of samplesPercent shareStratum size (Million km2)Weights (size of one sample, km2)
Calculated weights needed for computing accuracy indexes are also shown.    
1 (0%)5001.3984.01168026
2 (0–25%)1096030.5618.761712
3 (25–75%)1598444.5714.66917
4 (>75%)842223.4816.471955
The sampling unit was a frame/pixel of 300 m×300 m corresponding to the grid of PROBA-V images and the final number of sampling units was 35,866.

Data collection using Geo-Wiki

The reference data were acquired through a dedicated Geo-Wiki interface (Fig. 2). Once a participant was registered and logged on, he/she could see a sample location where a semi-transparent 300×300 m frame subdivided in 25 grid cells is superimposed on Google Maps imagery (indicated by A in Fig. 2). Users were then asked to click (i.e., shade in yellow) all grid cells covered by more than 50% cropland. Thus, the final values for sampling units (i.e., a 300×300 m frame=one location) were cropland proportions ranging from 0 (absence of cropland) to 100%. When all sub-cells were examined, the user could either click the submit button or the skip button (indicated by B in Fig. 2) and was then shown the next randomly selected sample location. The user could also add comments regarding the observed location and then submit the validation. The cropland definitions were provided to the participants in an introductory video and through an info button in the Geo-Wiki interface. Additional tools and learning materials were provided to the participants to aid their interpretations. For example, in Geo-Wiki it is possible to switch between imagery from Google Maps and Microsoft Bing as well as viewing the location on OpenStreetMap (indicated by C in Fig. 2), which can provide additional useful information. The system registers whether a participant used imagery from Google Maps or not, which is included as a variable in the data set. Any location could also be saved as a keyhole markup language (kml) file for visualization using the desktop version of Google Earth (indicated by D in Fig. 2), which provides historical imagery, 3D viewing capabilities, geotagged photographs from Panoramio, etc. The usage of this feature was also registered in the data set. Participants were asked to use imagery from the latest date possible between Google Maps and Bing. Learning materials were compiled into an online gallery (Fig. 3), which provided the participants with different examples of cropland and non-cropland surfaces (http://www.geo-wiki.org/Application/modules/sigma_validation/sigma_gallery.html). Finally, it is possible to view different time series of vegetation indices, e.g., the NDVI (indicated by E in Fig. 2), obtained from different satellite sensors, i.e., Landsat 7, 8, MODIS and PROBA-V. These indices allowed participants to view the profiles of vegetation change over time at a particular location, which could help with satellite image interpretation, e.g., cropland is often characterized by a rapid increase in NDVI at growing stage after planting and a rapid decline near maturity stage or after harvesting.
Figure 2

The Geo-Wiki interface (http://www.geo-wiki.org) for collecting cropland information based on image interpretation.

(a) is the sub-grid of pixels that users must classify; (b) is the Submit button that users must press once they have completed their interpretation; (c) allows the user to change the background imagery; (d) shows the ‘View in Google Earth’ button, which users can press to be shown the location in Google Earth so that that they can view historical imagery; and (e) shows the NDVI profiles that can be viewed when the user clicks on a location.

Figure 3

Definition and examples of cropland (yellow shading) and areas of non-cropland as shown in a gallery of examples on Geo-Wiki (http://www.geo-wiki.org/Application/modules/sigma_validation/sigma_gallery.html).

Feedback was provided to participants as the campaign progressed using the Geo-Wiki Facebook page https://www.facebook.com/GeoWiki, which contained additional examples and a link to the YouTube explanatory video https://youtu.be/PR3xMPPyp-I showing how to use the interface. Participants could request help from experts for images that were difficult to classify and the answers were then posted to Facebook for all to view.

Quality control measures

Out of the total sample locations, 2,000 were randomly selected and validated by a group of three students trained in satellite imagery interpretation. The methodology for validation of control points was the same as for normal locations. These sample locations were compared for consistency, resulting in the removal of 207 sample units where there was disagreement in 3 or more grid cells/sub-pixels between the student validators. Additionally, independent verification was undertaken by experts at the International Institute for Applied Systems Analysis (IIASA) to ensure the quality of the control data set. Experts are members of IIASA staff with a background in remote sensing or geospatial sciences and considerable experience in image classification. This control data set was then used during the campaign, where participants received one control location for every 20 sample locations although this control location could appear at any point during the sequence of 20 samples. Each time a control location was viewed, the submission sent by the participant was compared with the control validation and a quality score was calculated for each participant as shown in Table 2. This, in combination with the amount of validations undertaken, was used to determine the participant’s ranking on the campaign leader board.
Table 2

Quality score calculation per location. Units for agreement are in number of grid cells/sub-pixels per 300 m×300 m location.

Agreement with controlPoints to the participantAgreement with controlPoints to the participant
252512−1
242311−3
232110−5
22199−7
21178−9
20157−11
19136−13
18115−15
1794−17
1673−19
1552−21
1431−23
1310−25
The campaign aimed to validate all sample locations at least 3 times by different participants. The final result achieved was that the majority of locations (32,287) were validated 4 to 7 times. Control points were validated more often, sometimes more than once by the same person to check for consistency. Despite a technical problem in the middle of the campaign, where some validations done in the middle of the campaign were not recorded, the full sample of validations was obtained.

Incentives and motivations

The top 30 participants (ranked by quality score) had the option to choose between becoming a co-author on a scientific paper or receiving an Amazon gift voucher ranging in value from 50 to 750 EUR (Table 3) depending upon the final position on the campaign leader board. A total of 26 participants chose to be co-author. They were also asked to fill out a survey providing some basic information about themselves and details regarding their motivation in participating in the campaign.
Table 3

Financial rewards offered according to the final ranking of the participants.

RankProposed financial prize
1€ 750
2€ 500
3€ 300
4€ 100
5€ 85
6€ 65
7–9€ 50
10–30€ 25
The same survey as that sent to the top 30 was also sent to the other participants where they were offered the following incentive: they were entered into a draw in which they could win one of two Amazon vouchers of €50 euros. A total of 20 additional answers were received. From the 1,793 control locations, a further sub-sample of 60 locations was selected and then evaluated independently by three land cover experts at IIASA following the same methodology as a normal participant. These locations were then reviewed for consensus between the experts, creating a gold standard data set. Although the gold standard was not used to calculate the quality score, it is provided here as an additional data set for independent quality and reliability assessment. These 60 locations were evaluated by all participants sequentially in the middle of the campaign, although no notice was given to the participants and no changes to the Geo-Wiki interface were made.

Data Records

The data are presented in six different data records. The first three data records contain all of the grid cells marked as cropland by either the campaign participants (Data record 1, n=1,086,485), the controls from the trained students (Data Record 2, n=8,918) or the gold standard (Data record 3, n=582) and can be found in crop_all.txt, crop_con.txt and crop_exp.txt (Data Citation 1), respectively. The format and information contained in these first three data records is shown in Table 4. Note that when these data correspond to the control data or to data from the experts, the following fields are not present: comment, timestamp, used_gmaps, viewed_ge, and skip_reason. The userid field in Data Record 2 is the number 111,111 and Data Record 3 is the number 222,222.
Table 4

The format and field descriptions of data records containing all grid cells marked as cropland.

VariableTypeDescriptionExample
location_idNumeric, continuousUnique number identifying each location in the campaign.47286
useridNumeric, continuousNumeric field used to uniquely identify participants/users11182
sub_idNumeric, continuousSequentially assigned number identifying every submission done in the system383725
commentTextComments entered by the participantApparent pastures
timestampDate and timeExact time and date when the submission was entered into the system2016-09-16 13:20:19
used_gmapsYes=‘t’No=‘f’Registers whether the participant was viewing the Google background imagery when the submission was donet
viewed_geYes=‘t’No=‘f’Registers whether the participant pressed the button labelled View in Google Earthf
skip_reasonNumeric, categoricalRegisters whether the participant did not skip the point (Skip=0), skipped the point and used the reason ‘no img. available/ low resolution/ clouds’ (Skip=1), or skipped the point and used the reason ‘too difficult’ (Skip=2)0
sub_item_idNumeric, continuousUnique identifier of each grid cell classified as cropland at a given location by a given user10579829
sub_item_xNumeric, continuousLongitude of each grid cell centroid inside a frame/location (decimal degrees)30.95357144
sub_item_yNumeric, continuousLatitude of each grid cell centroid inside a frame/location (decimal degrees)−20.75119048
Additionally, data records 4 to 6 show the information compiled per 300 m×300 m frame and per user, i.e., one record shows the average (mean) cropland from the 25 grid cells from a given user at a given location. Data Record 4 (n=203,515) contains data from all participants, data record 5 (n=1,793) contains the control data from the trained students while Data Record 6 contains the expert data (n=60). These data sets can be found in loc_all.txt, loc_con.txt and loc_exp.txt (Data Citation 1), respectively, while the format and field descriptions are provided in Table 5. As in data sets 2 and 3, the userid field in data record 5 is the number 111,111 and in data record 6 it is the number 222,222.
Table 5

Format and field descriptions of data records containing average (mean) cropland per frame/location and user.

VariableTypeDescriptionExample
location_idNumeric, continuousUnique number identifying each location in the campaign.47286
useridNumeric, continuousNumeric field used to uniquely identify the participants/users.5
sumcropNumeric, continuousAverage (mean) cropland at a given location in percentage80
loc_cent_XNumeric, continuousLongitude of a frame/location centroid (decimal degrees)−39.75
loc_cent_YNumeric, continuousLatitude of a frame/location centroid (decimal degrees)−8.047619048

Technical Validation

Figure 4 illustrates the origin of the 50 participants who provided information on the post-campaign survey and their familiarity with the regions validated as well as general information. It is clear that the majority of participants were male (68%) with a background in research (62%), highly educated (92%), and between 20 and 39 years of age (72%). The largest number of participants were from India (17) although more than 20 countries were represented. Participants had varying knowledge of different parts of the world although there was no area where participants had zero familiarity. This may reflect the geographical spread of the participants and their backgrounds.
Figure 4

Geographical location, previous knowledge and general information from the participants who filled in the survey at the end of the cropland validation campaign (n=50).

Figure 5a shows data collected during the campaign, expressed as the average (mean) cropland percentage per location and its global distribution. Figure 5b contains the IIASA-IFPRI hybrid cropland percentage map[13], and it is provided as a reference; in general, the patterns of cropland between the two maps are similar. Figure 5c shows the number of times a location was validated, where the majority of locations were classified at least 3 to 5 times.
Figure 5

Cropland validation campaign and worldwide spatial distribution of cropland.

The (a) presents cropland data collected during the cropland validation campaign, showing the mean cropland percentage per location and on the (b) the IIASA-IFPRI hybrid cropland map is shown for comparison. The third (c) shows the number of validations at each location during the campaign.

Usage Notes

The primary use of this reference data set is to validate global cropland maps generated using remote sensing that range from 60 to 300 m in resolution. More specifically, the data allows for an extensive spatially explicit validation of the cropland layer due to the rich amount of reference data. A validation exercise is planned for a 300 m cropland map that has been created for agricultural monitoring purposes as part of the FP7-funded SIGMA project (http://www.geoglam-sigma.info/). The data can also be used to train classification algorithms in developing new cropland maps based on remote sensing or to create hybrid cropland maps by fusing together existing cropland products[13]. Finally, it would be possible to use the data for studies about the quality of crowdsourced data.

Additional Information

How to cite this article: Laso Bayas, J.C. et al. A global reference database of crowdsourced cropland data collected using the Geo-Wiki platform. Sci. Data 4:170136 doi: 10.1038/sdata.2017.136 (2017). Publisher’s note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
  3 in total

1.  Mapping global cropland and field size.

Authors:  Steffen Fritz; Linda See; Ian McCallum; Liangzhi You; Andriy Bun; Elena Moltchanova; Martina Duerauer; Fransizka Albrecht; Christian Schill; Christoph Perger; Petr Havlik; Aline Mosnier; Philip Thornton; Ulrike Wood-Sichra; Mario Herrero; Inbal Becker-Reshef; Chris Justice; Matthew Hansen; Peng Gong; Sheta Abdel Aziz; Anna Cipriani; Renato Cumani; Giuliano Cecchi; Giulia Conchedda; Stefanus Ferreira; Adriana Gomez; Myriam Haffani; Francois Kayitakire; Jaiteh Malanding; Rick Mueller; Terence Newby; Andre Nonguierma; Adeaga Olusegun; Simone Ortner; D Ram Rajak; Jansle Rocha; Dmitry Schepaschenko; Maria Schepaschenko; Alexey Terekhov; Alex Tiangwa; Christelle Vancutsem; Elodie Vintrou; Wu Wenbin; Marijn van der Velde; Antonia Dunwoody; Florian Kraxner; Michael Obersteiner
Journal:  Glob Chang Biol       Date:  2015-01-16       Impact factor: 10.863

2.  Downgrading recent estimates of land available for biofuel production.

Authors:  Steffen Fritz; Linda See; Marijn van der Velde; Rachel A Nalepa; Christoph Perger; Christian Schill; Ian McCallum; Dmitry Schepaschenko; Florian Kraxner; Ximing Cai; Xiao Zhang; Simone Ortner; Rubul Hazarika; Anna Cipriani; Carlos Di Bella; Ahmed H Rabia; Alfredo Garcia; Mar'yana Vakolyuk; Kuleswar Singha; Maria E Beget; Stefan Erasmi; Franziska Albrecht; Brian Shaw; Michael Obersteiner
Journal:  Environ Sci Technol       Date:  2013-01-11       Impact factor: 9.028

3.  A global dataset of crowdsourced land cover and land use reference data.

Authors:  Steffen Fritz; Linda See; Christoph Perger; Ian McCallum; Christian Schill; Dmitry Schepaschenko; Martina Duerauer; Mathias Karner; Christopher Dresel; Juan-Carlos Laso-Bayas; Myroslava Lesiv; Inian Moorthy; Carl F Salk; Olha Danylo; Tobias Sturn; Franziska Albrecht; Liangzhi You; Florian Kraxner; Michael Obersteiner
Journal:  Sci Data       Date:  2017-06-13       Impact factor: 6.444

  3 in total
  3 in total

1.  Pervasive cropland in protected areas highlight trade-offs between conservation and food security.

Authors:  Varsha Vijay; Paul R Armsworth
Journal:  Proc Natl Acad Sci U S A       Date:  2021-01-26       Impact factor: 12.779

2.  A national-scale land cover reference dataset from local crowdsourcing initiatives in Indonesia.

Authors:  Ping Yowargana; Muhammad Thoha Zulkarnain; Fathir Mohamad; Bunga K Goib; Paul Hultera; Tobias Sturn; Mathias Karner; Martina Dürauer; Linda See; Steffen Fritz; Adis Hendriatna; Afi Nursafingi; Dian Nuraini Melati; F V Astrolabe Sian Prasetya; Ita Carolita; Muhammad Iqbal Firdaus; Muhammad Rosidi; Florian Kraxner
Journal:  Sci Data       Date:  2022-09-17       Impact factor: 8.501

3.  A crowdsourced global data set for validating built-up surface layers.

Authors:  Linda See; Ivelina Georgieva; Martina Duerauer; Thomas Kemper; Christina Corbane; Luca Maffenini; Javier Gallego; Martino Pesaresi; Flavius Sirbu; Rekib Ahmed; Kateryna Blyshchyk; Brigitte Magori; Volodymyr Blyshchyk; Oleksandr Melnyk; Roman Zadorozhniuk; Marian-Traian Mandici; Yuan-Fong Su; Ahmed Harb Rabia; Ana Pérez-Hoyos; Roman Vasylyshyn; Chandra Kant Pawe; Svitlana Bilous; Serhii B Kovalevskyi; Sergii S Kovalevskyi; Kusumbor Bordoloi; Andrii Bilous; Kripal Panging; Valentyn Bilous; Reinhard Prestele; Dhrubajyoti Sahariah; Anjan Deka; Nityaranjan Nath; Rui Neves; Viktor Myroniuk; Mathias Karner; Steffen Fritz
Journal:  Sci Data       Date:  2022-01-20       Impact factor: 6.444

  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.