Literature DB >> 31509880

Pan-European Data Harmonization for Biobanks in ADOPT BBMRI-ERIC.

Sebastian Mate1, Marvin Kampf1, Wolfgang Rödle2, Stefan Kraus2, Rumyana Proynova3, Kaisa Silander4, Lars Ebert5, Martin Lablans5, Christina Schüttler2, Christian Knell1, Niina Eklund4, Michael Hummel6,7, Petr Holub7, Hans-Ulrich Prokosch1,2.   

Abstract

BACKGROUND: High-quality clinical data and biological specimens are key for medical research and personalized medicine. The Biobanking and Biomolecular Resources Research Infrastructure-European Research Infrastructure Consortium (BBMRI-ERIC) aims to facilitate access to such biological resources. The accompanying ADOPT BBMRI-ERIC project kick-started BBMRI-ERIC by collecting colorectal cancer data from European biobanks.
OBJECTIVES: To transform these data into a common representation, a uniform approach for data integration and harmonization had to be developed. This article describes the design and the implementation of a toolset for this task.
METHODS: Based on the semantics of a metadata repository, we developed a lexical bag-of-words matcher, capable of semiautomatically mapping local biobank terms to the central ADOPT BBMRI-ERIC terminology. Its algorithm supports fuzzy matching, utilization of synonyms, and sentiment tagging. To process the anonymized instance data based on these mappings, we also developed a data transformation application.
RESULTS: The implementation was used to process the data from 10 European biobanks. The lexical matcher automatically and correctly mapped 78.48% of the 1,492 local biobank terms, and human experts were able to complete the remaining mappings. We used the expert-curated mappings to successfully process 147,608 data records from 3,415 patients.
CONCLUSION: A generic harmonization approach was created and successfully used for cross-institutional data harmonization across 10 European biobanks. The software tools were made available as open source. Georg Thieme Verlag KG Stuttgart · New York.

Entities:  

Mesh:

Year:  2019        PMID: 31509880      PMCID: PMC6739205          DOI: 10.1055/s-0039-1695793

Source DB:  PubMed          Journal:  Appl Clin Inform        ISSN: 1869-0327            Impact factor:   2.342


Background and Significance

An important task in the medical research of the postgenomic era is the investigation of diseases by linking “-omics” data with phenotypes. 1 Widespread conditions, such as cancer, diabetes, or Alzheimer's disease, are influenced by a variety of small, often cumulative effects 2 that result from a combination of genes, 3 lifestyle, 4 and environmental factors. 5 The complexity of molecular disease patterns requires many forms of therapeutic intervention, tailored to the individual characteristics of a particular patient. 6 Clinical data, associated with well-characterized biomaterials, both of the highest quality, are key for the development of new drugs and diagnostic tests and are expected to be the building blocks of personalized medicine in the near future. 7 The Biobanking and Biomolecular Resources Research Infrastructure - European Research Infrastructure Consortium (BBMRI-ERIC) aims to facilitate access to biological resources. 8 Via its Common Service Information Technology (CS-IT) platform, a distributed computing environment, BBMRI-ERIC is working on establishing a pan-European search engine for biomaterial samples to support biomolecular and biomedical research. As described in more detail in Proynova et al, 9 it is planned that the CS-IT platform follows the principles of the decentralized search approach , 10 developed by the German Cancer Consortium. 11 In contrast to centralized search approaches (as implemented, e.g., in Schröder et al 12 ), data never leaves the biobanks, which enables them to retain data sovereignty and protect their patients' privacy. To support the implementation of the CS-IT platform, the Implementation and Operation of the Gateway for Health into BBMRI-ERIC (ADOPT BBMRI-ERIC) project was launched. One of its tasks was the collection and aggregation of colorectal cancer data across European biobanks. 13 Colorectal cancer was selected as it is one of the most prevalent neoplasms, 14 and most BBMRI-ERIC national nodes already had expertise in this area through ongoing research programs. The national nodes are the national biobank networks in each European country participating in the BBMRI-ERIC network. 15 Besides sample data (such as material type or preservation mode), the data collection covered patient information regarding diagnostic and surgical procedures, histopathology, molecular markers, demographics, planned and performed therapy (e.g., radiation or chemotherapy), and outcome. The goal was to make these anonymous data available on the BBMRI-ERIC platform after collection and validation. This process was known as the Colon Cancer Data Collection (CCDC) within ADOPT BBMRI-ERIC. 16

Objectives

Consolidating the heterogeneous biobank data into the common format required the data to be processed in a systematic way. One of the ADOPT BBMRI-ERIC work packages was focusing on the development of a data integration and harmonization toolkit to support this extract-transform-load process (ETL, also see refs. 17–19). It had to be able to convert the biobank data semiautomatically into a format ready for import into the central CCDC database. In this article, we describe this data harmonization approach and report on the experiences gained while integrating the data from 10 European biobanks. In particular, we analyze the quality of the semiautomatic mapping approach and report on the results of the ETL process.

Methods

Overview

The ADOPT BBMRI-ERIC project used an installation of the software Open Source Registry System for Rare Diseases (OSSE, Open-Source-Registersystem für Seltene Erkrankungen ) 20 as collection system for the CCDC data. OSSE is a case report form (CRF) system that generates forms for data entry by using standardized data element definitions from the Samply.MDR metadata repository (MDR). 21 Both applications are part of the Samply software environment, 22 for which expertise was available within the BBMRI-ERIC community and which is also used in other projects. 11 23 24 25 It allows for manual data entry via the CRFs to enable the participation of those biobanks, which are only able to provide data manually, but also provides an application programmable interface (API) for the automatic import of XML-encoded data. The latter method was preferred as it helps avoiding errors that may occur when copying data manually. The ETL process for this task should fulfill the following requirements: The overall ETL process has to provide a generic input interface for the data originating from the biobanks to be compatible with the wide range of data formats. It has to support the semantic translation of data elements and value sets from the biobank data into the target CCDC terminology as defined in the MDR. It has to convert the biobank data into the XML import format as required by OSSE. Since biobanks might provide data in a completely different format, complex syntactic conversions need to be supported. To address the first requirement, we specified two easy-to-generate input file formats for the ETL process, in which the biobanks could provide their data. The first one is a comma separated value file format, which follows the principles of the Entity-Attribute-Value (EAV) data modeling approach, as shown in Table 1 and described in more detail in refs. 26 and 27. It enables the provision of multiple instances of individual data elements, such as multiple locations of metastases. The second one is a flat file format. As shown in Table 2 , it uses one row per patient and encodes multiple instances within single cells, separated by semicolons (see highlighting).
Table 1

Example for the EAV table format (mock-up data)

Patient IDConceptValueInstance
1TNM-TT11
1TNM-NN01
1TNM-MM11
1Age451
1Metastases Hepatic 1
1Metastases Osseus 2
1GenderMale1
1Date of Surgery21.07.20131
2TNM-TT41
2TNM-NN01

Abbreviation: EAV, Entity-Attribute-Value.

Table 2

Example for the flat file format (mock-up data)

PatientTNM-TTNM-NTNM-MAgeMetastasesGenderDate of surgery
1T1N0M145 Hepatic; Osseus Male21.07.2013
2T4N0M034Female23.11.2018
3T2N0M121OsseusMale04.04.2017
4T1N1M143Brain; OsseusMale19.03.2012
5T2N2M176HepaticFemale10.09.2013
Abbreviation: EAV, Entity-Attribute-Value. The goal of the second and third requirement was to convert the biobank data into the target XML format, while adhering to the semantics of the target terminology with a mapping and transformation process that requires minimal human intervention. As already mentioned, the MDR contained a detailed, machine-readable description of the CCDC data elements (see ref. 28), which not only serves the automatic OSSE form generation, but also defines the contents and data types that may be utilized in an OSSE import XML file. We asked the biobanks to provide the same standardization by using the MDR for the description of their data items while adhering to the MDR data types. This approach helped us to achieve our third requirement by facilitating the comparison and translation between the two data sets, semantically and syntactically: A semiautomatic mapping process could help find correlations (semantically) between both metadata definitions, and a data transformation process could use this mapping information to transform the original biobank data (syntactically) into the target OSSE XML representation. Fig. 1 outlines the architecture of our ETL approach and its integration with the Samply.MDR (on top) and OSSE (on the right). As indicated with the two big arrows, it processes either an EAV or a flat file and generates an XML import file, compatible with the OSSE system. The smaller arrows represent the data flows processed by our implemented software tools: MDRExtractor , TablePreprocessor , MDRMatcher , MappingGUI , and ETLHelper . The following section describes these programs and their role in the ETL process.
Fig. 1

The extract-transform-load (ETL) pipeline as configured for the ADOPT BBMRI-ERIC project, shown for two exemplary biobanks, one contributing an Entity-Attribute-Value (EAV) and the other a flat file. The files are extracted from the Biobank Information Management Systems (BIMS, left), processed by our tools into an XML file, and finally loaded into the OSSE system (right). CMF, central metadata definition file; LMF, local metadata definition file.

The extract-transform-load (ETL) pipeline as configured for the ADOPT BBMRI-ERIC project, shown for two exemplary biobanks, one contributing an Entity-Attribute-Value (EAV) and the other a flat file. The files are extracted from the Biobank Information Management Systems (BIMS, left), processed by our tools into an XML file, and finally loaded into the OSSE system (right). CMF, central metadata definition file; LMF, local metadata definition file.

The BBMRI-ERIC ETL Approach

Metadata and Data Extraction

As a first step, we implemented the MDRExtractor program, which recursively extracts data elements, data types, value sets, and hierarchical information from MDR project namespaces via the MDR's REST API. The extracted information is stored in what we call metadata definition files . These are cached representations of the MDR's content and are faster to process by the tools compared with using the REST API of the MDR. Each namespace is either associated with a biobank's local terminology, or the central CCDC terminology. For the biobank namespaces, the MDRExtractor generates local metadata definition files , whereas for the latter, it produces a central metadata definition file (see Fig. 1 ). We introduced support for the flat file format when it became clear that many biobanks would not be able to provide their data in the EAV format. To allow for the integration of such data, we implemented the program TablePreprocessor . It analyzes the flat files and converts their contents into the EAV format. It also generates a simplified local metadata definition file by extracting the information from the file's column headings. These simplified metadata definition files served as a fallback solution in such cases where the biobanks would not enter their local metadata into the MDR.

Lexical Matching

To create the semantic bridge between the biobanks' local and the CCDC's central data elements, it is necessary to create mappings between the two terminologies. To assist the user in this task, we implemented MDRMatcher , a tool that matches the local and central metadata definition files to find similar data elements. This results in a mapping file . This file contains automatically generated, initial proposals, which we call matches . For each term in a biobank's source terminology, MDRMatcher creates a list of possible matches to the target CCDC terminology. If MDRMatcher or a human expert considers one of these matches to be correct, we call it a mapping . To put it in other words, a match is a mapping candidate. The implementation of MDRMatcher follows the assumption that if two metadata items contain the same or similar words, the two must be related. Consequently, the program compares all items from the source with all items from the target terminology. One such comparison is shown in Fig. 2 . After selecting two metadata items (1), one from the local and one from the central metadata definition file, MDRMatcher performs a string normalization (2), where the program reformats the strings to uppercase. It also removes nonalphanumeric characters and duplicate words.
Fig. 2

Illustration of the bag-of-words algorithm in MDRMatcher. After normalization and semantic expansion, it compares all n -grams from the source and the target item and computes a similarity score.

Illustration of the bag-of-words algorithm in MDRMatcher. After normalization and semantic expansion, it compares all n -grams from the source and the target item and computes a similarity score. MDRMatcher then performs what we call semantic expansion (3). By inserting additional words into the strings to be compared, we augment the metadata entries and support the program in comparing abbreviations or differently named concepts. For example, based on the synonym definition “CT = Computer Tomography,” the string “CT DONE” is transformed into “CT COMPUTER TOMOGRAPHY DONE.” Other synonyms defined are, for example, “Sex = Gender,” “Osseous = Bone = Skeleton,” or “Pulmonary = Lung.” We based the small database on the CCDC terminology, but also integrated other common medical abbreviations. 29 The same technique is also used for tagging metadata items with positive or negative sentiments. Theoretically, this improves matches among different positive (e.g., “Yes,” “True”) or negative value sets (e.g., “No,” “Not,” “False”). As an example, consider the match between the source data element “Patient is alive” with the value set {“No,” “Yes”} and the target data element “Patient is living” with the value set {“False,” “True”}. Even though both value sets are different, the sentiment tagging may support the lexical matcher in finding the correct mappings, for example, where “Patient is alive = No Neg ” is mapped to “Patient is living = False Neg ” and “Patient is alive = Yes Pos ” is mapped to “Patient is living = True Pos .” In the example in Fig. 2 , the word “NOT” was augmented with “NEG.” MDRMatcher then performs the actual lexical matching (4). The algorithm used is based on the bag-of-words approach. 30 31 It considers sequences of n successive words, called n-grams . 31 For this, it first determines g min , which is the lower word count of the two normalized strings being compared (in the example, 4). It extracts the sets with all n -grams with n  ≤  g min from both sides and puts these into two “bags,” one for the source and one for the target item. In the next step, the algorithm compares both bags via exact and fuzzy partial matches (with the latter not shown in Fig. 2 ). Each of these matches contributes to a final similarity score. Our approach is based on the assumption that longer words and in particular word sequences carry more information than short, individual words; this is why we calculate the scores using the product of the number of characters (red) and the gram count (blue). Fuzzy matches are computed only for single words ( n  = 1) by calculating the normalized Levenshtein distance 32 (0...1, with 0 = no similarity and 1 = same string), which is then multiplied with the half-length of the longer term. As a result, fuzzy matches are weighted less than exact matches. The lists of the best matches with a similarity score above a configurable threshold are then stored in a mapping file. The correct choice of the threshold prevents, on the one hand, that too many unnecessary (wrong) proposals are shown (a threshold of 0 would propose all target terms, sorted by relevance) and, on the other hand, that a correct proposal is not shown. When the method was prototyped on the data supplied by the first biobank, it was found that a threshold of one-third of the highest similarity score provided the best results. This value was then used for processing the data of all other biobanks. MDRMatcher also automatically suggests the best match as a mapping if the highest similarity score is higher than that of the second best match. If multiple top matches share the same similarity score, the program is unable to decide which one is better, and leaves the decision to the human expert.

Curation of Mappings by Human Experts

Human experts, typically the persons responsible for the data at the biobanks, have to complete, verify, and potentially correct the mapping file generated by MDRMatcher. To support this process, we developed a graphical user interface, MappingGUI , as shown in Fig. 3 .
Fig. 3

The MappingGUI program, which is used to curate mappings between source and target terms and values.

The MappingGUI program, which is used to curate mappings between source and target terms and values. The program displays the source terminology from the biobank on the left, and the mapping proposals from MDRMatcher on the right side of the window. The tool displays mappings curated by human experts in green, and mappings proposed by MDRMatcher in yellow. Otherwise, the tool highlights entries in red to indicate that only lexical matches from MDRMatcher are available. The expert can then either approve or correct a mapping by selecting another term from the right side. If the correct term was not among the proposals, the user can open a search window and search for the correct term. It is also possible to remove mappings via the “Remove mapping” button for cases where no mapping exists.

Transformation of Instance Data

The last program in the ETL pipeline is ETLHelper , which processes the mapping and EAV file containing the patients' instance data and then generates the OSSE XML import file. This step comprises replacing the source value sets with those from the target terminology, as well as converting between different data types according to the expert-curated mapping rules. The implementation of the ETLHelper uses an internal SQLite database 33 to efficiently sort, filter, and align the data. A single SQL join statement merges the data from the EAV file, the mapping file, and the metadata definition files, thereby creating the semantic bridge between the biobank's data and the CCDC's target terminology. Finally, the program steps through the SQL results, performs the data type conversions, and populates the OSSE XML file with the transformed biobank data. The generated XML file can then be imported into the OSSE system. As our ETL approach is based on the foundation of Samply.MDR's semantics, we only had to consider the conversion between eight different data types, as shown in Table 3 . Translations between simple value sets, represented in the MDR via the data type “Enumerated,” are handled in the above-mentioned SQL statement by replacing the original values with those of the target terminology according to the mapping rules in the mapping file. This is the case for example for the mapping “VITAL_STATUS = ALIVE => Vital status and survival information / Vital status = person is still alive.” Here, the translation method between “Enumerated” and “Enumerated” from Table 3 applies, and ETLHelper replaces the original “ALIVE” cell entries in the biobank data with “person is still alive.” When populating the XML file with the instance data, ETLHelper puts these entries into the respective location in the XML file, as defined in the CCDC namespace in the MDR (Vital status and survival information / Vital status). The translations between other data types are done following common sense, as shown in Table 3 . If, for example, a float is to be converted into an integer, we round it to the nearest integer. To translate a Boolean to integer, we convert the Boolean “True” to 1 and the Boolean “False” to 0.
Table 3

Supported data type transformations, with source data types shown on the left and target data types on top

From/ToEnumeratedIntegerFloatBoolean (True)Boolean (False)StringDateDateTime
Enumerated Target value True False Cast to String
Integer Pass Through Cast to Float if = 1 True if = 0 False Cast to String
Float Round to n. Int. Pass Through if = 1.0 True if = 0.0 False Cast to String
Boolean (True) Target value 1 1.0 Pass Through False Cast to String
Boolean (False) Target value 0 0.0 True Pass Through Cast to String
String Pass Through
Date Cast to String Pass Through Date Only
DateTime Cast to String Date Only Pass Through

Note: Italics indicates the type of handling, bold the return value. Blank indicates no transformation rule.

Note: Italics indicates the type of handling, bold the return value. Blank indicates no transformation rule. The implementation is robust in the sense that it can automatically parse dirty (but correct) data in different date or number formats. However, as indicated by the empty cells in Table 3 , we do not support all types of transformations. For example, we do not attempt to recognize non-string data (such as time stamps) in strings, and therefore do not offer converting a string into another data type. Whenever the parsing of a data record fails or the mapping contains an unimplemented transformation rule, the user is notified with an error message.

Results

The ETL approach described above was used to implement the data harmonization for the CCDC in the ADOPT BBMRI-ERIC project. From 14 participating biobanks (as of November 2018), 10 were unable to generate the OSSE XML file themselves. Instead, they extracted and anonymized relevant data from their source systems and converted them into the flat file format. These files were then collected via a secured Web portal, the CCDC Uploader , and the ADOPT BBMRI-ERIC IT personnel was notified about new uploads. It downloaded the files and performed initial data quality checks to assess the overall suitability for the CCDC. They also resolved easily correctable problems, such as minor formatting or encoding errors; however, when major changes in data were necessary, the biobanks were consulted. The IT personnel then created the mappings in close collaboration with the biobank experts. In most cases, it was clear how the mappings were to be created; however, in other cases the biobank data experts had to be consulted. The final mappings, which we also used as the “gold standard” for the evaluation of MDRMatcher as described below, were finally approved by the biobanks. The biobanks were constantly informed about interim ETL results. The XML files produced by ETLHelper were XSD-validated and uploaded into the OSSE system. Finally, the biobanks were given the opportunity to review their data in the OSSE system.

Data Received

During this task, we observed different types of problems with the biobank data. Many of them were identified directly by the BBMRI-ERIC IT personnel upon receiving the data files, but most were detected later in the process through mechanisms implemented into the harmonization tools. Shifted tabular data was the most common problem, which we attribute to copying and pasting of data. This could also have been the cause of other formatting problems, such as inconsistent upper and lower case or spaces at the beginning or end of cell entries. Character encoding errors, which were potentially caused by office software, also occurred frequently. The 10 biobanks made great efforts to provide as much information as possible. Unfortunately, this also means that they were reluctant to leave table cells empty, even where this would have been appropriate. Entries, such as “Unknown,” “Not available,” or “N/A” were used frequently, but also more uncommon values, such as “ND” for “not done.” The biobanks also used textual entries in numerical and date fields, which, as expected, led to data rejections in ETLHelper. This is the desired behavior of the software, because such entries do not contribute any relevant information.

Analysis of Lexical Matching

We used MDRMatcher to match the metadata items from the 10 biobanks to the target CCDC terminology. To assess the quality of this process and to be able to improve our approach in the future, we compared the program's output against the final, expert-curated mappings. We designed a four-axial classification scheme to categorize the behavior of MDRMatcher on a per-source-item basis. It is comprised of the four dimensions conceptual , mapping , correctness , and matching , as illustrated in Fig. 4 .
Fig. 4

A four-axial classification scheme to assess the mapping quality of MDRMatcher.

A four-axial classification scheme to assess the mapping quality of MDRMatcher. The conceptual dimension describes whether a source data item has an equivalent in the target terminology or not. In other words, it describes on a conceptual level whether a mapping can be created. This can be easily determined in the final mapping file (the “gold standard”) by checking whether the human experts left a mapping blank or checking whether the expert kept, modified, or removed a proposed mapping. If there was an equivalent item in the target terminology, it can be analyzed whether MDRMatcher has created a mapping or not, which we capture in the mapping dimension . The correctness dimension evaluates whether this mapping is correct or not by comparing the proposed mapping with the expert-curated mapping. The matching dimension determines whether the correct item from the target terminology was among the list of matches, or whether it has been cut off because its score was below the implemented cut-off threshold. Each of these four dimensions can be answered with “Yes” or “No” (see Fig. 4 ), and as such, a combination would result in 2 4  = 16 classes. However, only seven of these are applicable in practice, as some of them do not make sense. For example, an automatically created mapping cannot be correct if there is no equivalent item in the target terminology. The seven valid classes are, ordered from “good” to “bad”: A mapping was created and was correct. No mapping was created, which is correct because there is no equivalent item in the target terminology to map to. No mapping was created, but should have been. At least the correct term was in the list of proposals, which later support the curation process by the human expert. A mapping was created, but it was wrong. The correct item was in the list of proposals. A mapping was created, but it was wrong. The correct item was not in the list of proposals. A mapping was created, which is wrong, because there is no equivalent item in the target terminology. No mapping was created, but should have been. The correct term was not in the list of proposals. In the next step, we classified the mappings generated by MDRMatcher according to these seven classes. The MappingGUI program kept track of all modifications to the mapping files and thus allowed us to automatically compare the software-generated mappings with the final, expert-curated ones. The result is shown in Table 4 . In total, we matched 1,492 biobank source items to the 226 target items of the CCDC target terminology. As shown, 77.21% + 1.27% = 78.48% of all automatically generated mappings by MDRMatcher were correct (classes 1 and 2). For 20.24%, the correct mapping was among the suggestions and could be selected easily in MappingGUI by the human experts (classes 3 and 4). In other words, 98.72% of the mappings were correct or easily correctable (classes 1–4) with our tools. In only 1.27% (classes 5–7) of all cases, the correct target item was not among the list of suggestions and had to be looked up manually.
Table 4

Assessment of the automatic mapping quality for metadata items received

PropertyClass
1234567
Dimension Conceptual (Equivalent in Target?) YesNoYesYesYesNoYes
Mapping (Created A Mapping?) YesNoNoYesYesYesNo
Correctness (Correct Behavior?) YesYesNoNoNoNoNo
Matching (Among Proposals?) YesNoYesYesNoNoNo
Source Item Counts Biobank 1163101711220
Biobank 212112117010
Biobank 38373614212
Biobank 4940128200
Biobank 58901413010
Biobank 6117079000
Biobank 714201718100
Biobank 811701812201
Biobank 99711218000
Biobank 1012901612110
Total1152191701321063
Percent77.211.2711.398.850.670.400.20

Note: The numbers are per source item, which is comprised of concept and value, e.g., “UICC_STAGE = Not known.”

Note: The numbers are per source item, which is comprised of concept and value, e.g., “UICC_STAGE = Not known.” Based on Table 4 , it is possible to derive additional standard metrics to evaluate MDRMatcher's role as an automatic mapper. Based on a confusion matrix as shown in Table 5 , it is possible to calculate the precision, recall, and F 1 score:
Table 5

Confusion matrix for the evaluation of MDRMatcher as an automatic mapper

TP + FN Should create mapping1171 FP + TN Should not create mapping321
TP + FP Created mapping 1300 TP Created mapping, correct1152 FP Created mapping, wrong132 + 10 + 6 = 148
FN + TP Did not create mapping 192 FN Did not create mapping, wrong170 + 3 = 173 TN Did not create mapping, correct19
We then analyzed why MDRMatcher did not succeed in finding the correct mappings for the terms in classes 3 to 7, with a deeper analysis of classes 3 to 5. The results for the classes 3 to 5 are summarized in Table 6 .
Table 6

Analysis of wrong behavior of MDRMatcher per target data element in classes 3 to 5 for metadata items

Type of problemClass 3Class 4Class 5Total
Missing synonyms2222
No down-ranking2121
N -gram matching across hierarchy/concept/value 36413
Wrong up-ranking due to bad fuzzy matches3811
Unfavorable removal of redundant words3811
Inability to compare Roman with Arabic number189
Use of different languages55
Spelling errors or inconsistent naming55
Unable to map something to “Other”22
Other41218
The most common type of wrong behavior in MDRMatcher could be traced back to a lack of synonym definitions, for example, for “ND” (“Not Done”) or “X” (“Lung Imaging,” as in “X-ray”). Defining these commonly used words in the synonyms database would have resulted in correct automatic mappings. Similarly, the missing ability of down-ranking metadata items in the matching algorithm has often proven to be a problem. MDRMatcher uses only matches to increase (or up-rank) the similarity score, but it does not consider “nonmatches”; that is, text passages that do not occur on both sides. For example, for the source item “MM_ MISMATCH _ REPAIR _GE =  EXPRESSION ,” the matcher could not decide if “Molecular markers / Mismatch repair gene expression =  Expression ” or “Molecular markers / Mismatch repair gene expression = Loss of expression ” is the correct match. In both cases, the algorithm found the same partial matches (as indicated in bold) and thus calculated the same similarity score, but the additional information (“Loss of …”) was not considered. The implemented n -gram-based matching approach theoretically allows hierarchical information to be taken into account, because all metadata items to be matched are built using strings in the form “Hierarchy Path / … / Concept = Value,” from generic to specific. In practice, however, this sometimes led to the exact opposite effect. As the slashes (/) and equal signs (=) were removed from the data during preprocessing, misleading matches across these components occurred. For example, “DIAG_ MRI_DONE  = Not done” was incorrectly matched to “Diagnostic exam / MRI = Done , data not available” instead of the correct “Diagnostic exam / MRI = Not done.” Wrong up-rankings due to incorrect additional fuzzy matches were another common problem. For example, the International Classification of Diseases, Tenth Revision (ICD-10) code “C 19” was incorrectly mapped to “C 19.9” instead of the correct “C 19,” because in addition to the correct 2-gram match between “C 19” and “C 19,” the tool incorrectly created a second fuzzy match between “19” and “9.” The removal of redundant words during the preprocessing, which we believed prevented overfitting, also caused problems in practice. For example, MDRMatcher incorrectly proposed mappings between NRAS and KRAS, due to a fuzzy match between “NRAS” and “KRAS” and the removal of duplicate words in the hierarchy during text preprocessing. “NRAS” and “KRAS” were part of the hierarchical grouping (e.g., “Molecular markers / KRAS mutation status / NRAS exon 2 [codons 12 or 13] = Not mutated”) in the MDR. This combined type of error affected 9 biobanks and could have been prevented if the text preprocessing would not have removed duplicate words. MDRMatcher proposed some wrong mappings because it failed to match Roman to Arabic numbers (e.g., for the UICC stages). Bad spelling or inconsistent naming (also within the CCDC target terminology) also led to incorrect mappings. For example, the ICD-10 code “C18.0” was mixed up for surgery and histopathology due to inconsistent writing of “caecum” (vs. “cecum”) in the CCDC target terminology. Similarly, the misspelled value “3th” was mapped to “4th,” because for the Levenshtein algorithm “3th” is closer to “4th” than “3th” is to “3rd.” Finally, MDRMatcher was not able to map certain biobank terms to “Other” values in the target terminology. This was the case, for example, when a biobank specified metastasis localizations for which there were no equivalent entries in the target terminology. This is where the software simply lacks the intelligence to make the correct assignment to “Other.” As a result, the human expert must complete the mappings. Class 6 errors (mapping despite no representation in target terminology) were incurred, because the lexical matcher always proposes the “best” match as a mapping if its score is higher than the second-best matching. For example, for “UICC_STAGE = Not known” it incorrectly proposed “Histopathology / UICC staging / UICC version = Not known.” Class 7 did not contain enough mappings to discern any recurring pattern.

Analysis of Data Transformation

We used the user-validated mappings to transform the biobank data into the XML import format. Table 7 summarizes the whole ETL process for the data of the 10 biobanks. In total, we were able to process 98.7% of 147,608 data records from 3,415 patients. As we will describe later, we consider the remaining 1.3% as correct rejections.
Table 7

ETL results for the actual facts data received from 10 biobanks participating in the CCDC pilot as reported by ETLHelper

PropertyBiobank
12345678910Total
Input data
1Patients1,06621850553006003003082183003415
2Data records50,0208,7562,0672,47711,98422,61713,22513,7299,16013,573147,608
3Different concepts53474947474349524056483
4Different source items2051611451161171331781501291581,492
Mapping rules
5Mappings between source and target items1931591341161161331781501271581,464
6Data records that should have a mapping50,0208,7562,0672,47711,98422,61711,32513,7299,16013,573145,708
7Data records that do have a mapping49,2638,3311,9312,47711,98322,61711,32513,7299,14813,573144,377
8Percentage of data records that have a mapping98.5%95.1%93.4%100.0%100.0%100.0%100.0%100.0%99.9%100.0%99.1%
9Data records that don't have a mapping 757 425 136 0 1 000 12 0 1331
Data type transformations
10Total number of transformations49,2638,3311,9312,47711,98322,61713,22513,7299,14813,573146,277
11 Good Enumerated ⇒ Enum. 35,7866,5441,3831,9079,43619,35010,29410,7987,62210,234113,354
12 Enumerated ⇒ Boolean 1,11910028308218911,864
13 Integer ⇒ Integer 7,0938152303009952,5682,1401,6274361,85118,055
14 Float ⇒ Integer 102300218620
15 String ⇒ String 2,65443512655636981913802187975,590
16 Date ⇒ Date 2,132388921106526006006164366006,226
17Total number of good transformations48,7848,2841,9312,40011,98322,61613,22513,7299,14813,573145,673
18 Bad Integer ⇒ Integer 479 00 77 0 1 0000 557
19 Date ⇒ Date 0 47 0000000 47
20Total number of bad transformations 479 47 0 77 0 1 0000 604
21Percentage of good transformations99.0%99.4%100.0%96.9%100.0%100.0%100.0%100.0%100.0%100.0%99.6%
Summary
22Percentage of input data records that could be mapped and transformed97.5%94.6%93.4%96.9%100.0%100.0%100.0%100.0%99.9%100.0%98.7%

Abbreviations: CCDC, Colon Cancer Data Collection; ETL, extract-transform-load.

Note: Rejected data records are printed in bold.

Abbreviations: CCDC, Colon Cancer Data Collection; ETL, extract-transform-load. Note: Rejected data records are printed in bold. Table 7 consists of four sections. The first one describes the input data. When loading the biobank data, ETLHelper first counts the number of patients (row 1) and the number of data records, the number of nonempty cells in the original flat file (row 2). The number of different concepts (row 3) corresponds to the number of columns in a flat file, whereas the source items count (row 4) is the number of concept-value combinations. The second section (rows 5–9) describes the availability of mapping-rules, where ETLHelper evaluates for how many of the data records (not metadata items) a mapping rule can be utilized to transform the actual patient data. A low percentage in row 8 indicates that a biobank's data did not match the target terminology well, or that the mapping needs to be revised. The third section (rows 10–21) summarizes the number and kind of data type validations and transformations utilized. Finally, row 22 outlines the total percentage of records that we were able to transform into the target CCDC representation. Although the ETL process delivered overall good results, where most biobanks achieved a transformation rate of or close to 100%, we were interested in what happened in the remaining cases. We therefore analyzed two aspects: First, why the human expert did not map a source data item, and second, why certain data type castings resulted in errors. As mentioned above, many biobanks (1, 2, 3, and 9) provided data with “unknown” values that did not contribute any useful information. Examples are entries such as “UICC_STAGE = Not known” or “WHO_GRADE = Not known.” Additionally, biobank 1 provided data for a data element “Targeted Therapy Scheme” with nine enumerated values, which was not collected in the CCDC. Similarly, biobank 5 contributed entries with “Localization of primary tumor = C18.9,” which were outside of the inclusion criteria of the colorectal cancer cohort. Such entries were deliberately unmapped by the human experts and thus appear in Table 7 as data records that do not have a mapping (row 9). ETLHelper then rejected these unmapped data. Similarly, the root cause for the data type casting errors (rows 18–20), which affected “INTEGER” and “DATE” entries, was the utilization of string values for numeric data types, in most cases entries such as “Unknown.” As these could not be parsed as numeric values or timestamps, ETLHelper correctly rejected them.

Discussion and Outlook

Explanation of the Results

ADOPT BBMRI-ERIC was facing the challenge of supporting a cross-intuitional data collection project, the CCDC. From the beginning it was not clear in which format the biobank data would be provided, therefore we defined two simple file formats. We sought to convert these data into a representation that could be imported into a central research database via a generic ETL approach. We successfully supported this process via a semiautomatic matching approach, as implemented in MDRMatcher. It correctly mapped 78.48% of the 1,492 local biobank terms from 10 European biobanks, achieving a precision of 0.89, a recall of 0.87, and an F 1 score of 0.88. With ETLHelper, the actual instance data transformation tool, we successfully processed 147,608 data records from 3,415 patients. These numbers must be read with caution, however. For example, it has to be considered that the biobanks were informed in advance which data elements are collected in the CCDC. We do not know to what extent they had preprocessed their data, but in many cases the designators of the data elements and value sets were largely identical to those in the CCDC. Differences in data preprocessing could also explain the different mapping results for the individual biobanks. Similarly, the achieved data transformation rate of the instance data with ETLHelper is not surprising, as experts curated the mappings in an iterative process. Rather, these figures confirm that the implemented ETL approach succeeded in successfully transforming the biobank data into the OSSE import format.

Comparison with Related Research

Supporting information integration via automation has a well-established background in research. In particular, schema and instance matching techniques have been thoroughly investigated in the past (e.g., see refs. 34–38). Some use advanced background knowledge to improve matching results, such as ontologies. 39 These approaches aim to provide generic frameworks and may therefore require the utilization of additional techniques, such as metadata discovery. 40 41 Similar techniques are used in the field of ontology matching. 42 43 According to Euzenat and Shvaiko, 43 MDRMatcher can be classified as a string-based matcher that incorporates some informal resources, such as the synonyms database. The evaluation results of MDRMatcher are similar to those of other matchers (see, e.g., refs. 44 and 45). However, as already stated above, these numbers must be interpreted with caution. To fully compare the results, we would have to evaluate MDRMatcher with a known data set, as done in, e.g., Kock et al 46 and Achichi et al. 44 In the biobanking field, we are aware of only a few similar works. Pang et al presented two related methods for mapping biobank data. 45 47 Both are based on lexical matching and use synonyms. Compared with our work, the authors start supporting the data collection and harmonization process at an earlier stage by implementing an initial search of the desired research data elements in the source systems. The authors do this by matching from the target terminology to the source terminology (another reason why their metrics are not directly comparable to ours). Adopting this approach could also be useful for the future of BBMRI-ERIC, where larger data sets need to be recognized for future project purposes from biobank databases. However, it must be clarified whether and in which format biobanks could provide their complete metadata. Based on our experiences from the CCDC, this could prove to be very challenging.

Limitations and Future Work

The matching algorithm in MDRMatcher was not modified throughout the execution of the CCDC to generate consistent results. However, certain peculiarities in the behavior of the matching component were identified in our analysis, and these need to be taken into account in future revisions of the program. For example, we will have to extend the program's synonyms database, integrate functionality for down-ranking, and prevent matches across the components of metadata items (hierarchy, concept, and value). As identified in our analysis, simple improvements of the fuzzy matcher (e.g., preventing fuzzy matches across numbers) and deactivating the removal of redundant words should also improve the results further. With regard to missing synonyms, a potential solution could be to use BioPortal ontologies 48 as a source for synonym definitions, as implemented by Pang et al. The analysis of MDRMatcher' results in the “Analysis of Lexical Matching” section is very detailed, but still has its limitations. For example, we did not analyze in detail the impact of synonyms and sentiment tagging—we only noticed afterwards that synonym definitions were missing. If we succeed in integrating further terminological resources as described above, such aspects should also be analyzed in future research. Consideration must also be given to the data transformation part as implemented in ETLHelper. Although our approach was designed to process unstandardized data, it makes sense to investigate how it can efficiently handle data that is already standardized (e.g., ICD-10 or ICD-O-3 codes). Here, one could possibly reuse already existing mappings from ontologies (e.g., UMLS, 49 OxO 50 ) and other mapping resources (such as between SNOMED CT and ICD-10 51 ). This would avoid the lexical matching of already standardized, mapped terms. ETLHelper does not yet support automatic unit conversion. This feature was not required in the CCDC, but support for it would be useful, for example, for supporting laboratory parameters. Because Samply.MDR supports storing units, conversion rules could potentially integrated into ETLHelper. The ETL method presented in this article only supports 1:1, but no n :1 mappings. In other words, there is currently no possibility to merge information from multiple data elements in the source terminology into one data element of the target terminology. This was sufficient for the CCDC, because small calculations could be done directly in the flat files (such as the calculation of the patient age at diagnosis). The definition of machine-interpretable, complex mappings has already been investigated in the past (see, e.g., Mate et al 52 ). The German Biobank Node and German Biobank Alliance projects, 10 23 53 which are working on establishing the German national node for BBMRI-ERIC and whose members are active contributors to the Samply software environment, are currently discussing how the MDR could be extended to store such transformation rules.

Generalizability of Our Method

The approach presented here is not limited to the biobanking domain. The individual steps of the ETL process were implemented using modular software components to enable the logical separation of the individual tasks. Each tool can be used independently from the others. As such, the MIRACUM project 25 is currently investigating whether MDRMatcher can be used to map local laboratory terms to the LOINC standard. 54 As part of this work, MDRMatcher has been extended with multithreading capabilities and a new matching algorithm that considers word frequencies to up-rank rare and down-rank frequent words. The implementation of ETLHelper is still geared toward generating an OSSE XML import file, but modifying the source code allows for generating other output formats as well. ETLHelper uses the EAV format internally to represent the instance data, and as such, the processed data can be transformed into any output format. During the CCDC it became apparent that the creation of metadata in the MDR was a challenging task for the biobanks. We attribute this to the fact that the use of the MDR is initially complicated and time-consuming. Since it was hardly used by the biobanks (and if it was, then often incorrectly), we developed the workaround of extracting the reduced metadata from the flat files. Although this has worked well for the CCDC, it is unlikely to be sufficient in a future context. Due to the increased complexity and heterogeneity of upcoming research scenarios, semantically richer metadata will be required. An important finding is that organizational and administrative aspects must be given even more attention in advance. For BBMRI-ERIC this means that it will have to ensure that biobanks routinely maintain their metadata in the MDR. In addition, as target terminologies will continue to evolve, a workflow will be required to enable biobanks to keep their mappings up to date.

Conclusion

The reuse of biomedical data for research is a challenging task, especially when data from multiple sources are to be merged. These challenges arise, one the one hand, from syntactic and semantic differences in the source data, and on the one hand, from potential data quality issues. We presented an ETL approach for the integration of heterogeneous data that provided good results in the CCDC use case. On the one hand, the lexical matching process in MDRMatcher did most of the work of identifying the correct mappings, with the bag-of-words algorithm achieving a fully automatic, correct mapping almost 80% of the time. On the other hand, the generic approach of data type conversion in ETLHelper proved to be suitable for performing the subsequent data transformations and for detecting errors in the biobank data. The approach has thus provided considerable support to the CCDC of ADOPT BBMRI-ERIC to date. The source code of our tools, including full documentation and a small demo data set with artificial data, is available under the GPL3 license on GitHub: https://github.com/sebmate/ADOPT-BBMRI-ERIC-ETL-Tools .

Clinical Relevance Statement

The collective analysis of patient data and biomaterials is becoming increasingly important. The use of uniform metadata simplifies this process, as the data can be uniformly checked and processed by only a few software components. In this article, we report on a working, largely automated approach for merging heterogeneous biobank data. When collecting the ADOPT CCDC data from the different European biobanks, what was the most common problem associated with the biobanks' data quality? Missing or erroneous data. Missing synonym definitions. Copy and paste errors. Encoding errors. Correct Answer: The correct answer is option c. Shifted tabular data was the most common problem, which we attribute to copying and pasting of data. Sentiment tagging can improve the matching quality by Up-ranking positive and down-ranking negative terms. Up-ranking negative and down-ranking positive terms. Creating matches between different positive and negative terms. Creating matches between different positive or negative terms. Correct Answer: The correct answer is option d. Sentiment tagging “forces” the matcher to create matches only between two positive or between two negative terms. This reduces the likelihood of a match being created between a positive and a negative term.
  26 in total

1.  Organization of heterogeneous scientific data using the EAV/CR representation.

Authors:  P M Nadkarni; L Marenco; R Chen; E Skoufos; G Shepherd; P Miller
Journal:  J Am Med Inform Assoc       Date:  1999 Nov-Dec       Impact factor: 4.497

2.  Safeguarding donors' personal rights and biobank autonomy in biobank networks: the CRIP privacy regime.

Authors:  Christina Schröder; Karsten R Heidtke; Nikolaus Zacherl; Kurt Zatloukal; Jochen Taupitz
Journal:  Cell Tissue Bank       Date:  2010-07-15       Impact factor: 1.522

Review 3.  Methods in biomedical ontology.

Authors:  Alexander C Yu
Journal:  J Biomed Inform       Date:  2005-12-07       Impact factor: 6.317

4.  Beyond synonymy: exploiting the UMLS semantics in mapping vocabularies.

Authors:  O Bodenreider; S J Nelson; W T Hole; H F Chang
Journal:  Proc AMIA Symp       Date:  1998

5.  Exploiting Distributed, Heterogeneous and Sensitive Data Stocks while Maintaining the Owner's Data Sovereignty.

Authors:  M Lablans; D Kadioglu; M Muscholl; F Ückert
Journal:  Methods Inf Med       Date:  2015-07-21       Impact factor: 2.176

Review 6.  Interactions Between Genetics, Lifestyle, and Environmental Factors for Healthcare.

Authors:  Yuxin Lin; Jiajia Chen; Bairong Shen
Journal:  Adv Exp Med Biol       Date:  2017       Impact factor: 2.622

Review 7.  [Requirements for a cross-location biobank IT infrastructure : Survey of stakeholder input on the establishment of a biobank network of the German Biobank Alliance (GBA)].

Authors:  C Schüttler; N Buschhüter; C Döllinger; L Ebert; M Hummel; J Linde; H-U Prokosch; R Proynova; M Lablans
Journal:  Pathologe       Date:  2018-07       Impact factor: 1.011

8.  [Registries for rare diseases : OSSE - An open-source framework for technical implementation].

Authors:  Holger Storf; Jannik Schaaf; Dennis Kadioglu; Jens Göbel; Thomas O F Wagner; Frank Ückert
Journal:  Bundesgesundheitsblatt Gesundheitsforschung Gesundheitsschutz       Date:  2017-05       Impact factor: 1.513

9.  BBMRI-ERIC as a resource for pharmaceutical and life science industries: the development of biobank-based Expert Centres.

Authors:  Gert-Jan B van Ommen; Outi Törnwall; Christian Bréchot; Georges Dagher; Joakim Galli; Kristian Hveem; Ulf Landegren; Claudio Luchinat; Andres Metspalu; Cecilia Nilsson; Ove V Solesvik; Markus Perola; Jan-Eric Litton; Kurt Zatloukal
Journal:  Eur J Hum Genet       Date:  2014-11-19       Impact factor: 4.246

Review 10.  Biobanks in the era of personalized medicine: objectives, challenges, and innovation: Overview.

Authors:  Judita Kinkorová
Journal:  EPMA J       Date:  2016-02-22       Impact factor: 6.543

View more
  4 in total

1.  Sync for Genes: Making Clinical Genomics Available for Precision Medicine at the Point-of-Care.

Authors:  Stephanie J Garcia; Teresa Zayas-Cabán; Robert R Freimuth
Journal:  Appl Clin Inform       Date:  2020-04-22       Impact factor: 2.342

2.  The Future of Biobanking: What Is Next?

Authors:  Luciana Caenazzo; Pamela Tozzo
Journal:  BioTech (Basel)       Date:  2020-11-23

Review 3.  Understanding the Nature of Metadata: Systematic Review.

Authors:  Hannes Ulrich; Ann-Kristin Kock-Schoppenhauer; Noemi Deppenwiese; Robert Gött; Jori Kern; Martin Lablans; Raphael W Majeed; Mark R Stöhr; Jürgen Stausberg; Julian Varghese; Martin Dugas; Josef Ingenerf
Journal:  J Med Internet Res       Date:  2022-01-11       Impact factor: 5.428

Review 4.  Guidelines for Biobanking of Bone Marrow Adipose Tissue and Related Cell Types: Report of the Biobanking Working Group of the International Bone Marrow Adiposity Society.

Authors:  Stephanie Lucas; Michaela Tencerova; Benoit von der Weid; Thomas Levin Andersen; Camille Attané; Friederike Behler-Janbeck; William P Cawthorn; Kaisa K Ivaska; Olaia Naveiras; Izabela Podgorski; Michaela R Reagan; Bram C J van der Eerden
Journal:  Front Endocrinol (Lausanne)       Date:  2021-09-27       Impact factor: 5.555

  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.