| Literature DB >> 27648088 |
Luca Tagliaferri1, György Kovács2, Rosa Autorino1, Ashwini Budrukkar3, Jose Luis Guinot4, Guido Hildebrand5, Bengt Johansson6, Rafael Martìnez Monge7, Jens E Meyer8, Peter Niehoff9, Angeles Rovirosa10, Zoltàn Takàcsi-Nagy11, Nicola Dinapoli1, Vito Lanzotti12, Andrea Damiani13, Tamer Soror2, Vincenzo Valentini1.
Abstract
PURPOSE: Aim of the COBRA (Consortium for Brachytherapy Data Analysis) project is to create a multicenter group (consortium) and a web-based system for standardized data collection.Entities:
Keywords: ENT-COBRA; consortium; data collection; head and neck cancer
Year: 2016 PMID: 27648088 PMCID: PMC5018530 DOI: 10.5114/jcb.2016.61958
Source DB: PubMed Journal: J Contemp Brachytherapy ISSN: 2081-2841
Minimal requirements of each Centre to participate to the COBRA consortium
| In order to participate in the consortium, sign the agreement |
| To have an Electronic Medical Record (EMR) for brachytherapy to record patient's information |
| To be able to ‘translate’ local data into an ontology based archives |
| To be able to anonymize local data |
| To use technology able to developed advanced multicentre researches |
| To provide patient's written informed consents according to local national legislation |
COBRA framework
| The development and validation of multi-factorial prediction models requires the availability of a large amount of data pathology-bounded considered significant for present and futures studies |
| Each variable has to be included into a terminological system; adding more variable in the future is possible, if everything about the data is correctly specified (e.g. denomination, measurement units, measurement modality) |
| Collected data has to be reusable both in time (e.g. in the future) and in the space (across different institutions or research groups); reusability of legacy data is possible, at the condition that suitable semantic remapping functions from old to new data are provided |
| Appropriate mathematical and statistical methods are needed in order to learn from a large collection of data (Large Database) and help to suggest new modelling hypotheses to be tested |
| Patients privacy protection has to be protected; this can be accomplished in two ways: by anonymizing data before they leave the collecting institutions walls, making sure that no inverse remapping is available (“cloud” solution) by exploiting so called “Distributed Learning” solution, in which no data ever leaves the collecting institution but a regressive or classifying predictive model can be learned exactly as if all data had been collected in the same place |
Forms
| 1) Registry and history |
| 2) Histology |
| 3) Staging |
| 4) Protocol |
| 5) Surgery |
| 6) Radiotherapy |
| 7) Neoadjuvant chemotherapy (CT) |
| 8) Concomitant CT |
| 9) Adjuvant CT |
| 10) Brachytherapy |
| 11) Follow-up (repeated) |
| 12) Outcome (automatically calculated based on follow-up) |
| 13) Images and treatment files |
Fig. 1BOA physically separates privacy relevant information from registry level data splitting this two pieces of information into two databases: “Local Patient Index Archive” and “Pathology Archive”. It sends only clinical data to Cloud Large Database, destroying the inverse mapping, HUB extracts and harmonizes legacy data while making them available for BOA, Local Research Proxy makes local queries on its own pathology database, Cloud Research Proxy run queries on the cloud large database and computes outcomes for each consortium member to use
Fig. 2BOA physically separates privacy relevant information from registry level data splitting this two pieces of information into two databases: “Local Patient Index Archive” and “Pathology Archive”. It sends only clinical data to Cloud Large Database, destroying the inverse mapping, HUB (optional module of BOA) extracts and harmonizes legacy data while making them available for BOA, Local Research Proxy (optional module of BOA) makes local queries on its own pathology database. Learning Analyzer Proxy (module of BOA only in distributed mode) sends algorithms directly to Local Research Proxies, taking back from them only the results of each iteration step, with no need to work with shared data in the Cloud anymore. In this mode, Local Research Proxies do not move data around: they only apply iterative algorithms that the Supervisor will use to build consensus and estimate the model's parameters