Literature DB >> 33573429

Machine learning algorithms trained with pre-hospital acquired history-taking data can accurately differentiate diagnoses in patients with hip complaints.

Michiel Siebelt1, Dirk Das1, Amber Van Den Moosdijk1, Tristan Warren1, Peter Van Der Putten2, Walter Van Der Weegen1.   

Abstract

Background and purpose - Machine learning (ML) techniques are a form of artificial intelligence able to analyze big data. Analyzing the outcome of (digital) questionnaires, ML might recognize different patterns in answers that might relate to different types of pathology. With this study, we investigated the proof-of-principle of ML-based diagnosis in patients with hip complaints using a digital questionnaire and the Kellgren and Lawrence (KL) osteoarthritis score.Patients and methods - 548 patients (> 55 years old) scheduled for consultation of hip complaints were asked to participate in this study and fill in an online questionnaire. Our questionnaire consists of 27 questions related to general history-taking and validated patient-related outcome measures (Oxford Hip Score and a Numeric Rating Scale for pain). 336 fully completed questionnaires were related to their classified diagnosis (either hip osteoarthritis, bursitis or tendinitis, or other pathology). Different AI techniques were used to relate questionnaire outcome and hip diagnoses. Resulting area under the curve (AUC) and classification accuracy (CA) are reported to identify the best scoring AI model. The accuracy of different ML models was compared using questionnaire outcome with and without radiologic KL scores for degree of osteoarthritis.Results - The most accurate ML model for diagnosis of patients with hip complaints was the Random Forest model (AUC 82%, 95% CI 0.78-0.86; CA 69%, CI 0.64-0.74) and most accurate analysis with addition of KL scores was with a Support Vector Machine model (AUC 89%, CI 0.86-0.92; CA 83%, CI 0.79-0.87).Interpretation - Analysis of self-reported online questionnaires related to hip complaints can differentiate between basic hip pathologies. The addition of radiological scores for osteoarthritis further improves these outcomes.

Entities:  

Year:  2021        PMID: 33573429      PMCID: PMC8231380          DOI: 10.1080/17453674.2021.1884408

Source DB:  PubMed          Journal:  Acta Orthop        ISSN: 1745-3674            Impact factor:   3.717


Use of artificial intelligence (AI) techniques like data mining, machine learning (ML), and deep learning are now starting to erupt within healthcare, with first applications aimed at cancer diagnostics (Nguyen et al. 2018, Codari et al. 2019), cardiology (Nirschl et al. 2018) and image recognition in radiology (Wang et al. 2017, Fourcade and Khonsari 2019). AI is also emerging within the field of orthopedic surgery (Duffield et al. 2017). Earlier work using AI in orthopedic studies showed the ability of ML to classify knee osteoarthritis (OA) subjects versus healthy patients. Based on kinematic data Kotti et al. (2017) achieved an accuracy of 73%. In comparison with that study, which collected its data in a laboratory setting, Dolatabadi et al. (2017) used kinematic data from more unobtrusive sensors and were also able to distinguish OA subjects from healthy patients. Other ML-related publications in orthopedics report on spine pathology detection, fracture detection, and bone and cartilage image segmentation (Ashinsky et al. 2015). However, to our knowledge, no studies in orthopedics have developed ML algorithms for predicting a clinical diagnosis. In this paper we used information from digital intake forms, which were completed online by our patients before initial consultation with an orthopedic surgeon. We sought to determine (1) the accuracy of different ML algorithms to predict a pre-hospital diagnosis in patients suffering from hip complaints based on history-taking questions only, and (2) how much radiographic imaging results contribute to accurately predicting a diagnosis in these patients.

Patients and methods

For the development of an ML algorithm we designed a prospective cohort study that included patient data from a single hospital (St Anna Hospital, Geldrop, The Netherlands). All patients aged > 55 years with hip complaints were eligible for inclusion. Immediately after contacting the hospital to schedule an appointment, all participating patients received our questionnaire by e-mail, which had a hyperlink embedded, leading to a secure online environment (Interactive Studios, Rosmalen, the Netherlands). Here, patients were able to answer all questions before initial consultation, which was usually within 1 to 2 weeks. For this purpose we used our online patient reported outcome measurement (PROM) system. This system is normally used to collect standardized PROMs before and after surgery to track orthopedic healthcare outcomes from a patient’s perspective. Within this online environment, 2 authors (MS and WvdW) created a new questionnaire for the purpose of this study, which was verified by a third author (DD). This new questionnaire included standard history-taking questions for suspected hip pathology (i.e., location of pain, severity and duration of symptoms; an overview of all questions is presented in Supplementary data 1—complete questionnaire). These questions were combined with well-validated PROM questionnaires: the Oxford Hip Score (OHS) (Dawson et al. 1996, de Groot et al. 2007), and severity of pain measured with a Numeric Rating Scale (NRS) (Salaffi et al. 2004). Questionnaires of patients who responded to our digital intake form were checked. Incomplete questionnaires were excluded, except for missing answers in the Oxford Hip Score (OHS). As advocated for this specific hip score, a maximum of 2 missing items is allowed and can be dealt with by replacing missing scores with the average score of completed items (Dawson et al. 1996, Murray et al. 2007). After history taking, physical examination and radiographic evaluation, all patients were informed of their diagnosis by their consulting orthopedic surgeon. We retrieved this diagnosis from the medical file and linked it to the questionnaire for that specific patient. This diagnosis was assigned to 1 of 3 categorical outcomes: (1) osteoarthritis (OA) of the hip; (2) bursitis or tendinitis around the hip; or defined as (3) other pathology. These 3 diagnoses were chosen since they represent a large portion of hip complaints. For this proof-of-principle we did not want to start with more diagnoses, since ML techniques will have more trouble differentiating between many possible outcome options and therefore would require larger numbers of patient-reported questionnaires. This dataset was imported into Orange Workflow (version 3.22, Ljubljana, Slovenia), which is an open-source AI software system using different ML techniques. Using Orange, a data file was created to train and test the algorithms in a 10-fold stratified cross-validation loop. The 27 variables were ranked for their ability to differentiate between the 3 diagnosis groups by averaging the outcomes of multiple ranking techniques (Information Gain, Information Gain Ratio, Gini Decrease, X2, ReliefF, and Fast Correlation Based Filter [FCBF]). We trained and tested all ML models available in Orange Workflow (Constant, CN2 rule induces, k Nearest Neighbour [kNN], Tree, Random Forest, Support Vector Machine (SVM), Logistic Regression, Naïve Bayes, AdaBoost, and Neural Network). Orange does not have hyperparameter tuning capabilities so hyperparameters were selected by hand (for an overview of the hyperparameters see Table 1). Resulting area under the curve (AUC) and classification accuracy (CA) outcomes were used to identify the best scoring model (Duffield et al. 2017). 95% confidence intervals (CI) were calculated for AUC and AC with each ML model.
Table 1.

Selected hyperparameters for each evaluated algorithm

AlgorithmHyperparameterValue
SVMEpsilon0.1
  Cost (C)1 
  KernelRBF 
Decision treeMin. number of 
  instances in leaves2 
  Do not split subsets smaller than5 
  Max. tree depth100 
Logistic regressionRegularizationL2
  Cost (C)1 
Neural networkHidden neurons100
  ActivationReLu 
  SolverAdam 
  Alpha0.0001 
KNNNumber of neighbors5
  MetricEuclidean 
  WeightUniform 
Random ForestNumber of trees10
  Number of attributes  
  considered at each split5 
  Max. tree depth3 
  Do not split subsets smaller than5 
Selected hyperparameters for each evaluated algorithm Each model was first trained and evaluated on the questionnaire data with all questions included. Next, we investigated the possibility of achieving similar performance with fewer questions included in the dataset. For this purpose we evaluated the performance of predictive models that were trained on data only including the top 5 ranking questions, and in a second experiment only including the top 10 ranking questions. To analyze the contribution of radiographic imaging results to the diagnosis process, we scored the pelvic radiograph for each included patient using the Kellgren–Lawrence (KL) scoring method (Kellgren and Lawrence 1957) and trained and tested the algorithms again with this score KL added to the full dataset. With inclusion of KL scores, the model was again retested with all questions of the questionnaire, with only the top 5 questions and using only the top 10 questions.

Ethics, registration, funding, and potential conflicts of interest

This study was reviewed by the regional medical ethical committee and was considered to be exempt from full review (registration number N19.066) according to Dutch law. The study protocol was registered in the Dutch Trial Register (Trial registration number NL8229). No external funding was obtained. The authors report no conflicts of interest. Study flow.

Results

Questionnaires of 517 participating patients were received, but after checking for completeness of answers 336 patients could be included in the study (Figure). The collection of this data resulted in a dataset with 283 observations of 27 variables from the questionnaire (see Supplementary material) and 1 target variable (diagnosis group). The distribution of the target variables is as follows. 191 (68%) patients were diagnosed with OA, 61 (22%) patients were diagnosed with bursitis or tendinitis around the hip, and 31 (11%) were diagnosed with other pathology. There is a clear imbalance in the distribution of the target variables with OA being the overrepresented class. The Random Forest algorithm with 20 folds trained on the full dataset (all answers to all questions included) resulted in the highest AUC (82%, CI 0.78–0.86) and CA (69%, CI 0.64–0.74). The 5 most differentiating questions were (in decreasing order of differentiating power): OHS 4: Have you been able to put on a pair of socks, stockings, or tights? Do you experience pain in the groin area? Does your hip feel stiff during the first steps you take when walking? Would you be willing to undergo surgery if needed? OHS 7: Are you able to walk up and down the stairs? All ML models were tested to see if the AUC and/or CA improved by leaving out possibly less important questions. Using only the top 5 questions in the training set, logistic regression (10 folds) gave the highest results of 81% AUC (CI 0.77–0.85) and 73% CA (CI 0.67–0.77). When we selected the top 10 ranking questions, Neural Network (10 folds) gave the highest results with 74% AUC (CI 0.69–0.78) and 67% CA (CI 0.62–0.72) (Table 2). Adding the radiographic data to the dataset increased both the AUC and the CA of the ML models. The distribution of the KL scores accross the 3 diagnosis groups is presented in Table 3. Under this condition, SVM resulted in the highest AUC and CA scores of 89% (CI 0.86–0.92) and 83% (CI 0.79–0.87) respectively (Table 2). A full overview of all ML algorithms is described in Supplementary data.
Table 2.

Artificial intelligence analysis using machine learning (ML) algorithms on pre-hospital-acquired patient history-taking form for patients aged > 55 years with hip complaints. Values are ML algorithm accuracy in percent

 History-taking only
KL score added
   Prediction  Prediction
DatasetAUCCAmodelAUCCAmodel
All questions8269RF8983SVM
Top 5 questions only8273SVM8579SVM
Top 10 questions only7870SVM7979SVM

RF = Random Forest.

SVM = Standard Vector Machine.

Table 3.

Distribution (%) of the KL scores accross the 3 diagnosis groups

 KL score
Diagnosis01234
Bursitis/tendinitis24552010
Osteoarthritis17264818
Other1152111511
Artificial intelligence analysis using machine learning (ML) algorithms on pre-hospital-acquired patient history-taking form for patients aged > 55 years with hip complaints. Values are ML algorithm accuracy in percent RF = Random Forest. SVM = Standard Vector Machine. Distribution (%) of the KL scores accross the 3 diagnosis groups

Discussion

Computer algorithms which use patients’ answers to digital history-taking questions are capable of differentiating a hip complaint related diagnosis with fairly good accuracy (AUC 74% and CA 67%). Adding radiographic information results in even higher accuracy and improves AI performance (AUC 89% and CA 83%). Obviously, there are clear logistical problems that need to be solved in order to achieve integration of conventional radiological examination into pre-hospital AI analysis, but this study shows the proof-of-principle for ML techniques in orthopedics. Our approach using ML may help improve patient care in many ways. With accurate prediction of diagnosis and related treatment, patients can be educated about their condition in advance of their hospital visit, which is easily managed by using a smartphone app. Such an app with supporting information may help patients to increase knowledge and understanding of underlying hip pathology related to their complaint. Subsequently, patients might experience a more in-depth first consultation with the orthopedic surgeon during their hospital visit (Timmers et al. 2018). In order to test this hypothesis, we are currently enrolling a prospective randomized controlled trial using pre-hospital AI diagnosis and its effect on patient knowledge and satisfaction levels during hospital consultation. Besides patient satisfaction, ML diagnosis may also increase outpatient clinic efficiency. First, patients who are more likely to have a diagnosis that is treated nonoperatively can be grouped together when outpatient clinic appointments are scheduled. Other supporting healthcare providers (i.e., physician assistants, physical therapists) can be scheduled to join these consultations and patients’ complaints may be dealt with by a multidisciplinary team. Second, patients who are more likely to be treated with surgery (e.g., hip arthroplasty) can also be grouped together and simultaneously planned for preoperative screening, reducing the number of visits needed to the hospital to a minimum. This predictive analytic study has several limitations. Most importantly, our questionnaire is of course in need of validation in other hospitals. Next, we grouped multiple hip pathologies in 3 categorical groups. This does not cover clinical reality in which orthopedic surgeons make a much more detailed diagnosis. Since this study is a first exploration of ML applied to the clinical diagnostic process in orthopedic surgery and history-taking in particular, we consider our approach justifiable for now. Larger datasets should allow further explorations using more detailed diagnostic outcomes. Furthermore, our resulting accuracy of 82% is high, but could be insufficient in daily clinical practice since it still results in approximately 2 out of 10 patients receiving an incorrect diagnosis. The most important consideration is related to the number of patients receiving a wrong prediction (either a false-positive or a false-negative prediction. However, these computer algorithms should not be considered a substitute for the diagnostic process, but rather an aid to educate patients pre-hospital and organize outpatient clinic logistics. In conclusion, ML algorithms are capable of making a clinical diagnosis for selected patients who suffer from hip complaints using online questionnaires. This first study yields an accuracy of 82% using outcome from our digital questionnaire only, which improved to 89% in combination with radiological osteoarthritis scores. Consultation of patients with complaints using ML techniques can therefore be considered as a valuable tool to aid the orthopedic surgeon in many practical ways, but should not yet be considered as a substitution for human made diagnosis. Click here for additional data file.
  14 in total

1.  The use of the Oxford hip and knee scores.

Authors:  D W Murray; R Fitzpatrick; K Rogers; H Pandit; D J Beard; A J Carr; J Dawson
Journal:  J Bone Joint Surg Br       Date:  2007-08

2.  Questionnaire on the perceptions of patients about total hip replacement.

Authors:  J Dawson; R Fitzpatrick; A Carr; D Murray
Journal:  J Bone Joint Surg Br       Date:  1996-03

3.  Deep learning in medical image analysis: A third eye for doctors.

Authors:  A Fourcade; R H Khonsari
Journal:  J Stomatol Oral Maxillofac Surg       Date:  2019-06-26       Impact factor: 1.569

4.  Artificial Intelligence for Breast MRI in 2008-2018: A Systematic Mapping Review.

Authors:  Marina Codari; Simone Schiaffino; Francesco Sardanelli; Rubina Manuela Trimboli
Journal:  AJR Am J Roentgenol       Date:  2019-01-02       Impact factor: 3.959

5.  Validation of the Dutch version of the Hip disability and Osteoarthritis Outcome Score.

Authors:  I B de Groot; M Reijman; C B Terwee; S M A Bierma-Zeinstra; M Favejee; E M Roos; J A N Verhaar
Journal:  Osteoarthritis Cartilage       Date:  2006-08-04       Impact factor: 6.576

Review 6.  The contribution of musculoskeletal disorders in multimorbidity: Implications for practice and policy.

Authors:  Stephen J Duffield; Benjamin M Ellis; Nicola Goodson; Karen Walker-Bone; Philip G Conaghan; Tom Margham; Tracey Loftis
Journal:  Best Pract Res Clin Rheumatol       Date:  2017-11-01       Impact factor: 4.098

7.  Machine learning classification of OARSI-scored human articular cartilage using magnetic resonance imaging.

Authors:  B G Ashinsky; C E Coletta; M Bouhrara; V A Lukas; J M Boyle; D A Reiter; C P Neu; I G Goldberg; R G Spencer
Journal:  Osteoarthritis Cartilage       Date:  2015-06-09       Impact factor: 6.576

8.  Machine learning applications for the differentiation of primary central nervous system lymphoma from glioblastoma on imaging: a systematic review and meta-analysis.

Authors:  Anthony V Nguyen; Elizabeth E Blears; Evan Ross; Rishi R Lall; Juan Ortega-Barnett
Journal:  Neurosurg Focus       Date:  2018-11-01       Impact factor: 4.047

9.  Assessing the Efficacy of an Educational Smartphone or Tablet App With Subdivided and Interactive Content to Increase Patients' Medical Knowledge: Randomized Controlled Trial.

Authors:  Thomas Timmers; Loes Janssen; Yvette Pronk; Babette C van der Zwaard; Sander Koëter; Dirk van Oostveen; Stefan de Boer; Keetie Kremers; Sebastiaan Rutten; Dirk Das; Rutger Ci van Geenen; Koen Lm Koenraadt; Rob Kusters; Walter van der Weegen
Journal:  JMIR Mhealth Uhealth       Date:  2018-12-21       Impact factor: 4.773

10.  A deep-learning classifier identifies patients with clinical heart failure using whole-slide images of H&E tissue.

Authors:  Jeffrey J Nirschl; Andrew Janowczyk; Eliot G Peyster; Renee Frank; Kenneth B Margulies; Michael D Feldman; Anant Madabhushi
Journal:  PLoS One       Date:  2018-04-03       Impact factor: 3.240

View more
  1 in total

1.  Editorial: One Step at a Time: Advances in Osteoarthritis.

Authors:  Ali Mobasheri; Troy N Trumble; Christopher R Byron
Journal:  Front Vet Sci       Date:  2021-07-16
  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.