Literature DB >> 32156739

Predictive models of medication non-adherence risks of patients with T2D based on multiple machine learning algorithms.

Xing-Wei Wu1,2, Heng-Bo Yang3, Rong Yuan4, En-Wu Long5,2, Rong-Sheng Tong5,2.   

Abstract

OBJECTIVE: Medication adherence plays a key role in type 2 diabetes (T2D) care. Identifying patients with high risks of non-compliance helps individualized management, especially for China, where medical resources are relatively insufficient. However, models with good predictive capabilities have not been studied. This study aims to assess multiple machine learning algorithms and screen out a model that can be used to predict patients' non-adherence risks.
METHODS: A real-world registration study was conducted at Sichuan Provincial People's Hospital from 1 April 2018 to 30 March 2019. Data of patients with T2D on demographics, disease and treatment, diet and exercise, mental status, and treatment adherence were obtained by face-to-face questionnaires. The medication possession ratio was used to evaluate patients' medication adherence status. Fourteen machine learning algorithms were applied for modeling, including Bayesian network, Neural Net, support vector machine, and so on, and balanced sampling, data imputation, binning, and methods of feature selection were evaluated by the area under the receiver operating characteristic curve (AUC). We use two-way cross-validation to ensure the accuracy of model evaluation, and we performed a posteriori test on the sample size based on the trend of AUC as the sample size increase.
RESULTS: A total of 401 patients out of 630 candidates were investigated, of which 85 were evaluated as poor adherence (21.20%). A total of 16 variables were selected as potential variables for modeling, and 300 models were built based on 30 machine learning algorithms. Among these algorithms, the AUC of the best capable one was 0.866±0.082. Imputing, oversampling and larger sample size will help improve predictive ability.
CONCLUSIONS: An accurate and sensitive adherence prediction model based on real-world registration data was established after evaluating data filling, balanced sampling, and so on, which may provide a technical tool for individualized diabetes care. © Author(s) (or their employer(s)) 2020. Re-use permitted under CC BY. Published by BMJ.

Entities:  

Keywords:  adherence; personality; prediction and prevention; type 2 diabetes

Year:  2020        PMID: 32156739      PMCID: PMC7064141          DOI: 10.1136/bmjdrc-2019-001055

Source DB:  PubMed          Journal:  BMJ Open Diabetes Res Care        ISSN: 2052-4897


Medication adherence is a key to diabetes care. Proper care can help improve medication adherence. The newly developed model has a good performance to predict the risk of non-medication adherence in diabetics. This model can be used to filter patients with high risk of non-medication adherence and precise educational interventions can be conducted.

Introduction

Type 2 diabetes (T2D), one of the most popular disorders all over the world,1 is a chronic condition that requires long-term treatment to improve life quality and reduce the probability of related disability and death.2–4 Studies have confirmed that good medication adherence is important for patients with T2D to improve glycemic control, avoid complications, and reduce overall health expenditure.5 6 Interventions such as integrative health coaching can significantly improve medication adherence.7–9 Considering the large crowd and complex features of patients with T2D, identifying high-risk groups with poor medication adherence and carrying out precise educational interventions are feasible measures to improve the overall disease control. Studies found that numerous factors may be associated with medication adherence in patients with T2D, including the duration of the disease, mental state, anxiety, depression, irritability, smoking, cost, and so on.5 10 11 These factors vary with study designs and regions. That may be resulted from the differences in the selection of variables, patients’ income levels, education levels, cultural characteristics, living habits, and so on. Therefore, it is necessary to explore the main influencing factors of local patients’ medication compliance and to establish a model that can accurately predict therapeutic compliance in specific regions.12 13 Despite several compliance-related predictive models that have been reported recently in patients with tuberculosis14 and heart failure,15 studies on patients with T2D based on multivariate machine learning algorithms have not been retrieved. In this study, a rigorous and comprehensive questionnaire survey was conducted and dozens of machine learning-based models were built and assessed. In addition, we also examined the impacts of data preprocessing, modeling methods and different sample sizes on prediction performance.

Research design and methods

Study design, study area, and participants

A survey was conducted in the outpatient clinic of Sichuan Provincial People’s Hospital from April 2018 to March 2019. The hospital mainly serves patients from Sichuan Province, a populous province in southwestern China. A total of 630 patients were approached, and 401 completed the survey. Researchers conducted a face-to-face questionnaire survey and filled out questionnaires according to the answers from the patients who participated in the survey. Patients with T2D were selected according to the criteria: (1) examined HbA1c on the day of the questionnaire, and (2) willing to take part in the survey and to provide information to the investigators. Patients were eliminated if he or she (1) was a non-T2D patient, (2) did not receive hypoglycemic agency treatment, (3) less than 18 years old, and (4) with a short life expectancy.

Data collection

The data in this study were collected from electronic medical records (EMR) and face-to-face questionnaires. Clinical laboratory results were collected according to EMRs. The results of the previous examination and the duration from the last test were recalled by patients and, if necessary, confirmed by telephone after they returned home.

Questionnaires

The questionnaire consists of five parts. The first part is about basic characteristics, including age, gender, occupation, family history, and so on. The second part is involved in the information related to diabetes, including self-glycemic monitoring, diabetes course, medication regimen, duration of medication regimen, the test results of HbA1c and blood glucose at the time of the questionnaire. The third part is referred to the other clinical information, including the use of Chinese traditional medicine products, surgical histories, comorbidities, and concomitant medications. The fourth part is about exercise, diet, and mental state. The last part is information related to adherence, in which we recorded how many medications should be taken, how many were prescribed and how many were left. The adherence status, the medication possession ratio (the proportion of medication’s available days, higher than 80% is regarded as good medication compliance16 17), was determined as the target variable.

Data blindness

All variable names were encoded to X1–X44 to achieve statistical blindness in this study. Data analysis was performed using the encoded variable names. The variables were unblinded after the model evaluation processes.

Data cleaning

Variables with more than 50% missing or a certain category’s proportion greater than 80% were excluded. The maximum likelihood ratio method was used to assess the correlations between the input variables and the target variable. Variables with p value >0.1 were considered unimportant and were excluded after the likelihood ratio test. Data filling was performed using mean value (for numerical variables), median (for ordinal variables) or mode (for nominal variables).18 Outlier values were modified as the maximum or minimum of normalized values. Because of the proportional imbalance between good and non-adherence patients, sampling methods, including oversampling (four times for patients with poor compliance) and undersampling (50% for patients with good adherence), were taken to make up the shortage caused by the imbalanced sample size between the different levels of the target variable. Unbalanced data were analyzed simultaneously to evaluate the risk of overfitting on account of balancing dispose.

Data partition

There were two data partitioning processes in this study for two-way cross-validation. The first was conducted before data cleaning. In this process, the original data were randomly divided into two subsets (named set 1 and set 2) in a ratio of 8:2, which would be used for model establishment and verification, respectively. During the second partition, set 1 was divided into a training set and a testing set by 7:3 after data cleaning. The training set would be used to build machine learning models, and the testing set would be used to evaluate the fitting effectiveness of the models.

Variable selection

The forward, backward and stepwise methods of logistic regression were used to screen significant variables. The continuous variables were grouped (or not) according to the IQR, in which process was named binning, aimed to facilitate interpretation.

Model validation

The predictive performances of the models were evaluated by the area under the receiver operating characteristic curve (AUC). AUC of every data set (training set of set 1, testing set of set 1, and set 2) was calculated to examine the sensitivity and specificity of models. Due to oversampling dispose, a number of duplicate records would be generated, which means some records might be used both to build the models and to validate the models, and cause the risk of overly optimistic estimation of the models’ prediction effects. So, AUC of set 2, which was partitioned before oversampling, would be used as the best model evaluation indicator. The overfitting risks of the models were assessed by OF1, which was calculated using the formula: AUCSet 1 training set/AUCSet 1 testing set, and OF2, AUCSet 1 testing set/AUCSet 2. Higher levels of OF1 and OF2 indicate more serious overfitting risks. Tenfold independent repeated values of the above indexes were generated by changing the random seed number in the first data partition process 10 times.

Model building

More than a dozen classification algorithms were applied and assessed in this study. The proposed models included C 5.0 model (marked as $C), logistic regression model (marked as $L), decision list, Bayesian network (marked as $B), discriminant model (marked as $D), KNN algorithm (marked as $KNN), LSVM, random forest, SVM (marked as $S), Tree-AS, CHAID (marked as $R), Quest, C&R Tree (marked as $R), Neural Net (marked as $N), and the ensemble model (marked as $XF). The ensemble models would summarize the output of the best five models (assessed by AUC) and generate their outputs according to the voting principle.

Sample size assessment

The model with the largest AUC was selected, and 10 subsets, 10%–100% in a step of 10% of the total sample size randomly extracted, were used to build models in order to evaluate the influence of different sample sizes on the predictive ability. Every subset was divided into a training set and a testing set by 7:3 and the AUC calculated from the testing set was used for sample size examination. Ten independent replicate results were generated for each model by transforming random sampling seeds.

Statistical description and hypothesis testing

Continuous variables were expressed as mean±SD and counting variables were expressed in terms of frequency. Student’s t-test and signed-rank tests were used to test the difference between paired quantitative data. T-test and general linear models for analysis of variance were used as parameter test approaches. Non-parametric tests were implemented in Wilcoxon rank-sum analysis and Kruskal-Wallis test. If the data were normally distributed and the variances were equal, an appropriate parameter test was used, otherwise, a non-parametric test was used to realize hypothesis testing. Spearman correlation analysis was used as correlation analysis approach between two continuous variables. Multiple linear regression was used for multivariate analysis, and during which process variance inflation factor and standardized estimate (SE) were used to assess the multicollinearity risk and the weight of the affecting to the dependent variable. Excel 2016 was used to summarize the data. Data cleaning and modeling were completed using IBM SPSS Modeler V.18.0 software. Variable screening, hypothesis testing, and regression analysis were performed using SAS V.9.21 (SAS Institute). The figure in the sample size verification section was plotted using Prism for Windows 6 software (GraphPad Software).

Results

Respondent population

A total of 401 patients out of 630 candidates completed the survey, among which 244 were male and 157 were female. The mean age was 58.9±12.2 years. Eighty-five patients were defined as poor medication adherence (21.20%). For detailed patients’ characteristics see table 1. The rate of data missing was 10.67% and the main reasons were the missing of patients’ memory or that the respondents were not willing to provide certain information.
Table 1

Demographic and clinical data of participants

ParameterValue (n=401)
Age (years)n401
Mean±SD58.9±11.89
Median58
Minimum, maximum27, 85
Gendern401
Male244 (60.8%)
Female157 (39.1%)
Weight (kg)n397
Mean±SD65.0±10.28
Median65
Minimum, maximum42, 110
Marital statusn396
Married/living as married/civil partnership393 (99.2%)
Single/never married2 (0.4%)
Divorced or separated2 (0.4%)
Employment statusn399
Unemployed58 (14.5%)
Employed149 (37.3%)
Retirement191 (47.9%)
Other1 (0.2%)
Highest level of educationn399
Illiteracy41 (10.3%)
Junior middle school128 (32.1%)
High school or special secondary school130 (32.6%)
College and above100 (25.1%)
Family history of diabetesn391
Yes122 (31.2%)
No269 (68.8%)
Previous HbA1c valuen264
<7%97 (36.7%)
7%–9%126 (47.7%)
>9%41 (15.5%)
The interval between the last HbA1c measurement and the present (days)n267
Mean±SD227.9±271.52
Median150
Minimum, maximum2, 2920
The course of diabetes (months)n401
Mean±SD89.7±76.44
Median72
Minimum, maximum1, 480
Regular monitoring of fasting blood glucose frequencyn401
Irregular monitoring71 (17.7%)
Two or three times a week156 (38.9%)
Three or four times a month129 (32.1%)
Two or three times in 3 months45 (11.2%)
Fasting blood glucose value (mmol/L)n325
3.8–6.123 (7.1%)
6.1–796 (29.5%)
≥7206 (63.3%)
Complicationsn401
Yes42 (10.5%)
No359 (89.5%)
Exercise intensityn401
No exercise44 (10.9%)
Low-intensity exercise (eg, walking)269 (67.0%)
Medium-intensity exercise (eg, fast walking, jogging)57 (14.2%)
High-intensity exercise (eg, fitness, cycling, dancing)31 (7.7%)
Exercise time (min)n401
Mean±SD63.3±71.66
Median60
Minimum, maximum0, 600
Eat reasonablyn401
Yes294 (73.3%)
No107 (26.7%)
Sleep statusn401
Good214 (53.3%)
Ordinary120 (29.9%)
Lose sleep67 (16.7%)
Psychological statusn401
Optimistic247 (61.6%)
Ordinary144 (35.9%)
Depressed10 (2.5%)
Compliancen401
Good316 (78.8%)
Poor85 (21.2%)

n, number of respondents.

Demographic and clinical data of participants n, number of respondents.

Data stream

A process framework of the data flow is shown in figure 1. Data flow through each node according to a predetermined schedule.
Figure 1

The data flowed into the ‘Partition’ node after feature selection. The ‘Auto Data Prep’ node was used for data filling, the ‘Balance’ node performed a data balanced sampling process, and the ‘Binning’ node was applied for data binning. The ‘Partition’ node divided set 1 into a training set and a testing set, used the ‘Auto Classifier’ node to build various classification models, and used the ‘Analysis’ and ‘Evaluation’ nodes to output the AUC values and curve figure of each model. Use the ‘Select’ node to select the ‘Set 2’ data set. The set 2 set was concatenated with the models established above and the ensemble model of them, and the AUC values and graphs of all models were output using the ‘Analysis’ and ‘Evaluation’ nodes. AUC, area under the receiver operating characteristic curve.

The data flowed into the ‘Partition’ node after feature selection. The ‘Auto Data Prep’ node was used for data filling, the ‘Balance’ node performed a data balanced sampling process, and the ‘Binning’ node was applied for data binning. The ‘Partition’ node divided set 1 into a training set and a testing set, used the ‘Auto Classifier’ node to build various classification models, and used the ‘Analysis’ and ‘Evaluation’ nodes to output the AUC values and curve figure of each model. Use the ‘Select’ node to select the ‘Set 2’ data set. The set 2 set was concatenated with the models established above and the ensemble model of them, and the AUC values and graphs of all models were output using the ‘Analysis’ and ‘Evaluation’ nodes. AUC, area under the receiver operating characteristic curve.

Feature selection

Three variables were excluded due to too many missing values. Four variables were excluded because the proportion of certain categories was too large, and 18 variables were excluded because of the low correlation with the target variable. Thereby, a total of 16 variables were used for modeling, including the last HbA1c value, fasting glucose, age, diet adjustment or not, weight, cost of hypoglycemic drugs, duration of current treatment regimen, body mass index (BMI), working status, the duration since the prior blood glucose test, dyslipidemia, and so on.

Model establishment and evaluation

In this study, 300 machine learning models were developed based on 30 algorithms varying from whether imputing or not, sampling methods (unbalanced, oversampling and undersampling) and variable screening methods (forward, backward and stepwise). By changing the seed value when data partition, each modeling algorithm would be built and tested using 10 separate training and testing data sets, and generate 10 models. Consequently, we got 10 independent duplicates for each modeling algorithm. AUC and overfitting values for 30 machine learning modeling algorithms were shown in table 2. Among 30 modeling algorithms, the minimum and maximum of AUC in set 2 were 0.557 (SD 0.051) and 0.866 (SD 0.082), respectively. The best algorithm was an ensemble one from five models that used oversampling for data balance after data imputing, and without data binning (table 3). Nine variables were used to build this model, which were age, gender, whether the prior fasting blood glucose was under control, duration of the current treatment regimen, diet adjustment or not, the daily cost of medications, fasting blood glucose value, hyperlipidemia and BMI after the backward variable selection process.
Table 2

AUC and overfitting values for 30 machine learning models

Imputing or notSamplingmethodsBinning or notScreening methodsModel methodsNum of VariablesNum of SamplesAUC_TRAUC_TEAUC_Set 2OF1OF2
NotNotNotNot$D161670.693±0.0250.621±0.0540.737±0.0621.125±0.1220.849±0.111
NotNotNotForward and stepwise$S31670.551±0.0330.577±0.0780.557±0.0510.974±0.1621.039±0.131
NotNotNotBackward$L51670.659±0.0190.658±0.0390.687±0.0521.005±0.0720.964±0.102
NotNotYesForward and stepwise$S31670.551±0.0330.577±0.0780.577±0.0510.974±0.1621.039±0.131
NotNotYesBackward$S31670.551±0.0330.577±0.0780.577±0.0510.974±0.1621.039±0.131
NotUndersamplingNotNot$XF16980.778±0.0250.827±0.0800.744±0.0850.949±0.1021.130±0.198
NotUndersamplingNotForward and stepwise$L3980.679±0.0320.660±0.0750.664±0.0671.044±0.1421.008±0.192
NotUndersamplingNotBackward$L3980.679±0.0320.660±0.0750.664±0.0671.044±0.1421.008±0.192
NotUndersamplingYesForward and stepwise$KNN4980.725±0.0280.674±0.0740.715±0.0701.086±0.1220.955±0.169
NotUndersamplingYesBackward$XF5980.755±0.0470.753±0.0740.725±0.0581.013±0.1361.044±0.125
NotOversamplingNotNot$R162630.781±0.0210.770±0.0400.758±0.0701.017±0.0601.028±0.150
NotOversamplingNotBackward$XF52630.814±0.0310.799±0.0370.761±0.0941.021±0.0731.066±0.155
NotOversamplingNotForward and stepwise$XF42630.716±0.0200.726±0.0260.690±0.0680.987±0.0481.064±0.133
NotOversamplingYesForward and stepwise$XF42630.834±0.0300.821±0.0310.782±0.0801.018±0.0681.060±0.112
NotOversamplingYesBackward$XF72630.864±0.0280.856±0.0220.813±0.1271.010±0.0401.088±0.261
YesNotNotNot$D163150.725±0.0120.678±0.0470.703±0.0511.074±0.0870.973±0.131
YesNotYesForward and stepwise$XF53150.812±0.0240.760±0.0480.757±0.0561.073±0.0891.008±0.091
YesNotNotForward and stepwise$XF43150.752±0.0200.701±0.0700.711±0.0551.084±0.1310.994±0.144
YesNotNotBackward$XF63150.742±0.0190.734±0.0630.734±0.0661.017±0.0941.012±0.160
YesNotYesBackward$B63150.729±0.0190.718±0.0990.714±0.1001.034±0.1511.034±0.266
YesUndersamplingNotNot$B161990.785±0.0320.811±0.0870.778±0.0630.980±0.1261.052±0.170
YesUndersamplingYesForward and stepwise$XF41990.701±0.0270.665±0.0740.722±0.0501.067±0.1350.927±0.130
YesUndersamplingNotForward and stepwise$S41990.685±0.0220.658±0.0690.702±0.0531.053±0.1370.946±0.146
YesUndersamplingYesBackward$S51990.699±0.0150.754±0.0830.733±0.0520.938±0.1131.034±0.143
YesUndersamplingNotBackward$KNN51990.740±0.0290.738±0.0820.736±0.0651.017±0.1431.013±0.165
YesOversamplingNotNot$XF165130.916±0.0300.869±0.0410.862±0.1231.056±0.0521.039±0.243
YesOversamplingYesForward and stepwise$B75130.857±0.0230.824±0.0390.849±0.0721.042±0.0520.978±0.118
YesOversamplingNotForward and stepwise$XF85130.907±0.0310.861±0.0390.843±0.1151.054±0.0491.049±0.230
YesOversamplingNotBackward$XF95130.907±0.0240.871±0.0300.866±0.0821.041±0.0361.017±0.134
YesOversamplingYesBackward$B95130.865±0.0320.823±0.0500.839±0.1071.054±0.0701.003±0.191

OF1 was calculated using the formula: AUCSet 1 training set /AUCSet 1 testing set, and OF2, AUCSet 1 testing set /AUCSet 2.

The bold value was the maximum AUCSet 2 of 30 algorithms.

AUC, area under the receiver operating characteristic curve; AUC_Set 2, AUC of set 2; AUC_TE, AUC of set 1 testing set; AUC_TR, AUC of set 1 training set; $B, Bayesian network; $D, discriminant model; $KNN, KNN algorithm; $L, logistic regression model; $R, CHAID; $S, SVM; $XF, the ensemble model.

Table 3

Assessment of models by different algorithms of the selected data governance

ModelsAUCPrecisionRecallF1 score (Online supplementary file 1)
Bayesian network0.764±0.0290.729±0.0550.717±0.0420.721±0.025
KNN0.838±0.0180.813±0.0460.636±0.0620.712±0.039
SVM0.765±0.0440.728±0.0590.589±0.0780.647±0.046
C&R Tree0.755±0.0300.739±0.0260.669±0.0660.700±0.033
CHAID0.770±0.0410.792±0.0350.655±0.0490.716±0.031
Ensemble0.866±0.0820.824±0.0430.732±0.0610.773±0.032
P valuep<0.0001p<0.0001p<0.0001p<0.0001

The bold value indicate the performance parameters of the best algorithm.

AUC, area under the receiver operating characteristic curve.

AUC and overfitting values for 30 machine learning models OF1 was calculated using the formula: AUCSet 1 training set /AUCSet 1 testing set, and OF2, AUCSet 1 testing set /AUCSet 2. The bold value was the maximum AUCSet 2 of 30 algorithms. AUC, area under the receiver operating characteristic curve; AUC_Set 2, AUC of set 2; AUC_TE, AUC of set 1 testing set; AUC_TR, AUC of set 1 training set; $B, Bayesian network; $D, discriminant model; $KNN, KNN algorithm; $L, logistic regression model; $R, CHAID; $S, SVM; $XF, the ensemble model. Assessment of models by different algorithms of the selected data governance The bold value indicate the performance parameters of the best algorithm. AUC, area under the receiver operating characteristic curve.

Overfitting risk of the models

OF1 and OF2 were used to evaluate the overfitting risk of the model. The overall OF1 value was 1.028±0.113, and its difference from 1 was statistically significant (Student’s t-test, p<0.0001). The OF2 value was 1.015±0.165, and no significant difference was observed between OF2 and 1 (signed-rank test, p=0.6268). Despite the statistical differences between OF1 and 1, we believe that the overfitting risks of all models were negligible because their mean values were very close to 1, and their differences between 1 were Gaussian distributed and SD values were small (which means few variances).

Impact of modeling methods on predictive performance

The impacts of algorithms on predictive performance are shown in table 4. Data imputing can significantly improve the AUC of set 2 (imputing, 0.770±0.096 vs not imputing, 0.694±0.106). Oversampling and undersampling will help get a better AUC (oversampling, 0.806±0.107 and undersampling, 0.718±0.069 vs not sampling, 0.671±0.097). As the number of variables and samples increases, the prediction performances of the models were significantly improved. In addition, different predictive powers were shown among various methods.
Table 4

The impact of modeling approaches on predictive indicators

ApproachesAUC_TRAUC_TEAUC_Set 2OF1OF2
Univariate analysisMultivariate analysis*Univariate analysisMultivariate analysisUnivariate analysisMultivariate analysisUnivariate analysisMultivariate analysisUnivariate analysisMultivariate analysis
P valueMMD/RP valueSEP valueMMD/RP valueSEP valueMMD/RP valueSEP valueMMD/RP valueSEP valueMMD/RP valueSE
Imputing or not<0.00010.07930.00190.3020<0.00010.06050.0813−0.2082<0.00010.07590.1719−0.16870.03940.02310.6812−0.06350.1084†0.02030.6829−0.0634
Sampling methods<0.00010.16950.01410.2757<0.00010.16200.2980−0.1436<0.00010.13490.0840−0.24710.7884†0.01450.3667−0.16150.5764†0.04400.66170.0786
Binning or not0.6441†0.00530.7135−0.01490.7837†0.00090.8834−0.00730.8258†0.00280.75780.01600.9188†0.00660.9212−0.00640.6672†0.00360.7668−0.0193
Screening methods0.01190.04890.03380.11020.00420.05410.00240.19500.00910.05130.03520.13940.2277‡0.02420.1343−0.12420.6512†0.02130.44840.0631
Model methods<0.00010.2015<0.00010.2734<0.00010.1654<0.00010.2864<0.00010.1739<0.00010.20250.01430.11660.5617−0.03860.02710.12740.16160.0936
Num of Variables<0.0001§0.6121<0.00010.2905<0.0001§0.5197<0.00010.3360<0.0001§0.5147<0.00010.31340.2593§0.06530.4035−0.07390.2219§−0.07070.74250.0292
Num of Samples¶<0.0001§0.6716<0.00010.9020<0.0001§0.50830.00220.5584<0.0001§0.4949<0.00010.36550.0722§0.10390.19720.30240.5946§−0.03080.8326−0.0497

Oversampling leading to abnormal distribution in all five indexes.

The bold values indicate the parameters of approaches which would significantly affect predictive indicators.

*Multiple linear regression was used for multivariate analysis.

†Kruskal-Wallis test.

‡General linear models for analysis of variance (ANOVA).

§Spearman correlation analysis.

¶The variance inflation factor (VIF) of variable ‘Num of Samples’ in multiregression model is 16.4146 (which is greater than 10), indicates multicollinearity that maybe exists and may make the model unstable; this variable may be severely collinear with imputing, binning, and sampling, so the multiple linear regression (MLR) model was re-established after the three variables were eliminated.

AUC, area under the receiver operating characteristic curve; AUC_Set 2, AUC of set 2; AUC_TE, AUC of set 1 testing set; AUC_TR, AUC of set 1 training set; MMD, maximum mean difference among levels; R, correlation coefficient; SE, standardized estimate.

The impact of modeling approaches on predictive indicators Oversampling leading to abnormal distribution in all five indexes. The bold values indicate the parameters of approaches which would significantly affect predictive indicators. *Multiple linear regression was used for multivariate analysis. †Kruskal-Wallis test. ‡General linear models for analysis of variance (ANOVA). §Spearman correlation analysis. ¶The variance inflation factor (VIF) of variable ‘Num of Samples’ in multiregression model is 16.4146 (which is greater than 10), indicates multicollinearity that maybe exists and may make the model unstable; this variable may be severely collinear with imputing, binning, and sampling, so the multiple linear regression (MLR) model was re-established after the three variables were eliminated. AUC, area under the receiver operating characteristic curve; AUC_Set 2, AUC of set 2; AUC_TE, AUC of set 1 testing set; AUC_TR, AUC of set 1 training set; MMD, maximum mean difference among levels; R, correlation coefficient; SE, standardized estimate. Multilinear regression analysis showed that variable screening methods, algorithms, number of variables and sample size will affect the AUC of set 2 remarkably. The factor ‘sample size’ had the strongest impact on AUC of set 2 according to SE value (see table 3). As the size of sample data incorporated into the model from small to large, the AUC values of the testing sets continued to increase. When the sample size is extremely small (less than or equal to 30%), the SDs of AUC are large, and the values show significant discrete distribution trends. As the sample size increases, the above situation is alleviated.

Discussion

The development of patient’s medication adherence predictive models facilitates individualized diabetes education and care.19 20 In our research, 30 machine learning-based modeling algorithms were developed for adherence prediction, which at most 16 variables were used to build the model. After examining data filling, sampling, binning, and variable screening methods, the best algorithm, whose mean AUC was 0.866±0.082, was selected. The model was validated by sensitivity, specificity, and overfitting risk, and excellent predictive capability was shown to identify non-adherence patients. Although studies related to medication adherence predictive models are found in tuberculosis14 and heart failure,15 few studies have been retrieved that established and examined the predictive models in patients with type 2 diabetes mellitus (T2DM), for whom medication adherence is a key factor to treatment outcomes.5 21 22 Kumamaru H et al23 established a logistic regression model for the prediction of future drug compliance. In their study, the patient’s current drug dependence characteristic played the role of the main predictor. Data from 90 000 patients were analyzed in the study. However, the AUC of their model was 0.666 (95% CI 0.661 to 0.671), which was smaller than ours (0.866±0.082). That may be because they focused on patients newly initiating statins, used retrospective data and evaluated fewer variables while we focused on a smaller population (patients with T2DM), conducted a prospective study and examined more targeted variables. The proportion of patients with good adherence in this study is 78.8%, which is comparable to data from Ethiopia24 (70.5%), the USA25 (70.0%) and white patients who were newly prescribed diabetes medications26 (62.5%). A systematic review showed that the rate of adherence to oral hypoglycemic agents ranged from 36%–93% and 62%–64% to insulin in patients with T2DM.27 Twenty-seven percent of Palestine refugees who have diabetes living in Jordan were not adherent to their regimens.28 Many variables are reportedly considered to be related to patient medication compliance such as gender, dyslipidemia, medication, and so on.29 30 To our knowledge, this study is the first to examine the factors and develop predictive models based on the factors such as the duration of the current medication regimen, whether the last fasting blood glucose was under control, and whether a reasonable diet plan had been implemented. Data filling is an important way to solve data loss problems that frequently appear in real-world studies. Every record needs a value for each variable during the multivariate analysis process in spite of missing data are inevitable in real-world data sets. Deleting the entire patient’s record due to small parts of data missing will result in a large amount of information loss. It could be necessary to delete the variables and records under severe missing circumstances (eg, more than 70% of data missing) to reduce noise in the whole data. However, there is still no consensus about whether the absence of the remaining data should be filled or how to perform effective data padding. Several studies have reported interpolation methods under the situation of data missing.31–33 Specially, multiple imputation (MI) methods are a wonderful approach to handle data missing issues for the consideration of the imputation uncertainty.34 MI generates m data sets, obtains results from each data set and combines m results at the last step. In this study, 300 models were built by using one data set and one algorithm was selected for use. We have not figured out how to combine m×300 models and find one algorithm out. So we used simple imputation here. The results of our study showed that the predictive abilities of moles were significantly improved (p<0.0001) after data filling. Univariate and multivariate analyses also showed that the increase in the available sample size would significantly enhance the models’ predictive performances which maybe also partly due to data imputation. Undersampling and oversampling are the main methods used to solve the issues that result from sample imbalance which also frequently happen in medical-related research data,35 such as the occurrence of adverse reactions and non-adverse reactions.36 Undersampling reduces the number of samples at the level of more samples while oversampling increases the number of samples at the level of fewer samples. Overestimation of AUC may occur on account of oversampling because of the existence of a large number of duplicate samples and some records are used to test the model after being used to train the model. Therefore, in this study, the data partitioning process was designed twice to produce a non-repeating set of verification samples, which is set 2, in order to ensure the credibility of the results. In addition, this study shows that oversampling is a key measure to improve predictive sensitivity and specificity (p<0.0001).

Strengths and limitations

There are several innovations in this study. First, we are more focused on collecting and evaluating variables that may be relevant to medication compliance in the real-world environment, which may increase the prediction accuracy of the model based on a limited sample size. Second, the model with the best prediction effect was found based on as many as 300 models built according to a variety of data processing methods and ten times randomly partitioning of data. Finally, we have innovatively tried a sample size posterior verification method based on the relative relationship between sample size and AUC, which can provide a reference for sample size analysis of predictive studies. However, limitations also exist in this study. First, in the sample size verification results, there is no inflection point in the AUC curve as the sample size increases, indicating that there is still a need for a further increase in the sample size. The second one is that we need to work with more centers to collect more data to optimize the model and verify the applicability of the model in healthcare facilities. Lastly, although we classified the patients’ last test results to reduce accuracy, for some variables, recall bias still exists.

Conclusion

In Sichuan Province, the southwest region of China, 21.20% of patients with T2D were not adherent to their medication regimens. The duration of the current medication regimen, former blood glucose level, and eating habits can be key factors in predicting patient medication compliance. By performing data padding when the data have a small number of missing values, balanced sampling in the imbalanced data, and model training using a variety of algorithms, the better predictive models will be obtained. After establishing and evaluating a large number of models, we got a predictive model with potential clinical value for preventing non-adherence risks of patients with T2D.
  35 in total

Review 1.  Impact of improved glycemic control on quality of life in patients with diabetes.

Authors:  Alan M Jacobson
Journal:  Endocr Pract       Date:  2004 Nov-Dec       Impact factor: 3.443

2.  The development of an integrated behavioural model of patient compliance with diabetes medication: a mixed-method study protocol.

Authors:  Yogarabindranath Swarna Nantha; Shamsul Haque; Hemanath Swarna Nantha
Journal:  Fam Pract       Date:  2019-10-08       Impact factor: 2.267

3.  Using Machine Learning for Personalized Patient Adherence Level Determination.

Authors:  Maksim Taranik; Georgy Kopanitsa
Journal:  Stud Health Technol Inform       Date:  2019

4.  Comparison of medication adherence and persistence in type 2 diabetes: A systematic review and meta-analysis.

Authors:  Andrew McGovern; Zayd Tippu; William Hinton; Neil Munro; Martin Whyte; Simon de Lusignan
Journal:  Diabetes Obes Metab       Date:  2017-12-12       Impact factor: 6.577

5.  Multiple imputation for time series data with Amelia package.

Authors:  Zhongheng Zhang
Journal:  Ann Transl Med       Date:  2016-02

6.  Medication adherence by Palestine refugees living in Jordan who have diabetes: a cross-sectional study.

Authors:  Giulia Canali; Victoria Tittle; Akihiro Seita
Journal:  Lancet       Date:  2018-02-21       Impact factor: 79.321

Review 7.  Synthesizing and Appraising the Quality of the Evidence on Factors Associated with Medication Adherence in Diabetes: A Systematic Review of Systematic Reviews.

Authors:  Myriam Jaam; Ahmed Awaisu; Mohamed Izham Ibrahim; Nadir Kheir
Journal:  Value Health Reg Issues       Date:  2017-09-28

8.  Integrative health coaching: a behavior skills approach that improves HbA1c and pharmacy claims-derived medication adherence.

Authors:  Ruth Q Wolever; Mark H Dreusicke
Journal:  BMJ Open Diabetes Res Care       Date:  2016-05-09

9.  Medication adherence and glycemic control among newly diagnosed diabetes patients.

Authors:  Lee-Kai Lin; Yan Sun; Bee Hoon Heng; Daniel Ek Kwang Chew; Phui-Nah Chong
Journal:  BMJ Open Diabetes Res Care       Date:  2017-07-31

10.  Associations between adherence and outcomes among older, type 2 diabetes patients: evidence from a Medicare Supplemental database.

Authors:  Kristina Secnik Boye; Sarah E Curtis; Maureen J Lage; Luis-Emilio Garcia-Perez
Journal:  Patient Prefer Adherence       Date:  2016-08-16       Impact factor: 2.711

View more
  3 in total

1.  Integrated Digital Health Solutions in the Management of Growth Disorders in Pediatric Patients Receiving Growth Hormone Therapy: A Retrospective Analysis.

Authors:  Vincenzo Tornincasa; David Dixon; Quentin Le Masne; Blaine Martin; Lilian Arnaud; Paula van Dommelen; Ekaterina Koledova
Journal:  Front Endocrinol (Lausanne)       Date:  2022-06-30       Impact factor: 6.055

2.  Medication Adherence and Associated Factors in Patients With Type 2 Diabetes: A Structural Equation Model.

Authors:  Jing Huang; Shenglan Ding; Shuyuan Xiong; Zhiping Liu
Journal:  Front Public Health       Date:  2021-11-04

3.  A machine learning-based risk warning platform for potentially inappropriate prescriptions for elderly patients with cardiovascular disease.

Authors:  Wu Xingwei; Chang Huan; Li Mengting; Qin Lv; Zhang Jiaying; Long Enwu; Zhu Jiuqun; Tong Rongsheng
Journal:  Front Pharmacol       Date:  2022-08-11       Impact factor: 5.988

  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.