Literature DB >> 30451063

Data Analytics and Modeling for Appointment No-show in Community Health Centers.

Iman Mohammadi1, Huanmei Wu1, Ayten Turkcan2, Tammy Toscos3, Bradley N Doebbeling4.   

Abstract

OBJECTIVES: Using predictive modeling techniques, we developed and compared appointment no-show prediction models to better understand appointment adherence in underserved populations. METHODS AND MATERIALS: We collected electronic health record (EHR) data and appointment data including patient, provider and clinical visit characteristics over a 3-year period. All patient data came from an urban system of community health centers (CHCs) with 10 facilities. We sought to identify critical variables through logistic regression, artificial neural network, and naïve Bayes classifier models to predict missed appointments. We used 10-fold cross-validation to assess the models' ability to identify patients missing their appointments.
RESULTS: Following data preprocessing and cleaning, the final dataset included 73811 unique appointments with 12,392 missed appointments. Predictors of missed appointments versus attended appointments included lead time (time between scheduling and the appointment), patient prior missed appointments, cell phone ownership, tobacco use and the number of days since last appointment. Models had a relatively high area under the curve for all 3 models (e.g., 0.86 for naïve Bayes classifier). DISCUSSION: Patient appointment adherence varies across clinics within a healthcare system. Data analytics results demonstrate the value of existing clinical and operational data to address important operational and management issues.
CONCLUSION: EHR data including patient and scheduling information predicted the missed appointments of underserved populations in urban CHCs. Our application of predictive modeling techniques helped prioritize the design and implementation of interventions that may improve efficiency in community health centers for more timely access to care. CHCs would benefit from investing in the technical resources needed to make these data readily available as a means to inform important operational and policy questions.

Entities:  

Keywords:  access to care; appointment non-adherence; community health centers; electronic health records; predictive modeling

Mesh:

Year:  2018        PMID: 30451063      PMCID: PMC6243417          DOI: 10.1177/2150132718811692

Source DB:  PubMed          Journal:  J Prim Care Community Health        ISSN: 2150-1319


Introduction

Community health centers (CHCs) are safety-net clinics providing primary care for underserved and uninsured populations. For individuals at or below the US federal poverty level, CHCs provide a vital safety health care net. CHCs provide primary care services for acute and chronic diseases, injuries, and preventive services. High missed appointment rates have been identified as one of the most significant barriers to access to care for these populations.[1,2] In semistructured interviews conducted at CHCs, clinic staff and providers agreed that a high missed appointment rate is a major problem.[3] Given financial challenges of delivering quality health care in the United States, finding ways to improve performance is critical in the plight to provide greater access to care. Optimizing scheduling systems has been identified as one system level approach to address access needs. For example, reducing the number of missed appointments is crucial as when appointment slots go unused it effectively reduces access to others in need of an appointment.[4] In addition to underutilizing providers’ time, missed appointments impact waits and delays for others, increase health care costs, and increase possibility for adverse health outcomes.[5,6] Research has shown that lowering missed appointment rates can improve clinical efficiency and utilization, reduce waste, improve provider satisfaction and lead to better health outcomes for patients.[7,8] Missed appointment rates range from 10% to 50% across healthcare settings in the world with an average rate of 27% in North America.[6] Patients with higher missed appointment rates are significantly more likely to have incomplete preventive cancer screening, worse chronic disease control and increased rates of acute care utilization.[9] In previous studies, missed appointments have been due to logistical issues, lack of understanding of the scheduling system, patients not feeling respected by healthcare providers or the health system, affordability, timeliness, patients forgetting appointment and patient severity of illness.[6,10] To understand the complexity of appointment adherence in different health care settings, different datasets, variables, and data volumes have been studied. Medium-scale studies (ranging from 6,000 to 8,000 patients) focused on a few patient characteristics or a single (eg, time) component.[11-13] For example, a large-scale no-show modeling of a Veterans Affairs (VA) outpatient clinic included 555,183 patients, which scheduled 25,050,479 appointments; however, the study only considered a few variables such as the patient gender, the date of the appointment, and new versus established patients.[14] Most studies developed regression models to predict appointment nonadherence.[12,15] Most similar to the present study, one study identified predictors of missed clinic appointments among an underserved population.[16] These results revealed predictors for a missed appointment included percentage of no-shows in patients previous appointments (no-show or cancellation within 24 hours), wait time from scheduling to appointment, season, day of the week, provider type, and patient age, sex, and language proficiency. In other studies of predictive modeling in health care arena using electronic health record (EHR) data, other predictive modeling techniques such as naïve Bayes classifier[17] and neural network[18] were used to predict hospital readmissions. In this study, we apply and build on these techniques to predict appointment no-show in CHCs. Here, we test missed appointment prediction models by analyzing EHR and scheduling data. We aim to exploit predictive modeling to improve understanding of the complexity of appointment adherence in underserved populations. Information about patients, providers, appointments and time are used to predict patients’ adherence to appointments. The main contributions of this study are to (a) build on previous no-show modeling in community health centers by expanding the focus on various outpatient specialties and underserved population specific predictors; (b) compare different predictive modeling methodologies, namely logistic regression, naïve Bayes classifier, and artificial neural networks (specifically multilayer perceptron); and (c) investigate the impact of clinic characteristics on predictors of the no-show.

Materials and Methods

Participants

Data for this project were collected from a large urban multisite community health center, involving 10 locations in Indianapolis, most of which are considered federally qualified health centers (FQHC). This CHC has provided care for more than 100,000 patients during 2014 to 2016. Health care services provided by this CHC include but not limited to primary care, pediatrics, family practice, internal medicine, obstetrics/gynecology, dental care, vision care, behavioral health services, and preventive care. The goal of the no-show modeling was to focus on primary care, so data on dental and vision care visits was not considered. All study methods were approved by our institutional review board.

Data Collection and Sample Size

We extracted and deidentified semistructured data from over 17 tables in the CHC’s database from 2010 to 2016 to address the study aim. EHR data, including clinic (ie, operational and financial data) and patient (ie, patient demographics and clinical characteristics) information, were included and linked at the patient level. The data was stored in a secure Microsoft SQL Server with limited access. For this study, we created a dataset of patients’ encounters from January 1, 2014 to April 30, 2016. The dataset included 599,636 appointments by 76,453 unique patients (Table 1).
Table 1.

Distribution of Patient Characteristics Versus Appointment Adherence.

Appointment Adherence
Patient CharacteristicsAttended (n = 61,419)Missed (n = 12,392) P [a]
Categorical variables, Percentages
New patientYes2.12.4.0455
Translator neededYes15.28<.0001
EthnicityHispanic or Latino19.611.9<.0001
Not Hispanic or Latino7580.2
Unspecified5.47.9
RaceAmerican Indian or Alaska Native0.10.1<.0001
Asian4.22
Black30.337.7
Multiple races3.93.7
Native Hawaiian and Other Pacific Islander1.10.7
White60.455.7
GenderFemale61.464.8<.0001
Marital statusDivorced3.33.1<.0001
Legally separated1.31.7
Married12.89.5
Partner0.40.3
Single80.883.4
Widowed1.20.8
Cell phone ownershipNo18.226.4<.0001
Email availabilityNo70.674.5<.0001
Using patient portalNo78.283.5<.0001
Employment statusEmployed full-time1310.8<.0001
Employed part-time5.15.5
Not employed79.682.4
Retired1.50.4
Self-employed0.50.3
InsuranceCommercial14.88.4<.0001
Marketplace0.60.3
Medicaid66.869
Medicare5.63.6
Self-pay12.218.7
Tobacco useCurrent every day smoker22.835.5<.0001
Current some day smoker2.83.4
Former smoker1312
Never smoker61.349.1
Continuous variables, Mean (SD)
Age (years)21.1 (19.4)21.4 (16.9).1393
Annual income$2748 (8421)$2046 (7109)<.0001
Prior no-show Rate0.11 (0.2)0.2 (0.3)<.0001

T test for continuous variables and chi-square for categorical variables.

Distribution of Patient Characteristics Versus Appointment Adherence. T test for continuous variables and chi-square for categorical variables.

Data Preprocessing

Appointment compliance field was the dependent variable in this analysis, which included the categories of checkout (ie, complete) appointment, no-show, cancelled, rescheduled, and others. A no-show appointment is defined as a patient who did not keep the prescheduled appointment and did not cancel the appointment at least 24 hours ahead of the appointment time. We focused on appointments scheduled with medical doctor, nurse practitioner, or certified nurse-midwife. All other nurse visit appointments were excluded from analyses. We performed the following data filtering steps: Filtering appointment categories: To create the binary outcome variable in this study, we only included no-show and checkout appointments in the final analysis, and observations having other appointment compliance, such as rescheduled, cancelled, and so on, were censored from the dataset. Ensuring appointment independences: To ensure observations are independent from each other, we only included the last appointment of each patient in the final analysis. Handling missing information: unstructured free text fields, such as schedulers’ notes, were used to complete any missing values in fields, such as appointment type, patient age or gender. Simple rules were used to find visit types from scheduler notes. For example, if the note contained “acute” and visit type field was missing, visit type field was filled by “Acute care”, and other types can be seen in Table 3. All other observations with missing information were removed from the dataset.
Table 3.

Distribution of Visit Characteristics Versus Appointment Adherence.

Appointment Adherence
VariablesAttended (n = 61,419)Missed (n = 12,392) P [a]

Categorical variables, Percentages

Appointment duration (minutes)100.80.1<.0001
1568.360.3
2014.314.7
3015.622.1
450.51.7
600.51.1
Lead timeSame day31.48.4<.0001
Next day97.1
Within 2 weeks31.635.4
Between 2 weeks and 1 month1320.7
More than 1 month1528.5
Days since last appointmentWithin a week1.41.9<.0001
Between 1 and 2 weeks1.11.8
Between 2 weeks and 1 month2.34.1
Between 1 and 3 months5.69.3
Between 3 and 6 months6.710.2
Between 6 months and a year14.516.2
More than a year53.739.6
No prior appointment since 201414.816.9
Appointment timeAM43.844.5.1294
SeasonFall18.119.8<.0001
Spring29.928.9
Summer15.118.3
Winter36.933
WeekdayMonday22.323.4<.0001
Tuesday21.922.1
Wednesday20.119.1
Thursday18.819.2
Friday15.815.3
Saturday1.11
Visit typeAcute care27.712.1<.0001
Adult routine/Follow-up1724.4
Behavioral health24.8
Podiatry0.71.4
Pediatric37.637.5
Pregnant4.56.5
Women10.513.4

T test for continuous variables and chi-square for categorical variables.

Out of 76,453 unique patients, 2,642 patients were removed because they had observations with missing information that could not be found in the data. The final dataset included 73,811 observations of unique individuals, and whether they showed for their last appointment during the study period. Data imputation was not necessary as we had sufficient number of observations for our analyses.

Variable Preparation

Data fields included visit characteristics (facility/clinic type, date of visit, date contacted the clinic for scheduling the visit, time of visit, visit duration, and visit type), patient characteristics (patient pseudo-ID, age, race, ethnicity, gender, marital status, cell phone ownership, email availability, whether using patient portal, employment status, tobacco use, income, needing translator and primary insurance), provider characteristics (whether seeing the patients’ primary care practitioner [PCP] or not, specialty and medical license) and appointment compliance (“no-show” or “check out”). In addition to the existing variables in the EHR, we created the following variables to consider in our no-show modeling: Lead time, which is the time difference (in days) between the date of visit and date the patient had contacted the clinic to arrange an appointment. Prior no-show rate, which is the number of no-shows for a given patient prior to the last appointment, divided by the patient’s total number of appointments prior to the last appointment. We used this to test the effect of patient no-show behavior on appointment adherence. Days since the last appointment, which is the difference between the date of the last visit and the date of appointment before the last visit.

Statistical Analyses

We hypothesized that patient and provider characteristics and visit features were all predictors of appointment no-show in CHCs. We tested variables individually for relationships with the appointment adherence using a chi-square test for categorical variables and t test for continuous variables. We chose variables with a P value less than .2 to enter into the model development step. Tables 1-3 list variables that were included in the modeling. The dataset included 73,811 observations, 83% arrived and 17% no-show.
Table 2.

Distribution of Provider Characteristics Versus Appointment Adherence.

Appointment Adherence
VariablesAttended (n = 61,419)Missed (n = 12,392) P [a]
Categorical variables, Percentages
Provider specialtyBehavioral Health1.84.3<.0001
Certified Nurse-Midwife9.512.7
Family Medicine17.114.7
Internal Medicine11.511.7
Nurse Practitioner9.97.3
Obstetrics/Gynecology4.35.9
Pediatrics33.630.3
Podiatry0.71.4
Patient’s primary care practitioner?No83.286.6<.0001

T test for continuous variables and chi-square for categorical variables.

Prediction Model Development

We randomly split the dataset into 2 samples: 70% for the training (or derivation) set and 30% for the test (or validation) set. This training and test set selection was repeated 10 times to overcome selection bias. To decrease potential bias of learning algorithm for training set, we randomly selected training subsets with no-show to checkout ratio of 2 to 1 and repeated this randomization for 10 times. We used the training subsets to develop the no-show prediction model using 3 methodologies: Logistic regression: We used logistic regression in SAS 9.4. to develop the prediction model with a stepwise selection and significance level of α = .01. All the variables, shown in Tables 1-3 and their interactions, were included in the model development. Artificial neural network: The large number of features and observations in this study led us to use more complex machine learning algorithms such as multilayer perceptron. Multilayer perceptron consists of multiple linear regression models are advantageous when there is a large number of features (variables) with complex relations among them.[19] Categorical variables were transformed to numeric variables. For example, if a patient is a “New Patient,” the numeric variable of New Patients would be created with a value of 1. Continuous, binary, and numeric variables were used as inputs for a multilayer perceptron and 1 binary variable (No-show = 1 or 0) was used as output. Matlab software was used to develop the multilayer perceptron in this project having 3 layers of the input layer, hidden layer including 25 nodes and output layer. The training data subsets were used to train the network by minimizing the mean-square error (MSE) between the desired output and the actual output of the network. The value of the output node determined the classification using a range (between 0 and 1) of cutoff thresholds. Here, we used absolute value of weights for input layer nodes to identify and rank the most important variables contributing to no-show prediction. Naïve Bayes classifier: The majority of predictors in our datasets were categorical; hence, we applied a naïve Bayes classifier that is appropriate to categorical data.[20] This classifier computes a conditional probability of each category in each variable given the outcome. Then, Bayes rules are applied to calculate the probability of the outcome given different categories of variables in the data. We applied the naïve Bayes classifier algorithm implemented in “scikit-learn” in Python over the randomly selected train and test datasets. The smoothing value of 0.1 provided the best performance for the classifier.

Model Validation

Models were assessed by calculating the area under the curve for the receiver operating characteristic (AUC-ROC) curve. Test dataset was used to validate models’ ability to discriminate between patients who no-showed versus those who attended. Ten-fold cross-validation was used to validate the 3 models, and average AUCs, sensitivities to predicting no-show and overall model accuracy were the key indicators of model validation.

Results

The final dataset included 73,811 observations with 12,392 missed appointments. Comparative analyses of patient characteristics revealed that black, non-Hispanic or non-Latino, female, single, not employed, Medicaid, self-pay, or smoker patients had a higher chance of missed appointments (P < .0001; see Table 1). The average annual income is lower, and the average prior missed appointment rate is higher in patients who no-showed in their last appointment (P < .0001). Patients without a cell phone, email, or patient portal had a higher chance of a missed appointment (P < .0001). The comparative analysis of the provider characteristics showed that patients scheduled with behavioral health or OB-GYN providers or not scheduled with their primary care providers have higher missed appointment rates compared with other appointment types (P < .0001), as demonstrated in Table 2. Distribution of Provider Characteristics Versus Appointment Adherence. T test for continuous variables and chi-square for categorical variables. The appointment duration, the time between appointment days, and the day appointment requested, the time (daytime, weekday, or season) of an appointment, and the type of an appointment are statistically significantly different between checkout and missed appointment patients (P < .0001), as shown in Table 3. Distribution of Visit Characteristics Versus Appointment Adherence. Categorical variables, Percentages T test for continuous variables and chi-square for categorical variables. Table 4 shows characteristics of 10 facilities within this CHC system. Clinics are different in terms of missed appointment rates and distributions of patient type, visit type, and provider type.
Table 4.

Clinic Characteristics.

FacilityTotal No. of PatientsNo-show
Clinic CharacteristicsPercentage/Mean Among All Clinics
FrequencyPercentage
Clinic 110,6332,24821• Large number (23%) of patients needing translator• Large number (20%) of Asian patients• Highest mean lead time (28.6 days)• 14%, P < .0001• 4%, P < .0001• 17 days, P < .0001
Clinic 23,68066018• Higher percentage of new patients (10.3%)• Dominantly pregnant and woman patients (98%)• Dominantly certified nurse-midwife and obstetrics/gynecology providers (95%)• Dominantly female patients (98%)• Dominantly adult patients (95%)• Patients with lower prior no-show rates (0.08)• 2.1%, P < .0001• 15.8%, P < .0001• 15%, P < .0001• 62%, P < .0001• 43%, P < .0001• 0.12, P < .0001
Clinic 33,20639212• Mostly scheduled with patients’ primary care practitioners (56%)• Patients with lower prior no-show rates (0.08)• 16.2%, P < .0001• 0.12, P < .0001
Clinic 46,73180312• Majority Black (77%)• Mostly same-day appointments (67%)• Higher number of acute care appointments (46%)• 32%, P < .0001• 27%, P < .0001• 25%, P < .0001
Clinic 52,21648022• Highest no-show rate• 17%, P < < .0001
Clinic 67,8701,54320• Mostly 20-minute appointments (79%)• Dominantly children (97%)• Majority black (63%)• Dominantly not employed (98%)• 14%, P < .0001• 55%, P < .0001• 32%, P < .0001• 80%, P < .0001
Clinic 710,7031,91618• Large number (23%) of patients needing translator• Higher number of Hispanic or Latino (34%)• 14%, P < .0001• 18%, P < .0001
Clinic 812,0161,65914• Large number (22%) of patients needing translator• Higher number of Hispanic or Latino (31%)• Highest income level ($4553/year)• 14%, P < .0001• 18%, P < .0001• $2665, P < .0001
Clinic 911,5211,94217• Dominantly white (85%)• 60%, P < .0001
Clinic 105,23574914• Patients with lower prior no-show rates (0.08)• 0.12, P < .0001
Clinic Characteristics.

Predictive Modeling

As shown in Table 4, clinics had different population sizes, characteristics, and no-show rates. Therefore, we developed a separate logistic regression model for each clinic. Supplementary Table S1 (available in the online version of the article) shows the results from regression model development. These separate models corresponding to individual clinics yielded different predictors for missed appointments. Notably, lead time, prior missed appointment rate, age, insurance type, tobacco use, days since the last appointment, and cell phone ownership were consistent significant factors across clinics.

Patient Characteristics

Table 4 demonstrates that clinic 2 patients had lower prior missed appointment rates compared with other clinics. In all clinics except clinic 6, patients between 18 and 64 years old were 1.6 (99% CI 1.5-1.6) and 3.7 (99% CI 2.9-4.6) times more likely to no-show their next appointments compared with patients between 0 to 17 years old and 65 years and older patients, respectively. Notably, clinic 6 is a pediatric clinic and patients are dominantly between 0 and 17 years old. Patients who needed a translator in their appointments, particularly in clinic 7 (with a high proportion of Hispanic or Latinos), were 0.5 times less likely to no-show in their next appointments (99% CI 0.4-0.5). In 2 clinics, the interaction between age and gender also influenced no-shows. Insurance status was another significant predictor of missed appointments, such that insured patients were less likely to keep their appointments. In most clinics, patients insured by commercial, marketplace, Medicaid, and Medicare plans were 0.4 (99% CI 0.3-0.4), 0.3 (99% CI 0.2-0.5), 0.7 (99% CI 0.6-0.7), and 0.4 (99% CI 0.37-0.50) times as likely to miss appointments, compared with their uninsured counterparts. Smoking daily increased the likelihood of missed appointments by 95%, compared with patients who never smoked (odds ratio OR = 2, 99% CI 1.8-2.1). Patients using for their clinics patient portal (web-enabled) were less likely to no-show in their appointments (OR = 0.7, 99% CI 0.7-0.8). In clinic 5, patients without an email address recorded in the EHR system are 1.2 times more likely to no-show (99% CI 1.21-1.23). Patients without a cell phone number available in the records were 1.6 times more likely to no-show (99% CI 1.52-1.71).

Scheduling Characteristics

Lead time was the most consistent significant factor across all the clinics. Longer lead time provides greater opportunity for a missed appointment (P < .0001). Appointments made more than 1 month in advance are 7.1 (99% CI 6.5-7.5), 2.4 (99% CI 2.2-2.7), 1.7 (99% CI 1.6-1.9), and 1.2 (99% CI 1.1-1.3) times more likely to become a no-show, compared with appointments made on same day, 1 day, 2 weeks, and between 2 weeks and 1 month in advance, respectively. Next day appointments were 2.9 times more likely to become a missed appointment than same day appointments (99% CI 2.6-3.3). Patients with a history of missed appointments were 4.9 times more likely to miss their next appointments (99% CI 4.4-5.8), in all clinics except clinic 2. Patients who had an appointment between 1 and 2 weeks prior to their last appointment were more likely to miss that last appointment compared with patients who had a prior appointment in the last 6 to 12 months (OR = 1.5, 99% CI 1.2-1.8), more than 12 months (OR = 2.2, 99% CI 1.8-2.7), or patients who had no prior appointments (OR = 1.4, 99% CI 1.1-1.7).

Clinic Visit Characteristics

In one-half of the clinics, type of visit predicted appointment adherence. Supplementary Table S1 shows that acute visits had lower missed appointment rates than all other visit types, while behavioral health visits had the highest missed appointment rates. Seasonality of the appointments predicted missed appointments such that appointments occurring during spring or summer had higher missed appointment rates than winter appointments. Notably, patients scheduled with their own PCP were less likely to miss the appointment than the ones scheduled with other providers (OR = 0.8, 99% CI 0.7-0.8). Appointment duration was also a significant factor (particularly in clinics 3 and 5). Longer durations such as 1 hour or 45 minutes were more likely to be no-show than shorter durations such as 15 or 20 minutes. The ranking of variables contributing to prediction of no-show in the multilayer perceptron are shown in Supplementary Table S2. The ranking is based on the weights nodes in the input layer of multilayer perceptron. The top 10 predictors of the no-show in our multilayer perceptron analyses included: lead-time, provider specialty, race, employment status, days since last appointment, prior no-show rate, cell phone ownership, tobacco use, marital status, and gender. Similarly, there were multiple variables contributing to no-show (Supplementary Table S3) using the naïve Bayes classifier. prior no-show rate, age group, visit type, lead-time, days since last appointment, duration, insurance, cell phone ownership, tobacco use, and ethnicity are the top 10 factors predicting next appointment no-show. Those variables important in all three types of models included: lead time, patient prior no-show behavior, cell phone ownership, tobacco use, and the number of days since the last appointment of patient. Logistic regression and naïve Bayes classifier have commonly identified visit type, age, and insurance as top 10 predictors. Table 5 shows the validation results for 3 models. Overall accuracy in Table 5 is the correct classification ratio for the model. The AUC for logistic regression and the naïve Bayes classifier are, respectively, 0.81 and 0.86, which are considered excellent for discriminating between 2 outcomes.[6] Multilayer perceptron had low AUC of 0.66.
Table 5.

Validation and Comparison of Prediction Models.

Modeling MethodTrain Set
Test Set
AUCSensitivityPositive (No-show) Predictive ValueOverall Accuracy (%)AUCSensitivityPositive (No-show) Predictive ValueOverall Accuracy (%)
Logistic regression0.910.840.58800.810.720.5473
Multilayer perceptron0.770.730.43790.660.630.3571
Naïve Bayes classifier0.960.820.67920.860.730.4582

Abbreviation: AUC, area under the curve for the receiver operating characteristic curve.

Validation and Comparison of Prediction Models. Abbreviation: AUC, area under the curve for the receiver operating characteristic curve.

Discussion

We studied missed appointments in 10 separate clinics within one urban community health care system. Our study shows that clinics have different population characteristics, specialties, and patient demographics; thus, it is not surprising that appointment adherence varies across geographic sites. For example, specialty clinics such as pediatric or woman clinics have higher missed appointment rates than the ones providing acute or general primary care. Appointment lead time, past missed appointments, and age group of patients are the common important factors differentiating clinics’ overall missed appointment rate. Our study suggests that any attempt to create a missed appointment prediction model or to design interventions for reducing missed appointment rates should be clinic/facility specific and tailored based on clinic, facility, or department characteristics. Our study has 4 major findings. First, patient, scheduling, and visit characteristics differ across missed and arrived appointments. These characteristics should be of interest to managers and policy makers, in order to better design interventions and policies to reduce missed appointments. Second, the consensus of the logistic regression, multilayer perceptron, and naïve Bayes classification was that lead-time, patient prior missed appointments, cell phone ownership, tobacco use, and the number of days since the last appointment of a patient are the most significant predictors of missed appointments. Other factors were important in certain clinics, even after control for these factors. These findings should help managers in health care systems prioritize the design and implementation of interventions to reduce missed appointments. Third, patient appointment adherence had different determinants in different clinics or facilities within a single health care system. This finding makes sense in a large urban area, where neighborhood, population and clinic characteristics, as well as policies and procedures differ. It also underlines the importance of looking at data at the clinic level, because different clinics, even within the same system may have an important population and organizational differences. Fourth, according to the accuracy of the predictions, logistic regression and Bayes classifiers concluded similarly and perform better in missed appointment modeling than a multilayer perceptron. This might be because of categorical nature of our data. Studies have reported that the discrimination ability of neural networks (such as multilayer perceptron) versus other statistical modeling techniques is data specific.[21]

Poverty, Employment, and Access to Health Information Technology

One key social determinant of health in populations is economic stability; this includes measures such as education, poverty, and employment status.[22] We found that lower income and unemployment were associated with more missed medical appointments that would likely impair the health and/or health outcomes of patients. Studies found that socioeconomic characteristics have negative impact on health outcomes.[23] The role of poverty and employment are obviously complex and multifactorial across the United States. Our findings point to the need for social, financial, and educational interventions to help indigent people prosper and communities thrive. Access to emerging technologies such as cell phones, the Internet and social media is another social and financial determinant. We found that patients without access to cell phone, email, and a patient portal were more likely to miss their medical appointments. Therefore, lack of access to these technologies may affect health outcomes. Future research should examine if the provision of these consumer health technologies alone can enhance access to health for individuals in poverty or if our finding is more directly related to financial status alone. Our results show that patients without insurance for medical services are at risk of not adhering to their appointments and consequently their care plans. This factor is highly correlated with unemployment, which was very high (approximately 80%) in our study population.

Patient Engagement, Tobacco Use, and Promoting Patient Appointment Adherence

In our study, smoking was one of the most significant factors related to missing medical appointments. We hypothesize that this variable as a health behavior, which may be highly related to other health practices, including adherence to scheduled clinic visits. It is beyond the scope of this study to determine whether this variable is a marker for adherence with recommendations or a confounder. Regardless, its importance underscores the importance of engagement of the underserved populations in their care and the role of individual health behaviors, attitudes, and practices. Common reasons for missed appointments found in prior research include forgetting about the appointment, competing priorities, and demands (such as the need to work or inability to leave work), availability of transportation, or feeling better at the time of the appointment.[24] These reasons can be magnified if the lead time (the most important predictor in our study) for appointments is elongated. Interventions such as increasing number of open access (same-day) hours and decreasing number of appointments made more than 1 month in advance should be considered to improve access to care in community health centers. Past missed appointment is an important predictor of future appointment adherence. Our findings are consistent with other research that operationalized passed missed appointments using clinicians’ notes containing phrases like “no-show,” “did not present,” “failed to attend,” and “missed appointment.” These researchers found that patients that previously missed appointments were more likely to miss future appointments.[25] Further investigation of this problem should focus on extracting important information available as free text in patient complaint and reason to seek health.[26] Our study found that behavioral health patients were more likely to miss their next appointments than any other type of patients. Differences in adherence with appointments here could either be related to different systems for scheduling and reminding patients of appointments between medical and behavioral health systems, or related to intrinsic differences in practices, attitudes or adherence among behavioral health patients. Further investigation of this problem should focus on differences between the practices and policies for such patients, before efforts to make special accommodations for the population.

Application to Medical Practice

Our study used large patient datasets with multiple potential explanatory variables in order to develop prediction models using various clinics within a health care system. We also used multiple methods to develop and compare the models. Access to health care can affect individuals’ health status and quality of life. Missed appointments are one of the most important factors determining access to care. High levels of no-shows are not only an expensive waste of limited provider resources, but they can also lead to unmet health needs and delays in receiving appropriate care. Therefore, predicting and preventing missed appointments can potentially improve access to care.[27] The outcomes of this study could help clinicians predict appointment no-shows that can potentially reduce no-show rates in CHCs. Researchers have reported lower no-show rates can improve clinical efficiency and utilization, reduce waste, improve provider satisfaction and lead to better health.[28] Redesigning and testing the alternate scheduling processes will help patients get appointments in a timelier manner. These better scheduling systems will improve access for acute patients, increase continuity of care for chronic patients and essentially positively affect health outcomes. There are 2 possible real-world applications of this study. First, the methodologies and findings of this study can be used to redesign scheduling systems in CHCs to reduce the number of no-show appointments. Second, no-show predictions models can be implemented in EHR systems as decision support systems that would identify patients with a high risk of appointment no-show. Appointments with high risk of no-show may be double booked, or patients with high risk of no-show may be reminded more rigorously.

Limitations

One of the limitations of this study is that it includes only patients from 1 CHC system in Indianapolis. However, this CHC system involves multiple geographic sites and is very diverse from the patient characteristics perspective. Another limitation of this study is that the dataset used in this study did not have information on the clinical, physical, and functional status of patients (eg, diabetes, depression, congestive heart failure, etc). These attributes can be significant predictors of the no-show. However, visit type variable in our dataset did relate to a patient’s clinical characteristics. Findings of this study are drawn from FQHC clinics providing primary care to underserved populations. Whether these results are generalizable to other patient populations will need to be addressed in other studies. Another limitation of this study is that the dataset did not include information about new patients who no-showed in their first appointments; however, sufficient number of observations did not significantly impact the outcomes of this study.

Future Work

These results demonstrate the value of using existing clinical and operational data to address important operational issues. Further resources are needed in CHCs to make these data readily available and to inform important operational and policy questions. Future work might also focus on linking billing information and claims data with EHR to extract important information about patients and appointments. One example could be using evaluation and management codes to adequately identify provider type or provider time spent with patients.

Conclusion

This project developed the statistical model and machine learning models that can be used to predict patients’ chance of no-showing to their next medical appointment. Logistic regression, multilayer perceptron, and naïve Bayes classifiers were used to develop and compare the no-show prediction models that resulted in finding lead time, patient prior no-show behavior, cell phone ownership, tobacco use, and the number of days since the last appointment of a patient as significant predictors of appointment adherence. The application of these findings may be used to design new interventions to improve scheduling processes and other policies and practices for better and timelier access to care. We suggest that redesigned operations and policies, from scheduling practices to reminder systems and other technological tools to improve adherence can improve clinic revenues, utilization of resources, and ultimately improve health outcomes. Click here for additional data file. Supplemental material, Supplement_Tables for Data Analytics and Modeling for Appointment No-show in Community Health Centers by Iman Mohammadi, Huanmei Wu, Ayten Turkcan, Tammy Toscos and Bradley N. Doebbeling in Journal of Primary Care & Community Health
  20 in total

1.  PREDICTIVE MODELING OF HOSPITAL READMISSION RATES USING ELECTRONIC MEDICAL RECORD-WIDE MACHINE LEARNING: A CASE-STUDY USING MOUNT SINAI HEART FAILURE COHORT.

Authors:  Khader Shameer; Kipp W Johnson; Alexandre Yahi; Riccardo Miotto; L I Li; Doran Ricks; Jebakumar Jebakaran; Patricia Kovatch; Partho P Sengupta; Sengupta Gelijns; Alan Moskovitz; Bruce Darrow; David L David; Andrew Kasarskis; Nicholas P Tatonetti; Sean Pinney; Joel T Dudley
Journal:  Pac Symp Biocomput       Date:  2017

2.  Risk factor model to predict a missed clinic appointment in an urban, academic, and underserved setting.

Authors:  Orlando Torres; Michael B Rothberg; Jane Garb; Owolabi Ogunneye; Judepatricks Onyema; Thomas Higgins
Journal:  Popul Health Manag       Date:  2014-10-09       Impact factor: 2.459

3.  Using no-show modeling to improve clinic performance.

Authors:  Joanne Daggy; Mark Lawley; Deanna Willis; Debra Thayer; Christopher Suelzer; Po-Ching DeLaurentis; Ayten Turkcan; Santanu Chakraborty; Laura Sands
Journal:  Health Informatics J       Date:  2010-12       Impact factor: 2.681

4.  Comparison of logistic regression and neural networks to predict rehospitalization in patients with stroke.

Authors:  K J Ottenbacher; P M Smith; S B Illig; R T Linn; R C Fiedler; C V Granger
Journal:  J Clin Epidemiol       Date:  2001-11       Impact factor: 6.437

5.  Appointment "no-shows" are an independent predictor of subsequent quality of care and resource utilization outcomes.

Authors:  Andrew S Hwang; Steven J Atlas; Patrick Cronin; Jeffrey M Ashburner; Sachin J Shah; Wei He; Clemens S Hong
Journal:  J Gen Intern Med       Date:  2015-03-17       Impact factor: 5.128

6.  No-show to primary care appointments: why patients do not come.

Authors:  Emma Kaplan-Lewis; Sanja Percac-Lima
Journal:  J Prim Care Community Health       Date:  2013-07-26

Review 7.  Clinical information extraction applications: A literature review.

Authors:  Yanshan Wang; Liwei Wang; Majid Rastegar-Mojarad; Sungrim Moon; Feichen Shen; Naveed Afzal; Sijia Liu; Yuqun Zeng; Saeed Mehrabi; Sunghwan Sohn; Hongfang Liu
Journal:  J Biomed Inform       Date:  2017-11-21       Impact factor: 6.317

8.  Time dependent patient no-show predictive modelling development.

Authors:  Yu-Li Huang; David A Hanauer
Journal:  Int J Health Care Qual Assur       Date:  2016-05-09

Review 9.  Leveraging the Social Determinants of Health: What Works?

Authors:  Lauren A Taylor; Annabel Xulin Tan; Caitlin E Coyle; Chima Ndumele; Erika Rogan; Maureen Canavan; Leslie A Curry; Elizabeth H Bradley
Journal:  PLoS One       Date:  2016-08-17       Impact factor: 3.240

10.  Large-Scale No-Show Patterns and Distributions for Clinic Operational Research.

Authors:  Michael L Davies; Rachel M Goffman; Jerrold H May; Robert J Monte; Keri L Rodriguez; Youxu C Tjader; Dominic L Vargas
Journal:  Healthcare (Basel)       Date:  2016-02-16
View more
  10 in total

1.  Application of Machine Learning to Predict Patient No-Shows in an Academic Pediatric Ophthalmology Clinic.

Authors:  Jimmy Chen; Isaac H Goldstein; Wei-Chun Lin; Michael F Chiang; Michelle R Hribar
Journal:  AMIA Annu Symp Proc       Date:  2021-01-25

2.  Impact of IMPACT: Longitudinal Analysis of an Integrated Participant Scheduling System in a Clinical Research Setting.

Authors:  Alex Butler; Junghwan Lee; Yat So; Linda Busacca; Karen Marder; Henry N Ginsberg; Dianne Frederick; Ismael Castaneda; Elizabeth Guerridoi; Chunhua Weng
Journal:  AMIA Annu Symp Proc       Date:  2021-01-25

3.  Word Embedding and Clustering for Patient-Centered Redesign of Appointment Scheduling in Ambulatory Care Settings.

Authors:  Iman Mohammadi; Saeed Mehrabi; Bryce Sutton; Huanmei Wu
Journal:  AMIA Annu Symp Proc       Date:  2022-02-21

4.  Improving intervention design to promote cervical cancer screening among hard-to-reach women: assessing beliefs and predicting individual attendance probabilities in Bogotá, Colombia.

Authors:  David Barrera Ferro; Steffen Bayer; Sally Brailsford; Honora Smith
Journal:  BMC Womens Health       Date:  2022-06-07       Impact factor: 2.742

5.  Evaluating the reasons for nonattendance to outpatient consultations: is waiting time an important factor?

Authors:  Bernadeta Zykienė; Vytenis Kalibatas
Journal:  BMC Health Serv Res       Date:  2022-05-09       Impact factor: 2.908

6.  Machine learning approaches to predicting no-shows in pediatric medical appointment.

Authors:  Dianbo Liu; Won-Yong Shin; Eli Sprecher; Kathleen Conroy; Omar Santiago; Gal Wachtel; Mauricio Santillana
Journal:  NPJ Digit Med       Date:  2022-04-20

Review 7.  Patient No-Show Prediction: A Systematic Literature Review.

Authors:  Danae Carreras-García; David Delgado-Gómez; Fernando Llorente-Fernández; Ana Arribas-Gil
Journal:  Entropy (Basel)       Date:  2020-06-17       Impact factor: 2.524

8.  How machine-learning recommendations influence clinician treatment selections: the example of the antidepressant selection.

Authors:  Maia Jacobs; Melanie F Pradier; Thomas H McCoy; Roy H Perlis; Finale Doshi-Velez; Krzysztof Z Gajos
Journal:  Transl Psychiatry       Date:  2021-02-04       Impact factor: 6.222

9.  Sensitivity of Psychosocial Distress Screening to Identify Cancer Patients at Risk for Financial Hardship During Care Delivery.

Authors:  J Alberto Maldonado; Shuangshuang Fu; Ying-Shiuan Chen; Chiara Acquati; K Robin Yabroff; Matteo P Banegas; Shine Chang; Rena M Conti; Cristina M Checka; Susan K Peterson; Pragati Advani; Kimberly Ku; Reshma Jagsi; Sharon H Giordano; Robert J Volk; Ya-Chen T Shih; Grace L Smith
Journal:  JCO Oncol Pract       Date:  2021-05-27

10.  Efficient Prediction of Missed Clinical Appointment Using Machine Learning.

Authors:  Zeeshan Qureshi; Ayesha Maqbool; Alina Mirza; Muhammad Zubair Iqbal; Farkhanda Afzal; Deborah Dormah Kanubala; Tauseef Rana; Mir Yasir Umair; Abdul Wakeel; Said Khalid Shah
Journal:  Comput Math Methods Med       Date:  2021-10-22       Impact factor: 2.238

  10 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.