Literature DB >> 35384134

Comparing risk adjustment estimation methods under data availability constraints.

Marica Iommi1, Savannah Bergquist2, Gianluca Fiorentini3, Francesco Paolucci4,5.   

Abstract

The Italian National Healthcare Service relies on per capita allocation for healthcare funds, despite having a highly detailed and wide range of data to potentially build a complex risk-adjustment formula. However, heterogeneity in data availability limits the development of a national model. This paper implements and ealuates machine learning (ML) and standard risk-adjustment models on different data scenarios that a Region or Country may face, to optimize information with the most predictive model. We show that ML achieves a small but generally statistically insignificant improvement of adjusted R2 and mean squared error with fine data granularity compared to linear regression, while in coarse granularity and poor range of variables scenario no differences were observed. The advantage of ML algorithms is greater in the coarse granularity and fair/rich range of variables set and limited with fine granularity scenarios. The inclusion of detailed morbidity- and pharmacy-based adjustors generally increases fit, although the trade-off of creating adverse economic incentives must be considered.
© 2022 The Authors. Health Economics published by John Wiley & Sons Ltd.

Entities:  

Keywords:  data granularity; formula funding; health expenditure; machine learning; risk-adjustment

Mesh:

Year:  2022        PMID: 35384134      PMCID: PMC9320950          DOI: 10.1002/hec.4512

Source DB:  PubMed          Journal:  Health Econ        ISSN: 1057-9230            Impact factor:   2.395


INTRODUCTION

In middle‐ and high‐income countries, health expenditure has steadily increased over time, making its containment a major issue for governments: policymakers must control healthcare resources to efficiently address population aging and public budget constraints. Risk‐adjustment schemes are one such tool for managing the efficiency and fairness of healthcare spending (Cid et al., 2016). The primary statistical purpose of risk‐adjustment models, both in a health plan payment or provider reimbursement, is to accurately predict individual healthcare costs. When designing a risk‐adjustment program, policymakers should aim to optimize the choice of statistical methods based on their data availability in the short‐medium term, pending a longer‐term investment to expand the range of variables and level of detail. The performance of a risk‐adjustment algorithm depends not only on the model or functional form, but also on the underlying data. In a setting with decentralized data administration, differences in data collection – and thus in the data available for prediction – stem from several sources, including: the accessibility of distinct databases (inpatient, pharmaceutical, outpatient services), the range of data elements collected, the ability to merge information at an individual level (due to privacy regulations), and different quality standards or level of detail in data collection (rudimentary vs. advanced information systems) according to the local context (we refer to the level of detail or aggregation as the “granularity” of the data). The Italian National Healthcare Service (NHS) is an example of a healthcare system with substantial heterogeneity in data collection. With its decentralized units (21 health regions) all subject to the same national legislation regarding data collection duties, the Italian NHS has the potential for using a highly detailed and wide‐ranging set of information to predict individual health expenditure. However, data collection takes place at the local level, where there are differing capacities to combine healthcare administrative databases, heterogeneous information systems, and variability in data granularity and quality control (Skrami et al., 2019). Thus, despite this potential richness of individual health information, the allocation of Italian NHS resources to Regions is still based on age‐weighted capitation. Per capita allocation using synthetic indices methods or actuarial cells formulas is common among national health systems (e.g., Spain, Denmark, Scotland, Portugal) to distribute resources in a large geographical area for its simplicity of construction and low cost information (Cid et al., 2016). However, person‐based formulae are increasingly used as an alternative method (e.g., United Kingdom), and they offer considerable scope for ensuring equitable allocations of health budgets to geographical areas (Radinmanesh et al., 2021; Smith, 2006). In this paper, we develop a framework for designing an effective risk‐adjustment algorithm to predict total healthcare expenditure by using routinely available, person‐based Italian data on health needs. We explore different data scenarios by examining how the predictive performance of conventional and machine learning (ML) risk‐adjustment estimation varies across by data granularity and range of variables settings. Our primary contribution is to develop general guidelines for policymakers who have various scenarios of data granularity and/or range of variables and to provide a set of applicable statistical tools. For countries with national health systems this study may serve as a basis for budget setting using person‐based regression models, which are more predictive than synthetic indices methods or actuarial cells formulas because they capture more variability in patient needs (Ellis, 2007). Countries with data constraints may use our framework to optimize the information available to improve efficiency and fairness in allocating resources or to reduce incentives for risk‐selection (Henríquez at al., 2020). Countries with less limiting data constraints can implement more sophisticated ML techniques to improve predictive performances and fit efficiency (Geruso & McGuire, 2016; Kronick & Welch, 2014; Rose, 2016).

EMPIRICAL FRAMEWORK

In Table 1 we present six typical data scenarios to predict total healthcare expenditure, which varies in two dimensions: range of variables (the number of health databases), and data granularity (the detail of the information in each database). These scenarios are based on real‐world implementations of risk‐adjustment programs and data collection; it is common to begin with age‐sex cells, then add in hospitalization information, and finally pharmaceutical data. Similarly, a coarser level of detail may be more amenable to basic data collection and simpler prediction methods. The Italian health information system is theoretically located in the cell at the bottom right (fine data/rich range of variables), but as long as the reliability and consistency of the administrative databases is uncertain, the first cell at the top left (coarse data/poor range of variables) could be the first plausible scenario in the transition from synthetic indices methods to more predictive models.
TABLE 1

Data setting scenarios to predict total health care expenditure

Poor range of variablesFair range of variablesRich range of variables
Demographic (DEM)DEM + hospital discharge records (HDR)DEM + HDR + pharmacy database (PD)
Coarse granularity (sparse detail) Total health care costs = f (sex * age groups) Total health care costs = f (sex * age groups + n. of hospitalizations) Total health care costs = f (sex * age groups + n. of hospitalizations + n. of drug prescriptions) 
Age groups: 0, 1–4, 5‐year classes up to 90, 90+ HDR: N. of hospitalizations HDR: N. of hospitalizations
PD: N. of prescriptions
Fine granularity (high detail) Total health care costs = f (sex * age groups + citizenship + degree of urbanization + income) Total health care costs = f (sex * age groups + citizenship + municipality of residence + income + group of ICD‐9 codes)  Total health care costs = f (sex * age groups + citizenship + municipality of residence + income + group of ICD‐9 codes+ group of ATC codes)
Age groups: 0, 1–4, 5‐year classes up to 90, 90+
Citizenship: Italian (yes/no) HDR: ICD‐9‐CM codes which can be classified into diagnosis‐related groups a HDR: ICD‐9‐CM codes; may be classified into DCGs
Degree of urbanization: Urban, peri‐urban, rural PD: ATC codes; may be classified into PCGs b
Income: low, medium‐low, medium high, high

Abbreviations: ESRD, End‐Stage Renal Disease; HCC, Hierarchical Condition Category; HDR, hospital discharge records; PD, pharmacy database; RxHCC, Prescription Drug Hierarchical Condition Category.

For example, possible coding classification systems include the Clinical Classifications Software for ICD‐9‐CM (https://www.hcup‐us.ahrq.gov/toolssoftware/ccs/ccs.jsp) or the 2015 Risk Adjustment model software HCC, RxHCC, ESRD of Centers for Medicare & Medicaid Services (https://www.cms.gov/Medicare/Health‐Plans/MedicareAdvtgSpecRateStats/Risk‐Adjustors) or the DCG.

For example, Pharmacy Cost‐Group (PCG).

Data setting scenarios to predict total health care expenditure Abbreviations: ESRD, End‐Stage Renal Disease; HCC, Hierarchical Condition Category; HDR, hospital discharge records; PD, pharmacy database; RxHCC, Prescription Drug Hierarchical Condition Category. For example, possible coding classification systems include the Clinical Classifications Software for ICD‐9‐CM (https://www.hcup‐us.ahrq.gov/toolssoftware/ccs/ccs.jsp) or the 2015 Risk Adjustment model software HCC, RxHCC, ESRD of Centers for Medicare & Medicaid Services (https://www.cms.gov/Medicare/Health‐Plans/MedicareAdvtgSpecRateStats/Risk‐Adjustors) or the DCG. For example, Pharmacy Cost‐Group (PCG). We use “range of variables” to refer to the data availability with respect to different health areas, and with “data granularity” we refer to the detail available. Age and sex are optimal exogenous risk‐adjusters because they cannot be influenced by health providers' actions. However, they do not accurately reflect the epidemiological variability and health needs among geographical areas (Atella et al., 2018). Since the late 1990s, the United States (US) and many European health systems expanded the range of variables to reduce risk selection with increasingly sophisticated risk‐adjustment formulas, including socio‐economic, pharmacy‐ and diagnosis‐based indicators (Ellis et al., 2018). Diagnosis information may be endogenous to some extent but is generally far more predictive than age and sex alone. In the UK, factors capturing demography, morbidity, deprivation, and unavoidable cost of providing services in different areas have been included in a complex person‐based formula to improve the allocation of the “fair share” for each area (Dixon et al., 2011). Early studies in the US demonstrated the inclusion of diagnosis cost groups (DCGs) improved the R 2 of the demographic model (age‐sex data) by 40% points in a concurrent model and by 10% points in a prospective model (Ellis et al., 1996; Pope et al., 2000). A study for allocating commissioning funds to general practices in England, showed that from a model with age and sex only, adding diagnostic‐related variables improved predictive power from 3.7% to 12.6% at individual level (Dixon et al., 2011). In 2002, the Dutch health system was the first to expand the range of variables with pharmaceutical data through the Pharmacy‐based Cost Group (PCG), a risk‐adjuster based on pharmaceutical use, which increased the R 2 from 4% in the demographic model to almost 9% (Lamers, 1999). In 2017, the sophisticated prospective Dutch risk‐adjustment model reached an R 2 of almost 32% (van Kleef et al., 2018). Other countries with competitive insurance markets have introduced morbidity indicators in their risk adjustment modeling, albeit with some limitations. Switzerland introduced PCGs into risk‐adjustment modeling after a study found that the inclusion of pharmaceutical indicators in the Swiss prospective formula increases the R 2 by 13% points (Schmid & Beck, 2016), while diagnostic information is still incorporated as a dummy variable (yes/no prior hospitalizations). Germany uses pharmaceutical data to validate or impute missing diagnoses, while the Netherlands does not include outpatient diagnoses in the risk‐adjustment systems (Bauhoff et al., 2017). Most US risk‐adjustment systems do not use pharmaceutical data as predictors for health care expenditure (except Marketplaces), although Wagner et al. showed the DxCG system (supported by Verisk, a commonly used proprietary risk adjustment systems for cost data), which computes risk score using pharmacy and hospital data, increased the R 2 estimates by 8 to 20% points compared to the base models without pharmacy data (Wagner et al., 2016). Despite the development of indicators based on diagnosis and pharmaceutical data, which have substantially improved the predictive performance of risk‐adjustment algorithms over the past 2 decades, data constraints are still present in many countries. In Italy, approximately 35% (outpatient service and half of hospital expenditure) of the Italian National Health Fund is distributed to Regions based on an age‐weighting system, while the remainder is allocated based on the unweighted resident population (Ferre et al., 2014). Ireland's health system relies on age, gender, and level of coverage as risk adjusters to determine transfers under risk equalization (Armstrong, 2018); Australian and Chilean risk equalization formulas are based on age, gender, and location (Paolucci et al., 2018; Velasco et al., 2018).

NEW RISK‐ADJUSTMENT ESTIMATION TECHNIQUES

Ordinary least squares (OLS) is the tool most commonly used in practice to predict healthcare spending (Ellis et al., 2018; Iezzoni, 2012). Recent literature has explored deploying ML methods, particularly in health systems with fine data and rich range of variables. The exponential growth of the size and complexity of electronic health information and the advancement in computational capacity has shifted the attention of health statisticians and econometricians towards newer techniques (Rose, 2016). These techniques allow for the automated investigation of all possible covariates available and to leverage the complex structure of population‐level health data (Kan et al., 2019). Rose (2016) compared OLS with 8 ML algorithms (lasso, ridge, elastic net, neural net, single tree, random forests (RF), and a discrete and weighted super learner (SL)) using the inputs of the official US individual Marketplace risk‐adjustment formula, including age, sex, geographic area, 5 inpatient diagnosis categories, and 74 Hierarchical Condition Category (HCC) variables, to predict total annual expenditures in a prospective formula. The results highlighted the minor improvement in performance of the SL with respect to cross‐validated R 2, but also showed that reducing the number of covariates to 10 only marginally reduced the predictive performance, providing preliminary evidence that few variables could yield effective plan payment risk‐adjustment. McGuire, Zink, and Rose (McGuire et al., 2020) examined payment system fit with a concurrent formula in the setting of the US individual Marketplaces. The study found a penalized regression with age and sex cells and 30 HCCs performed similarly to the baseline linear regression with age and sex cells and 90+ HCC‐based variables, while pharmaceutical covariates did not improve fit over the diagnosis‐based risk adjustors. In recent work, Kan et al. (2019) compared OLS and penalized linear regressions for predicting Medicare Advantage health care expenditure (patients aged ≥65 years). The authors emphasized the advantages of penalized regressions, particularly the lasso, which yielded improvements over OLS in a single year of data and a 4‐year pooled‐data setting. In non‐US settings, exploration of interaction terms has been common. Buchner et al. employed a regression tree analysis to improve the German risk‐adjustment formula with significant interactions between variables (Buchner et al., 2017). The inclusion of morbidity interaction terms showed a marginal improvement of R 2, with an insubstantial loss in accuracy. Examining the 2014 Dutch risk‐adjustment formula, van Veen et al. (van Veen et al., 2018) found that a similar approach improved the adjusted R 2, but concluded it may also increase the possibility of risk‐selection for some subgroups.

DATA

Our study population was drawn from the 2016 administrative databases of Emilia, a Northern Italian Region with about 4.4 million beneficiaries with universal access to the Italian NHS. The main databases used were: Hospital Discharge Records: Admissions and discharge dates; primary and up to five secondary diagnoses and up to six interventions (ICD‐9‐CM coding system) are included. Expenditure is registered using the diagnosis‐related group tariffs system since 1995 (Emilia‐Romagna, 2014). Outpatient Pharmaceutical Database (OPD): Contains drugs reimbursed by the NHS (prescribed by the primary care physician or a specialist, or directly dispensed by the hospital pharmacies) and details on substance name, Anatomical Therapeutic Chemical (ATC) classification system code‐V.2013, brand name, date of prescription filling, number of unit doses and number of packages and prescribers. Outpatient Speciality Database (OSD): Contains individual expenditures of laboratory test and specialistic visits. The linkage among these databases was possible through a unique anonymized patient identifier. We include only individuals who were residents in 2016 of one of the 8 Local Health Authorities of Emilia‐Romagna.

Socio‐demographic covariates

We included gender, age (divided into 20 classes: 0, 1‐4, 5‐year classes up to 90 years and a class open over 90 years), citizenship (dichotomous variable: Italian yes/no), the degree of urbanization of residence (classified using the Eurostat's Degree of Urbanization (DEGURBA) (Eurostat, 2014) classification system into sparsely populated areas, intermediate density areas, and densely populated areas) and the income derived from the exemptions (if an individual presented more than one exemption they were assigned to the lowest income class).

Hospital covariates

The hospital covariates included the total number of hospitalization and ICD‐9‐CM codes, which were grouped into DCGs. This method, originally proposed by Ash and colleagues (Ash et al., 1989; Ellis & Ash, 1995), first aggregates the diagnostic codes into 78 clinically homogeneous subgroups, then further aggregates these subgroups into 9 large groups on the basis of their equivalence in terms of resource absorption. Individuals with multiple diagnoses were classified only in their most expensive DCG while individuals without any diagnoses were grouped into DCG 0. Appendix Table A contains the full list of DCGs.

Pharmacy covariates

From the OPD, we included the total number of drug prescriptions and ATC codes, which we grouped into PCGs. Through the ATC coding system, individuals with pharmaceutical prescriptions in 1 year are marked with a medical condition, which is then used as risk adjuster for the distribution of the resources thereafter. For the construction of the PCGs, we combined the Dutch model and other approaches that use the ATC codes to identify patients with chronic condition from the pharmaceutical databases defining 36 distinct classes (Chini et al., 2011; Corrao et al., 2017; Trottmann et al., 2010; Trottmann et al., 2015). An individual is classified into a PCG if the year amount of defined daily doses (DDD) in a given pharmaceutical class was greater than 180 DDD (cancer drugs required only >15 DDD). The full list of PCGs is shown Appendix Table B.

Dependent variable

The outcome of interest is annual total expenditure, calculated by summing up the expenditures of hospital, pharmaceutical, and outpatient care services provided to residents of Emilia‐Romagna Region during 2016. We implemented a concurrent risk‐adjustment approach: health care expenditure was predicted by individual characteristics from the same year. The concurrent method, compared to the prospective, reduces the data burden by using only a single year of data and does not require a separate formula for individuals with no information from the prior year. Concurrent risk models generally yield higher predictive power than prospective models. The compromise of this higher predictiveness is a reduced incentive to control healthcare spending and a longer timeline for payment, adding uncertainty, administrative burdens, and planning challenges (Geruso & McGuire, 2016).

Characteristics of the population

A summary of socio‐demographic variables is presented in Table 2. There were slightly more women than men in the sample (52%) and the largest age groups were 50–54 (9.2%) and 45–49 (9.1%). Mean total healthcare expenditure was 834€. Women had a slightly lower mean expenditure compared to men (832 vs. 841€) but a much higher median expenditure (114 vs. 59€). Italians had higher mean expenditure than non‐Italians (861 vs. 530€) and people who lived in rural areas had higher expenditure (876€) than those living in urban (858€) or peri‐urban areas (796€). Low‐income exempted individuals had the highest mean and median expenditure (1161 and 226€) compared with other income groups. Average spending increased with more hospitalizations or prescriptions. Similarly, individuals with DCG were more expensive (5631€) than individuals with no DCG (227€), as well for individuals with at least one PCG (1729 €) compared to those with no PCG (406 €).
TABLE 2

Description of the study population (n = 4,262,982)

Total expenditure in 2016
N% ColMeanS. D.25th percMedian75th perc
Observations4,262,982100%834.03306.01387405
Male2,054,69548.2%838.23618.0659344
Female2,208,28751.8%829.82986.924114460
Age groups
048,6321.1%1358.14832.4128340416
1–4119,8042.8%234.61875.0113082
5–9157,5983.7%167.71339.272165
10–14153,2153.6%201.91448.762999
15–19144,9993.4%244.11873.152498
20–24148,6953.5%286.81870.232291
25–29184,7744.3%322.51795.3426111
30–34225,9615.3%358.41730.4531131
35–39279,3426.6%374.31925.0531138
40–44353,2498.3%384.82267.9232141
45–49389,1689.1%424.22359.5448172
50–54391,4619.2%528.02883.7457223
55–59313,1617.3%743.23579.018103337
60–64266,9006.3%1008.03978.940172497
65–69273,6826.4%1378.14351.7101318809
70–74224,1785.3%1771.54675.01734591145
75–79226,0815.3%2077.45020.82375681414
80–84173,6444.1%2277.64929.62876381690
85–89116,7052.7%2325.14592.52756172082
90+71,7331.7%2225.84002.12055062713
Italian citizenship
No351,8158.3%529.92761.0850220
Yes3,911,16791.7%861.23349.51392425
Degree of urbanization
Urban1,548,23136.3%857.83418.71184406
Peri‐urban1,896,50144.5%796.13191.11383385
Rural818,25019.2%875.93350.317102450
Income
Exempt low income1,746,67741.0%1161.13576.454226735
Exempt middle‐low income583,62413.7%720.92767.824114378
Exempt middle‐high income98,8322.3%696.82777.11794334
Exempt high income397,2339.3%797.93817.2038235
Exempt other reason1,436,61633.7%501.22994.1122119
No. of hospitalization
03,777,07388.6%226.3728.4961245
1380,6268.9%3768.65440.2136222974272
271,4251.7%9101.38904.13974684111,434
3+33,8580.8%18,174.615,304.0914514,35522,410
No. of pharmaceutical prescription
0931,69721.9%205.51726.902175
1876,85020.6%158.31244.20528
2326,0767.6%350.61895.81534116
3+2,128,35949.9%1461.34313.9115322904
Diagnostic cost group (DCG)
No3,784,03988.8%226.7730.9961245
Yes478,94311.2%5631.18194.2161430336466
Pharmacy cost group (PCG)
No2,884,26067.7%406.02302.7534131
Yes1,378,72232.3%1728.84639.11864461109

Abbreviations: DCG, Diagnostic cost group; PCG, Pharmacy cost group.

Description of the study population (n = 4,262,982) Abbreviations: DCG, Diagnostic cost group; PCG, Pharmacy cost group.

STATISTICAL ANALYSIS

Model specifications

As our baseline method, we implemented an OLS regression. To account for the highly skewed distribution of healthcare spending, we also estimated generalized linear models, using the frequently applied log‐link function and two distribution families (Gaussian and Gamma; Jones, 2010). We deployed the following ML algorithms: penalized regressions (lasso, ridge, and elastic net penalties), a generalized additive model (GAM), RF, and, finally, we constructed a SL, an ensembling method that takes a weighted average of predicted values from underlying candidate algorithms (such as penalized regressions, GAMs, and RFs) to produce a single best prediction function. (M. J. van der Laan & Rose, 2011; Mark J. van der Laan et al., 2007). For further background on these methods and their application to risk adjustment we refer the interested reader to Rose (2016).

Model performance

We randomly sampled 100,000 observations from our total sample. We divided this subset a training set (70%), to develop and validate the algorithms, and test set (30%) to evaluate the performance of the models (Appendix C reports summary statistics for the train and test sets). To choose the optimal tuning parameters of the penalized linear regressions, 10‐fold cross‐validation was used within the training set. For the RF, 500 trees were produced (node size = 250) and p/3 predictors, for each specification formula, were randomly chosen at each step of the tree‐building. Model fit was judged using the adjusted‐R2 and mean squared error (MSE). Predictive Ratios (PR) were computed as the ratio between mean predicted spending and mean observed spending in each observed quintile. We also calculated mean under/overcompensation, computed as the difference between mean predicted spending and the mean observed spending in each quintile. For MSE, a smaller value indicates better performance. An adjusted‐R2 a value closer to one indicates more explanatory power, while PR values between 0.9 and 1.1 are considered reasonable prediction accuracy (Kautter et al., 2012). The Relative Efficiency (RE) of each algorithm compared to OLS, with respect to both cross‐validated MSE (RE = cv MSE /cv MSE , k = other algorithms) and R2 values (RE = cv R2 /cv R2 , k = other algorithms), was evaluated. A total of 54 models were estimated to predict healthcare expenditure (based on six data scenarios and nine algorithms). Statistical analyses were performed using SAS Enterprise Guide (version 7.1) and R (version 3.6.3).

RESULTS

Evaluation of the models: Adjusted‐R2 and MSE

Tables 3 and 4report the best and worst performing models compared to OLS according to adjusted‐R2 and MSE, respectively. In all scenarios, SL displayed the best performance, except in the coarse granularity/poor range of variables setting, in which the lasso model performed best (in both adjusted‐R2 and MSE measures). However, in this setting, the performance difference between models was trivial, such that the RE was 1 across all models. Focusing on the adjusted‐R2, the gain in relative efficiency compared to OLS was higher in the coarse granularity/rich range of variables setting (SL adj.‐R2 = 49.9%) and in the fine granularity/poor range of variables (SL adj.‐R2 = 4.6%), but in both these settings there was no gain in RE. In the fine granularity/fair and rich range of variables scenarios, the SL RE was 1.0, with an adjusted‐R2 of 43.6% and 45.1% when DCG and DCG+PCG were added, respectively. We obtained inconsistent values (the predictions were so poor that the adjusted‐R2 were negative and the MSEs were extremely high) for the GLMs, denoting their poor performances. The lowest adjusted‐R2 of OLS was 4.3% in the coarse granularity/poor range of variables data while the highest was 48.5%, obtained in the coarse granularity/rich range of variables data.
TABLE 3

Adjusted‐R2 of the models compared to OLS

 Poor range of variablesFair range of variablesRich range of variables
Models rankingAdj. R 2 REModels rankingAdj. R 2 REModels rankingAdj. R 2 RE
Coarse granularity Lasso 4.3%1.0 SL 48.8%1.0 SL 49.9%1.0
ElNet 4.3%1.0 GAM 48.0%1.0 RF 49.1%1.0
RF 4.3%1.0 OLS 47.6% 1.0 GAM 48.9%1.0
SL 4.3%1.0 Lasso 47.6%1.0 OLS 48.5% 1.0
GLM l‐g 4.3%1.0 ElNet 47.6%1.0 Lasso 48.5%1.0
GLM l‐n 4.3%1.0 RF 47.5%1.0 ElNet 48.5%1.0
OLS 4.3% 1.0 Ridge 47.3%1.0 Ridge 48.2%1.0
GAM 4.3%1.0 GLM l‐n Inc. value GLM l‐n Inc. value
Ridge 4.3%1.0 GLM l‐g Inc. value GLM l‐g Inc. value

Abbreviations: GAM, generalized additive model; MSE, mean squared error; OLS, Ordinary least squares; RF, random forest; SL, super learner.

TABLE 4

Average cross‐validated MSE of the models compared to OLS

 Poor range of variablesFair range of variablesRich range of variables
Models rankingMSEREModels rankingMSEREModels rankingMSERE
Coarse granularity Lasso 10,657,7931.0 SL 5,705,0971.0 SL 5,583,5581.0
ElNet 10,657,7941.0 GAM 5,791,3261.0 RF 5,663,5591.0
RF 10,657,8131.0 OLS 5,830,424 1.0 GAM 5,689,2201.0
SL 10,657,8621.0 Lasso 5,830,7181.0 OLS 5,734,502 1.0
OLS 10,657,875 1.0 ElNet 5,830,8671.0 Lasso 5,735,4061.0
GAM 10,657,8751.0 RF 5,842,1051.0 ElNet 5,735,4681.0
GLM l‐n 10,657,8751.0 Ridge 5,863,9161.0 Ridge 5,766,7631.0
GLM l‐g 10,657,8751.0 GLM l‐n 13,446,6900.4 GLM l‐n 41,287,6820.1
Ridge 10,658,0651.0 GLM l‐g Inc. value GLM l‐g Inc. value

Abbreviations: GAM, generalized additive model; MSE, mean squared error; OLS, Ordinary least squares; RF, random forest; SL, super learner.

Adjusted‐R2 of the models compared to OLS Abbreviations: GAM, generalized additive model; MSE, mean squared error; OLS, Ordinary least squares; RF, random forest; SL, super learner. Average cross‐validated MSE of the models compared to OLS Abbreviations: GAM, generalized additive model; MSE, mean squared error; OLS, Ordinary least squares; RF, random forest; SL, super learner. Concerning MSE, in the coarse granularity/poor range of variables and fine granularity/fair and rich range of variables scenarios, there were no gains in RE of using SL compared to OLS. ML technique produced slight gain in the coarse granularity/fair range and rich range of variables scenarios, where SL reduced the MSE by 2.1% (MSE = 5,705,097; RE = 1.02) and by 2.6% (MSE = 5,583,558; RE = 1.03). The GLMs yielded extremely high MSE values and the worst performance in many scenarios. The worst MSE of OLS was in the coarse granularity/poor range of variables data (MSE = 10,657,875) while the best was in the coarse granularity/rich range of variables data (MSE = 5,734,502). We carried out further analyses, in which we compared the performance of the ML models, estimated on a subsample of the population as above, with that of the OLS model, estimated on the entire population of 4,262,982 individuals (“OLS total”). The results are presented in the appendix D and E. In the coarse and fine granularity/poor range of variables scenarios, OLS total outperformed the best ML algorithm for both adjusted‐R2 and MSE. Concerning MSE, the gap between the best ML model and the OLS total model narrows, until it reverses the sign in the fine granularity/rich range of variables scenario.

Under/overcompensation and predictive ratios

Figure 1 shows the mean under/overcompensation by observed spending quintile. Complete tables for each data scenario of mean under/overcompensation and PRs are reported in Appendix F and G. In the coarse granularity/poor range of variables scenario all models highly overcompensate the expenditure of the individuals in the 80th percentile group (best: ridge = 822.8; worst: OLS = 826.5) as well as in the fine granularity/poor range of variables (best: SL = 859.7; worst: OLS = 868.9). In the 99th percentile all models highly undercompensate. In contrast, the 99th percentile group is (in a lesser extent) overcompensated in the fair and rich range of variables settings, for both coarse and fine granularity. Finally, in the coarse granularity/rich range of variables setting we also see the lowest under/overcompensation relative to the other scenarios for the 20th percentile.
FIGURE 1

Under/overcompensation by quintiles for each data scenario. The ridge and elastic net mean compensations are not shown because they were very similar to lasso. GLMs are not shown because some of their mean compensation's values were completely out of range. Top 1% spenders were not included in the graph representation due to the higher range of values compared to quintiles (mean under compensation range from −24,971 to −14,606 €)

Under/overcompensation by quintiles for each data scenario. The ridge and elastic net mean compensations are not shown because they were very similar to lasso. GLMs are not shown because some of their mean compensation's values were completely out of range. Top 1% spenders were not included in the graph representation due to the higher range of values compared to quintiles (mean under compensation range from −24,971 to −14,606 €)

DISCUSSION

The recent developments in ML techniques to estimate healthcare spending have shown promising results, although as implemented in this study their performance is not uniformly dominant compared to OLS. The goal of this study is to provide a framework and toolkit for policymakers, therefore we selected well‐established ML algorithms that required minimal tuning or could be implemented “off the shelf.” ML methods are still relatively unfamiliar for many policymakers, and some algorithms, like those based on neural networks, suffer from problems of interpretability that may limit their application (Kan et al., 2019). This study provided an in‐depth analysis of the advantages and disadvantages of implementing ML techniques compared to standard regression (OLS) considering various scenarios of data granularity and range of variables and making use of healthcare administrative dataset from a large, unselected Italian population. Our findings show that ML techniques, particularly SL, outperformed OLS in all data scenarios, although the adjusted‐R2 RE ranges between 0.02% points in the coarse granularity/poor range of variables scenario and 2.8% points in the coarse granularity/rich range of variables scenarios, indicating no statistically significant gain in our sample of 100,000 observations. Performance based on MSE also showed consistent results. Concerning under/overcompensation measures, OLS showed the least, albeit negligible, overcompensation compared to ML techniques in the lowest observed spending percentiles in the coarse granularity/poor and fair range of variables setting. The top 1% spenders are better predicted with GAM or SL, except in the coarse granularity/poor range of variables and fine granularity/rich range of variables settings, where OLS was superior. The mean under/overcompensation also highlighted how the increase in the range of variables can considerably reduce the overestimation of the group of individuals in the 20th percentile and at the same time reduce the gap of the underestimation for the top of the spending distribution. However, the gap between the best ML and OLS is relatively narrow, indicating that the implementation of more complex and sophisticated models does not lead to a significant increase in under/overcompensation reduction. Unlike prior, US‐based studies (McGuire et al., 2020; Rose, 2016), we did not find a consistent relationship between the increase in data granularity and/or range of variables and the advantage of ML over OLS. However, our most detailed data scenario (fine granularity, rich range of variables) included only 9 diagnosis‐based risk adjustors, suggesting the ability of ML algorithms to outperform OLS with fewer variables may be limited to settings with even more granular diagnosis groupings, such as the US formulas that use 90+ diagnosis‐based risk adjustors. Increasing the range of variables produced a significant improvement in the ability of all models (excepting the GLMs) to predict healthcare expenditure, more so in the coarse than fine granularity scenarios. The jump in the adjusted‐R2 from the coarse granularity/poor range of variables to the coarse granularity/fair range of variables due the addition of the number of hospitalizations in the risk‐adjustment formula is greater than the increase moving to the fine granularity/rich range of variables scenario where PCG and DCG are added. We expected this result because the number of hospitalizations is much more predictive than demographic variables and it is highly correlated with the total expenditure in the concurrent approach. However, we encourage caution regarding its use in risk‐adjustment models since it may lead to inappropriate treatment decisions (Geruso & Layton, 2015). Furthermore, the underfunding of high‐cost morbidities that do not require hospitalization (e.g., outpatient dialysis for chronic kidney disease patients) may still present a challenge for policymakers. Similar to McGuire, Zink, and Rose (McGuire et al., 2020), we find the addition of pharmaceutical‐based indicators (moving from fair to rich range of variables in either granularity level) does not have a large impact on the adjusted‐R2. Although the fine granularity variables DCG and PCG reduce the adjusted‐R2, they may be regarded as reasonable trade‐offs since they generally increase efficiency and fairness, and also reduce incentives for risk selection by including more risk‐classes and are better predictors in prospective approaches. Nevertheless, in addition to the upcoding behavior that morbidity‐based indicators may induce, Geruso and McGuire (Geruso & McGuire, 2016) observed that HCC‐based concurrent risk‐adjustment systems increase fit but at the same time they reduce the power (the term power describes the share of costs at the margin born by the health plan (Geruso & McGuire, 2016) in cost containment by about 30%, meaning that concurrent diagnosis‐based risk‐adjusters weaken incentives for cost control. The inclusion of pharmaceutical‐based indicators in the risk‐adjustment model may incentivize health care providers to prescribe unnecessary medication (van Kleef et al., 2014) but, in the long‐term, PCGs may direct providers' effort to cost‐containing innovations (Beck et al., 2010). We note some specific limitations to our study. There were many alternative fine granularities diagnosis risk‐adjustors that we could have selected for this comparison. We limited our choice to DCG and PCG because they are well known in the risk‐adjustment literature and commonly adopted for utilization and cost outcomes. However, we believe the methods we present can guide future comparisons of other risk‐adjustors. Similarly, many other ML algorithms could be compared, for example, neural networks, other decision tree‐based methods, or different choices could have been made regarding the tuning parameters (Hastie et al., 2009; James et al., 2013). Ultimately, our choice of algorithms was driven by a desire for both interpretability and potential familiarity for practitioners. Concerning data sources, the OPD includes only drug dispensations that are reimbursed by the NHS. The same caveat holds for OSD. Thus, we did not capture healthcare use paid by patients either directly or funded by private insurers. Finally, we relied on combining multiple data sources to select our base population, because we lacked access to the central Resident's Registry. The choice of the model depends crucially on the performance measures; thus, it is of foremost importance to compare various measures of fit and interpret the results very carefully. It is more appropriate to rely on the adjusted‐R2 and MSE to maximize the variance explained and reduce the overall prediction error. However, if the primary goal is to minimize the differences between the predicted values and the observed values in specific subgroups that may be at risk of selection and to incentivize cost control, then the under/overcompensation or the PR are more targeted measures. Policymakers should consider other evaluation criteria in designing the risk‐adjustment model in addition to models' predictive performance, such as time/computing power, availability of data or appropriateness for incentives for risk selection and efficiency (van Veen et al., 2015). Generally, ML techniques – particularly the SL – out‐performed OLS. From an operational perspective, we suggest implementing ML methods to increase variance explained by the model in predicting the expenditure to further decrease risk selection, to reduce upcoding actions by selecting a subset of variables over all possible covariates, and to explore possible interaction terms with automatic processes (Rose, 2016). Retaining only a small number of important variables could reduce the opportunities to inflate the number of diagnosis and facilitate care management by emphasizing key risk‐factors (Kan et al., 2019; Kronick & Welch, 2014). Conversely, if algorithm performance differences on the desired evaluation metrics are considered negligible and the investment in these tools too onerous in terms of computational time, OLS may be used with large populations. In our further analyses, when the entire population is considered rather than a subsample to estimate the OLS model, we indeed observed a narrowing of the gap in the MSE measure between ML models and the OLS. To summarize, it is well known that in designing a risk‐adjustment model it is important to evaluate not only the predictive performance but also the trade‐offs with adverse economic behaviors (e.g., risk selection, moral hazard, cost containment). Our analysis shows a similar trade‐off when deciding between the investment in increasing the range of variables and the granularity of the data, and in the search of the more appropriate statistical techniques to implement in conjunction with a given data scenario.

CONFLICT OF INTEREST

The authors have declared no conflict of interest. Supplementary Material S1 Click here for additional data file.
  28 in total

1.  Risk adjustment in health insurance and its long-term effectiveness.

Authors:  Konstantin Beck; Maria Trottmann; Peter Zweifel
Journal:  J Health Econ       Date:  2010-04-02       Impact factor: 3.883

2.  Regression Trees Identify Relevant Interactions: Can This Improve the Predictive Performance of Risk Adjustment?

Authors:  Florian Buchner; Jürgen Wasem; Sonja Schillo
Journal:  Health Econ       Date:  2015-10-26       Impact factor: 3.046

3.  Super learner.

Authors:  Mark J van der Laan; Eric C Polley; Alan E Hubbard
Journal:  Stat Appl Genet Mol Biol       Date:  2007-09-16

4.  Improvements in Medicare Part D risk adjustment: beneficiary access and payment accuracy.

Authors:  John Kautter; Melvin Ingber; Gregory C Pope; Sara Freeman
Journal:  Med Care       Date:  2012-12       Impact factor: 2.983

5.  Measuring coding intensity in the Medicare Advantage program.

Authors:  Richard Kronick; W Pete Welch
Journal:  Medicare Medicaid Res Rev       Date:  2014-07-17

6.  Tradeoffs in the design of health plan payment systems: Fit, power and balance.

Authors:  Michael Geruso; Thomas G McGuire
Journal:  J Health Econ       Date:  2016-02-10       Impact factor: 3.883

7.  A person based formula for allocating commissioning funds to general practices in England: development of a statistical model.

Authors:  Jennifer Dixon; Peter Smith; Hugh Gravelle; Steve Martin; Martin Bardsley; Nigel Rice; Theo Georghiou; Mark Dusheiko; John Billings; Michael De Lorenzo; Colin Sanderson
Journal:  BMJ       Date:  2011-11-22

8.  Principal inpatient diagnostic cost group model for Medicare risk adjustment.

Authors:  G C Pope; R P Ellis; A S Ash; C F Liu; J Z Ayanian; D W Bates; H Burstin; L I Iezzoni; M J Ingber
Journal:  Health Care Financ Rev       Date:  2000

9.  Diagnosis-based risk adjustment for Medicare capitation payments.

Authors:  R P Ellis; G C Pope; L Iezzoni; J Z Ayanian; D W Bates; H Burstin; A S Ash
Journal:  Health Care Financ Rev       Date:  1996

10.  Adjusting Medicare capitation payments using prior hospitalization data.

Authors:  A Ash; F Porell; L Gruenberg; E Sawitz; A Beiser
Journal:  Health Care Financ Rev       Date:  1989
View more
  1 in total

1.  Comparing risk adjustment estimation methods under data availability constraints.

Authors:  Marica Iommi; Savannah Bergquist; Gianluca Fiorentini; Francesco Paolucci
Journal:  Health Econ       Date:  2022-04-05       Impact factor: 2.395

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.