Literature DB >> 35965972

A study on quality control using delta data with machine learning technique.

Yufang Liang1, Zhe Wang2, Dawei Huang1,3, Wei Wang4, Xiang Feng2, Zewen Han2, Biao Song2, Qingtao Wang1,5, Rui Zhou1,5.   

Abstract

Background: In the big data era, patient-based real-time quality control (PBRTQC), as an emerging quality control (QC) method, is expanding within the clinical laboratory industry. However, the main issue of current PBRTQC methodology is data stability. Our study is aimed to explore a novel protocol for data stability by combining delta data with machine learning (ML) technique to improve the capacity of QC event detection.
Methods: A data set of 423,290 laboratory results from Beijing Chao-yang Hospital 2019 patient results were used as a training set (n = 380960, 90%) and internal validation set (n = 42330, 10%). A further 22,460 results from Beijing Long-fu Hospital 2019 patient results were used as a test set. Three-type data (1) Single-type data processed by truncation limits; (2) delta-type data processed by truncation limits and (3)delta-type data processed by Isolated Forest (IF) algorithm were evaluated with accuracy, sensitivity, NPed, etc., and compared with previously published statistical methods.
Results: The optimal model was based on Random Forest (RF) algorithm by using delta-type data processed by IF algorithm. The model had a better accuracy (0.99), sensitivity (0.99) specificity (0.99) and AUC (0.99) with the dependent test set, surpassing the critical bias of PBRTQC by over 50%. For the LYMPH#, HGB, and PLT, the cumulative MNPed of MLQC were reduced by 95.43%, 97.39%, and 97.97% respectively when compared to the best of the PBRTQC.
Conclusion: Final results indicate that by integrating an innovative ML algorithm with the overall data processing protocol the detection of QC events is improved.
© 2022 Published by Elsevier Ltd.

Entities:  

Keywords:  Data processing; Delta data; Machine learning; Patient-based real-time quality control; Random forest

Year:  2022        PMID: 35965972      PMCID: PMC9363967          DOI: 10.1016/j.heliyon.2022.e09935

Source DB:  PubMed          Journal:  Heliyon        ISSN: 2405-8440


Introduction

Laboratory test results play a crucial role in disease screening, diagnosis, prognosis evaluation, treatment monitoring. Traditional QC, due to non-commutable control materials measured, is not treated as real patient samples [1, 2], leading to lower error detection and higher false alarms. PBRTQC, as a newly QC method, because of real testing results used for QC, the commutability issue reduces, however, data stability is outstandingly cumbersome. Several studies have reported methods for data stability. Xincen Duan et al. [3] used the residual of the regression model as the input for improving univariate statistical process control (SPC) algorithms. Ng et al [4]. developed subpopulation protocols for hospitalized and ambulatory patients by adding additional clinical information to improve method performance. Ichihara et al. [5]. set up the weighted cumulative delta-check (wCDI) method in steps of data transformation, data standardization and index adjustment, for data pre-processing used for detecting specimen mix-up. Cembrowski GS, et al. [3, 6] combined delta check (DC) and moving average (MA) algorithms, developed an average of deltas (AoD) strategy, used for monitoring the mean delta of consecutive, intra-patient results to detect systematic error. In their protocol, a simulated annealing algorithm was used to select the number of patient delta values to calculate the average delta and to determine truncation limits to eliminate the effect from large deltas. DC, a laboratory information system (LIS)-based quality tool, involves the calculation and evaluation of sequential patients differences, to detect errors derived from total testing process (TTP). The increase of delta values may mainly arise from intra-patient variation, from analytical variation (e.g., instrument) or other out-of analytical variation [7] Previous studies found that delta check was used for error detection, its capability was limited, exclusively suited to detect larger analytic errors or mislabeled specimens [8, 9, 10, 11]. Machine learning (ML), one branch of Artificial Intelligence (AI), is prevailing in various application fields in recent years. It enables users to construct a related learning system in accordance with specific tasks. To be exact, ML can learn useful information from the unknown characteristics of a process and environment, and apply the learned information to develop a prediction, a classification and inform the decision-making process for new unknown problem in the future. ML differs from traditional statistical methods in that ML can cross industry boundaries and summarize solutions to problems that cannot be resolved by simple functions. Our study is aimed to, using delta data, combined with ML technique, explore a new approach for data stability, thus improving the capacity of QC event detection. Our study is aimed to assess explore a novel protocol for data stability by combining delta data with machine learning (ML) technique to improve the capacity of QC event detection. To verify the improvement effect of data stability, we compared both delta-type data and single-type data which were processed by truncation limits in PBRTQC based on statistical method, and further compared delta-type data processed by Isolated Forest in ML and in PBRTQC.

Materials and methods

Data collection

Our data was divided into training set, validation set and test set. The validation set can be understood as a part of the training set used for monitoring the model training process. All data were obtained from two Beijing’s Hospitals. 423290 results measured on XN-9000 (Sysmex, Kobe, Japan) were extracted through the laboratory information system (LIS) of Beijing Chao-yang Hospital in 2018, of which the data in the first 10 months were used for model training and the latter two months were used for model validation. 22460 results measured on BC-5390 (Mindray, Shenzhen, China) in the same time interval were obtained from another hospital, Beijing Long-fu Hospital for model testing. The data were filtered by rules, including: (1) patients with only one result in the study interval were excluded; (2) according to Tukey’s standard [12] the values less than the overall 25% quantile or greater than the overall 75% quantile were removed as outliers; (3) after Tukey’s standard was applied, patients with less than two results were excluded; (4) that the age was 14–60 years old was included; (5) and delta check interval was defined as one year [13]. All data were collected in-control status. Only seven representative test items were selected, including lymphocyte count (lymph#), lymphocyte ratio (lymph%), hemoglobin volume (HGB), mean hemoglobin volume (MCH), mean hemoglobin concentration (MCHC), red blood cell volume distribution width (R–CV) and platelet count (PLT). The reason for selecting these test items was that they represented different degrees of variation in leukocyte, erythrocyte and platelet series, and the degree of variation was based on their respective GVi/CVg rates. For a pair of results for each patient, one was used for obtaining delta value by calculating the difference between the pair of results in experiment 2 and 3, the other for selecting the second result each patient in experiment 1. Experiment 1–3 were described as followed.

Data stimulation

The data filtered by rules were regarded as unbiased data, in order to simulate out-of-control status in the real setting, we artificially introduced 10 biases of different sizes according to the formula below:where n refers to different multiples (−3/2, -1, -1/2, -1/4, -1/6, 1/6, 1/4, 1/2, 1, 3/2), represents the mean of all data for each test item, TEa represents the total allowable error for each test item which is defined as the sum of random error (RE) and systematic error (SE). For delta and single data sets, the simulation methods of biased data were different: 1) for the delta data set, a goal bias was only introduced into the second result of each patient, and then the biased delta was obtained by calculating a difference with the first result of this pair of patient results; 2) for the single data set, a goal bias was directly introduced on unbiased data as biased data. In this paper, a bias represented a shift in the mean.

MLQC based on delta-type data

Data pre-processing

Isolation Forest (IF) algorithm was used for pre-processing of delta data. IF is an unsupervised anomaly detection method commonly used for continuous numerical data [14] Its principle is to set up multiple isolation trees (iTree), each of which belongs to a binary tree structure to cut in the data space. As usual, the probability of data occurrence is extremely rare in the area with sparse spatial distribution, thus if the data falling in these areas can be considered as an abnormal value [15].

Model construction

Unbiased data and biased data in our study were defined as two kinds of data in ML, and they were distinguished by Random Forest (RF). RF classifier, an integrated supervised learning, is consists of multiple decision trees, each decision tree will give its own prediction, and the final prediction is given by way of voting [14]. Patient data was partitioned as a block size. A block size was taken as a whole. which was called a “machine learning sample”. The input feature of RF algorithm model was a sequence composed of multiple values. If the step size was 10, it indicated that the input dimension was 10. This not only introduced the serialization feature, but also met the multi feature requirements of machine learning algorithm. The following described the process of determining the step size and RF algorithm parameters: Firstly, take an example of one test item, when RF algorithm parameters in the program were set as default values, a block size needed to be pre-defined. A traversal experiment in a range of 5–20 in step of 1 was carried on to determine a proper block size. Secondly, all data included were partitioned by the block size determined, then a new “ML sample” was formed. Thirdly, ML samples were standardized. The samples composed of unbiased data was labeled “0″ and the samples composed of biased data as “1”. The same experimental steps above were implemented as unbiased data and biased data. In the following, the RF algorithm parameters were optimized by adjusting the numbers of-trees (N-trees) and max depth of trees (Max Depth). The search scope for N-trees from 100 to 500 in step of 100, and for Max Depth from 50 to 300 in step of 50. The accuracy of training and testing results for each combination was analyzed, to select the best parameter combination.

PBRTQC based on statistics

Two PBRTQC algorithms, moving average (MA) and moving standard deviation (MovSD) by only using single-type patient data, were selected as comparative methods. Then, both PBRTQC by using delta-type patient data and ML QC using delta-type patient data were regarded as comparing methods, one to estimate the ability of QC event prediction by using single-type data and delta-type data separately, the other to compare the efficiency of data filtering by the way of transaction limits recommended by The International Federation of Clinical Chemistry and Laboratory (IFCC) [16, 17] and by ML IF algorithm. For PBRTQC, the optimization of parameters strictly followed the experimental process recommended by IFCC. In the first step, the input data for PBRTQC was filtered by truncation limits, like data filtering of IF in ML, the minimum and maximum values were removed after sorting the original data. In order to reduce the influence of noise caused by extreme values, six data truncation limits were explored (0%, 1%, 5%,15%, 20% and 40%); then the data filtered were dealt with or without Box-Cox transformation.

PBRTQC by using single-type data

For PBRTQC by using single-type patient data, the data were firstly filtered by reference intervals of test items. Both MA and MovSD [18]algorithms were experimented. The optimal range of block size was selected from 10, 30, 50, 90, 110, 130 to 150. the mean and standard deviation (SD) of the two PBRTQC algorithms were calculated for each block size. All patient results were divided into 20 virtual days according to the original time sequence, and 1150 data were allocated every day, inclu-ding the first 150 unbiased data and the last 1000 biased data. The control limits based on three calculation methods provided by IFCC, namely: symmetric, all PBRTQC and daily extremes were calculated for each test item [16]. For the symmetric method, we choose two distances: 2.5-time coefficient variance (CV) (equivalent to 2.5-time SD of PBRTQC results, 3-time CV (equivalent to 3-time SD of PBRTQC results). All parameters above mentioned were combined and experimented. The best combined results were obtained.

PBRTQC by using delta-type data

To ensure the comparability with PBRTQC based on single-type patient data, the same experimental steps as method 4.1 were followed. The only change in this experiment was that the input data was replaced by delta-type patient data.

Evaluation metrics

Four process indicators: true positive (TP), true negative (TN), false positive (FP) and false negative (FN), were recorded, and then five indicators in the confusion matrix commonly used were evaluated the performance of our model, namely: area under ROC curve (AUC), true positive rate (TPR), true negative rate (TNR), false positive rate (FPR), false negative rate (FNR), and accuracy (ACC), here TPR was equivalent to sensitivity, TNR was equivalent to specificity. When FPR <5%, the number of patients affected (NPed) from the beginning of a bias introduced to the bias detected was used to evaluate the clinical performance of PBRTQC and MLQC. The mean (ANPed), median (MNPed) and 95% quantile (95 NPed) of NPed on 20 virtual days were used as clinical performance indicators. The minimum value of accumulative MNPed was the optimal result. Data processing and model analysis were performed in Python 3.7.3 package. The model training process depends on “numpy”, “Pandas” and other tool kits. Figure 1 showed the integrated experimental process diagram.
Figure 1

Integrated experimental process diagram.

Integrated experimental process diagram.

Results

Data description

For the seven test items included, three different data pre-processing methods were used. As seen in Table 1, when single-type data was converted to delta-type data, the mean and SD of the data narrowed significantly, and the concentration and stability of the data were also increased. In terms of the values adopting different quartiles, the distribution was more uniform to the results of delta-type data than to that of single-type data, although the sign introduced by delta-type data increased the value threshold for some test items. Further, the data was more concentrated when filtered by IF based on delta-type data. Taking LYMPH# as an example, the SD of delta-type data after dealt with by IF was reduced by 67.03% compared with the single-type data processed by truncation limits, and the reduction rates of the remaining six test items were 55.62%, 72.45%, 84.67%, 66.99%, 87.98%, and 71.81%, respectively. To visualize of data distribution characteristics of the three types of data, principal component analysis (PCA) technique was used to show the difference between biased data at the critical level and unbiased data, when a block size was set as 10. In Figure 2, the 3 rows from left to right representing LYMPH #, HGB and PLT, and the 3 columns from top to bottom representing single data, delta delta, and delta by IF. Every point in each diagram represented a ML sample consisting of 10 patient data as a block size. The results proved that the separability increased when the single-type data sequentially was processed by delta and then IF.
Table 1

Data analysis of the 3 data types for the seven test items.

Test itemAlgorithmMeanSDMin25th50th75thMax
LYMPH#Single2.07220.98550.24001.44001.86002.45009.4700
Delta-0.01380.7819-3.6900-0.3600-0.02000.270011.0600
IF-0.00760.3249-0.6800-0.26000.00000.25000.6000
LYMPH%Single31.364511.77262.900023.300030.200037.300077.8000
Delta-1.80328.8492-40.9000-6.5000-1.50003.300034.2000
IF-0.21615.2249-10.6600-4.4000-0.16004.00009.5600
HGBSingle129.681519.562656.0000118.0000131.0000143.0000184.0000
Delta2.016211.2371-64.0000-4.00002.00007.000068.0000
IF0.88135.3897-10.4000-4.00001.00005.000011.6000
MCHSingle29.56752.709517.900028.300029.800031.200039.8000
Delta0.09410.9510-7.3000-0.30000.10000.50007.3000
IF-0.01400.4154-0.8600-0.30000.00000.30000.8000
MCHCSingle325.817813.1967268.0000318.0000326.0000335.0000364.0000
Delta0.859612.1156-48.0000-7.00001.00007.000075.0000
IF1.04804.3566-9.5000-3.00001.00005.00009.2000
RCVSingle13.56932.129710.920012.300013.040014.070030.4000
Delta-0.02111.4696-10.9000-0.49000.00000.400018.1600
IF0.02430.2560-0.4000-0.20000.00000.20000.5300
PLTSingle235.435982.772730.0000182.0000224.0000275.0000742.0000
Delta2.848957.2888-243.0000-21.00004.000027.0000456.0000
IF1.415123.3346-42.0000-17.00001.000020.000048.0000

Single - single-type data; Delta - delta-type data pre-processed by different truncation limits based on statistical method; IF - delta-type data pre-processed by IF based on ML method; Mean - average value; SD - standard deviation; Min - minimum value; 25 - 25 quartile; 50 - 50 quartile; 75 - 75 quartile; Max - maximum value.

Figure 2

Data separability between critical biased and unbiased data for three-type data by PCA. The 3 rows from top to bottom represented LYMPH #, HGB and PLT of Cell Blood Count, the 3 columns from left to right represented single-type data, delta-type delta, and delta-type data processed by IF. Every point in each diagram represented a ML sample with the same block size consisting of 10 patient raw data.

Data analysis of the 3 data types for the seven test items. Single - single-type data; Delta - delta-type data pre-processed by different truncation limits based on statistical method; IF - delta-type data pre-processed by IF based on ML method; Mean - average value; SD - standard deviation; Min - minimum value; 25 - 25 quartile; 50 - 50 quartile; 75 - 75 quartile; Max - maximum value. Data separability between critical biased and unbiased data for three-type data by PCA. The 3 rows from top to bottom represented LYMPH #, HGB and PLT of Cell Blood Count, the 3 columns from left to right represented single-type data, delta-type delta, and delta-type data processed by IF. Every point in each diagram represented a ML sample with the same block size consisting of 10 patient raw data.

MLQC results by using delta-type data

Take RCV as an example, the influencing degree of the block size to the ability of QC event detection was explored by RF algorithm in ML. Here RCV was selected as the arithmetic example from the seven test items included because its within-individual and between individual ratio (CVi/CVg) was the largest, and this indicated that the degree of variation of this test item was considered relatively complex. The biased data at the critical level and unbiased data were used for model training. When RF algorithm parameters were set as default values, the block size was gradually increased starting from 5, and the AUC value of the model also increased; however, the change trend in AUC was no longer significant for block sizes above 10, so 10 was used as the block size for all subsequent experiments. When further adjusting the RF parameters, the search range of the parameter of N-trees was from 100 to 500 in step of 100; for each N-trees, the Max Depth was set from 50 to 300 in step of 50, and the accuracy of corresponding training and testing was counted in each group of experiments. The following Table 2 showed some data of the traversal experiment process, and the results showed that when the N-trees was 300 and the Max Depth was 200 was the optimal parameter combination.
Table 2

AUC for different block sizes and the results of RF tuning for R–CV at the critical bias.

Block sizeAUCN-treesMax DepthTraining accuracyTesting accuracy
50.91221001500.900.89
60.93522001000.910.92
70.94492003000.920.93
80.95563001000.960.94
90.97683003000.930.93
100.98624001000.940.93
110.98414003000.940.93
120.98325001000.940.93
130.98655003000.950.92

N-trees - number of trees; Max Depth - max depth of trees.

AUC for different block sizes and the results of RF tuning for R–CV at the critical bias. N-trees - number of trees; Max Depth - max depth of trees.

The performance of QC event prediction for three-type data

For each test item, three types of data were experimented, namely: (1) PBRTQC by using single-type data processed by statistical truncation limits; (2) PBRTQC by using delta-type data processed by statistical truncation limits; and (3) MLQC by using delta-type data processed by IF, where (1) and (2) experiments included 2 different algorithms, MA and MovSD, respectively. For MA and MovSD algorithms, 280 permutations derived from 5 truncation limits, 7 block sizes, 4 control limits, and 2 data transformation, respectively, were experimented. Due to paper space limitation, take examples of LYMPH #, HGB and PLT of CBC, Figure 3A, B, C showed the distribution feature of the training set and the test set in the MLQC. The coverage of the two sets was consistent for the three items in the scatter plots of A-C.
Figure 3

The visualization of data distribution feature for the training and the test sets and the performance parameters of five experiments at critical bias for LYMPH #, HGB and PLT. A-C take examples of LYMPH #, HGB and PLT ordered from left to right, represented principal component analysis (PCA) plots of the training set and internal validation set. D represented the TPR, TNR, FPR, FNR and ACC of the five algorithms (TPR - true positive rate; TNR - true negative rate; FPR - false positive rate; FNR - false negative rate; ACC - accuracy). E represented ANPed, MNPed, 95NPed of them (ANped - average of Nped; MNped - median of Nped; 95Nped - 95 quantile of Nped).

The visualization of data distribution feature for the training and the test sets and the performance parameters of five experiments at critical bias for LYMPH #, HGB and PLT. A-C take examples of LYMPH #, HGB and PLT ordered from left to right, represented principal component analysis (PCA) plots of the training set and internal validation set. D represented the TPR, TNR, FPR, FNR and ACC of the five algorithms (TPR - true positive rate; TNR - true negative rate; FPR - false positive rate; FNR - false negative rate; ACC - accuracy). E represented ANPed, MNPed, 95NPed of them (ANped - average of Nped; MNped - median of Nped; 95Nped - 95 quantile of Nped). In Figure 3D, E, for the LYMPH#, HGB, and PLT, the experimental results for the positive bias at the critical level (equivalent to adding a new factor n with values between 1/4 and 1/2 to the previous simulation equation) showed that the model had a better accuracy (0.99), sensitivity (0.99), specificity (0.99) in the in dependent test set. Its performance surpassed over 50% by the optimal method by MA and MovSD. The MNPed of the ML QC results using delta-type data were all within 10, which were reduced by 93.06%, 90.72%, 93.20%, respectively, relative to the best results of PBRTQC. The remaining of test items, the tuning parameters, and their performance in different experimental conditions at the critical biases, were shown in Tables 3, 4, and 5.
Table 3

Test results of 5 algorithms at the critical level in leucocyte lineage.

Test itemAlgorithmTL (%)Transfor-mationBSCLCL_lCL_UTPRTNRFPRFNRACCANPedMNPed95NPed
LYMPH#Single-MA1BC130daily extremes1.69121.97150.08991.00000.00001.00000.5335736.669311001100
Single-MovSD5BC50daily extremes0.37410.63510.38760.99550.00450.99550.6741599.32285731100
Delta-MA10-110daily extremes-0.43560.38400.75720.99450.00550.99450.8704145.333398451
Delta-MovSD15-902.5CV0.83711.77680.97500.96100.03900.96100.968577.554272166
Delta-MLProcessing10RF-model--0.99501.00000.00001.00000.99575.000056
LYMPH%Single-MA10neat130daily extremes28.464032.64760.56140.99890.00110.99890.7676591.45674941100
Single-SD10neat110daily extremes4.75169.43910.36860.99890.00110.99890.6691635.51186571100
Delta-MA5-903CV-6.30655.25030.87711.00000.00001.00000.9343330.3810237951
Delta-SD1-130daily extremes9.727022.45940.95900.99660.00340.99660.976563.956557140
Delta-MLProcessing10RF-model--0.96780.98470.01530.98470.97004.000045

TL - truncation limit; BC - Box–Cox transformation; BS-block size; CL_l - Control limit_lower; CL_U - Control limit _upper; TPR - true positive rate; TNR - true positive rate; FPR - false positive rate; FNR - false negative rate; ACC - accuracy; ANped - average of Nped; MNped - median of Nped; 95Nped - 95 quantile of Nped.

Table 4

Test results of 5 algorithms at the critical level in erythrocyte lineage.

Test itemAlgorithmTL (%)Transfor-mationBSCLCL_lCL_UTPRTNRFPRFNRACCANPedMNPed95NPed
HGBSingle-MA10BC130daily extremes125.9305130.81000.14591.00000.00001.00000.5622842.086611001100
Single-MovSD5neat50daily extremes2.403514.53980.07091.00000.00001.00000.5286602.76195511100
Delta-MA15-110daily extremes-3.89185.14180.69430.95960.04040.95960.8193570.32284691100
Delta-MovSD5-30daily extremes11.440526.41590.91911.00000.00001.00000.9568103.146797236
Delta-MLProcessing10RF-model--0.99120.99970.00030.99970.99238.8500911
MCHSingle-MA15BC903CV29.238030.25430.43860.97220.02780.97220.7015643.42526431100
Single-MovSD15Neat30daily extremes1.25042.12720.47850.95580.04420.95580.7136547.81894541100
Delta-MA5-302.5CV-1.15581.24300.16580.99390.00610.99390.5759568.59524261100
Delta-MovSD1-203CV1.40544.10350.40860.99890.00110.99890.6900196.8989136494
Delta-MLProcessing10RF-model--0.99090.99230.00770.99230.99118.3500812
MCHCSingle-MA20BC130daily extremes329.3763334.20940.54850.99900.00100.99900.7704630.46465591100
Single-MovSD0Neat303CV6.795311.72940.58840.99900.00100.99900.7907567.58274291100
Delta-MA1-303CV-9.15369.66390.40660.99890.00110.99890.6890268.1905200669
Delta-MovSD5-90daily extremes5.672410.62320.95700.99770.00230.99770.976083.312583120
Delta-MLProcessing10RF-model--0.99130.99300.00700.99300.99158.7500911
R–CVSingle-MA10BC90daily extremes12.582313.14400.38660.96740.03260.96740.6697476.04722931100
Single-MovSD5BC503CV0.19281.16640.40260.95680.04320.95680.6756187.357194921
Delta-MA20-30daily extremes-0.45300.43320.22480.96300.03700.96300.5884473.93702831100
Delta-MovSD1-302.5CV0.89073.15280.79220.99780.00220.99780.890283.440476185
Delta-MLProcessing10RF-model--0.98970.98430.01570.98430.989010.30001012

TL - truncation limit; BC - Box–Cox transformation; BS-block size; CL_l - Control limit_lower; CL_U - Control limit_ upper; TPR - true positive rate; TNR - true positive rate; FPR - false positive rate; FNR - false negative rate; ACC - accuracy; ANped - average of Nped; MNped - median of Nped; 95Nped - 95 quantile of Nped.

Table 5

Test results of 5 algorithms at the critical level in platelet lineage.

Test itemAlgorithmTL (%)Transfor-mationBSCLCL_lCL_UTPRTNRFPRFNRACCANPedMNPed95NPed
PLTSingle-MA15neat1103CV212.2105242.04520.07591.00000.00001.00000.5312525.95284401100
Single-MovSD15BC30daily extremes4.792519.77850.07791.00000.00001.00000.5322498.08663721100
Delta-MA0-302.5CV-48.877346.94430.06691.00000.00001.00000.5266482.61903921100
Delta-MovSD1-303CV64.3340178.19320.82020.99660.00340.99660.9033126.3592103303
Delta-MLProcessing10RF-model--0.98940.99370.00630.99370.99007.4500711

TL - truncation limit; BC - Box–Cox transformation; BS-block size; CL_l - Control limit_lower; CL_U - Control limit upper; TPR - true positive rate; TNR - true positive rate; FPR - false positive rate; FNR - false negative rate; ACC - accuracy; ANped - average of Nped; MNped - median of Nped; 95Nped - 95 quantile of Nped.

Test results of 5 algorithms at the critical level in leucocyte lineage. TL - truncation limit; BC - Box–Cox transformation; BS-block size; CL_l - Control limit_lower; CL_U - Control limit _upper; TPR - true positive rate; TNR - true positive rate; FPR - false positive rate; FNR - false negative rate; ACC - accuracy; ANped - average of Nped; MNped - median of Nped; 95Nped - 95 quantile of Nped. Test results of 5 algorithms at the critical level in erythrocyte lineage. TL - truncation limit; BC - Box–Cox transformation; BS-block size; CL_l - Control limit_lower; CL_U - Control limit_ upper; TPR - true positive rate; TNR - true positive rate; FPR - false positive rate; FNR - false negative rate; ACC - accuracy; ANped - average of Nped; MNped - median of Nped; 95Nped - 95 quantile of Nped. Test results of 5 algorithms at the critical level in platelet lineage. TL - truncation limit; BC - Box–Cox transformation; BS-block size; CL_l - Control limit_lower; CL_U - Control limit upper; TPR - true positive rate; TNR - true positive rate; FPR - false positive rate; FNR - false negative rate; ACC - accuracy; ANped - average of Nped; MNped - median of Nped; 95Nped - 95 quantile of Nped. In Figure 4 A, B, C, the different colored lines indicated the MNPed for each experiment, and the corresponding colored area indicated ANPed and 95NPed. The top horizontal line of the MNPed curve indicated the current error is not detected. The PBRTQC for both single-type and delta-type data were inferior to MLQC with delta-type data and showed asymmetric error detection curves for both positive and negative errors. However, MLQC was not only more sensitive than both PBRTQC, but also the MNPed curves were symmetrically distributed, suggesting that MLQC showed stronger ability for error detection, especially at critical levels.
Figure 4

The curves for the comparison performance of 5 experiments. A,B,C corresponded to LYMPH #, HGB, and PLT, respectively. Colored lines represented MNPed for each bias, colored area represented the associated 95NPed. Parameters were displayed in the top corner (BS: block size; T: truncation limit, BC: with Box–Cox transformation).

The curves for the comparison performance of 5 experiments. A,B,C corresponded to LYMPH #, HGB, and PLT, respectively. Colored lines represented MNPed for each bias, colored area represented the associated 95NPed. Parameters were displayed in the top corner (BS: block size; T: truncation limit, BC: with Box–Cox transformation). For MLQC, all results of the MNPed to the seven test items for all biases (except for zero bias), were below 10 except for the MCHC item. The cumulative ANPed, MNPed and 95NPed corresponding to each of the 10 experiments for all biases (except for bias at critical level for each test item) showed significant differences among the methods. Taking LYMPH# as an example, the best performance for using single-type data was the MovSD algorithm with the above three cumulative values of 5044.92, 4871, and 8476.5, respectively; the best performance for using delta-type data was the MA algorithm with the three cumulative values of 1629.52, 1094, and 5213, respectively; and the three cumulative values for using delta-type data by ML QC were 45.07, 50, 60. Taking HGB as an example, the best performance for using single-type data was the MovSD algorithm with three cumulative values of 4550.80, 4751.5, and 8015.3; the best performance for using delta-type data was the MovSD algorithm with three cumulative values of 3515.60, 3504, and 5341.5; and the three cumulative values for delta-type data by ML QC were 89.42、91.5、104.55. Taking PLT as an example, the best performance for using single-type data was MovSD algorithm with 3 cumulative values of 4479.43, 3818, 8260.3; the best performance for using delta-type data was MovSD algorithm with 3 cumulative values of 3454.44, 3798, 5362.1; and the three cumulative values for delta-type data by ML QC were 70.33, 72, 102.2. It was obvious that the experimental results using delta-type data were better than single-type data, and the cumulative ANPed and MNPed of delta data ML QC for all 10 biases were reduced to within 100 from the previous 1000 or beyond. For the LYMPH#, HGB, and PLT, the cumulative MNPed of ML QC results relative to the best PBRTQC results were reduced by 95.43%, 97.39%, and 97.97%, respectively.

Discussion

Levey and Jennings introduced QC by using QC materials into the clinical laboratory in 1950 as the primary way to improve poor analytical performance [19]. The fundamental objective of QC program in the clinical laboratory is to characterize the analytical process accurately and thereby to provide information regarding the quality of results reported for clinical specimens [20]. Patient-based real-time quality control (PBRTQC) is a generic term for the use of patient results for real-time quality-control purpose as an alternative tool for insufficient or inefficient QC. Recent studies by Xincen Duan, et al. [3] used an additional regression adjustment before using a common algorithm in the RARTQC framework removed autocorrelation in the test results, and allowed researchers to add additional variables, and to improve data transformation; Ichihara et al. [5] set up the weighted cumulative delta-check (wCDI) method, applying a series of techniques for data stability. Otherwise, in real settings, clinical testing data takes on significant heterogeneous feature and contains a number of extreme data. And some independent variables from population impact patient data-oriented QC method performance, such as age, sex, patient type, within-or-between biological variability, sample mislabeling, patient misidentification, distribution patterns of test results. Delta check is a quantity of change expressed as a magnitude or ratio that can be determined by calculating continuous paired data from representative patients. It has been used for monitoring quality issues in total testing process, such as patient identification errors, sample identification errors or sample mishandling in the pre-analytical phase, as well as for QC in analytical phase [21]. While DC is still limited by a simple linear transformation for handling the noise from data. ML, as one of the main tools for data mining, can seek structural features of the data from complex dimensions and big data volume. In this paper, ML technique is introduced into QC, combined with delta data, to explore a newly overall protocol for data stability, thus improving patient data-based QC effectiveness. An overall protocol for stabilizing data is set up by using delta data in combination with IF algorithm in ML. First, inputs are changed from single-type data to delta-type data, the paired-data weakens the perturbation from single patient data although additional intra-individual variation information introduced, which is equivalent to denoising to the single data, making the characteristics of the bias expected to be identified more significant, thus improving the accuracy of QC event detection. The experimental results in Table 1 showed that the SD was reduced by 31.85% on average after the single data of the seven test items were converted to delta data. Second, for the stabilization of sample sources, ML IF shows a powerful advantage, which is essentially based on the distribution of samples in a high-dimensional space to remove outliers in samples. It transforms a sample pre-processing issue into a classification problem based on density and distance by the spatial location of the samples. The IF algorithm enables to recursively segment data set randomly until all sample points are completely isolated. With this segmentation strategy, every data point is effectively utilized, thus improving data utilization, reducing information loss, and preventing denoising failure caused by removing outlier data directly by means of setting statistical truncation limits. When the delta-type data processed by IF, the SD was reduced by an average of 72.36% compared to the single-type data processed only by a truncation limit on statistics. As while, RF algorithm is also used to further improve the effective utilization of data, which in turn improves the QC effectiveness. In this study, a ML RF model is established, which maps every single data point within a moving window length to a high-dimensional space, and the data points within the same moving window are regarded as a whole and mapped in the high-dimensional space with a divisible population effect. Here, the data points within each moving window are characterized as serialized information after feature engineering process in ML, and the data points within each moving window produce a horizontal cross-correlation into the RF model in a multidimensional parallel manner, instead of by the way of the mean calculated by using the data points in one block size data in PBRTQC. The RF model is trained through many data iterations, and there exists a technique similar to IF in its process, which summarizes a relatively reasonable delimited hyper-curve. And when a new unknown input sequence needs to be predicted, the sequence will be compared with the delimited hyper-curve, a final anomaly probability is output by calculating the size relationship of each element in it. Our experimental results showed that ANPed, MNPed and 95NPed of MLQC using delta data, were basically within 10, which were reduced by 96.39%, 95.34% and 96.37% respectively compared to the optimal results of PBRTQC. The sensitivity and specificity of the MLQC were also both better than the best results of PBRTQC. Here, TPR refers to the sensitivity, which represents out-of-control status of QC, and TNR refers to the specificity, which represents in-control status of QC. High sensitivity and specificity of our model indicates that the MLQC is rarely possible for misclassification and omission of QC event detection as well as without delay and labor-intensive to false alarm. In summary, by implementing an overall protocol for data processing, together with ML algorithm innovation, an effective tool for QC error detection is established.

Declarations

Author contribution statement

Yufang Liang, Rui Zhou: Conceived and designed the experiments, Analyzed and interpreted the data; Wrote the paper. Zhe Wang: Conceived and designed the experiments; Performed the experiments; Analyzed and interpreted the data; Wrote the paper. Qingtao Wang: Conceived and designed the experiments; Contributed reagents, materials, analysis tools or data. Xiang Feng, Zewen Han: Performed the experiments. Biao Song: Performed the experiments; Contributed reagents, materials, analysis tools or data. Dawei Huang: Analyzed and interpreted the data Contributed reagents, materials, analysis tools or data. Wei Wang: Contributed reagents, materials, analysis tools or data.

Funding statement

This work was supported by Excellence Project of key clinical specialty in Beijing; and the 1351 Talent Training Plan [grant numbers CYMY-2017-01], and provincial key fund of Science and Technology Development Special Project in Mongolia, entitled by Application of Big Data Technology for disease prediction only adopting routine laboratory data [grant numbers in 2021-Science and Technology Development Project in Mongolia-independent Innovation Region-01].

Data availability statement

Data included in article/supplementary material/referenced in article.

Declaration of interest’s statement

The authors declare no conflict of interest.

Additional information

No additional information is available for this paper.
  18 in total

1.  Commutability limitations influence quality control results with different reagent lots.

Authors:  W Greg Miller; Aybala Erek; Tina D Cunningham; Olajumoke Oladipo; Mitchell G Scott; Robert E Johnson
Journal:  Clin Chem       Date:  2010-11-19       Impact factor: 8.327

2.  Quality control in clinical chemistry: characterization of reference materials.

Authors:  R Rej; R W Jenny; J P Bretaudiere
Journal:  Talanta       Date:  1984-10       Impact factor: 6.057

3.  The use of control charts in the clinical laboratory.

Authors:  S LEVEY; E R JENNINGS
Journal:  Am J Clin Pathol       Date:  1950-11       Impact factor: 2.493

4.  Understanding Patient-Based Real-Time Quality Control Using Simulation Modeling.

Authors:  Andreas Bietenbeck; Mark A Cervinski; Alex Katayev; Tze Ping Loh; Huub H van Rossum; Tony Badrick
Journal:  Clin Chem       Date:  2020-08-01       Impact factor: 8.327

5.  Assessment of patient-based real-time quality control algorithm performance on different types of analytical error.

Authors:  Xincen Duan; Beili Wang; Jing Zhu; Wenqi Shao; Hao Wang; Junfei Shen; Wenhao Wu; Wenhai Jiang; Kwok Leung Yiu; Baishen Pan; Wei Guo
Journal:  Clin Chim Acta       Date:  2020-10-28       Impact factor: 3.786

6.  Relationship between biological variation and delta check rules performance.

Authors:  Rui Zhen Tan; Corey Markus; Tze Ping Loh
Journal:  Clin Biochem       Date:  2020-04-02       Impact factor: 3.281

7.  Application of 3-D Delta check graphs to HbA1c quality control and HbA1c utilization.

Authors:  David V Tran; George S Cembrowski; Terrence Lee; Trefor N Higgins
Journal:  Am J Clin Pathol       Date:  2008-08       Impact factor: 2.493

8.  Optimization of a Moving Averages Program Using a Simulated Annealing Algorithm: The Goal is to Monitor the Process Not the Patients.

Authors:  David Ng; Frank A Polito; Mark A Cervinski
Journal:  Clin Chem       Date:  2016-08-18       Impact factor: 8.327

9.  Average of Patient Deltas: Patient-Based Quality Control Utilizing the Mean Within-Patient Analyte Variation.

Authors:  George S Cembrowski; Qian Xu; Mark A Cervinski
Journal:  Clin Chem       Date:  2021-07-06       Impact factor: 8.327

10.  How useful are delta checks in the 21 century? A stochastic-dynamic model of specimen mix-up and detection.

Authors:  Katie Ovens; Christopher Naugler
Journal:  J Pathol Inform       Date:  2012-02-29
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.