Literature DB >> 31532490

Detecting conversation topics in primary care office visits from transcripts of patient-provider interactions.

Jihyun Park1, Dimitrios Kotzias1, Patty Kuo2, Robert L Logan Iv1, Kritzia Merced2, Sameer Singh1, Michael Tanana3, Efi Karra Taniskidou1, Jennifer Elston Lafata4,5, David C Atkins6, Ming Tai-Seale7, Zac E Imel2, Padhraic Smyth1.   

Abstract

OBJECTIVE: Amid electronic health records, laboratory tests, and other technology, office-based patient and provider communication is still the heart of primary medical care. Patients typically present multiple complaints, requiring physicians to decide how to balance competing demands. How this time is allocated has implications for patient satisfaction, payments, and quality of care. We investigate the effectiveness of machine learning methods for automated annotation of medical topics in patient-provider dialog transcripts.
MATERIALS AND METHODS: We used dialog transcripts from 279 primary care visits to predict talk-turn topic labels. Different machine learning models were trained to operate on single or multiple local talk-turns (logistic classifiers, support vector machines, gated recurrent units) as well as sequential models that integrate information across talk-turn sequences (conditional random fields, hidden Markov models, and hierarchical gated recurrent units).
RESULTS: Evaluation was performed using cross-validation to measure 1) classification accuracy for talk-turns and 2) precision, recall, and F1 scores at the visit level. Experimental results showed that sequential models had higher classification accuracy at the talk-turn level and higher precision at the visit level. Independent models had higher recall scores at the visit level compared with sequential models.
CONCLUSIONS: Incorporating sequential information across talk-turns improves the accuracy of topic prediction in patient-provider dialog by smoothing out noisy information from talk-turns. Although the results are promising, more advanced prediction techniques and larger labeled datasets will likely be required to achieve prediction performance appropriate for real-world clinical applications.
© The Author(s) 2019. Published by Oxford University Press on behalf of the American Medical Informatics Association.

Entities:  

Keywords:  classification; communication; patient care; supervised machine learning

Mesh:

Year:  2019        PMID: 31532490      PMCID: PMC6857514          DOI: 10.1093/jamia/ocz140

Source DB:  PubMed          Journal:  J Am Med Inform Assoc        ISSN: 1067-5027            Impact factor:   4.497


INTRODUCTION

Background

Appropriate documentation of the clinical visit is critical for communication among medical professionals,, enabling quality assurance, and accurate billing and reimbursement. The traditional way of documenting a clinical visit in the electronic health record, namely physicians’ notes, provide a source of valuable information on what occurred during the interaction and what physicians consider to be important. Electronic health records have improved the accessibility of medical information, but patients are demanding access to information at a greater scale. Pressure to quickly document the medical visit may lead providers to type into the record during the medical visit, which can negatively impact patient-provider communication. Primary care providers spend about half of their time working on computers,, which appears partly responsible for growing concerns of physician burnout across a wide range of physician specialties. In addition, physician generated notes do not always provide an accurate representation of what occurred during the visits., Technologies that could reduce the burden of documentation on providers, and increase convergence between documentation of a visit and the content of the clinical interaction are greatly needed. Natural language processing (NLP) technologies combined with advances in automatic speech recognition, offer potentially promising solutions., Information extraction and summarization technologies built on top of resulting transcripts will be needed to take the next step in reducing the burden of documentation on physicians and in providing clinical decision support to both patients and physicians. If successful, machine learning–enabled automatic speech recognition and charting could free up valuable time for physicians to talk to their patients, rather than typing extensively during clinical encounters., Patient-provider conversations are complex, multidimensional, and multifunctional. Patients present multiple issues during an office visit requiring clinicians to divide time and effort during a visit to address competing demands,, such as a patient could be concerned about blood pressure, knee pain, and blurry vision in a single appointment. Moreover, visit content does not solely focus on biomedical issues, but also on psychosocial matters, personal habits, mental health, patient-physician relationship, and small talk., Health communications researchers analyze the content of patient-provider communication by directly labeling the interaction using trained raters to label the topical content of clinical interactions. For example, during periodic health examinations with their primary care physicians, only one-third of patients with mental health needs had a discussion about mental health., These findings suggest that quality improvement efforts that evaluate the content of clinical interactions might help address gaps in service delivery. However, labeling each talk-turn using coding systems designed to capture the content of medical visits, such as the Multi-Dimensional Interaction Analysis system,,, is labor intensive and costly. It requires training a team to label the text and establish and maintain reliability. Depending on the extensiveness of the labeling system, it can take several hours to label one patient-provider interaction., This level of effort means that even in research settings with resources for detailed evaluation, studies of patient-provider interaction are often limited in scale. As a result, the direct evaluation of clinical interactions is not feasible in routine care settings for quality improvement purposes., Automated methods capable of extracting the topical content of clinical encounters could support providers who overlook asking patients about critical issues (eg, suicide, blood pressure medication) when it is clinically indicated, reduce the burden of documentation currently placed on providers, and facilitate large-scale research on the quality of patient-provider communication.

Related work

The past decade has seen an explosion of interest in machine learning and NLP in medical contexts, targeting problems such as automated extraction of information from clinical notes and electronic health records. Of direct relevance to the present article is a growing body of work dedicated to applying methods from machine learning and NLP to the automatic annotation of conversations between providers and patients. The typical approach in such studies begins with labeling a corpus of transcript data (eg, providing human-generated labels for each utterance or talk-turn in each visit in the corpus). Machine learning techniques are then used to learn a classification model from a subset of the corpus, the training data, and the model’s predictions are evaluated by comparing its predictions with the known human labels on unseen test transcripts. For example, Mayfield et al analyzed patient-provider transcripts from the ECHO (Enhancing Communication and HIV Outcomes) study and employed a logistic regression model to classify utterances into the categories “information-giving” or “information-requesting.” More recently, Kotov et al analyzed transcripts from motivational interviewing related to pediatric obesity and developed probabilistic machine learning models for classifying patient utterances into classes of behavioral responses. In later work on the same dataset, Hasan et al compared probabilistic models with more recent recurrent neural network approaches and found the latter to be generally more accurate on that dataset. There has been less prior work on the problem of automated classification of topical content in patient-provider dialog. Wallace et al developed machine learning models to classify dialog utterances into 1 of 6 high-level discussion topics: biomedical, logistics, psychosocial, antiretroviral (ARV), missing/other, and socializing. Using the same ECHO dataset as used in the Mayfield et al study, they evaluated the performance of conditional random fields (CRFs) for this prediction task and concluded that the results showed promise for automated classification of patient-provider interactions into clinically relevant topics. Gaut et al proposed the use of labeled topic models for classifying psychotherapy sessions with 161 possible topic labels, using a dataset published by Alexander Street Press, and again finding that these models showed promise in terms of predictive ability. These earlier studies demonstrate that machine learning systems can generate plausible annotations of medical dialogues—our work in this article pursues this line of research further. We differ from earlier work on topic classification in a number of aspects. For example, we compare both probabilistic and neural network classification methods for topic classification of talk-turns, whereas Wallace et al and Gaut et al only focused on probabilistic approaches. We also evaluate performance for a more detailed set of 27 topics compared with the 6 high-level topics used in the Wallace et al study. The Gaut et al study also differs from our work in that it primarily focused on session-level labels and did not investigate the use of sequential information across talk-turns for talk-turn–level predictions as we do here. To our knowledge, this is the first study that systematically compares sequential and nonsequential classification methods, for both probabilistic and neural network models, on the problem of talk-turn topic classification from transcripts of patient-provider dialog.

MATERIALS AND METHODS

Dataset

The source data include transcripts of audio-recordings of primary care office visits from the MHD (Mental Health Discussion) study. Each transcript corresponds to a visit between a patient and a provider—a small fraction of the dialog corresponds to other participants in the conversation (such as a nurse and family member). Data collection occurred from 2007 to 2009 in a health system in Michigan with 26 ambulatory care clinics. Patients were 50-80 years of age, all had insurance, and were due for a colorectal cancer screening at the time of appointment. All aspects of the research protocol were approved by relevant organizations’ institutional review boards. Each visit is comprised of a series of talk-turns, with 122 083 talk-turns in total across 279 visits (median and mean of 408 and 438 talk-turns per visit, respectively, with upper and lower quartiles of 312 and 522) from 59 providers. Each talk-turn was manually assigned by a human coder (labeler) to 1 of 39 different topic labels that were modified from the Multi-Dimensional Interaction Analysis coding system. A topic is defined as an issue raised in a conversation that required a response from the other member of the dyad and had at least 2 exchanges between the dyad. A small number of talk-turns were split into 2 if the turn straddled 2 topics. This resulted in a few of the original talk-turns being represented as 2 separate talk-turns after coding. Figure 1 illustrates how different topics were assigned to talk-turns during a short portion of a particular visit. Topic labels that occurred in less than 20 visits were merged into a single topic denoted as Other, resulting in a total of 27 unique topics in the corpus. Table 1 provides the names of the topic labels, a brief description of each, as well as the percentage of talk-turns assigned to each by the labelers across the corpus. The topic label distribution is skewed toward topics relevant to periodic health exams—the 3 most frequent topics (BiomedHistory, PreventiveCare, and MusSkePain) account for more than half of all talk-turns.
Figure 1.

A short excerpt from an annotated dialogue transcript. Topic labels are assigned to each talk-turn. MD and PT indicate the speaker for each talk-turn, where MD stands for “medical doctor” or “provider,” and PT stands for “patient.”

Table 1.

Name and brief description of each topic ordered by the percentage of each topic in talk-turns

Short topic nameBrief descriptionTalk-turns assigned (%)
Biomed HistoryBiomedical history, symptoms, and medical condition29.85
Preventive CarePreventive medical measures14.67
Mus Ske PainMusculoskeletal pain8.30
Visit Flow MgmtAgenda setting, opening of visit, closing of visit6.38
Gyn Genito UrinaryGynecological and genitourinary problem4.72
Physical ExamPhysical exam3.41
FamilyFamily and significant other3.10
Health Care SystemHealth care system2.89
Work LeisureWork and leisure activities2.73
TestsTests and diagnostic procedures2.59
CigaretteCigarette2.43
WeightWeight2.38
Dizzy Dent Hear VisionDizzyness, vision, hearing, dental issues2.03
OtherOther (various rare topics)1.94
ExerciseExercise1.89
DepressionDepression1.86
MedicationMedications1.84
SmallTalkSmall talk1.72
General AnxietiesGeneral anxieties and worries1.38
MDLifePhysician personal life1.04
DietDiet, food (exclude supplements)0.69
AlcoholAlcohol0.57
SleepSleep0.53
Therapeutic InterventionTherapeutic intervention0.33
Risky BehaviorRisky behaviors (eg, international travel, weapons at home) and risk avoidance preventive practices (eg, safe sex, wearing seatbelt or bike helmet)0.32
OtherAddictionsCaffeine, or other addictions0.21
AgeAge0.17
Name and brief description of each topic ordered by the percentage of each topic in talk-turns A short excerpt from an annotated dialogue transcript. Topic labels are assigned to each talk-turn. MD and PT indicate the speaker for each talk-turn, where MD stands for “medical doctor” or “provider,” and PT stands for “patient.”

Text preprocessing

We applied a number of preprocessing steps to convert the dialog text into a set of tokenized words. We first replaced the patient names and numbers with -NAME- and -NUMBER- tokens to remove potentially identifiable information. After removing symbols, other than a set of punctuation symbols, such as “.”, “?”, “-”, the sentences in each talk-turn were tokenized into a list of words using the standard Python NLTK tokenizer. For the models that used bag-of-words encoding, the vocabulary included all unigrams and bigram noun phrases that occurred at least 5 times in the corpus, except for 2 sets of stopwords (see Supplementary Appendix A for more information), resulting in a vocabulary of size . For our neural network models, the vocabulary consisted of all unigrams, with neither set of stopwords removed (as is customary in neural network models) with a final vocabulary size of including an unknown token.

Representing talk-turns for classification models

The data for each visit , , is represented as a sequence of labeled talk-turns , with talk-turns in the th visit, . Let and represent the list of word tokens and the topic label, respectively, for the th talk-turn in the th visit. As mentioned earlier, can take values from 1 to 27, corresponding to each of the 27 unique topics. Each word in is encoded as a binary vector (“one-hot encoding”) of length , where is the vocabulary size. For example, if a word occurs in a talk-turn and the ID in the vocabulary for the word is 10, then the binary/one-hot-encoded vector becomes a vector of length , where the 10th entry of the binary word vector is set to 1 and all other entries are set to zero. For the majority of the classification models we evaluated, the binary word vectors in each talk-turn were aggregated into a single talk-turn vector by adding the individual binary word vectors (also known as a bag-of-words encoding) and reweighting using tf-idf weights, a common text preprocessing step that downweights common and uninformative words. The vector represents each talk-turn and has dimensionality equal to the size of the vocabulary . For our neural network models, we used a different representation as follows. We generated one vector per talk-turn using a network composed of an embedding layer and a bidirectional set of gated recurrent units (GRUs). The embedding layer, initialized with pretrained GloVe vectors, takes each binary word vector in the talk-turn and maps it to a dense embedded vector representation. The GRU component takes the sequence of embedded vectors within a talk-turn (1 embedded vector per word) and produces a single fixed-dimensional vector to represent the talk-turn. This approach has been shown to be useful in NLP applications for encoding variable-length sequential information from words in a sentence (talk-turn) into a fixed-dimensional vector that can be used as a feature vector for downstream classification. We used 128 for the GRU unit size, and the output talk-turn vector had size 256 because the GRU outputs in 2 directions are concatenated. Given the talk-turn vectors , we classify each talk-turn either independently or by using sequential information across talk-turns. Figure 2A provides a high-level overview of the 3 primary different types of models we explored: 1) independent models that classify each talk-turn only using the words in talk-turn , 2) window-based models that also use words from a window of talk-turns both before and after talk-turn , and 3) fully sequential models that also consider the topic labels (or predictions) of talk-turns before and after when predicting a topic for talk-turn . In addition, we consider another type of sequential model that uses the talk-turn–level GRUs on top of word level GRU outputs, which is depicted in Figure 2B. We describe each of the 3 approaches in more detail below. Additional implementation details can be found in Supplementary Appendix C.
Figure 2.

(A) A high-level diagram of the various models discussed in the article. i, j stands for talk-turn j in session i. W is the list of tokenized words in talk-turn j. For each talk-turn j, we first generate the vectorized talk-turn representation , and the talk-turn representation is used as an input to different classifiers to predict the topic label y, which is the topic label for talk-turn j. Windowed models use the adjacent talk-turns to create the talk-turn–level representation, and the fully sequential models make use of the sequential dependencies between the topic labels. (B) Simplified diagram of the hierarchical gated recurrent units (Hier-GRU). Each one-hot-encoded word vector in talk-turn j in visit i (shown as for the k-th word in talk-turn j) is fed into the word level encoder to get a talk-turn representation , which becomes the input to the talk-turn–level encoder. The model has dependencies at the hidden state of talk-turn–level gated recurrent units (GRUs). In our experiments, both encoders were bidirectional.

(A) A high-level diagram of the various models discussed in the article. i, j stands for talk-turn j in session i. W is the list of tokenized words in talk-turn j. For each talk-turn j, we first generate the vectorized talk-turn representation , and the talk-turn representation is used as an input to different classifiers to predict the topic label y, which is the topic label for talk-turn j. Windowed models use the adjacent talk-turns to create the talk-turn–level representation, and the fully sequential models make use of the sequential dependencies between the topic labels. (B) Simplified diagram of the hierarchical gated recurrent units (Hier-GRU). Each one-hot-encoded word vector in talk-turn j in visit i (shown as for the k-th word in talk-turn j) is fed into the word level encoder to get a talk-turn representation , which becomes the input to the talk-turn–level encoder. The model has dependencies at the hidden state of talk-turn–level gated recurrent units (GRUs). In our experiments, both encoders were bidirectional.

Independent models

Independent models classify each talk-turn by using only the words in that talk-turn, , independently of the other talk-turns. The independent models used in our study were the logistic regression (LR) classifiers, support vector machines (SVMs), and feedforward neural networks with a single hidden layer. The LR and SVM classifiers used the bag-of-words vectors with tf-idf weights as input to predict the topic label . For the feedforward neural network, the output talk-turn vectors of the bidirectional GRU units were used as inputs, and the softmax function was used for the activation function in the perceptron. The parameters (weights) of the embedding layer, the GRU, and the feedforward neural network were all trained together as a single network, and we refer to this model as independent GRU.

Window-based models

Instead of using a single talk-turn vector as an input (as in the independent models), for the window-based models, we concatenated the adjacent vectors before and after each talk-turn. The input vector was defined as , with dimension , where is the size of the vocabulary for bag-of-words representations. This windowed input vectors were then used as input to either the LR or SVM classifier. In Figure 2, the window-based approach is indicated with dashed lines at the input level. The window-based representation captures potentially useful sequential information across talk-turns in a manner not available to the independent models—but at the cost of having on the order of times as many parameters. The approach is particularly useful when making topic predictions for short talk-turns with very little information—information from neighboring talk-turns can be used to help to make predictions in such cases. We used in our experiments based on the knowledge that the transcripts were labeled with the convention that each topic label must span at least 2 exchanges (4 talk-turns).

Fully sequential models

To model the sequential dependencies between the topic labels, we used linear-chain CRFs and hidden Markov models (HMMs). The linear-chain CRF is widely used for sequence labeling tasks such as named-entity recognition or part-of-speech tagging,—here, we applied it to the problem of predicting the topic label of a talk-turn, given a sequence of talk-turns. The HMMs are constructed by using the output class label probabilities from each of the independent models discussed above. We converted the class probabilities from the classifiers, , to emission probabilities, (needed by the HMM), by using the fact that , given that the talk-turn probabilities do not depend on a particular topic . The emission probabilities are combined with the Markov transition probabilities (which can be directly estimated from label sequences in the training data) via the Viterbi algorithm to compute the sequence of topic labels that has the highest joint probability across talk-turns, conditioned on the observed talk-turn words. We also found that using speaker-specific transition matrices improved the accuracy of our sequential models. Not surprisingly, providers tend to start new topics during a conversation more than the patients do. The figure in Supplementary Appendix B shows the percentage of time that a particular speaker (provider, patient, or other) starts each topic. To incorporate speaker information in the HMM approach, we augmented the standard HMM to use 2 topic transition matrices, 1 for the provider and 1 for the patient or other speakers. Each transition matrix corresponds to the speaker of the state that the HMM is transitioning to (eg, one transition matrix for transitioning to provider and the other for transitioning to all others). The decoding process in the Viterbi algorithm is modified so that it uses the appropriate matrix depending on the speaker. Another type of sequential model that is entirely neural-network-based does not have direct dependencies between the topic labels, but has bidirectional connections between the hidden states of the talk-turn–level GRUs. The model, which we also refer to as Hier-GRU, has a hierarchical structure having 2 different GRUs, one at the word level to generate talk-turn–level representation, and the other which takes the talk-turn vectors as inputs to predict the output label for each talk-turn . Similar to independent GRU, the talk-turn–level GRU output is connected to a fully-connected layer and then a softmax to make prediction.

RESULTS

Experimental methods

We evaluated all models using 10-fold cross-validation and computed evaluation metrics both at the talk-turn level and at the visit level. At the talk-turn level, we computed the classification accuracy by comparing 1) the predicted topic from a model with 2) the human-generated topic for the talk-turn. To obtain results at the visit level, we aggregated the predictions from the individual talk-turns within each visit to generate a visit-level binary-valued prediction vector of dimension 27 (the number of topic labels) with 1 for topic label if the model predicted topic for 1 or more talk-turns in the visit, and 0 otherwise. Using such a vector for each visit, we calculated the accuracy, precision, recall, and F1 scores. The metrics were micro-averaged by globally counting the true positives, false positives, and so on, for all the topic labels. To evaluate the performance of the classifiers, we compared the results with those of simple baseline models that predict the most common topics. The baseline at the talk-turn level just predicts the most common topic in the corpus, BiomedHistory (as shown in Table 1). At the visit level, the baseline always predicts the set of topic labels that occur in 50% or more of all visits, irrespective of the words within each visit.

Summary of experimental results

Table 2 shows the classification accuracies at the talk-turn level for independent models, windowed models, and sequential models. The most accurate independent model is the GRU and the most accurate windowed model is the windowed LR. The models with sequential information clearly outperform independent models with the Hier-GRU yielding the highest accuracy of 61.77% over all the models. The improvement in accuracy of Hier-GRU and HMM-GRU are both statistically significant, with , relative to each of the independent GRU and windowed LR models (using dependent t tests for paired samples across the 10 folds of cross-validation).
Table 2.

Accuracies for topic prediction at the level of talk-turns for different prediction models

ModelTalk-turn level accuracy (%)
Baseline 29.85
Independent models LR 37.00
SVM 36.64
GRU 38.85a
Window-based models Windowed LR 51.12a
Windowed SVM 50.46
Fully sequential models CRF 48.37
HMM-LR 56.89
HMM-SVM 51.52
HMM-GRU 57.60b
Hier-GRU 61.77ab

Micro-averaged precision and recall scores are the same as accuracy.

CRF: conditional random field; Hier-GRU: hierarchical gated recurrent units; HMM-GRU: hidden Markov model gated recurrent units; LR: logistic regression; SVM: support vector machine;

Highest talk-turn level accuracy in each model type.

Scores from the two best models.

Accuracies for topic prediction at the level of talk-turns for different prediction models Micro-averaged precision and recall scores are the same as accuracy. CRF: conditional random field; Hier-GRU: hierarchical gated recurrent units; HMM-GRU: hidden Markov model gated recurrent units; LR: logistic regression; SVM: support vector machine; Highest talk-turn level accuracy in each model type. Scores from the two best models. The visit-level evaluation scores are shown in Table 3. Interestingly, the gap in performance (as measured by the micro-averaged F1 score) between independent, windowed, and sequential models is much smaller in the visit-level scores. The primary reason for this is that the independent models tend to have relatively high recall scores, whereas sequential models have relatively high precision scores.
Table 3.

Micro-averaged accuracy, precision, recall, and F1 scores, at the visit level, for different prediction models

Visit level (%)
ModelAccuracyPrecisionRecallF1
Baseline 72.2973.7962.2267.42
Independent models LR 75.1967.9184.2575.15a
SVM 72.4563.4090.4074.50
GRU 72.4764.1986.5973.68
Window-based models Windowed LR 77.2869.8286.4777.21
Windowed SVM 79.8175.0682.1378.37a
Fully sequential models CRF 74.1980.4358.4267.64
HMM-LR 80.0080.1673.3176.55
HMM-SVM 75.2178.7060.9068.63
HMM-GRU 79.0074.9879.3577.06
Hier-GRU 78.9673.6982.4377.78a

CRF: conditional random field; Hier-GRU: hierarchical gated recurrent units; HMM-GRU: hidden Markov model gated recurrent units; LR: logistic regression; SVM: support vector machine;

Highest F1 score in each model type.

Micro-averaged accuracy, precision, recall, and F1 scores, at the visit level, for different prediction models CRF: conditional random field; Hier-GRU: hierarchical gated recurrent units; HMM-GRU: hidden Markov model gated recurrent units; LR: logistic regression; SVM: support vector machine; Highest F1 score in each model type. The prediction models also can be evaluated at the level of individual topics to understand variability in prediction accuracy across topics. Precision, recall, and F1 scores were calculated by treating each topic label separately at the talk-turn level. Scores from the 2 best-performing models (Hier-GRU and HMM-GRU) are shown in Table 4, sorted by the percentage of each topic. The more common topics (that occur for example in at least 5% of the talk-turns) generally have higher F1 scores. However, there are some less common but highly specific topics, such as the cigarette topic (in 2.43% of talk-turns), that also have relatively high F1 scores.
Table 4.

Precision, recall, and F1 scores of each topic, calculated at the talk-turn level using Hier-GRU and HMM-GRU prediction results.

Hier-GRU
HMM-GRU
LabelPrecisionRecallF1PrecisionRecallF1Assigned topic (%)
BiomedHistory65.8076.6170.79a74.2456.2664.0129.85
PreventiveCare73.3483.4178.05a77.9872.4975.13a14.67
MusSkePain67.4869.1468.30a67.6864.2765.93a8.30
VisitFlowMgmt63.6463.5863.61a64.9764.5464.76a6.38
GynGenitoUrinary67.6254.7660.5172.1557.6664.09a4.72
PhysicalExam48.9152.3350.5650.2163.2055.963.41
Family49.3147.5648.4242.7954.2147.833.10
HealthCareSystem37.8128.8532.7346.1039.4942.542.89
WorkLeisureActivity47.2646.2046.7252.9053.1453.022.73
TestDiagnostics49.6732.4339.2437.2452.8843.702.59
Cigarette71.3885.8477.94a81.3872.7376.81a2.43
Weight49.5452.6051.0247.2046.7746.982.38
Other21.429.5913.2514.4515.2014.812.03
Depression50.8264.2756.7645.7144.2444.961.94
DizzyDentHearVision23.0814.0117.4446.9650.8348.821.89
Medication38.9727.0931.9622.4664.2933.291.86
Exercise56.9459.2358.0645.4267.0154.141.84
SmallTalk20.6118.7019.6132.7931.0931.921.72
GeneralAnxieties32.5118.6323.6816.3632.8621.841.38
MDLife33.8114.1919.9912.7911.6912.221.04
Diet27.1519.6222.7827.5257.1037.140.69
Alcohol56.0836.1243.9452.1964.9157.860.57
Sleep13.402.333.9729.8040.3234.270.53
TherapeuticIntervention0.000.000.002.045.302.950.33
RiskyBehavior33.905.8710.0044.8741.0642.880.32
OtherAddictions0.000.000.0023.5541.4030.020.21
Age0.000.000.0013.2320.0015.930.17

The rows are sorted by the percentage of talk-turns of each topic. In general, the more frequently discussed topics have higher F1 scores.

Hier-GRU: hierarchical gated recurrent units; HMM-GRU: hidden Markov model gated recurrent units.

Highest F1 scores.

Precision, recall, and F1 scores of each topic, calculated at the talk-turn level using Hier-GRU and HMM-GRU prediction results. The rows are sorted by the percentage of talk-turns of each topic. In general, the more frequently discussed topics have higher F1 scores. Hier-GRU: hierarchical gated recurrent units; HMM-GRU: hidden Markov model gated recurrent units. Highest F1 scores.

DISCUSSION

Using sequential information across talk-turns in predictive models systematically leads to more accurate predictions, particularly when predicting topic labels for talk-turns. To illustrate this point in more detail, Figure 3 shows sequences of predicted and human-assigned topic labels, for one particular visit, where different colors represent different topic labels. The top plot is from the independent GRU model. The lack of sequential information in the model leads to predictions that are noisy and lack the sequential smoothness of the human-assigned labels (bottom plot). The second plot is from the Hier-GRU model, and the third plot shows the predicted sequence of topics from the Viterbi parse of the HMM-GRU model (ie, the probabilistic predictions from the same GRU model in the top plot, but which are now sequentially smoothed by the HMM transition matrices). It is visually apparent (not only for this visit but for all visits) that the sequential models (second and third) are much more similar to the human labeling (bottom) than the independent model (top).
Figure 3.

Sequences of color-coded topic labels for one of the visits in our dataset. The upper plot shows the predicted topic labels from an independent model, and the center 2 plots show those from fully sequential models. The lower plot corresponds to the human-coded labels. The segments for the MusSkePain topic (1) had lengths of 23 talk-turns for hierarchical gated recurrent units (Hier-GRU), 27 for hidden Markov model gated recurrent units (HMM-GRU), and 26 for human labeled. Similarly, (2) the PreventiveCare segments had lengths 30, 28, and 27, and (3) the TestDiagnostics segment had lengths 10, 31, and 22 in talk-turns, respectively, for Hier-GRU, HMM-GRU, and human labeled. See Supplementary Appendix E for the boxplots of topic segment lengths for the 4 sequences of labels.

Sequences of color-coded topic labels for one of the visits in our dataset. The upper plot shows the predicted topic labels from an independent model, and the center 2 plots show those from fully sequential models. The lower plot corresponds to the human-coded labels. The segments for the MusSkePain topic (1) had lengths of 23 talk-turns for hierarchical gated recurrent units (Hier-GRU), 27 for hidden Markov model gated recurrent units (HMM-GRU), and 26 for human labeled. Similarly, (2) the PreventiveCare segments had lengths 30, 28, and 27, and (3) the TestDiagnostics segment had lengths 10, 31, and 22 in talk-turns, respectively, for Hier-GRU, HMM-GRU, and human labeled. See Supplementary Appendix E for the boxplots of topic segment lengths for the 4 sequences of labels. Figures 4 and 5 provide a more detailed look at portions of the transcript corresponding to Figure 3. We see for example that there are quite a few short talk-turns that have no words with topic-relevant information, such as “Yeah” (talk-turns 5 and 7 in Figure 4 and 222 in Figure 5) and “Okay” (talk-turn 224 in Figure 5). The sequential models are able to use the context information to assign these talk-turns to the same topic as the human labeler. The independent GRU model, however, does not have any context and assigns these talk-turns by default to the topic with the highest marginal probability (BiomedHistory).
Figure 4.

The beginning part of the visit shown in Figure 3. Each talk-turn is presented with predicted labels from 3 different models (independent gated recurrent units [GRU], hierarchical gated recurrent units [Hier-GRU], and hidden Markov model gated recurrent units [HMM-GRU]) and the human-coded labels. For the short talk-turns the BiomedHistory topic label is predicted quite often by the independent GRU, while the 2 other models produce label sequences that are more similar to human-coded labels. MD: medical doctor or provider; PT: patient.

Figure 5.

Another excerpt from the same visit in Figure 3. Topics that are semantically similar are confusable (PhysicalExam and MusSkePain in talk-turns 233-242, and SmallTalk and WorkLeisure in talk-turns 249-254). GRU: gated recurrent units; Hier-GRU: hierarchical gated recurrent units; HMM-GRU: hidden Markov model gated recurrent units; MD: medical doctor or provider; PT: patient.

The beginning part of the visit shown in Figure 3. Each talk-turn is presented with predicted labels from 3 different models (independent gated recurrent units [GRU], hierarchical gated recurrent units [Hier-GRU], and hidden Markov model gated recurrent units [HMM-GRU]) and the human-coded labels. For the short talk-turns the BiomedHistory topic label is predicted quite often by the independent GRU, while the 2 other models produce label sequences that are more similar to human-coded labels. MD: medical doctor or provider; PT: patient. Another excerpt from the same visit in Figure 3. Topics that are semantically similar are confusable (PhysicalExam and MusSkePain in talk-turns 233-242, and SmallTalk and WorkLeisure in talk-turns 249-254). GRU: gated recurrent units; Hier-GRU: hierarchical gated recurrent units; HMM-GRU: hidden Markov model gated recurrent units; MD: medical doctor or provider; PT: patient. While the smoothing in sequential models helps to improve prediction accuracy, it can also produce errors due to oversmoothing. In particular, we found that the HMM-GRU model tends to predict longer topic segments relative to the human-labeled results, as can be seen in talk-turns 100 to 200 of the visit in Figure 3. The human labels contain short bursts of topics GynGenitoUrinary and BiomedHistory that are not detected by the HMM-GRU model. This is further quantified by the visit-level results in Table 3, where the recall scores of the fully sequential models are systematically lower than the independent models, and the reverse for the precision scores. We also observed that some topics are semantically similar and easily confusable. For example, in Figure 5, from talk-turn 233 to 242, the 2 sequential models predict the topic PhysicalExam, while the human labeled MusSkePain—from the corresponding transcript text either prediction seems reasonable. Similarly, from talk-turn 249 to 254 the HMM-GRU predicts SmallTalk, while the human labeled WorkLeisure—from the text the corresponding talk-turns appear to be a mixture of both. We also found other examples across the corpus where the model frequently gets confused among small groups of related topics (eg, GeneralAnxieties and Depression; Weight and Diet). The full confusion matrices are shown in Figure 6. There is inevitably a subjective aspect to the human labeling, suggesting that there is likely to be a performance ceiling in terms of the accuracy of any algorithm relative to human labels on this data.
Figure 6.

Confusion matrices generated by (A) hierarchical gated recurrent units (Hier-GRU) and (B) hidden Markov model gated recurrent units (HMM-GRU), where the intensity of each cell shows the conditional probabilities of p(predicted label | human-coded label) and each row sums to 1. A number of subsets of topics have high confusion probabilities, including Diet/Weight/Exercise, Depression/GeneralAnxieties, and Family/MDLife/WorkLeisure.

Confusion matrices generated by (A) hierarchical gated recurrent units (Hier-GRU) and (B) hidden Markov model gated recurrent units (HMM-GRU), where the intensity of each cell shows the conditional probabilities of p(predicted label | human-coded label) and each row sums to 1. A number of subsets of topics have high confusion probabilities, including Diet/Weight/Exercise, Depression/GeneralAnxieties, and Family/MDLife/WorkLeisure. From the topic-specific results in Table 4 we can see that while the predictions are relatively accurate for some topics, for others (eg, Age, TherapeuticIntervention, OtherAddictions, Other, MDLife), the scores are quite low. The broad nature of these topics is a likely contributor to the low accuracies, but the relative lack of training data per topic may also be another contributing factor. These topics account for roughly 1% (or less) of talk-turns in the corpus and many of these talk-turns are relatively short with little topical content, leading to relatively less signal, particularly for training neural network models. One possible approach to improve accuracy would be to incorporate additional external information relevant to these topics, such as incorporating lists of relevant words from ontological sources such as Unified Medical Language System into the training of prediction models, or adding relevant information from sources such as physician or specialist notes.

CONCLUSION

Patient-provider communication is an essential component of health care. In this context, prediction models that annotate patient-provider transcripts can in principle provide both useful information about the nature of topics discussed in specific conversations as well as contribute to a broader understanding of patient-provider communication. We have demonstrated that machine learning methods show promise for building models that can automatically predict discussion topics in dialog at the talk-turn and visit level. In particular, using a large real-world patient-provider dialog corpus, we investigated the performance of a variety of classification models including LR, SVMs, feedforward neural network model with GRUs, hierarchical GRUs, CRFs, and HMMs. We found that sequential models (eg, Hier-GRU and HMM-GRU) are more accurate compared with nonsequential models for predicting topic labels for talk-turns. In addition, we found that semantic similarity of discussion topics can be a significant contributor to prediction error. While additional research and model improvement is needed, our results show promise for a number of medical topics that are critical quality indicators in primary care (eg, cigarette smoking, pain). Potential applications might include exploring systems that incorporate prior information from a list of problem areas or prior diagnoses found in the medical record. For example, the presence or absence of smoking cessation counseling during primary care encounters may inform population health management programs aimed at helping patients quit smoking. Deployment of systems such as this in real-world primary care may also be useful for obtaining the scale of data needed to improve model performance.

FUNDING

This work was supported by Patient-Centered Outcomes Research Institute number ME-1602-34167. Zac E Imel is the one who received the PCORI grant as an overall PI of the project.

AUTHOR CONTRIBUTIONS

All authors assisted in the writing of the manuscript. JP carried out the analyses and was responsible for writing the Materials and Methods and Results sections, with assistance from DK, RLL, SS, and EK-T. PS supervised the analyses. MT and PK were responsible for data management and cleaning along with JP. JEL and MT-S were responsible for collection of source data and description of data and human labeling. ZI, MT-S, PS, and DCA jointly conceived of the project, and ZEI, KM, and PK wrote the introduction.

SUPPLEMENTARY MATERIAL

Supplementary material is available at Journal of the American Medical Informatics Association online. Click here for additional data file.
  44 in total

1.  Assessing competence in the use of motivational interviewing.

Authors:  Theresa B Moyers; Tim Martin; Jennifer K Manuel; Stacey M L Hendrickson; William R Miller
Journal:  J Subst Abuse Treat       Date:  2005-01

2.  Physicians' Well-Being Linked To In-Basket Messages Generated By Algorithms In Electronic Health Records.

Authors:  Ming Tai-Seale; Ellis C Dillon; Yan Yang; Robert Nordgren; Ruth L Steinberg; Teresa Nauenberg; Tim C Lee; Amy Meehan; Jinnan Li; Albert Solomon Chan; Dominick L Frosch
Journal:  Health Aff (Millwood)       Date:  2019-07       Impact factor: 6.301

3.  Automatically Charting Symptoms From Patient-Physician Conversations Using Machine Learning.

Authors:  Alvin Rajkomar; Anjuli Kannan; Kai Chen; Laura Vardoulakis; Katherine Chou; Claire Cui; Jeffrey Dean
Journal:  JAMA Intern Med       Date:  2019-06-01       Impact factor: 21.873

4.  Periodic health examinations and missed opportunities among patients likely needing mental health care.

Authors:  Ming Tai-Seale; Laura A Hatfield; Caroline J Wilson; Cheryl D Stults; Thomas G McGuire; Lisa C Diamond; Richard M Frankel; Lisa MacLean; Ashley Stone; Jennifer Elston Lafata
Journal:  Am J Manag Care       Date:  2016-10-01       Impact factor: 2.229

5.  Electronic Health Record Logs Indicate That Physicians Split Time Evenly Between Seeing Patients And Desktop Medicine.

Authors:  Ming Tai-Seale; Cliff W Olson; Jinnan Li; Albert S Chan; Criss Morikawa; Meg Durbin; Wei Wang; Harold S Luft
Journal:  Health Aff (Millwood)       Date:  2017-04-01       Impact factor: 6.301

6.  Interpretable Probabilistic Latent Variable Models for Automatic Annotation of Clinical Text.

Authors:  Alexander Kotov; Mehedi Hasan; April Carcone; Ming Dong; Sylvie Naar-King; Kathryn BroganHartlieb
Journal:  AMIA Annu Symp Proc       Date:  2015-11-05

7.  A study of the effectiveness of machine learning methods for classification of clinical interview fragments into a large number of categories.

Authors:  Mehedi Hasan; Alexander Kotov; April Carcone; Ming Dong; Sylvie Naar; Kathryn Brogan Hartlieb
Journal:  J Biomed Inform       Date:  2016-05-13       Impact factor: 6.317

8.  Time allocation in primary care office visits.

Authors:  Ming Tai-Seale; Thomas G McGuire; Weimin Zhang
Journal:  Health Serv Res       Date:  2007-10       Impact factor: 3.402

9.  Sharing Annotated Audio Recordings of Clinic Visits With Patients-Development of the Open Recording Automated Logging System (ORALS): Study Protocol.

Authors:  Paul J Barr; Michelle D Dannenberg; Craig H Ganoe; William Haslett; Rebecca Faill; Saeed Hassanpour; Amar Das; Roger Arend; Meredith C Masel; Sheryl Piper; Haley Reicher; James Ryan; Glyn Elwyn
Journal:  JMIR Res Protoc       Date:  2017-07-06

10.  Large-scale evaluation of automated clinical note de-identification and its impact on information extraction.

Authors:  Louise Deleger; Katalin Molnar; Guergana Savova; Fei Xia; Todd Lingren; Qi Li; Keith Marsolo; Anil Jegga; Megan Kaiser; Laura Stoutenborough; Imre Solti
Journal:  J Am Med Inform Assoc       Date:  2012-08-02       Impact factor: 4.497

View more
  7 in total

1.  Assessing the accuracy of automatic speech recognition for psychotherapy.

Authors:  Adam S Miner; Albert Haque; Jason A Fries; Scott L Fleming; Denise E Wilfley; G Terence Wilson; Arnold Milstein; Dan Jurafsky; Bruce A Arnow; W Stewart Agras; Li Fei-Fei; Nigam H Shah
Journal:  NPJ Digit Med       Date:  2020-06-03

2.  An Automated Quality Evaluation Framework of Psychotherapy Conversations with Local Quality Estimates.

Authors:  Zhuohao Chen; Nikolaos Flemotomos; Karan Singla; Torrey A Creed; David C Atkins; Shrikanth Narayanan
Journal:  Comput Speech Lang       Date:  2022-03-28       Impact factor: 3.252

3.  Assessing the accuracy of automatic speech recognition for psychotherapy.

Authors:  Adam S Miner; Albert Haque; Jason A Fries; Scott L Fleming; Denise E Wilfley; G Terence Wilson; Arnold Milstein; Dan Jurafsky; Bruce A Arnow; W Stewart Agras; Li Fei-Fei; Nigam H Shah
Journal:  NPJ Digit Med       Date:  2020-06-03

4.  Utilization of Time Series Tools in Life-sciences and Neuroscience.

Authors:  Harshit Gujral; Ajay Kumar Kushwaha; Sukant Khurana
Journal:  Neurosci Insights       Date:  2020-12-08

5.  A patient-centered digital scribe for automatic medical documentation.

Authors:  Jesse Wang; Marc Lavender; Ehsan Hoque; Patrick Brophy; Henry Kautz
Journal:  JAMIA Open       Date:  2021-02-17

6.  Exploring the Discursive Emphasis on Patients and Coaches Who Participated in Technology-Assisted Diabetes Self-management Education: Clinical Implementation Study of Health360x.

Authors:  Muhammed Y Idris; Ernest Alema-Mensah; Elizabeth Olorundare; Mohammad Mohammad; Michelle Brown; Elizabeth Ofili; Priscilla Pemu
Journal:  J Med Internet Res       Date:  2022-03-18       Impact factor: 7.076

Review 7.  Challenges of developing a digital scribe to reduce clinical documentation burden.

Authors:  Juan C Quiroz; Liliana Laranjo; Ahmet Baki Kocaballi; Shlomo Berkovsky; Dana Rezazadegan; Enrico Coiera
Journal:  NPJ Digit Med       Date:  2019-11-22
  7 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.