Published on in Vol 13 (2024)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/56729, first published .
Improving Triage Accuracy in Prehospital Emergency Telemedicine: Scoping Review of Machine Learning–Enhanced Approaches

Improving Triage Accuracy in Prehospital Emergency Telemedicine: Scoping Review of Machine Learning–Enhanced Approaches

Improving Triage Accuracy in Prehospital Emergency Telemedicine: Scoping Review of Machine Learning–Enhanced Approaches

Review

1Department of Family Practice, Faculty of Medicine, The University of British Columbia, Vancouver, BC, Canada

2Department of Emergency Medicine, Faculty of Medicine, The University of British Columbia, Vancouver, BC, Canada

3Faculty of Health Sciences, Simon Fraser University, Burnaby, BC, Canada

4School of Biomedical Engineering, Faculty of Applied Science, The University of British Columbia, Vancouver, BC, Canada

5Faculty of Medicine, The University of British Columbia, Vancouver, BC, Canada

6Computer Science, Faculty of Science, The University of British Columbia, Vancouver, BC, Canada

7Faculty of Applied Health Sciences, Brock University, St. Catharines, ON, Canada

Corresponding Author:

Kendall Ho, MD

Department of Emergency Medicine

Faculty of Medicine

The University of British Columbia

818 W10 Ave

3rd floor

Vancouver, BC, V5Z1M9

Canada

Phone: 1 6048220327

Email: kendall.ho@ubc.ca


Background: Prehospital telemedicine triage systems combined with machine learning (ML) methods have the potential to improve triage accuracy and safely redirect low-acuity patients from attending the emergency department. However, research in prehospital settings is limited but needed; emergency department overcrowding and adverse patient outcomes are increasingly common.

Objective: In this scoping review, we sought to characterize the existing methods for ML-enhanced telemedicine emergency triage. In order to support future research, we aimed to delineate what data sources, predictors, labels, ML models, and performance metrics were used, and in which telemedicine triage systems these methods were applied.

Methods: A scoping review was conducted, querying multiple databases (MEDLINE, PubMed, Scopus, and IEEE Xplore) through February 24, 2023, to identify potential ML-enhanced methods, and for those eligible, relevant study characteristics were extracted, including prehospital triage setting, types of predictors, ground truth labeling method, ML models used, and performance metrics. Inclusion criteria were restricted to the triage of emergency telemedicine services using ML methods on an undifferentiated (disease nonspecific) population. Only primary research studies in English were considered. Furthermore, only those studies using data collected remotely (as opposed to derived from physical assessments) were included. In order to limit bias, we exclusively included articles identified through our predefined search criteria and had 3 researchers (DR, JS, and KS) independently screen the resulting studies. We conducted a narrative synthesis of findings to establish a knowledge base in this domain and identify potential gaps to be addressed in forthcoming ML-enhanced methods.

Results: A total of 165 unique records were screened for eligibility and 15 were included in the review. Most studies applied ML methods during emergency medical dispatch (7/15, 47%) or used chatbot applications (5/15, 33%). Patient demographics and health status variables were the most common predictors, with a notable absence of social variables. Frequently used ML models included support vector machines and tree-based methods. ML-enhanced models typically outperformed conventional triage algorithms, and we found a wide range of methods used to establish ground truth labels.

Conclusions: This scoping review observed heterogeneity in dataset size, predictors, clinical setting (triage process), and reported performance metrics. Standard structured predictors, including age, sex, and comorbidities, across articles suggest the importance of these inputs; however, there was a notable absence of other potentially useful data, including medications, social variables, and health system exposure. Ground truth labeling practices should be reported in a standard fashion as the true model performance hinges on these labels. This review calls for future work to form a standardized framework, thereby supporting consistent reporting and performance comparisons across ML-enhanced prehospital triage systems.

Interact J Med Res 2024;13:e56729

doi:10.2196/56729

Keywords



Surging emergency department (ED) visits lead to overcrowding in the ED setting, contributing to adverse patient outcomes, staffing challenges, and health system constraints [1]. Challenges in maintaining ED capacity are estimated to cost millions in health care expenditures [2]. Various interventions have been proposed to improve ED conditions, and prehospital telemedicine triage systems are among the most promising, having the potential to prioritize patients based on their likelihood of requiring emergency versus community-based care [3], potentially alleviating the influx of low-acuity patients that would otherwise be managed in the high-cost, resource-intensive ED setting [4,5]. Such systems, including emergency medical dispatch, nurse-staffed telephone lines, and symptom checkers (chatbots), share the common goal to triage patients based on the information that is provided at the first contact for an urgent health concern. These prehospital services often do not include physician assessments, instead using either rule-based algorithms or health personnel for patient triage [6]. In telemedicine, defined as the delivery of health care services at a distance [7], the inherent scarcity of objective or physical measures such as vital signs has spurred efforts to improve risk prediction using machine learning (ML) models applied to a wide array of information sources such as free text from patient intake calls and vital symptoms monitoring [6,8].

Machine learning has been recognized as one option for improving the accuracy of prehospital telemedicine triage systems. To date, ML has commonly been applied in areas of precision medicine (ie, prediction of the success of treatment regimens), though it is rapidly expanding into diverse sectors of health care [9]. In the ED setting, ML models show promise in their ability to accurately predict inpatient admissions and sepsis [10,11]. Incorporating contextual information into ML models can improve prediction of prehospital emergency services [5]. However, there is a lack of understanding and evidence-based practices regarding how ML can optimally be implemented in remote prehospital settings as compared with more data-rich, in-person settings such as the ED [5,10,11].

In supervised learning, ML models learn from labeled data that serve as the “ground truth.” Ground truth refers to the nature of the problem that is the target of the ML model; in the context of prehospital triage, ground truth is the “correct” triage outcome of a patient. The exact process for defining ground truth is complex and substantially varies across studies. Unlike clear binary ground truths, such as “alive” or “deceased,” determining the “correct” triage outcome is more complex, potentially involving subsequent chart reviews or nuanced clinician assessments, thus introducing ambiguity.

A review of the predictors, labels, and models used in prehospital triage systems needs to establish an evidence base for efficient ML methods of triaging patients and identification of gaps not included in the existing models. Without a review, it is unclear whether existing methods and models may generalize to certain settings or be free from biases [12]. Thus, we conducted a scoping review to support our understanding of ML applications in remote prehospital settings. Scoping reviews are useful in the research fields of social sciences and health care, especially when a topic has not yet been comprehensively reviewed or exhibits a large, complex, or heterogeneous nature not amenable to a more thorough systematic review [13]. While there have been literature reviews outlining ML methods in emergency settings [14-16], no specific review touches on prehospital telemedicine triage services. Related reviews such as Sánchez-Salmerón et al [15] focused on in-person triage, as opposed to prehospital and remote triage. Through this scoping review, we aim to explore what evidence exists to compare the effectiveness of ML-enabled strategies with conventional triage methods in improving outcomes for patients seeking care through telemedicine services. We also aim to explore (1) what data sources and approaches are used for extracting meaningful predictors and labels, (2) what models and performance metrics are used, and (3) the processes of telemedicine triage systems where ML has been applied.


Search Strategy

Elements of the PRISMA-ScR (Preferred Reporting Items for Systematic reviews and Meta-Analyses extension for Scoping Reviews) protocol [17] and the Population, Intervention, Comparison, and Outcome (PICO) framework [18,19] were used to guide our search strategy. The search included all articles published prior to February 24, 2023. Methodological frameworks for scoping reviews from Arksey and O’Malley [20] and Levac et al [21] were followed.

The search strategy began with initial searches conducted through MEDLINE to extract terms based on article titles and abstracts. Keywords and expressions included both regular and Medical Subject Headings (MeSH) terms. The following search expression was developed and applied: “Telephone/ or telephone.mp. OR phone.mp. OR telemedicine.mp. OR Telemedicine/ OR Emergency Medical Service Communication Systems/ OR Emergency Medical Dispatch/ OR dispatch.mp. OR hotline.mp. OR Hotlines/ OR prehospital.mp. OR pre-hospital.mp. OR remote consultation.mp. OR Remote Consultation/” AND “machine learning.mp. OR Artificial Intelligence/ OR Machine Learning/ OR artificial intelligence.mp. OR natural language processing.mp. OR Natural Language Processing/ OR chatbot.mp.” AND “triage/ OR triage.mp”

Our search followed the 3 phases listed from the PRISMA-ScR protocol: identification, screening, and inclusion [17] (Multimedia Appendix 1). Using the search expression, articles were retrieved from the following databases: MEDLINE, PubMed, Scopus, and IEEE Xplore. These databases were selected for this review due to their ability to cover the most scientific information in fields such as telemedicine and health care decision-making while also being previously used by several health and technology-related reviews [15,16,22]. In order to limit bias and ensure reproducibility, we elected to exclusively include articles identified through the predefined search strategy.

Article Selection

All identified records were combined (n=296) and duplicates (n=131) removed resulting in 165 records remaining. The results of the initial search after duplicate removal are found in Multimedia Appendix 2. Articles needed to meet a set of inclusion and exclusion criteria (Table 1), which were developed to ensure that articles were relevant to prehospital telemedicine triage, conducted remotely as opposed to in the ED. During an initial piloting phase, 3 researchers (DR, KS, and MY) piloted these criteria to ensure consistency across them. Each researcher independently screened each title based on the inclusion and exclusion criteria, with discrepancies reaching consensus through discussion, which resulted in 32 records remaining (n=133 excluded). Articles that passed the title screening were then screened for relevant abstracts following a similar process resulting in 15 records remaining (n=17 excluded). The remaining articles were then read in their entirety and all 15 records were deemed eligible and included in the results of the review. Of the articles excluded, most were excluded for the provision of in-person care (as opposed to data collected remotely), were not primary research studies, or were not for general purpose triage (eg, stroke-specific triage).

Table 1. Inclusion and exclusion criteria of screening strategy.
PICOa/other elementInclusion criteriaExclusion criteria
PopulationUndifferentiated population seeking emergency services (including COVID-19 assessments).Specialty-specific population (eg, stroke, heart disease).
ConceptTriage of emergency telemedicine services enhanced by any MLb method that includes only data collected remotely.Only conventional triage methods used or ML models that include predictors derived from physical assessments (eg, vital signs). Internet of medical things devices requiring physical or in-person assessments were excluded (eg, home blood pressure equipment).
ContextProvision of emergency telemedicine services, including web-based symptom checkers, clinician-staffed telephone line, or emergency medical dispatch.Provision of in-person emergency care.
Evidence typePrimary research studies.Literature reviews, protocols, guidelines, letters, gray literature, and qualitative studies excluded.
LanguageStudies published in English.Studies in languages beyond English.
DatePublished before February 24, 2023.Published on or after February 24, 2023.

aPICO: Population, Intervention, Comparison, and Outcome.

bML: machine learning.

Data Extraction

We extracted data from the included articles with an aim to understand how research in this domain is conducted. A data extraction tool was developed using the JBI Manual for Evidence Synthesis template [23] with a pilot step on 2 sources conducted by 2 researchers (JS and DR). The data of interest fell in four main categories:

  1. Study characteristics: Author, year of publication, country of origin, the aim of the study, the result of the study, population assessed, dataset source, dataset size (the number of patient records), prospective/retrospective/deployed, and triage process.
  2. ML model predictors and labels: Number of predictors and data types; methods for determining ground truth labels.
  3. ML techniques and corresponding performance metrics:
  4. Comparators, dataset partitions, data-preprocessing methods, performance metrics and values, and data quality analysis approaches.
  5. Resources for future ML model development: Source code availability and software packages used.

This list of extraction items was supported by prior literature that involved emergency care–related triage [22,24] and was refined based on discussions among the authors. For each of the eligible articles included in this review, 1 author (JS) extracted and tabulated the relevant information, and the information was validated by another author (DR) with discrepancies further assessed by a last author (KS). All extracted information was analyzed by the authors to derive a narrative synthesis of the findings. Aligned with the scoping review methodology, articles were not assessed for quality or risk of bias and no statistical analyses were conducted. The data extraction tool and raw data extracted are available in Multimedia Appendix 2.

Triage Process

We found that the studies spanned 3 prehospital triage processes: emergency dispatch, telephone lines, and chatbot applications. When distinguishing between emergency dispatch and telephone line systems, context of the call and nature of interaction were the 2 main points of consideration. Emergency dispatch calls (eg, 9-1-1) are received and handled by trained emergency staff focused on collecting critical information on the emergency, such as the nature of the incident, location, and immediate risks. The dispatcher then makes decisions based on this information to allocate appropriate resources, such as ambulances or first responders. Conversely, telephone lines such as nurse-led helplines or crisis hotlines are designed to offer support, advice, and clinical guidance to individuals seeking health care information or experiencing a crisis [25]. The interaction is often more conversational and supportive, resulting in a more complex triage process that depends on the nature of the call, the expertise of the health care professionals involved, and the available resources or referrals.

Predictors

We classified the predictor variables into 4 domains (demographic factors, operational characteristics, clinical factors, and unstructured data such as free text) and extracted how these data were handled in the models. These domains were determined based on the prehospital triage field and iteratively refined as data were extracted.

Ground Truth Labels

Variability existed in how ground truth labeling methods were coded for ML processing due to diverse sources of training data and annotation methods. In our classification of the observed ground truth methods, we first identified whether the label was derived from (1) the usual clinical process, that is, subsequent clinician triage, or (2) outside of the usual clinical process, that is, post hoc. Within category 1, we further identified whether (1A) remote-only data were used, or (1B) physical, in-person data were used. Within category 2, we further classified whether (2A) labels were automatically derived or (2B) labels were manually derived. Automatically (systematically) derived labels are the results of a uniform and scalable application of a label-deriving algorithm across a dataset.


Characteristics of the Studies

Figure 1 shows a summary of the search protocol phases followed for this review using the PRISMA-ScR flow diagram. Of the 15 unique studies included, 33% (5/15) of the articles were published in 2022, 26% (4/15) in 2021, 20% (3/15) in 2020, and the remaining 20% (3/15) of the articles between 2014 and 2019. The most frequently occurring country of publication was the United States (4/15, 27%), followed by Japan (2/15, 13%), with the remaining being published in 9 distinct countries (Table 2).

Most studies (11/15, 73%) were retrospective, using historical patient outcomes to assess the performance of ML models in triage prediction. Among the retrospective articles, 2 reported a combination of retrospective results and the performance of deployed models. Of the other 4 nonretrospective articles, 13% (2/15) focused solely on deployed models and 13% (2/15) carried out a prospective study to explore the applications of ML triage in a specific health care setting.

Moreover, 47% (7/15) of articles investigated the use of ML in emergency dispatch calls, 33% (5/15) of articles focused on chatbot-style applications that could be accessed via the internet or within the ED, and 20% (3/15) of articles investigated ML in telephone lines, including 13% (2/15) nurse-led phone lines and 7% (1/15) crisis hotline.

Figure 1. PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses Extension for Scoping Reviews) flow diagram.
Table 2. List of 15 studies included in the review and their characteristics.
Author (year)Study setting (country)Dataset sizeaLabels includedAimTriage phase
Anthony et al (2021) [26]South Africa933 binaryClassify critical conditions in emergency callsDispatch
Ceklic et al (2022) [27]Australia11,9711 binaryClassify severity of traffic crash incidents in dispatch callsDispatch
Chin et al (2022) [28]Taiwan1141 binaryClassify severity of traffic crash incidents in dispatch callsDispatch
Cotte et al (2022) [29]Germany3851 multiclassClassify triage decisions using a symptom assessment appChatbot
Ferri et al (2021) [30]Spain1,244,6243 binary and multiclassClassify emergency incidents in dispatch callsDispatch
Gatto et al (2022) [31]United States5741 binaryClassify severity in patient’s text-based inquiriesChatbot
Inokuchi et al (2022) [32]Japan15,4421 binaryIdentify undertriage in prehospital telephone triageNurse-led phone line
Lai et al (2020) [33]United StatesbClassify triage for prehospital COVID-19 casesChatbot
Marchiori et al (2021) [25]Switzerland>900,000c1 multiclassEvaluate AId-powered chatbot for symptom-checker triageChatbot
Morse et al (2020) [34]United States26,646Evaluate AI-powered chatbot for symptom-checker triageChatbot
Pacula et al (2014) [35]United States4272 multiclassClassify triage and distress indicators in crisis hotline chatsCrisis hotline
Spangler et al (2019) [3]Sweden68,6681 continuousValidate MLe-generated risk scores for prehospital careDispatch (operated by nurses)
Tollinton et al (2020) [5]United Kingdom1,188,5091 binaryClassify triage of unconscious patients in dispatch callsDispatch
Veladas et al (2021) [36]Portugal269,6691 multiclassClassify clinical pathways from text dataNurse-led phone line
Yunoki et al (2014) [4]Japan61,9271 multiclassClassify triage categories from phone call dataDispatch

aNumber of patient records included.

bDataset size used for model development was not stated for this study or information on labels was not included.

cThe study stated that “more than 900,000 case records” were included.

dAI: artificial intelligence.

eML: machine learning.

The size of the dataset used across studies varied considerably, with a median sample size of 21,044 observations. The largest dataset comprised 1,224,624 anonymized patient records [30] and the smallest dataset included 93 call transcripts [28]. Distinct methods were reported to handle missing data; for example, Ferri et al [30] excluded all call records with missing values, while Inokuchi et al [32] performed imputation to account for missing data (3918/19,114, 20.5% of cases had missing data) using the k-nearest neighbors (k-NN) algorithm.

Predictors

Several articles used predictors that spanned multiple categories, as exemplified by Ferri et al [30], who extracted patient demographic data (age, gender), operational data (date, caller type), and unstructured data (clinical free-text observations). In contrast, Chin et al [28] used unstructured data (dispatch call transcript) as the only data source in their ML model development, citing the “higher expressiveness of patient condition than structured data.” The frequency of occurrence for each of these domains is shown in Figure 2.

The mapping and processing techniques varied across data types. Most commonly, many input variables were one-hot encoded, such as age groupings, sex, day of week, clinical indicators (symptoms, comorbidities), and question-answer pairs. Notably, distance to ED was one of the few continuous predictors used [3]. For unstructured data, various natural language processing (NLP) techniques were employed to transform these data to useful inputs, such as bag-of-words methods or text vectorization, which are strategies to transform text data into a numerical representation that can be processed by ML models (see Multimedia Appendix 3 for more information).

Figure 2. Percentage of predictor types included in the studies. ED: emergency department; Q&A: question and answer.

Data Labeling: Determining Ground Truth

Two studies (2/15, 13%) [33,34] did not provide sufficient information to determine their ground truth labeling method. Most of the ground truth methods (5/15, 33%) used physical assessments from the usual clinical process; however, within this group input data were from 3 different types of providers (nurses, physicians, and ambulance crew) with each employing a different triage or categorization system (Figure 3) [25]. This results in variability as similar information is outputted with different labels, which limits comparability (eg, patient conveyance [5] or “gold standard” ED triage protocols [4]). Notably, the 3 (20%) studies [3,27,30] that used automatic post hoc (2A) methods had larger sample sizes (mean 4,32,792 observations; range 15,550-1,244,624) versus studies [26,31,35] using manual (2B) methods (mean 364 observations; range 93-573). Across the 4 methods, 6 studies (40%) predicted labels that were binary and based on symptom or condition severity (eg, life-threatening [30], severe trauma [28]), while 7 studies (47%) predicted multiclass labels (eg, 1 of 53 clinical pathways [36]).

Figure 3. Label and ground truth methodologies by annotation method. *Marchiori et al used remote physician triage as ground truth for model training but manually derived labels for model testing. To ensure consistency in article classification, we have categorized all articles based on the ground truth method for model training. ED: emergency department; EMCI: emergency medical call incident; EMT: emergency medical technician; L&S: lights and sirens; MTS: Manchester Triage System.

Interrater Reliability

Methods to enhance interrater reliability were rarely employed, in contrast to published guidelines concerning manual data labeling [37]. Marchiori et al [25] did so by ensuring the adequate training of data collectors by selecting “only records generated by top-ranked doctors (based on years of experience and internal audits).” In the study by Gatto et al [31], data annotators “invested significant time to educate themselves on the symptoms” observed in the dataset; however, it was noted “that the use of nonmedical professionals limited the degree of granularity with which the dataset was labelled.” Pacula et al [35] used consensus among 3 evaluators when labeling data and measured the interannotator agreement using the Fleiss kappa statistic to confirm good agreement.

ML Algorithms

Most studies applied more than 1 ML technique to method development (11/15, 73%). However, 2 studies (13%) did not disclose the ML algorithm used in their triage prediction model. Three (20%) studies published the source code for various stages of their ML model development and validation (Table 3).

Overall, the most popular ML algorithm was support vector machine (SVM; 8/15, 53%) and tree-based methods, including random forest (RF; 6/15, 40%) and extreme gradient boosting (XGBoost; 4/15, 27%). Other popular algorithms were neural networks (NN; 5/15, 33%), with various implementations observed, such as deep NNs, bidirectional long short-term memory models, and ensemble of deep learning networks (ie, the model employed by Ferri et al [30] is described as a “Deep Ensemble Multitask Classifier for Emergency Medical Calls” composed of 4 different subnetworks). Naïve Bayes (NB) and k-NN were similarly popular, followed by regression methods.

Table 3. Type and frequency of ML algorithms used by the included studies.
MLa algorithmStudies, nAuthors (citations)
Support vector machine8Anthony et al [26], Ceklic et al [27], Chin et al [28], Gatto et al [31], Inokuchi et al [32], Pacula et al [35], Spangler et al [3]b, and Veladas et al [36]
Random forest6Anthony et al [26], Ferri et al [30], Inokuchi et al [32], Tollinton et al [5], Spangler et al [3], and Veladas et al [36]
Neural networks5Ceklic et al [27], Ferri et al [30], Gatto et al [31], Inokuchi et al [32], and Marchiori et al [25]
Extreme gradient boosting (XGBoost)4Ferri et al [30], Inokuchi et al [32], and Spangler et al [3]
Naïve Bayes4Ceklic et al [27], Chin et al [28], Ferri et al [30], and Veladas et al [36]
K-nearest neighbors4Anthony et al [26], Ceklic et al [27], Chin et al [28], and Gatto et al [31]
Logistic regression3Anthony et al [26], Ferri et al [30], and Spangler et al [3]
Bayesian network2Cotte et al [29] and Yunoki et al [4]
Decision treec2Chin et al [28] and Tollinton et al [5]
Ensembled2Ceklic et al [27] and Ferri et al [30]
LASSOe regression1Inokuchi et al [32]
Multilayer perceptron1Chin et al [28]
Hidden Markov model1Pacula et al [35]
Hierarchical attention network1Gatto et al [31]
Transformer-based1Gatto et al [31]
Unspecified algorithms2Lai et al [33] and Morse et al [34]

aML: machine learning.

bSpangler et al [3] did not include support vector machine, random forest, neural networks, or regression models in their “Methods” but stated in their “Discussion” that they investigated these algorithms.

cDecision trees here include methods such as gradient boosting but not random forest or extreme gradient boosting.

dCeklic et al [27] did not provide information about the specific models incorporated into their ensemble. Ferri et al [30] built their ensemble using a collection of deep learning subnetworks.

eLASSO: least absolute shrinkage and selection operator.

Model Performance Metrics

F1-score was the most used metric (6/15, 40%) to measure model performance, closely followed by accuracy (5/15, 33%) and area under the curve (AUC; 4/15, 27%). Furthermore, many studies tangentially used other quantitative measures, such as sensitivity or recall and specificity or precision. In the case where AUC scores were similar between the RF and XGBoost models in the study by Tollinton et al [5], analysis of other metrics showed that XGBoost outperformed RF models in terms of sensitivity (0.93 vs 0.62 in the combined model) but had lower specificity (0.17 vs 0.56 in the combined model). Moreover, several studies investigated how different combinations of predictors and NLP techniques affected model performance across ML techniques. For instance, Pacula et al [35] present an interesting discussion on the impact of various approaches to dialogue processing (turn-level classification and accounting for speaker role) on triage prediction outcomes.

A summary of the comparators and top-performing models for each study, alongside information on evaluation metrics used and data train or test split, is shown in Table 4. For models that were evaluated against non–ML-enhanced methods of triage, those established triage systems are indicated as well.

Table 4. Top-performing ML algorithms used in each study and corresponding performance metric and model training or testing split grouped by triage process.
Triage process and studyTop-performing modelModel comparatorsTriage comparatorsPerformance metricsTraining data split (%)
Emergency medical dispatch

Anthony et al [26]SVMaLRb, RFc, k-NNdeAccuracy80

Ceklic et al [27]Ensemblek-NN, SVM, NBf, deep NN +NLPgSJ-WAh dispatchPPVi, sensitivity (recall), F1-score60

Chin et al [28]Bernoulli NBk-NN, DTj, SVM, NB +NLPDispatcher evaluationAccuracy, PPV, NPVk, sensitivity (recall), specificity91

Ferri et al [30]EnsembleNB, LR, RF, GBl +NLPClinical decision treeAccuracy, sensitivity (recall), PPV, F1-score80

Spangler et al [3]XGBoostmSVM, LR, RF, deep NNDispatch priority, National Early Warning ScoresAUCn, PPV, sensitivity (recall)66

Tollinton et al [5]GB/RFoAUC, sensitivity (recall), specificity80

Yunoki et al [4]BNp,qAccuracy90
Symptom checker/chartbotr

Cotte et al [29]BNqMTSsCohen κN/At

Gatto et al [31]SBERTuBERTv, SVM, HANw, bi-LSTMx +NLPPPV, sensitivity (recall), F1-score80

Marchiori et al [25]Bi-LSTMConvolutional NN, recurrent NN +NLPPPV, sensitivity (recall), F1-score60
Telephone line

Inokuchi et al [32]RFqLASSOy regression, deep NN, XGBoostArea under the receiver operating characteristic curve, PPV, NPV, sensitivity (recall), specificity70

Pacula et al [35]SVMqHMMz +NLPAUC, F1-score82

Veladas et al [36]SVMRF, NB +NLPAccuracy, F1-score, PPV, sensitivity (recall)64

aSVM: support vector machine.

bLR: logistic regression.

cRF: random forest.

dK-NN: K-nearest neighbors.

eNo comparator.

fNB: naïve Bayes.

g“+NLP” indicates that model performance was evaluated across various natural language processing techniques.

hSJ-WA: St John Ambulance in Western Australia.

iPPV: positive predictive value (precision).

jDT: decision tree.

kNPV: negative predictive value.

lGB: gradient boosting.

mXGBoost: extreme gradient boosting.

nAUC: area under the curve.

oThe gradient boosting model scored better on sensitivity, but specificity was lower than RF.

pBN: Bayesian network.

qAlternate ML models were not compared.

rLai et al [33] and Morse et al [34] reported on symptom checker triage systems but did not provide details on the underlying ML model and comparators or development.

sMTS: Manchester Triage System.

tN/A: not applicable.

uSBERT: sentence bidirectional encoder representation from transformers.

vBERT: Bidirectional Encoder Representation From Transformers.

wHAN: hierarchical attention network.

xBi-LSTM: bidirectional long short-term memory.

yLASSO: least absolute shrinkage and selection operator.

zHMM: hidden Markov model.

Model Data Splitting

Thirteen studies (87%) reported some method for validation, including randomly partitioning data into training or testing sets (10/15, 67%) or training or validation or testing sets (3/15, 20%). Most studies used some form of cross-validation in the training sets for model construction (8/15, 53%). In addition, various types of resampling procedures were used, such as k-fold cross-validation, stratified k-fold cross-validation, repeated k-fold cross-validation, or repeated random test-train splits.


Principal Findings

We reviewed 15 primary research studies of ML-enhanced triage models in prehospital telemedicine settings where patients potentially require emergency care. We found that ML-enhanced triage systems typically outperformed conventional triage ones; however, there is likely a bias toward publishing these types of positive findings [38]. While the reviewed studies exhibited several commonalities in terms of predictor types, ML algorithms tested, and performance metrics, there were key discrepancies in how data were sourced and processed, particularly with regard to annotating ground truth labels. We discuss these similarities and differences in context and how the new evidence presented here relates to ED overcrowding and prehospital triage.

Predictor Variables

A critical limitation of prehospital telemedicine triage systems is the lack of access to objective measures of the patient condition typically obtained through physical assessment, vital signs being fundamental metrics of illness severity. As a result, our inclusion criteria ensured that no study had access to a physical assessment of their participants. Therefore, by investigating the range of predictors employed by the included articles, we identified how subjective and indirect indicators of patient condition are used in remote triage. Our review indicates that patient symptoms, age, sex, and comorbidities were the most frequently occurring predictor variables among structured data, with NLP techniques used to extract features from unstructured data. The structured and unstructured data were not combined in any of the included articles. This finding emphasizes the critical importance of considering these 4 clinical and demographic factors for prehospital telemedicine triage where physical assessment data are unavailable. The consistent use of patient symptoms, demographics, and comorbidities as predictor variables across all 3 triage processes reinforces that they are reliable indicators of a suitable triage outcome. However, while demographic factors were frequently used across the studies, only 2 specific factors, namely, age and sex, were considered.

Notably, we also highlight the absence of previously identified important inputs to ML models [12]. Race or ethnicity, region or geography, medication history, and health system exposure (hospitalizations, etc) were not represented in the corpus; however, they should be taken into consideration in prehospital triage both to improve performance and ensure algorithmic fairness [39]. The absence of these variables both as predictors and for post hoc evaluation of algorithmic bias suggests a significant gap in the extant literature and motivates further investigation into their potential to generate accurate and equitable triage outcomes for more diverse populations. Continuing to develop a comprehensive list of the most significant variables driving remote emergency triage is invaluable to improving equitable patient outcomes for all populations [12]. Identifying common patterns in predictor selection (and what is absent) can inform the development of standardized guidelines for building ML algorithms for triage using remote-only data. A future with consistent measurements of physiologic metrics, such as vital signs, would also be invaluable to strengthen prediction. Developing systems or technology whereby these critical data can be captured remotely is a future direction worth exploring.

Ground Truth Labeling Methods

Our review critically spotlights the variety of methods for annotating data labels from ground truth. Two papers did not provide sufficient information, and we classified the remaining 13 into 1 of 4 distinct methods, reflecting varying ML development philosophies. The most common method—(2A) post hoc automatic derivation of data labels—indicates preference for collaboration among domain experts to reduce human subjectivity and implement large-scale data mapping. Notably, Ferri et al [30] used a panel of 17 physicians to develop a mapping system, which was then automatically applied to more than 1 million records. In contrast, using the (2B) method, Pacula et al [35] reported that 3 psychologists manually annotated each of the 427 records. While this review cannot comment on the comparative efficiencies of these specific methods, we note that assembling large datasets with ground truth labels is an arduous and expensive task; thus, there may be scalability benefits to automatic methods [40].

While we observed trade-offs in data labeling, establishing methodologically robust ground truth is of paramount importance. Inadequate representation of ground truth can lead to misclassification issues within the ML model, thus reducing predictive accuracy on external data. This could result in unintended and potentially serious consequences, especially in the context of emergency care. The ground truth triage outcome has proved challenging to pinpoint, and this scoping review reveals 4 methods in determining it. However, each method exhibits limitations and potential manifestations of human subjectivity.

The accuracy of remote triage is inherently limited by the absence of physical data, even when carried out by individuals with domain expertise (eg, physicians). Furthermore, human bias and lack of experience can lead to either overtriage (ie, false positives; eg, sending low-acuity patients to the ED) or undertriage (ie, false negatives; eg, advising high-acuity patients to stay home), even in cases when physical data are available. To address this, the strategy of selecting data from only highly reputable clinicians [25] serves as an inspiration for developing a standardized evaluation system to identify qualified data annotators for health care settings. An additional method to enhance triage accuracy is inclusion of downstream triage with access to more comprehensive data, such as vital signs, detailed physical assessment, or a professional with higher-level training (eg, a physician conducting a home visit) [32].

The methodical classification of patients based on a predetermined list of outcomes minimizes the risk for interrater disagreement. However, relying solely on these selected outcomes, such as hospital admission or 2-day mortality [3], assumes that they are the only factors and of equal importance in determining the overall risks associated with the patient. This approach overlooks insights that can be revealed from a circumstantial and personalized analysis of patient condition. In addition, many of the inclusion studies selected highly specific outcomes, such as sepsis, myocardial infarction, and cardiac arrest [26], which are not generalizable to different remote triage processes. One potential solution, which no study in this review used, is using a weighted kappa index to consider different categories and disagreements and capture the rank magnitude of disagreement [41].

Retrospective evaluation of patient data by multiple annotators holds potential for highly accurate ground truth labels, as it is the most comprehensive approach observed in this review. However, we note that the included articles lack information regarding how evaluators determined such classifications. Subjectivity in human decision-making persists and variation in annotators’ levels of training and resolution strategies employed leaves room for further research. Again, implementing a standardized evaluation system to determine a qualified pool of data annotators becomes crucial to ensure reliability in the annotation of ground truth triage labels for ML-enhanced remote triage [37].

This analysis of subjectivity in ML systems underscores the need for nuance regarding objectivity and ground truth. To enhance remote triage algorithm generalizability, comprehensive datasets must be used. This entails capturing data from each of the predictor variables outlined in this review, as well as race or ethnicity, region or geography, and medical history. Furthermore, standardized protocols for data labeling across different clinical settings would facilitate consistency and enable meaningful comparisons and analyses. There is a need for more well-defined and externally valid procedures for determining the ground truth in the context of highly intricate and unstructured data such as remote patient triage.

ML Techniques

This review presents needed insights on the cutting-edge application of ML methods for remote prehospital triage systems: SVM, tree-based, and NN methods were commonly used. These observations are consistent with other reviews of ED-based implementations [16,42]. To our knowledge, this is the first review to provide a mapping between ML model development and 3 processes of triage in emergency medicine: emergency medical dispatch, chatbot or symptom checker, and telephone line. We uniquely charted the best-performing model of each article to 1 of the 3 triage processes and generated insight on the specific elements underpinning each model. By doing so, we provide insight into how predictive ML models can be developed for different stages of triage before physical assessment.

Limitations

We did not directly compare the varied performance metrics such as the F1-score, precision, recall, and AUC, as this could lead to misleading conclusions. This was not an objective of our review due to inconsistencies in reported metrics and variability in study designs. For instance, certain metrics can be optimized at the expense of other metrics; for example, precision and recall are inversely related. Not all metrics are universally applicable; the F1-score is unsuitable for multiclass problems found in a proportion of included studies. In addition, F1-score assumes false negatives and false positives are equally costly, an incorrect assumption in triage problems where undertriage due to a false negative would have serious consequences. While AUC is perhaps the optimal metric choice, it still has sensitivity to class imbalances, which varied across the included studies. Of the only 4 articles that reported AUC, the performance ranged from 0.64 for prediction of conveyance [5] to 0.88 [35]. A direct comparison among articles reporting AUC was avoided due to differences in study contexts, which included diverse triage labels, significant variations in sample sizes and populations, limited reporting on class imbalance, and the use of different validation strategies. As this review did not facilitate any quantitative analysis of ML model performance, insights into the accuracy of ML-enhanced triage compared with conventional remote triage, as well as the quantitative impact that ML triage may have on patient outcomes and overall health care systems, could not be derived and future systematic work is warranted.

Implications for Future Work

ML-enhanced triage presents an opportunity to alleviate the burden on EDs and support patients’ decision-making when seeking emergency versus community-based care. Based on the evidence synthesized here, our calls to the field are to determine a prioritized list of high-value predictor variables to consider, standardize ground truth labeling, and form a consensus on validation methods used, such that different health systems can continuously learn from new developments. This review provides a foundation for developing guidelines, which will also create opportunities for comparison across studies to quantitatively assess the effectiveness and benefits of ML-enhanced triage regarding patient outcomes and health system performance. While reducing resources spent on overtriage is also a priority, it is equally important to focus on the likelihood of undertriage, as it poses a significant risk to patients. Therefore, precision, recall, and specificity rates, as well as algorithmic bias [39], must be carefully monitored and improved in model development to ensure that safety and effectiveness are balanced in ML-enhanced triage systems.

Accurate triage recommendations do not guarantee that patients will follow them [43-45]. In the discussion of using telephone lines or chatbots during the triage process, further investigation is warranted into how delivery methods affect patients’ likeliness to adhere to the advice generated by ML-enhanced triage systems. Similarly, in the dispatch and telephone line triage processes, a question that arises is whether ML-generated triage results influence the decision-making process of dispatchers or providers. This area of research could provide insight on the most effective stage at which ML-based assistance should be introduced in the triage process. Exploring the interplay between ML-enhanced triage advice, patient behavior, and clinician decision-making will contribute to the optimization of prehospital telemedicine triage in emergency care.

Conclusions

Our scoping review of 15 recent studies of ML-enhanced prehospital telemedicine triage systems observed heterogeneity in dataset size, predictors, clinical setting (triage process), and reported performance metrics. Consequently, a comparison of ML performance across articles was not feasible, and we note that identifying the most efficient and accurate ML-enhanced triage system is valuable for future development and model deployment in prehospital settings, where a standardized performance metric such as the AUC would be important to facilitate comparisons. Standard structured predictors, including symptoms, age, sex, and comorbidities, across articles suggest the importance of these inputs; however, there was a notable absence of other potentially useful data, including medications and health system exposure. With advancing technology of transformer-based models [46,47], there exists the potential for combining structured and unstructured data; an approach that was also absent in the included articles. The lack of social variables leaves the potential for algorithmic bias critically unexplored. Ground truth labeling practices should be reported in a standard fashion as the true model performance hinges on these labels. This review establishes an evidence base for future investigations and an opportunity to form a consensus and standardized framework, thereby supporting consistent reporting, performance comparisons, and collaboratively developed ML-enhanced prehospital triage systems.

Authors' Contributions

DR, KS, SC, RT, JW, MCT, and KH developed the idea for the scoping review. DR and KS developed the search methodology. MCY conducted the literature search. MCY and JS wrote the initial draft. DR and KS completed the initial draft. All authors reviewed, edited, and approved the final manuscript.

Conflicts of Interest

KH is the lead of the UBC Digital Emergency Medicine Unit, which was provided grant funding from the BC Ministry of Health to implement the HealthlinkBC Virtual Physicians program, the BC prehospital emergency telemedicine program.

Multimedia Appendix 1

PRISMA-ScR (Preferred Reporting Items for Systematic reviews and Meta-Analyses Extension for Scoping Reviews) checklist.

DOCX File , 106 KB

Multimedia Appendix 2

Data extraction tool and search results.

ZIP File (Zip Archive), 196 KB

Multimedia Appendix 3

Additional findings.

DOCX File , 15 KB

  1. Morley C, Unwin M, Peterson GM, Stankovich J, Kinsman L. Emergency department crowding: a systematic review of causes, consequences and solutions. PLoS One. 2018;13(8):e0203316. [FREE Full text] [CrossRef] [Medline]
  2. Gidora H, Borycki EM, Kushniruk AW. Effects of telenursing triage and advice on healthcare costs and resource use. Stud Health Technol Inform. 2019;257:133-139. [Medline]
  3. Spangler D, Hermansson T, Smekal D, Blomberg H. A validation of machine learning-based risk scores in the prehospital setting. PLoS One. 2019;14(12):e0226518. [FREE Full text] [CrossRef] [Medline]
  4. Yunoki S, Hamagami T, Oshige K, Kawakami C, Suzuki N. High accuracy of call triage decision by Bayesian network. Electron Comm Jpn. 2014;97(1):62-69. [CrossRef]
  5. Tollinton L, Metcalf AM, Velupillai S. Enhancing predictions of patient conveyance using emergency call handler free text notes for unconscious and fainting incidents reported to the London ambulance service. Int J Med Inform. 2020;141:104179. [FREE Full text] [CrossRef] [Medline]
  6. Lidal IB, Holte HH, Vist GE. Triage systems for pre-hospital emergency medical services—a systematic review. Scand J Trauma Resusc Emerg Med. 2013;21:28. [FREE Full text] [CrossRef] [Medline]
  7. World Health Organization. Consolidated telemedicine implementation guide. 2022. URL: https://www.who.int/publications/i/item/9789240059184 [accessed 2023-07-24]
  8. Bohm K, Kurland L. The accuracy of medical dispatch—a systematic review. Scand J Trauma Resusc Emerg Med. 2018;26(1):94. [FREE Full text] [CrossRef] [Medline]
  9. Davenport T, Kalakota R. The potential for artificial intelligence in healthcare. Future Healthc J. 2019;6(2):94-98. [FREE Full text] [CrossRef] [Medline]
  10. King Z, Farrington J, Utley M, Kung E, Elkhodair S, Harris S, et al. Machine learning for real-time aggregated prediction of hospital admission for emergency patients. NPJ Digit Med. Jul 26, 2022;5(1):104. [CrossRef] [Medline]
  11. Adams R, Henry KE, Sridharan A, Soleimani H, Zhan A, Rawat N, et al. Prospective, multi-site study of patient outcomes after implementation of the TREWS machine learning-based early warning system for sepsis. Nat Med. 2022;28(7):1455-1460. [CrossRef] [Medline]
  12. Obermeyer Z, Powers B, Vogeli C, Mullainathan S. Dissecting racial bias in an algorithm used to manage the health of populations. Science. 2019;366(6464):447-453. [CrossRef] [Medline]
  13. Bichu YM, Hansa I, Bichu AY, Premjani P, Flores-Mir C, Vaid NR. Applications of artificial intelligence and machine learning in orthodontics: a scoping review. Prog Orthod. 2021;22(1):18. [FREE Full text] [CrossRef] [Medline]
  14. Shafaf N, Malek H. Applications of machine learning approaches in emergency medicine; a review article. Arch Acad Emerg Med. 2019;7(1):34. [FREE Full text] [Medline]
  15. Sánchez-Salmerón R, Gómez-Urquiza JL, Albendín-García L, Correa-Rodríguez M, Martos-Cabrera MB, Velando-Soriano A, et al. Machine learning methods applied to triage in emergency services: a systematic review. Int Emerg Nurs. 2022;60:101109. [CrossRef] [Medline]
  16. Gao F, Boukebous B, Pozzar M, Alaoui E, Sano B, Bayat-Makoei S. Predictive models for emergency department triage using machine learning: a systematic review. Obstet Gynecol Res. 2022;5(2):136-157. [CrossRef]
  17. Tricco AC, Lillie E, Zarin W, O'Brien KK, Colquhoun H, Levac D, et al. PRISMA Extension for Scoping Reviews (PRISMA-ScR): checklist and explanation. Ann Intern Med. 2018;169(7):467-473. [FREE Full text] [CrossRef] [Medline]
  18. Sackett DL. Evidence-based medicine. Semin Perinatol. 1997;21(1):3-5. [CrossRef] [Medline]
  19. Aslam S, Emmanuel P. Formulating a researchable question: a critical step for facilitating good clinical research. Indian J Sex Transm Dis AIDS. 2010;31(1):47-50. [FREE Full text] [CrossRef] [Medline]
  20. Arksey H, O'Malley L. Scoping studies: towards a methodological framework. Int J Soc Res Methodol. 2005;8(1):19-32. [CrossRef]
  21. Levac D, Colquhoun H, O'Brien KK. Scoping studies: advancing the methodology. Implement Sci. 2010;5:69. [FREE Full text] [CrossRef] [Medline]
  22. Fernandes M, Vieira SM, Leite F, Palos C, Finkelstein S, Sousa JMC. Clinical decision support systems for triage in the emergency department using intelligent systems: a review. Artif Intell Med. 2020;102:101762. [CrossRef] [Medline]
  23. Micah DP, Christina G, Patricia M, Zachary M, Andrea CT, Hanan K. Chapter 11: scoping reviews. In: JBI Manual for Evidence Synthesis JBI. New York, NY. JBI; 2020.
  24. Salman OH, Taha Z, Alsabah MQ, Hussein YS, Mohammed AS, Aal-Nouman M. A review on utilizing machine learning technology in the fields of electronic emergency triage and patient priority systems in telemedicine: coherent taxonomy, motivations, open research challenges and recommendations for intelligent future work. Comput Methods Programs Biomed. 2021;209:106357. [CrossRef] [Medline]
  25. Marchiori C, Dykeman D, Girardi I, Ivankay A, Thandiackal K, Zusag M, et al. Artificial intelligence decision support for medical triage. AMIA Annu Symp Proc. 2020;2020:793-802. [FREE Full text] [Medline]
  26. Anthony T, Mishra AK, Stassen W, Son J. The feasibility of using machine learning to classify calls to South African emergency dispatch centres according to prehospital diagnosis, by utilising caller descriptions of the incident. Healthcare (Basel). 2021;9(9):1107. [FREE Full text] [CrossRef] [Medline]
  27. Ceklic E, Ball S, Finn J, Brown E, Brink D, Bailey P, et al. Ambulance dispatch prioritisation for traffic crashes using machine learning: a natural language approach. Int J Med Inform. 2022;168:104886. [CrossRef] [Medline]
  28. Chin KC, Cheng YC, Sun JT, Ou CY, Hu CH, Tsai MC, et al. Machine learning-based text analysis to predict severely injured patients in emergency medical dispatch: model development and validation. J Med Internet Res. 2022;24(6):e30210. [FREE Full text] [CrossRef] [Medline]
  29. Cotte F, Mueller T, Gilbert S, Blümke B, Multmeier J, Hirsch MC, et al. Safety of triage self-assessment using a symptom assessment app for walk-in patients in the emergency care setting: observational prospective cross-sectional study. JMIR Mhealth Uhealth. 2022;10(3):e32340. [FREE Full text] [CrossRef] [Medline]
  30. Ferri P, Sáez C, Félix-De Castro A, Juan-Albarracín J, Blanes-Selva V, Sánchez-Cuesta P, et al. Deep ensemble multitask classification of emergency medical call incidents combining multimodal data improves emergency medical dispatch. Artif Intell Med. 2021;117:102088. [FREE Full text] [CrossRef] [Medline]
  31. Gatto J, Seegmiller P, Johnston G, Preum SM. Identifying the perceived severity of patient-generated telemedical queries regarding COVID: developing and evaluating a transfer learning-based solution. JMIR Med Inform. 2022;10(9):e37770. [FREE Full text] [CrossRef] [Medline]
  32. Inokuchi R, Iwagami M, Sun Y, Sakamoto A, Tamiya N. Machine learning models predicting undertriage in telephone triage. Ann Med. 2022;54(1):2990-2997. [FREE Full text] [CrossRef] [Medline]
  33. Lai L, Wittbold KA, Dadabhoy FZ, Sato R, Landman AB, Schwamm LH, et al. Digital triage: novel strategies for population health management in response to the COVID-19 pandemic. Healthc (Amst). 2020;8(4):100493. [FREE Full text] [CrossRef] [Medline]
  34. Morse KE, Ostberg NP, Jones VG, Chan AS. Use characteristics and triage acuity of a digital symptom checker in a large integrated health system: population-based descriptive study. J Med Internet Res. 2020;22(11):e20549. [FREE Full text] [CrossRef] [Medline]
  35. Pacula M, Meltzer T, Crystal M, Srivastava A, Marx B. Automatic detection of psychological distress indicators and severity assessment in crisis hotline conversations. IEEE; 2014. Presented at: 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP); 2014 July 14:4863-4867; Florence, Italy. [CrossRef]
  36. Veladas R, Yang H, Quaresma P, Gonçalves T, Vieira R, Sousa PC, et al. Aiding clinical triage with text classification. In: Marreiros G, Melo FS, Lau N, Lopes Cardoso H, Reis LP, editors. Progress in Artificial Intelligence. Cham. Springer International Publishing; 2021:83-96.
  37. de Hond AAH, Leeuwenberg AM, Hooft L, Kant IMJ, Nijman SWJ, van Os HJA, et al. Guidelines and quality criteria for artificial intelligence-based prediction models in healthcare: a scoping review. NPJ Digit Med. 2022;5(1):2. [FREE Full text] [CrossRef] [Medline]
  38. Ioannidis JPA. Why most published research findings are false. PLoS Med. 2005;2(8):e124. [FREE Full text] [CrossRef] [Medline]
  39. Chen RJ, Wang JJ, Williamson DFK, Chen TY, Lipkova J, Lu MY, et al. Algorithmic fairness in artificial intelligence for medicine and healthcare. Nat Biomed Eng. 2023;7(6):719-742. [FREE Full text] [CrossRef] [Medline]
  40. Cardoso SD, Da Silveira M, Pruski C. Construction and exploitation of an historical knowledge graph to deal with the evolution of ontologies. Knowledge Based Syst. 2020;194:105508. [CrossRef]
  41. de la Torre J, Puig D, Valls A. Weighted kappa loss function for multi-class classification of ordinal data in deep learning. Pattern Recognit Lett. 2018;105:144-154. [CrossRef]
  42. Huang Y, Talwar A, Chatterjee S, Aparasu RR. Application of machine learning in predicting hospital readmissions: a scoping review of the literature. BMC Med Res Methodol. 2021;21(1):96. [FREE Full text] [CrossRef] [Medline]
  43. Duncan R, Stewart K, Scheuermeyer F, Abu-Laban R, Ho K, Lavallee D, et al. Concordance between Virtual Physician advice and subsequent health service utilization for callers to a nurse-managed provincial health information telephone service: a time-to-event analysis. BMC Health Serv Res. 2023. [CrossRef]
  44. Blank L, Coster J, O'Cathain A, Knowles E, Tosh J, Turner J, et al. The appropriateness of, and compliance with, telephone triage decisions: a systematic review and narrative synthesis. J Adv Nurs. 2012;68(12):2610-2621. [CrossRef] [Medline]
  45. Tran DT, Gibson A, Randall D, Havard A, Byrne M, Robinson M, et al. Compliance with telephone triage advice among adults aged 45 years and older: an Australian data linkage study. BMC Health Serv Res. 2017;17(1):512. [FREE Full text] [CrossRef] [Medline]
  46. Khader F, Kather JN, Müller-Franzes G, Wang T, Han T, Tayebi Arasteh S, et al. Medical transformer for multimodal survival prediction in intensive care: integration of imaging and non-imaging data. Sci Rep. Jul 01, 2023;13(1):10666. [CrossRef] [Medline]
  47. Chen Z, Siltala-Li L, Lassila M, Malo P, Vilkkumaa E, Saaresranta T, et al. Predicting visit cost of obstructive sleep apnea using electronic healthcare records with transformer. IEEE J Transl Eng Health Med. 2023;11:306-317. [FREE Full text] [CrossRef] [Medline]


AUC: area under the curve
ED: emergency department
K-NN: K-nearest neighbors
ML: machine learning
NB: naïve Bayes
NLP: natural language processing
NN: neural network
RF: random forest
SVM: support vector machine
XGBoost: extreme gradient boosting


Edited by T de Azevedo Cardoso; submitted 24.01.24; peer-reviewed by T Aslanidis, L Yperzeele, O Salman; comments to author 23.04.24; revised version received 13.05.24; accepted 18.07.24; published 11.09.24.

Copyright

©Daniel Raff, Kurtis Stewart, Michelle Christie Yang, Jessie Shang, Sonya Cressman, Roger Tam, Jessica Wong, Martin C Tammemägi, Kendall Ho. Originally published in the Interactive Journal of Medical Research (https://www.i-jmr.org/), 11.09.2024.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Interactive Journal of Medical Research, is properly cited. The complete bibliographic information, a link to the original publication on https://www.i-jmr.org/, as well as this copyright and license information must be included.