<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "http://dtd.nlm.nih.gov/publishing/2.0/journalpublishing.dtd">
<article article-type="research-article" dtd-version="2.0" xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-id journal-id-type="publisher-id">JMIR</journal-id>
      <journal-id journal-id-type="nlm-ta">JMIR AI</journal-id>
      <journal-title>JMIR AI</journal-title>
      <issn pub-type="epub">2817-1705</issn>
      <publisher>
        <publisher-name>JMIR Publications</publisher-name>
        <publisher-loc>Toronto, Canada</publisher-loc>
      </publisher>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="publisher-id">v3i1e52190</article-id>
      <article-id pub-id-type="pmid"/>
      <article-id pub-id-type="doi">10.2196/52190</article-id>
      <article-categories>
        <subj-group subj-group-type="heading">
          <subject>Original Paper</subject>
        </subj-group>
        <subj-group subj-group-type="article-type">
          <subject>Original Paper</subject>
        </subj-group>
      </article-categories>
      <title-group>
        <article-title>Traditional Machine Learning, Deep Learning, and BERT (Large Language Model) Approaches for Predicting Hospitalizations From Nurse Triage Notes: Comparative Evaluation of Resource Management</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="editor">
          <name>
            <surname>El Emam</surname>
            <given-names>Khaled</given-names>
          </name>
        </contrib>
        <contrib contrib-type="editor">
          <name>
            <surname>Malin</surname>
            <given-names>Bradley</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Elbattah</surname>
            <given-names>Mahmoud</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Yao</surname>
            <given-names>Li-Hung</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib id="contrib1" contrib-type="author" corresp="yes" equal-contrib="yes">
          <name name-style="western">
            <surname>Patel</surname>
            <given-names>Dhavalkumar</given-names>
          </name>
          <degrees>MSc</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <address>
            <institution>Institute for Healthcare Delivery Science</institution>
            <institution>Icahn School of Medicine at Mount Sinai</institution>
            <addr-line>2nd Floor</addr-line>
            <addr-line>150 East 42nd Street</addr-line>
            <addr-line>New York, NY, 10017</addr-line>
            <country>United States</country>
            <phone>1 (212) 523 5555</phone>
            <email>pateldhaval021@hotmail.com</email>
          </address>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-6066-9844</ext-link>
        </contrib>
        <contrib id="contrib2" contrib-type="author">
          <name name-style="western">
            <surname>Timsina</surname>
            <given-names>Prem</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-6047-887X</ext-link>
        </contrib>
        <contrib id="contrib3" contrib-type="author">
          <name name-style="western">
            <surname>Gorenstein</surname>
            <given-names>Larisa</given-names>
          </name>
          <degrees>MD</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-7778-9332</ext-link>
        </contrib>
        <contrib id="contrib4" contrib-type="author">
          <name name-style="western">
            <surname>Glicksberg</surname>
            <given-names>Benjamin S</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff3" ref-type="aff">3</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-4515-8090</ext-link>
        </contrib>
        <contrib id="contrib5" contrib-type="author">
          <name name-style="western">
            <surname>Raut</surname>
            <given-names>Ganesh</given-names>
          </name>
          <degrees>MSc</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0009-0006-7914-7065</ext-link>
        </contrib>
        <contrib id="contrib6" contrib-type="author">
          <name name-style="western">
            <surname>Cheetirala</surname>
            <given-names>Satya Narayan</given-names>
          </name>
          <degrees>MSc</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0009-0008-5666-5593</ext-link>
        </contrib>
        <contrib id="contrib7" contrib-type="author">
          <name name-style="western">
            <surname>Santana</surname>
            <given-names>Fabio</given-names>
          </name>
          <degrees>BSc</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0009-0002-2973-2136</ext-link>
        </contrib>
        <contrib id="contrib8" contrib-type="author">
          <name name-style="western">
            <surname>Tamegue</surname>
            <given-names>Jules</given-names>
          </name>
          <degrees>BSc</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-3952-1711</ext-link>
        </contrib>
        <contrib id="contrib9" contrib-type="author">
          <name name-style="western">
            <surname>Kia</surname>
            <given-names>Arash</given-names>
          </name>
          <degrees>MD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-4699-7892</ext-link>
        </contrib>
        <contrib id="contrib10" contrib-type="author">
          <name name-style="western">
            <surname>Zimlichman</surname>
            <given-names>Eyal</given-names>
          </name>
          <degrees>MD</degrees>
          <xref rid="aff4" ref-type="aff">4</xref>
          <xref rid="aff5" ref-type="aff">5</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-6276-9937</ext-link>
        </contrib>
        <contrib id="contrib11" contrib-type="author">
          <name name-style="western">
            <surname>Levin</surname>
            <given-names>Matthew A</given-names>
          </name>
          <degrees>MD</degrees>
          <xref rid="aff6" ref-type="aff">6</xref>
          <xref rid="aff7" ref-type="aff">7</xref>
          <xref rid="aff8" ref-type="aff">8</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-6013-2684</ext-link>
        </contrib>
        <contrib id="contrib12" contrib-type="author">
          <name name-style="western">
            <surname>Freeman</surname>
            <given-names>Robert</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-4946-6533</ext-link>
        </contrib>
        <contrib id="contrib13" contrib-type="author" equal-contrib="yes">
          <name name-style="western">
            <surname>Klang</surname>
            <given-names>Eyal</given-names>
          </name>
          <degrees>MD</degrees>
          <xref rid="aff3" ref-type="aff">3</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-4567-3108</ext-link>
        </contrib>
      </contrib-group>
      <aff id="aff1">
        <label>1</label>
        <institution>Institute for Healthcare Delivery Science</institution>
        <institution>Icahn School of Medicine at Mount Sinai</institution>
        <addr-line>New York, NY</addr-line>
        <country>United States</country>
      </aff>
      <aff id="aff2">
        <label>2</label>
        <institution>Division of Diagnostic Imaging</institution>
        <institution>Sheba Medical Center</institution>
        <institution>Tel-Aviv University</institution>
        <addr-line>Tel Aviv</addr-line>
        <country>Israel</country>
      </aff>
      <aff id="aff3">
        <label>3</label>
        <institution>Division of Data-Driven and Digital Medicine (D3M)</institution>
        <institution>Icahn School of Medicine at Mount Sinai</institution>
        <addr-line>New York, NY</addr-line>
        <country>United States</country>
      </aff>
      <aff id="aff4">
        <label>4</label>
        <institution>Hospital Management</institution>
        <institution>Sheba Medical Center</institution>
        <institution>Tel-Aviv University</institution>
        <addr-line>Tel Aviv</addr-line>
        <country>Israel</country>
      </aff>
      <aff id="aff5">
        <label>5</label>
        <institution>ARC Innovation Center</institution>
        <institution>Sheba Medical Center</institution>
        <institution>Tel-Aviv University</institution>
        <addr-line>Tel Aviv</addr-line>
        <country>Israel</country>
      </aff>
      <aff id="aff6">
        <label>6</label>
        <institution>Department of Anesthesiology, Perioperative and Pain Medicine</institution>
        <institution>Icahn School of Medicine at Mount Sinai</institution>
        <addr-line>New York, NY</addr-line>
        <country>United States</country>
      </aff>
      <aff id="aff7">
        <label>7</label>
        <institution>Department of Genetics and Genomic Sciences</institution>
        <institution>Icahn School of Medicine at Mount Sinai</institution>
        <addr-line>New York, NY</addr-line>
        <country>United States</country>
      </aff>
      <aff id="aff8">
        <label>8</label>
        <institution>Windreich Department of Artificial Intelligence and Human Health</institution>
        <institution>Icahn School of Medicine at Mount Sinai</institution>
        <addr-line>New York, NY</addr-line>
        <country>United States</country>
      </aff>
      <author-notes>
        <corresp>Corresponding Author: Dhavalkumar Patel <email>pateldhaval021@hotmail.com</email></corresp>
      </author-notes>
      <pub-date pub-type="collection">
        <year>2024</year>
      </pub-date>
      <pub-date pub-type="epub">
        <day>27</day>
        <month>8</month>
        <year>2024</year>
      </pub-date>
      <volume>3</volume>
      <elocation-id>e52190</elocation-id>
      <history>
        <date date-type="received">
          <day>25</day>
          <month>8</month>
          <year>2023</year>
        </date>
        <date date-type="rev-request">
          <day>8</day>
          <month>2</month>
          <year>2024</year>
        </date>
        <date date-type="rev-recd">
          <day>15</day>
          <month>4</month>
          <year>2024</year>
        </date>
        <date date-type="accepted">
          <day>14</day>
          <month>6</month>
          <year>2024</year>
        </date>
      </history>
      <copyright-statement>©Dhavalkumar Patel, Prem Timsina, Larisa Gorenstein, Benjamin S Glicksberg, Ganesh Raut, Satya Narayan Cheetirala, Fabio Santana, Jules Tamegue, Arash Kia, Eyal Zimlichman, Matthew A Levin, Robert Freeman, Eyal Klang. Originally published in JMIR AI (https://ai.jmir.org), 27.08.2024.</copyright-statement>
      <copyright-year>2024</copyright-year>
      <license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/">
        <p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR AI, is properly cited. The complete bibliographic information, a link to the original publication on https://www.ai.jmir.org/, as well as this copyright and license information must be included.</p>
      </license>
      <self-uri xlink:href="https://ai.jmir.org/2024/1/e52190" xlink:type="simple"/>
      <related-article related-article-type="commentary" id="v4i1e67321" ext-link-type="doi" xlink:href="10.2196/67321" vol="4" page="e67321" xlink:type="simple">https://ai.jmir.org/2025/1/e67321/</related-article>
      <related-article related-article-type="commentary" id="v4i1e73342" ext-link-type="doi" xlink:href="10.2196/73342" vol="4" page="e[MS number]" xlink:type="simple">https://ai.jmir.org/2025/1/e73342/</related-article>
      <abstract>
        <sec sec-type="background">
          <title>Background</title>
          <p>Predicting hospitalization from nurse triage notes has the potential to augment care. However, there needs to be careful considerations for which models to choose for this goal. Specifically, health systems will have varying degrees of computational infrastructure available and budget constraints.</p>
        </sec>
        <sec sec-type="objective">
          <title>Objective</title>
          <p>To this end, we compared the performance of the deep learning, Bidirectional Encoder Representations from Transformers (BERT)–based model, Bio-Clinical-BERT, with a bag-of-words (BOW) logistic regression (LR) model incorporating term frequency–inverse document frequency (TF-IDF). These choices represent different levels of computational requirements.</p>
        </sec>
        <sec sec-type="methods">
          <title>Methods</title>
          <p>A retrospective analysis was conducted using data from 1,391,988 patients who visited emergency departments in the Mount Sinai Health System spanning from 2017 to 2022. The models were trained on 4 hospitals’ data and externally validated on a fifth hospital’s data.</p>
        </sec>
        <sec sec-type="results">
          <title>Results</title>
          <p>The Bio-Clinical-BERT model achieved higher areas under the receiver operating characteristic curve (0.82, 0.84, and 0.85) compared to the BOW-LR-TF-IDF model (0.81, 0.83, and 0.84) across training sets of 10,000; 100,000; and ~1,000,000 patients, respectively. Notably, both models proved effective at using triage notes for prediction, despite the modest performance gap.</p>
        </sec>
        <sec sec-type="conclusions">
          <title>Conclusions</title>
          <p>Our findings suggest that simpler machine learning models such as BOW-LR-TF-IDF could serve adequately in resource-limited settings. Given the potential implications for patient care and hospital resource management, further exploration of alternative models and techniques is warranted to enhance predictive performance in this critical domain.</p>
        </sec>
        <sec sec-type="registered-report">
          <title>International Registered Report Identifier (IRRID)</title>
          <p>RR2-10.1101/2023.08.07.23293699</p>
        </sec>
      </abstract>
      <kwd-group>
        <kwd>Bio-Clinical-BERT</kwd>
        <kwd>term frequency–inverse document frequency</kwd>
        <kwd>TF-IDF</kwd>
        <kwd>health informatics</kwd>
        <kwd>patient care</kwd>
        <kwd>hospital resource management</kwd>
        <kwd>care</kwd>
        <kwd>resource management</kwd>
        <kwd>management</kwd>
        <kwd>language model</kwd>
        <kwd>machine learning</kwd>
        <kwd>hospitalization</kwd>
        <kwd>deep learning</kwd>
        <kwd>logistic regression</kwd>
        <kwd>retrospective analysis</kwd>
        <kwd>training</kwd>
        <kwd>large language model</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec sec-type="introduction">
      <title>Introduction</title>
      <p>Efficient and effective patient triage within the emergency department (ED) plays a pivotal role in enhancing treatment outcomes and optimizing care delivery [<xref ref-type="bibr" rid="ref1">1</xref>-<xref ref-type="bibr" rid="ref3">3</xref>]. This process involves rapidly identifying patients who require immediate hospitalization upon their arrival. One of the resources for making these predictions are nurse triage notes, which provide a wealth of in-depth information about the patient’s condition at presentation [<xref ref-type="bibr" rid="ref4">4</xref>,<xref ref-type="bibr" rid="ref5">5</xref>].</p>
      <p>In the field of health care, machine learning has opened up new avenues for potential improvement in such complex classification tasks, thereby augmenting clinical decision-making processes [<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref7">7</xref>]. The recent developments in deep learning and natural language processing (NLP) techniques have further broadened this potential, bringing a new realm of possibilities for enhancing medical decision-making capabilities.</p>
      <p>Among these advanced algorithms is the Bidirectional Encoder Representations from Transformers (BERT) model [<xref ref-type="bibr" rid="ref8">8</xref>]. BERT has shown excellent performance in numerous NLP tasks [<xref ref-type="bibr" rid="ref9">9</xref>] and has inspired the development of more specialized versions tailored to particular fields, such as the Bio-Clinical-BERT model, which was designed to cater to the biomedical field [<xref ref-type="bibr" rid="ref10">10</xref>].</p>
      <p>The focus of this study is to delve into the potential of a fine-tuned Bio-Clinical-BERT model and compare it against a simpler, robust, and more traditional approach, mainly, the bag-of-words (BOW) logistic regression (LR) model complemented by the term frequency–inverse document frequency (TF-IDF) method. We also evaluated other approaches including the extreme gradient boosting (XGBoost) classifier and Word-2-Vec (W2V) embedding with bidirectional long short-term memory (Bi-LSTM) network. The primary objective of our research is to gauge the efficacy of these 2 methods in predicting hospital admissions using nurse triage notes.</p>
      <p>While it is true that Bio-Clinical-BERT could potentially offer improved accuracy in its predictions, it should be noted that it also requires a substantial investment in terms of computational resources. It necessitates the use of specialized hardware and demands a certain level of software expertise to operate effectively. On the other hand, the LR model paired with the TF-IDF method is more resource efficient and enjoys wide acceptance in the field of text classification due to its simplicity and effectiveness.</p>
      <p>We hypothesized that the Bio-Clinical-BERT model may surpass the performance of the BOW-LR model combined with the TF-IDF approach in the task of predicting triage outcomes. However, we also speculated that the incremental gains in performance might not necessarily justify the additional demands imposed by the large deep learning model in terms of computational resources and technical know-how. To test this hypothesis, we have undertaken an extensive study using over 1 million nurse triage notes collected from a large health system.</p>
      <p>The fundamental contribution of this paper is a comparison between these techniques for predicting hospital admission, which reflect different levels of computational requirements and cost implications. Our comparison not only looks at the accuracy of these models but also weighs the trade-offs between predictive accuracy and computational efficiency, a consideration that is often overlooked but is of prime importance in real-world settings when implementing models. Specifically, health systems may be able to use insights from this study to make informed decisions on which methodology may be right for their circumstances, with a clearer understanding of the limitations of each. Our aim is to equip health care practitioners, researchers, and decision makers with insights that could potentially aid in enhancing hospital resource management and improve the quality of patient care.</p>
    </sec>
    <sec sec-type="methods">
      <title>Methods</title>
      <sec>
        <title>Data Sources and Study Design</title>
        <p>For the construction and testing of our models, we used an extensive dataset from the Mount Sinai Health System (MSHS). This is a diverse health care provider based in New York City. In this study, the dataset included ED records spanning a 5-year period from 2017 to 2022. This dataset was gathered from 5 different MSHS hospitals, covering a broad range of population groups and diverse urban health settings.</p>
        <p>These 5 participating hospitals provided a rich source of data for our study, representing different communities in New York City. The hospitals include Mount Sinai Hospital, a health care institution located in East Harlem, Manhattan; Mount Sinai Morningside, situated in Morningside Heights, Manhattan; Mount Sinai West, operating in Midtown West, Manhattan; Mount Sinai Brooklyn, a community-focused health facility located in Midwood, Brooklyn; and Mount Sinai Queens (MSQ), based in Astoria, Queens. The dataset used for our study was compiled using the Epic Electronic Health Records software, a tool that aids in efficient data collection, management, and analysis. The dataset was made available by the diligent work of the Mount Sinai Hospital Clinical Data Science team.</p>
      </sec>
      <sec>
        <title>Model Development and Evaluation</title>
        <p>In the development and testing of our models, we leveraged data from 4 hospitals for training, validation, and hyperparameter tuning processes. We elected to use a distinct dataset from MSQ for external testing to ensure our model’s generalizability.</p>
        <p>The internal training and validation cohort underwent a procedure involving 5-fold cross-validation. Each fold contained 10,000 records, which were used for hyperparameter tuning. For the external dataset, we experimented with training sets of varying sizes: 10,000; 100,000; and roughly 1,000,000 patients, which represent the complete 4-hospital cohort. Subsequently, testing was carried out on 20% of these cohorts’ sizes, taken from the MSQ hospital cohort.</p>
        <p>Our study involved several models: Bio-Clinical-BERT and BOW-LR models using TF-IDF features. For further subanalyses using different machine and deep learning models, we also evaluated XGBoost with BOW and Bi-LSTM with a W2V pretrained embedding layer derived from bioclinical data (BioWordVec_PubMed_MIMICIII_d200).</p>
        <p>As a final subanalysis experiment, for the BERT model, we also experimented with up-sampling of the minority class to ensure balanced data representation, enhancing the stability and accuracy of our model predictions.</p>
        <p>These models were used to predict hospitalization outcomes from nurse triage notes. For Bio-Clinical-BERT, we adhered to text preprocessing and tokenization guidelines as outlined on the Hugging Face website [<xref ref-type="bibr" rid="ref11">11</xref>].</p>
        <p>For BOW-XGBoost, we evaluated 3 different numbers of estimators. Other XGBoost hyperparameters were set to default values, including a learning rate of 0.3, maximum depth of 6, and minimum child weight of 1.</p>
        <p>For W2V-Bi-LSTM, the network is comprised of a Bi-LSTM layer (256 hidden units), preceded by a pretrained embedding 200-dimensions W2V layer, with a fully connected layer followed by a sigmoid activation function.</p>
        <p>Further details on hyperparameter selection are elucidated in the <italic>Hyperparameter Tuning Results</italic> section. For BOW-LR-TF-IDF, we followed a similar methodology outlined in our previous publication [<xref ref-type="bibr" rid="ref12">12</xref>], covering both text preprocessing and hyperparameter selection processes.</p>
        <p>BERT is a model designed for NLP tasks. It learns from the context of both preceding and following words, making it “bidirectional.” This model is pretrained on large corpora and can be fine-tuned for specific tasks.</p>
        <p>The BOW model is a simple technique in NLP. It represents text data by counting the frequency of each word, disregarding the order in which they appear. Each unique word forms a feature, and the frequency of the word represents the value of that feature. However, this method can overlook context and semantics due to its simplicity.</p>
        <p>TF-IDF is a numerical statistic that reflects how important a word is to a document in a collection. It is a combination of 2 metrics: <italic>term frequency</italic>, which is the number of times a word appears in a document, and <italic>inverse document frequency</italic>, which diminishes the weight of common words and amplifies the weight of rare words across the entire dataset. This helps in reducing the impact of frequently used words and highlights more meaningful terms.</p>
        <p>XGBoost is an advanced gradient boosting framework known for its efficiency and performance in structured data classification and regression. It builds multiple decision trees sequentially to correct previous errors, excelling in handling diverse data types and preventing overfitting.</p>
        <p>Bi-LSTM is an artificial neural network that processes data in both directions to capture past and future context. This enhances its sequence understanding, making it suitable for text classification, sentiment analysis, and machine translation.</p>
      </sec>
      <sec>
        <title>Study Population</title>
        <p>The demographic for this study included adult patients aged 18 years and older. These were patients who made ED visits within the specified 5-year period from 2017 to 2022 across the 5 participating MSHS hospitals.</p>
      </sec>
      <sec>
        <title>Outcome Definition</title>
        <p>The primary outcome for our study was to ascertain our models’ effectiveness in predicting hospitalization. This prediction was based on 2 main types of data: tabular electronic health records and nurse triage notes.</p>
      </sec>
      <sec>
        <title>Model Evaluation and Comparison</title>
        <p>To assess the performance of our models, we used various metrics such as area under the receiver operating characteristic curve (AUC), sensitivity, specificity, and precision. These metrics allowed us to thoroughly evaluate the Bio-Clinical-BERT [<xref ref-type="bibr" rid="ref10">10</xref>] and BOW-LR models with TF-IDF features, as well as compare their capabilities in predicting hospitalization from nurse triage notes.</p>
      </sec>
      <sec>
        <title>Ethical Considerations</title>
        <p>This study, being retrospective in nature, was reviewed and approved by an ethical institutional review board committee from the MSHS (protocol: STUDY-18-00573). The institutional review board committee deemed that due to the retrospective nature of the study, the requirement for informed consent was waived.</p>
      </sec>
      <sec>
        <title>Statistical Analysis</title>
        <p>Our statistical analyses were conducted using Python (version 3.9.12; Python Software Foundation). We presented continuous variables as median (IQR) and categorical variables as percentages for better interpretability. To identify words linked to hospital admission within nurse triage notes, we calculated the odds ratio (OR) and mutual information (MI) [<xref ref-type="bibr" rid="ref12">12</xref>]. Statistical tests such as the chi-square test and 2-tailed Student <italic>t</italic> test were used for comparing categorical and continuous variables, respectively. A <italic>P</italic> value &lt;.05 was considered statistically significant. For evaluating our models, receiver operating characteristic (ROC) curves were plotted, and metrics including AUC, sensitivity (recall), specificity, and positive predictive value (precision) were derived.</p>
      </sec>
      <sec>
        <title>Technical Architecture</title>
        <p>The technical experiments involved in this study were conducted within a controlled hospital infrastructure that used an On-Premises Centos Linux environment in conjunction with Azure Cloud infrastructure. For the BOW-TF-IDF experiments, we elected to use the Centos Linux OS. In contrast, the BERT experiment was conducted using a Standard_NC6 GPU instance on Azure Cloud. This instance came with one 16-GB GPU and 6 vCPUs and incurred a cost of approximately US $80 during the training phase. <xref rid="figure1" ref-type="fig">Figure 1</xref> offers a detailed depiction of the fundamental technical architecture used for training the BERT and LR-TF-IDF models, across multiple patient datasets.</p>
        <fig id="figure1" position="float">
          <label>Figure 1</label>
          <caption>
            <p>Process flow of multiple patient datasets passing through 2 different models with GPU and non-GPU instances. BERT: Bidirectional Encoder Representations from Transformers; LR: logistic regression; TF-IDF: term frequency–inverse document frequency.</p>
          </caption>
          <graphic xlink:href="ai_v3i1e52190_fig1.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
    </sec>
    <sec sec-type="results">
      <title>Results</title>
      <sec>
        <title>Patient Population and Data</title>
        <p>Our study incorporated data from 1,745,199 patients drawn from the MSHS. Upon the exclusion of patients aged &lt;18 years, we had 1,391,988 participants in the study. These patients visited the ED between 2017 and 2022. <xref ref-type="table" rid="table1">Table 1</xref> presents a summary of the patient characteristics.</p>
        <p>The median number of words per triage note was 19.0 (IQR 12.0-31.0). Top 10 words associated with the highest MI score regarding hospital admission are outlined in <xref ref-type="table" rid="table2">Table 2</xref>.</p>
        <table-wrap position="float" id="table1">
          <label>Table 1</label>
          <caption>
            <p>Demographic distribution in the study.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="30"/>
            <col width="220"/>
            <col width="230"/>
            <col width="220"/>
            <col width="200"/>
            <col width="0"/>
            <col width="100"/>
            <thead>
              <tr valign="top">
                <td colspan="2">Demographics</td>
                <td>All patients (includes MSH<sup>a</sup>, MSM<sup>b</sup>, MSW<sup>c</sup>, MSB<sup>d</sup>, and MSQ<sup>e</sup>; N=1,391,988)</td>
                <td>4 hospitals (MSH, MSM, MSW, and MSB; n=1,110,272)</td>
                <td>MSQ (n=281,716)</td>
                <td colspan="2"><italic>P</italic> value</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td colspan="2">
                  Age (years), median (IQR)
                </td>
                <td>47.0 (31.0-75)</td>
                <td>48.0 (32.0-75.0)</td>
                <td>45.0 (30.0-75.0)</td>
                <td colspan="2">&lt;.001</td>
              </tr>
              <tr valign="top">
                <td colspan="6">
                  <bold>Sex, n (%)</bold>
                </td>
                <td>&lt;.001</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Female</td>
                <td>727,363 (52.3)</td>
                <td>586,224 (52.8)</td>
                <td>141,140 (50.1)</td>
                <td colspan="2">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Male</td>
                <td>664,625 (47.7)</td>
                <td>524,048 (47.2)</td>
                <td>140,576 (49.9)</td>
                <td colspan="2">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td colspan="6">
                  <bold>Race, n (%)</bold>
                </td>
                <td>&lt;.001</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Black</td>
                <td>428,594 (30.79)</td>
                <td>382,898 (34.5)</td>
                <td>45,696 (16.22)</td>
                <td colspan="2">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>White</td>
                <td>343,079 (24,65)</td>
                <td>265,457 (23.92)</td>
                <td>77,622 (27.56)</td>
                <td colspan="2">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Other</td>
                <td>620,315 (44,56)</td>
                <td>461,917 (41.58)</td>
                <td>158,398 (56.22)</td>
                <td colspan="2">
                  <break/>
                </td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table1fn1">
              <p><sup>a</sup>MSH: Mount Sinai Hospital.</p>
            </fn>
            <fn id="table1fn2">
              <p><sup>b</sup>MSM: Mount Sinai Morningside.</p>
            </fn>
            <fn id="table1fn3">
              <p><sup>c</sup>MSW: Mount Sinai West.</p>
            </fn>
            <fn id="table1fn4">
              <p><sup>d</sup>MSB: Mount Sinai Brooklyn.</p>
            </fn>
            <fn id="table1fn5">
              <p><sup>e</sup>MSQ: Mount Sinai Queens.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <table-wrap position="float" id="table2">
          <label>Table 2</label>
          <caption>
            <p>Odds ratios (OR) and mutual information (MI) values for words linked to admission to hospital wards, sorted by highest MI values.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="270"/>
            <col width="280"/>
            <col width="310"/>
            <col width="140"/>
            <thead>
              <tr valign="top">
                <td>Word</td>
                <td>OR for admission</td>
                <td>MI for admission</td>
                <td><italic>P</italic> value</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Sent</td>
                <td>3.6</td>
                <td>16.4</td>
                <td>&lt;.001</td>
              </tr>
              <tr valign="top">
                <td>Pt<sup>a</sup></td>
                <td>1.6</td>
                <td>15.8</td>
                <td>&lt;.001</td>
              </tr>
              <tr valign="top">
                <td>Per</td>
                <td>2.3</td>
                <td>15</td>
                <td>&lt;.001</td>
              </tr>
              <tr valign="top">
                <td>Of</td>
                <td>1.3</td>
                <td>12.7</td>
                <td>&lt;.001</td>
              </tr>
              <tr valign="top">
                <td>Home</td>
                <td>2.2</td>
                <td>11.5</td>
                <td>&lt;.001</td>
              </tr>
              <tr valign="top">
                <td>EMS<sup>b</sup></td>
                <td>2.2</td>
                <td>10.8</td>
                <td>&lt;.001</td>
              </tr>
              <tr valign="top">
                <td>Weakness</td>
                <td>3.6</td>
                <td>10.8</td>
                <td>&lt;.001</td>
              </tr>
              <tr valign="top">
                <td>Chest</td>
                <td>1.4</td>
                <td>8.9</td>
                <td>&lt;.001</td>
              </tr>
              <tr valign="top">
                <td>SOB<sup>c</sup></td>
                <td>2.1</td>
                <td>8.8</td>
                <td>&lt;.001</td>
              </tr>
              <tr valign="top">
                <td>BIBA<sup>d</sup></td>
                <td>2.1</td>
                <td>7.9</td>
                <td>&lt;.001</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table2fn1">
              <p><sup>a</sup>Pt: patient.</p>
            </fn>
            <fn id="table2fn2">
              <p><sup>b</sup>EMS: emergency medical services.</p>
            </fn>
            <fn id="table2fn3">
              <p><sup>c</sup>SOB: shortness of breath.</p>
            </fn>
            <fn id="table2fn4">
              <p><sup>d</sup>BIBA: brought in by ambulance.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
      <sec>
        <title>Hyperparameter Tuning Results</title>
        <p>A hyperparameter tuning process was performed. The best hyperparameters were identified for each model based on their performance during the 5-fold cross-validation on the training validation set. The results of the BERT hyperparameter tuning process can be found in <xref ref-type="table" rid="table3">Table 3</xref>.</p>
        <p>The results of the W2V-LSTM model hyperparameter tuning are presented in <xref ref-type="table" rid="table4">Table 4</xref>.</p>
        <p>The results of XGBoost hyperparameter tuning are presented in <xref ref-type="table" rid="table5">Table 5</xref>.</p>
        <table-wrap position="float" id="table3">
          <label>Table 3</label>
          <caption>
            <p>BERT<sup>a</sup> hyperparameter tuning in the internal training and validation cohorts using 5-fold experiments.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="260"/>
            <col width="180"/>
            <col width="180"/>
            <col width="180"/>
            <col width="200"/>
            <thead>
              <tr valign="top">
                <td>Batch size</td>
                <td>Max length</td>
                <td>Learning rate</td>
                <td>Epoch</td>
                <td>Value, mean (SD)</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>64</td>
                <td>—<sup>b</sup></td>
                <td>2×10<sup>–5</sup></td>
                <td>—</td>
                <td>0.78 (0.01)</td>
              </tr>
              <tr valign="top">
                <td>128</td>
                <td>—</td>
                <td>2×10<sup>–5</sup></td>
                <td>—</td>
                <td>0.80 (0.01)</td>
              </tr>
              <tr valign="top">
                <td>128</td>
                <td>128</td>
                <td>2×10<sup>–5</sup></td>
                <td>3</td>
                <td>0.80 (0.01)</td>
              </tr>
              <tr valign="top">
                <td>256</td>
                <td>64</td>
                <td>2×10<sup>–5</sup></td>
                <td>—</td>
                <td>0.79 (0.01)</td>
              </tr>
              <tr valign="top">
                <td>64</td>
                <td>—</td>
                <td>3×10<sup>–5</sup></td>
                <td>—</td>
                <td>0.79 (0.01)</td>
              </tr>
              <tr valign="top">
                <td>128</td>
                <td>—</td>
                <td>3×10<sup>–5</sup></td>
                <td>—</td>
                <td>0.79 (0.01)</td>
              </tr>
              <tr valign="top">
                <td>128</td>
                <td>128</td>
                <td>3×10<sup>–5</sup></td>
                <td>3</td>
                <td>0.78 (0.01)</td>
              </tr>
              <tr valign="top">
                <td>256</td>
                <td>64</td>
                <td>3×10<sup>–5</sup></td>
                <td>—</td>
                <td>0.78 (0.01)</td>
              </tr>
              <tr valign="top">
                <td>64</td>
                <td>—</td>
                <td>5×10<sup>–5</sup></td>
                <td>—</td>
                <td>0.79 (0.01)</td>
              </tr>
              <tr valign="top">
                <td>128</td>
                <td>—</td>
                <td>5×10<sup>–5</sup></td>
                <td>—</td>
                <td>0.80 (0.01)</td>
              </tr>
              <tr valign="top">
                <td>128</td>
                <td>128</td>
                <td>5×10<sup>–5</sup></td>
                <td>3</td>
                <td>0.79 (0.01)</td>
              </tr>
              <tr valign="top">
                <td>256</td>
                <td>64</td>
                <td>5×10<sup>–5</sup></td>
                <td>—</td>
                <td>0.79 (0.01)</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table3fn1">
              <p><sup>a</sup>BERT: Bidirectional Encoder Representations from Transformers.</p>
            </fn>
            <fn id="table3fn2">
              <p><sup>b</sup>Not applicable.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <table-wrap position="float" id="table4">
          <label>Table 4</label>
          <caption>
            <p>W2V<sup>a</sup>-LSTM<sup>b</sup> hyperparameter tuning in the internal training and validation cohorts using 5-fold experiments.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="250"/>
            <col width="250"/>
            <col width="250"/>
            <col width="250"/>
            <thead>
              <tr valign="bottom">
                <td>Batch size</td>
                <td>Learning rate</td>
                <td>Epochs</td>
                <td>AUC<sup>c</sup>, mean</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>16</td>
                <td>10<sup>−3</sup></td>
                <td>5</td>
                <td>0.768</td>
              </tr>
              <tr valign="top">
                <td>16</td>
                <td>10<sup>−3</sup></td>
                <td>10</td>
                <td>0.795</td>
              </tr>
              <tr valign="top">
                <td>16</td>
                <td>10<sup>−3</sup></td>
                <td>15</td>
                <td>0.765</td>
              </tr>
              <tr valign="top">
                <td>16</td>
                <td>10<sup>−4</sup></td>
                <td>5</td>
                <td>0.750</td>
              </tr>
              <tr valign="top">
                <td>16</td>
                <td>10<sup>−4</sup></td>
                <td>10</td>
                <td>0.781</td>
              </tr>
              <tr valign="top">
                <td>16</td>
                <td>10<sup>−4</sup></td>
                <td>15</td>
                <td>0.798</td>
              </tr>
              <tr valign="top">
                <td>32</td>
                <td>10<sup>−3</sup></td>
                <td>5</td>
                <td>0.797</td>
              </tr>
              <tr valign="top">
                <td>32</td>
                <td>10<sup>−3</sup></td>
                <td>10</td>
                <td>0.797</td>
              </tr>
              <tr valign="top">
                <td>32</td>
                <td>10<sup>−3</sup></td>
                <td>15</td>
                <td>0.777</td>
              </tr>
              <tr valign="top">
                <td>32</td>
                <td>10<sup>−4</sup></td>
                <td>5</td>
                <td>0.756</td>
              </tr>
              <tr valign="top">
                <td>32</td>
                <td>10<sup>−4</sup></td>
                <td>10</td>
                <td>0.728</td>
              </tr>
              <tr valign="top">
                <td>32</td>
                <td>10<sup>−4</sup></td>
                <td>15</td>
                <td>0.748</td>
              </tr>
              <tr valign="top">
                <td>64</td>
                <td>10<sup>−3</sup></td>
                <td>5</td>
                <td>0.661</td>
              </tr>
              <tr valign="top">
                <td>64</td>
                <td>10<sup>−3</sup></td>
                <td>10</td>
                <td>0.806</td>
              </tr>
              <tr valign="top">
                <td>64</td>
                <td>10<sup>−3</sup></td>
                <td>15</td>
                <td>0.795</td>
              </tr>
              <tr valign="top">
                <td>64</td>
                <td>10<sup>−4</sup></td>
                <td>5</td>
                <td>0.693</td>
              </tr>
              <tr valign="top">
                <td>64</td>
                <td>10<sup>−4</sup></td>
                <td>10</td>
                <td>0.767</td>
              </tr>
              <tr valign="top">
                <td>64</td>
                <td>10<sup>−4</sup></td>
                <td>15</td>
                <td>0.775</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table4fn1">
              <p><sup>a</sup>W2V: Word-2-Vec.</p>
            </fn>
            <fn id="table4fn2">
              <p><sup>b</sup>LSTM: long short-term memory.</p>
            </fn>
            <fn id="table4fn3">
              <p><sup>c</sup>AUC: area under the receiver operating characteristic curve.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <table-wrap position="float" id="table5">
          <label>Table 5</label>
          <caption>
            <p>XGBoost<sup>a</sup> hyperparameter tuning in the internal training and validation cohorts using 5-fold experiments.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="500"/>
            <col width="500"/>
            <thead>
              <tr valign="top">
                <td>Trees</td>
                <td>Value</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>100</td>
                <td>0.80 (0.01)</td>
              </tr>
              <tr valign="top">
                <td>200</td>
                <td>0.81 (0.01)</td>
              </tr>
              <tr valign="top">
                <td>1000</td>
                <td>0.80 (0.01)</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table5fn1">
              <p><sup>a</sup>XGBoost: extreme gradient boosting.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
      <sec>
        <title>Model Performance</title>
        <p>After training the Bio-Clinical-BERT and LR-TF-IDF models on the 4 hospitals’ data, we evaluated their performance on the held-out test data from MSQ. The AUC values were calculated for each model. The Bio-Clinical-BERT model achieved AUCs of 0.82, 0.84, 0.85, while the LR-TF-IDF model had AUCs of 0.81, 0.83, 0.84 for training on 10,000; 100,000; and ~1,000,000 patients, respectively.</p>
        <p><xref rid="figure2" ref-type="fig">Figure 2</xref> shows the ROC and AUC comparisons between the 2 models. The Bio-Clinical-BERT model consistently outperformed the LR-TF-IDF model in terms of AUC across the different training set sizes (10,000; 100,000; and ~1,000,000 patients), albeit by a small margin.</p>
        <p>In addition to the AUC comparisons, we also calculated other performance metrics, such as sensitivity, specificity, and precision, for both models (<xref ref-type="table" rid="table6">Table 6</xref> and <xref ref-type="table" rid="table7">Table 7</xref>).</p>
        <fig id="figure2" position="float">
          <label>Figure 2</label>
          <caption>
            <p>Receiver operating characteristic curves (ROC) of the 2 models tested on increasing training sample sizes. AUC: area under the receiver operating characteristic curve; BERT: Bidirectional Encoder Representations from Transformers; LR: logistic regression; MSQ: Mount Sinai Queens; TF-IDF: term frequency–inverse document frequency.</p>
          </caption>
          <graphic xlink:href="ai_v3i1e52190_fig2.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <table-wrap position="float" id="table6">
          <label>Table 6</label>
          <caption>
            <p>Metrics for the training and testing (external) cohorts for the Bio-Clinical-BERT<sup>a</sup> model.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="160"/>
            <col width="180"/>
            <col width="170"/>
            <col width="170"/>
            <col width="160"/>
            <col width="160"/>
            <thead>
              <tr valign="bottom">
                <td>Training data size</td>
                <td>AUC<sup>b</sup> score</td>
                <td>Sensitivity</td>
                <td>Specificity</td>
                <td>Precision</td>
                <td><italic>F</italic><sub>1</sub>-score</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>10,000</td>
                <td>0.82</td>
                <td>0.76</td>
                <td>0.74</td>
                <td>0.36</td>
                <td>0.49</td>
              </tr>
              <tr valign="top">
                <td>100,000</td>
                <td>0.84</td>
                <td>0.74</td>
                <td>0.77</td>
                <td>0.39</td>
                <td>0.51</td>
              </tr>
              <tr valign="top">
                <td>1,000,000</td>
                <td>0.85</td>
                <td>0.39</td>
                <td>0.96</td>
                <td>0.67</td>
                <td>0.50</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table6fn1">
              <p><sup>a</sup>BERT: Bidirectional Encoder Representations from Transformers.</p>
            </fn>
            <fn id="table6fn2">
              <p><sup>b</sup>AUC: area under the receiver operating characteristic curve.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <table-wrap position="float" id="table7">
          <label>Table 7</label>
          <caption>
            <p>Metrics for the training and testing (external) cohorts for the LR<sup>a</sup>-TF-IDF<sup>b</sup> model.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="160"/>
            <col width="180"/>
            <col width="170"/>
            <col width="170"/>
            <col width="160"/>
            <col width="160"/>
            <thead>
              <tr valign="bottom">
                <td>Training Data Size</td>
                <td>AUC<sup>c</sup> score</td>
                <td>Sensitivity</td>
                <td>Specificity</td>
                <td>Precision</td>
                <td><italic>F</italic><sub>1</sub>-score</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>10,000</td>
                <td>0.81</td>
                <td>0.66</td>
                <td>0.80</td>
                <td>0.40</td>
                <td>0.50</td>
              </tr>
              <tr valign="top">
                <td>100,000</td>
                <td>0.83</td>
                <td>0.75</td>
                <td>0.74</td>
                <td>0.37</td>
                <td>0.50</td>
              </tr>
              <tr valign="top">
                <td>1,000,000</td>
                <td>0.84</td>
                <td>0.71</td>
                <td>0.80</td>
                <td>0.42</td>
                <td>0.53</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table7fn1">
              <p><sup>a</sup>LR: logistic regression.</p>
            </fn>
            <fn id="table7fn2">
              <p><sup>b</sup>TF-IDF: term frequency–inverse document frequency.</p>
            </fn>
            <fn id="table7fn3">
              <p><sup>c</sup>AUC: area under the receiver operating characteristic curve.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <p>The metrics for the XGBoost and W2V-Bi-LSTM models are presented in <xref ref-type="table" rid="table8">Tables 8</xref> and <xref ref-type="table" rid="table9">9</xref>. The probability cutoff values for these metrics were calculated using the Youden index. These results further demonstrated the superior performance of the Bio-Clinical-BERT model compared to the LR-TF-IDF model.</p>
        <p>Further subanalysis for the BERT cohort using up-sampling of the minority class is presented in <xref ref-type="table" rid="table10">Table 10</xref>.</p>
        <table-wrap position="float" id="table8">
          <label>Table 8</label>
          <caption>
            <p>Metrics for the training and testing (external) cohorts for the W2V<sup>a</sup>-Bi-LSTM<sup>b</sup> model.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="160"/>
            <col width="180"/>
            <col width="170"/>
            <col width="170"/>
            <col width="160"/>
            <col width="160"/>
            <thead>
              <tr valign="bottom">
                <td>Training data size</td>
                <td>AUC<sup>c</sup> score</td>
                <td>Sensitivity</td>
                <td>Specificity</td>
                <td>Precision</td>
                <td><italic>F</italic><sub>1</sub>-score</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>10,000</td>
                <td>0.78</td>
                <td>0.32</td>
                <td>0.95</td>
                <td>0.59</td>
                <td>0.41</td>
              </tr>
              <tr valign="top">
                <td>100,000</td>
                <td>0.81</td>
                <td>0.42</td>
                <td>0.92</td>
                <td>0.52</td>
                <td>0.46</td>
              </tr>
              <tr valign="top">
                <td>1,000,000</td>
                <td>0.84</td>
                <td>0.46</td>
                <td>0.94</td>
                <td>0.62</td>
                <td>0.52</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table8fn1">
              <p><sup>c</sup>W2V: Word-2-Vec.</p>
            </fn>
            <fn id="table8fn2">
              <p><sup>c</sup>Bi-LSTM: bidirectional long short-term memory.</p>
            </fn>
            <fn id="table8fn3">
              <p><sup>c</sup>AUC: area under the receiver operating characteristic curve.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <table-wrap position="float" id="table9">
          <label>Table 9</label>
          <caption>
            <p>Metrics for the training and testing (external) cohorts for the XGBoost<sup>a</sup> model.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="160"/>
            <col width="180"/>
            <col width="170"/>
            <col width="170"/>
            <col width="160"/>
            <col width="160"/>
            <thead>
              <tr valign="bottom">
                <td>Training data size</td>
                <td>AUC<sup>b</sup> score</td>
                <td>Sensitivity</td>
                <td>Specificity</td>
                <td>Precision</td>
                <td><italic>F</italic><sub>1</sub>-score</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>10,000</td>
                <td>0.76</td>
                <td>0.21</td>
                <td>0.97</td>
                <td>0.69</td>
                <td>0.33</td>
              </tr>
              <tr valign="top">
                <td>100,000</td>
                <td>0.81</td>
                <td>0.27</td>
                <td>0.98</td>
                <td>0.73</td>
                <td>0.39</td>
              </tr>
              <tr valign="top">
                <td>1,000,000</td>
                <td>0.82</td>
                <td>0.33</td>
                <td>0.97</td>
                <td>0.69</td>
                <td>0.45</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table9fn1">
              <p><sup>a</sup>XGBoost: extreme gradient boosting.</p>
            </fn>
            <fn id="table9fn2">
              <p><sup>b</sup>AUC: area under the receiver operating characteristic curve.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <table-wrap position="float" id="table10">
          <label>Table 10</label>
          <caption>
            <p>Metrics for the training and testing (external) cohorts for BERT<sup>a</sup> with up-sampling of the minority class.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="160"/>
            <col width="180"/>
            <col width="170"/>
            <col width="170"/>
            <col width="160"/>
            <col width="160"/>
            <thead>
              <tr valign="bottom">
                <td>Training data size</td>
                <td>AUC<sup>b</sup> score</td>
                <td>Sensitivity</td>
                <td>Specificity</td>
                <td>Precision</td>
                <td><italic>F</italic><sub>1</sub>-score</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>10,000</td>
                <td>0.79</td>
                <td>0.81</td>
                <td>0.58</td>
                <td>0.30</td>
                <td>0.43</td>
              </tr>
              <tr valign="top">
                <td>100,000</td>
                <td>0.84</td>
                <td>0.81</td>
                <td>0.68</td>
                <td>0.34</td>
                <td>0.48</td>
              </tr>
              <tr valign="top">
                <td>1,000,000</td>
                <td>0.85</td>
                <td>0.75</td>
                <td>0.78</td>
                <td>0.41</td>
                <td>0.54</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table10fn1">
              <p><sup>a</sup>BERT: Bidirectional Encoder Representations from Transformers.</p>
            </fn>
            <fn id="table10fn2">
              <p><sup>b</sup>AUC: area under the receiver operating characteristic curve.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
    </sec>
    <sec sec-type="discussion">
      <title>Discussion</title>
      <p>In this study, we compared the performance of several predictive models, including Bio-Clinical-BERT and LR-TF-IDF, in predicting hospitalizations based on nurse triage notes. The findings of our study suggest that while Bio-Clinical-BERT does marginally outperform LR-TF-IDF in this predictive task, the difference in their performance is relatively minor.</p>
      <p>Such results echo the findings of previous studies in the field, which have often found BERT-based models to have a slight edge over more traditional methods such as LR-TF-IDF in various NLP tasks [<xref ref-type="bibr" rid="ref13">13</xref>,<xref ref-type="bibr" rid="ref14">14</xref>]. However, the marginal difference observed in our study suggests that, given certain limitations such as constraints on hardware, software expertise, or budget, hospitals might lean toward simpler methods. The rationale behind such a choice would lie in the ease of implementing these simpler methods, as well as their relatively less demanding computational requirements.</p>
      <p>The comparison of different models in the biomedical domain has been the focus of numerous previous studies. For instance, Chen et al [<xref ref-type="bibr" rid="ref15">15</xref>] conducted an assessment of transformer-based ChatGPT models in tasks such as reasoning and classification. Their study found that fine-tuning remained the most effective approach for 2 central NLP tasks. However, it is interesting to note that the basic BOW model managed to deliver comparable results to the more complex language model prompting. It should be noted that the creation of effective prompts required a substantial resource investment.</p>
      <p>In another study, Xavier and Chen [<xref ref-type="bibr" rid="ref16">16</xref>] compared 3 different model types for a multiclass text classification task, which involved the assignment of protocols for abdominal imaging computed tomography scans. These models spanned a range from conventional machine learning and deep learning to automated machine learning builder workflows. While the automated machine learning builder boasted the best performance with an <italic>F</italic><sub>1</sub>-score of 0.85 on an unbalanced dataset, the tree ensemble machine learning algorithm was superior on a balanced dataset, delivering an <italic>F</italic><sub>1</sub>-score of 0.80.</p>
      <p>A further study delved into the evaluation of machine learning multiclass classification algorithms’ performance in classifying proximal humeral fractures using radiology text data [<xref ref-type="bibr" rid="ref17">17</xref>]. Several statistical machine learning algorithms were performed, with a BERT model showcasing the best accuracy of 61%. In another relevant study conducted by Ji et al [<xref ref-type="bibr" rid="ref18">18</xref>], various models pretrained with BERT were compared for medical code assignment based on clinical notes. Interestingly, it was found that simpler artificial neural networks could sometimes outperform BERT in certain scenarios. This study, among others, offers further support to our recommendation for hospitals with limited resources to consider simpler, less resource-demanding methods for achieving comparable predictive performance.</p>
      <p>In the specific task of predicting hospitalization, both methods in our study effectively leveraged the rich information found within nurse triage notes. This finding aligns with those from other studies [<xref ref-type="bibr" rid="ref19">19</xref>-<xref ref-type="bibr" rid="ref21">21</xref>]. For instance, a study by Zhang et al [<xref ref-type="bibr" rid="ref19">19</xref>] that evaluated LR and neural network modeling approaches in predicting hospital admission or transfer after initial ED triage presentation found that the patient’s free-text data regarding referral improved overall predictive accuracy. Similarly, Raita et al [<xref ref-type="bibr" rid="ref20">20</xref>] used machine learning models to predict ED outcomes and demonstrated superior performance in predicting hospitalization.</p>
      <p>The results of our study carry practical implications for health care organizations. The ability to predict hospitalization from nurse triage notes could lead to improvements in patient care by facilitating efficient resource allocation, optimizing bed management, and improving patient flow.</p>
      <p>The choice between the use of Bio-Clinical-BERT and simpler methods, such as LR-TF-IDF, should be influenced by the specific context of the organization, including factors such as available computational resources, software expertise, and desired model performance.</p>
      <p>Our study is not without limitations. For instance, the data used for our study are specific to MSHS hospitals, which might not be representative of other health care systems, potentially limiting the generalizability of our findings. Despite using multisite data, representing the diverse New York City population, and an external validation site for our final analysis, we acknowledge the need for further studies with more diverse datasets, including those that are open source such as the Medical Information Mart for Intensive Care (MIMIC) dataset. We also recognize that we did not explore the potential of combining both methods and other potential techniques that could enhance these models’ performance. The BOW technique by nature does not consider context, which could have hindered performance. There is the possibility that more advanced deep learning models could have achieved a bigger difference in AUC performance compared to the shallow model. Moreover, the field of NLP is advancing fast, and some methodologies were not explored. Also, our study focused on comparative analysis using the Youden index, which may have caused several metrics to be lower than previous publications, such as the <italic>F</italic><sub>1</sub>-score. Despite this, the models demonstrated high specificity, suggesting potential for clinical use. Further exploration of thresholding methods is necessary to enhance model applicability and performance in real-world settings.</p>
      <p>Future research could focus on the exploration of BERT models that are pretrained and trained from scratch on a site’s entire textual data. Although such an approach may demand significant resources and be computationally intensive, it might yield better performance by capturing the unique characteristics and language patterns of a specific health care setting. The exploration of other pretrained language models or more advanced natural language processing techniques could also pave the way for the development of more effective hospitalization prediction methods based on nurse triage notes.</p>
      <p>In conclusion, our study demonstrates that while the Bio-Clinical-BERT model does marginally outperform the LR-TF-IDF model in predicting hospitalization from nurse triage notes, the difference is small enough to suggest that simpler methods might be viable for hospitals with limited resources. More research is needed to identify alternative methods that can enhance these models’ performance in predicting hospitalization, ultimately improving patient care and hospital resource management.</p>
      <p>Through an investigation of the Bio-Clinical-BERT and LR-TF-IDF models’ performance, our study contributes to the growing body of literature in the field of NLP and machine learning in health care. It emphasizes the importance of considering the trade-offs between model complexity and performance when deploying predictive tools in clinical settings, highlighting that sometimes, simpler methods can prove as effective as more complex ones.</p>
    </sec>
  </body>
  <back>
    <app-group/>
    <glossary>
      <title>Abbreviations</title>
      <def-list>
        <def-item>
          <term id="abb1">AUC</term>
          <def>
            <p>area under the receiver operating characteristic curve</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb2">BERT</term>
          <def>
            <p>Bidirectional Encoder Representations from Transformers</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb3">Bi-LSTM</term>
          <def>
            <p>bidirectional long short-term memory</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb4">BOW</term>
          <def>
            <p>bag-of-words</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb5">ED</term>
          <def>
            <p>emergency department</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb6">LR</term>
          <def>
            <p>logistic regression</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb7">MI</term>
          <def>
            <p>mutual information</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb8">MIMIC</term>
          <def>
            <p>Medical Information Mart for Intensive Care</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb9">MSHS</term>
          <def>
            <p>Mount Sinai Health System</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb10">MSQ</term>
          <def>
            <p>Mount Sinai Queens</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb11">NLP</term>
          <def>
            <p>natural language processing</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb12">OR</term>
          <def>
            <p>odds ratio</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb13">ROC</term>
          <def>
            <p>receiver operating characteristic</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb14">TF-IDF</term>
          <def>
            <p>term frequency–inverse document frequency</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb15">W2V</term>
          <def>
            <p>Word-2-Vec</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb16">XGBoost</term>
          <def>
            <p>extreme gradient boosting</p>
          </def>
        </def-item>
      </def-list>
    </glossary>
    <fn-group>
      <fn fn-type="conflict">
        <p>None declared.</p>
      </fn>
    </fn-group>
    <ref-list>
      <ref id="ref1">
        <label>1</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Chalfin</surname>
              <given-names>DB</given-names>
            </name>
            <name name-style="western">
              <surname>Trzeciak</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Likourezos</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Baumann</surname>
              <given-names>BM</given-names>
            </name>
            <name name-style="western">
              <surname>Dellinger</surname>
              <given-names>RP</given-names>
            </name>
            <collab>DELAY-ED study group</collab>
          </person-group>
          <article-title>Impact of delayed transfer of critically ill patients from the emergency department to the intensive care unit</article-title>
          <source>Crit Care Med</source>
          <year>2007</year>
          <month>06</month>
          <volume>35</volume>
          <issue>6</issue>
          <fpage>1477</fpage>
          <lpage>1483</lpage>
          <pub-id pub-id-type="doi">10.1097/01.CCM.0000266585.74905.5A</pub-id>
          <pub-id pub-id-type="medline">17440421</pub-id>
          <pub-id pub-id-type="pmcid">7440421</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref2">
        <label>2</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Rabin</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Kocher</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>McClelland</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Pines</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Hwang</surname>
              <given-names>U</given-names>
            </name>
            <name name-style="western">
              <surname>Rathlev</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Asplin</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Trueger</surname>
              <given-names>NS</given-names>
            </name>
            <name name-style="western">
              <surname>Weber</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <article-title>Solutions to emergency department 'boarding' and crowding are underused and may need to be legislated</article-title>
          <source>Health Aff (Millwood)</source>
          <year>2012</year>
          <month>08</month>
          <day>01</day>
          <volume>31</volume>
          <issue>8</issue>
          <fpage>1757</fpage>
          <lpage>1766</lpage>
          <pub-id pub-id-type="doi">10.1377/hlthaff.2011.0786</pub-id>
          <pub-id pub-id-type="medline">22869654</pub-id>
          <pub-id pub-id-type="pii">31/8/1757</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref3">
        <label>3</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Forero</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>McCarthy</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Hillman</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Access block and emergency department overcrowding</article-title>
          <source>Crit Care</source>
          <year>2011</year>
          <month>3</month>
          <day>22</day>
          <volume>15</volume>
          <issue>2</issue>
          <fpage>216</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://ccforum.biomedcentral.com/articles/10.1186/cc9998"/>
          </comment>
          <pub-id pub-id-type="doi">10.1186/cc9998</pub-id>
          <pub-id pub-id-type="medline">21457507</pub-id>
          <pub-id pub-id-type="pii">cc9998</pub-id>
          <pub-id pub-id-type="pmcid">PMC3219412</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref4">
        <label>4</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Sterling</surname>
              <given-names>NW</given-names>
            </name>
            <name name-style="western">
              <surname>Patzer</surname>
              <given-names>RE</given-names>
            </name>
            <name name-style="western">
              <surname>Di</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Schrager</surname>
              <given-names>JD</given-names>
            </name>
          </person-group>
          <article-title>Prediction of emergency department patient disposition based on natural language processing of triage notes</article-title>
          <source>Int J Med Inform</source>
          <year>2019</year>
          <month>09</month>
          <volume>129</volume>
          <fpage>184</fpage>
          <lpage>188</lpage>
          <pub-id pub-id-type="doi">10.1016/j.ijmedinf.2019.06.008</pub-id>
          <pub-id pub-id-type="medline">31445253</pub-id>
          <pub-id pub-id-type="pii">S1386-5056(19)30375-2</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref5">
        <label>5</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Patel</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Cheetirala</surname>
              <given-names>SN</given-names>
            </name>
            <name name-style="western">
              <surname>Raut</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Tamegue</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Kia</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Glicksberg</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Freeman</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Levin</surname>
              <given-names>MA</given-names>
            </name>
            <name name-style="western">
              <surname>Timsina</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Klang</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <article-title>Predicting adult hospital admission from emergency department using machine learning: an inclusive gradient boosting model</article-title>
          <source>J Clin Med</source>
          <year>2022</year>
          <month>11</month>
          <day>22</day>
          <volume>11</volume>
          <issue>23</issue>
          <fpage>634</fpage>
          <lpage>635</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.mdpi.com/resolver?pii=jcm11236888"/>
          </comment>
          <pub-id pub-id-type="doi">10.3390/jcm11236888</pub-id>
          <pub-id pub-id-type="medline">36498463</pub-id>
          <pub-id pub-id-type="pii">jcm11236888</pub-id>
          <pub-id pub-id-type="pmcid">PMC9740100</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref6">
        <label>6</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Deo</surname>
              <given-names>RC</given-names>
            </name>
          </person-group>
          <article-title>Machine learning in medicine</article-title>
          <source>Circulation</source>
          <year>2015</year>
          <month>11</month>
          <day>17</day>
          <volume>132</volume>
          <issue>20</issue>
          <fpage>1920</fpage>
          <lpage>1930</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/26572668"/>
          </comment>
          <pub-id pub-id-type="doi">10.1161/CIRCULATIONAHA.115.001593</pub-id>
          <pub-id pub-id-type="medline">26572668</pub-id>
          <pub-id pub-id-type="pii">CIRCULATIONAHA.115.001593</pub-id>
          <pub-id pub-id-type="pmcid">PMC5831252</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref7">
        <label>7</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Klang</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Barash</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Soffer</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Bechler</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Resheff</surname>
              <given-names>YS</given-names>
            </name>
            <name name-style="western">
              <surname>Granot</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Shahar</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Klug</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Guralnik</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Zimlichman</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Konen</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <article-title>Promoting head CT exams in the emergency department triage using a machine learning model</article-title>
          <source>Neuroradiology</source>
          <year>2020</year>
          <month>02</month>
          <day>10</day>
          <volume>62</volume>
          <issue>2</issue>
          <fpage>153</fpage>
          <lpage>160</lpage>
          <pub-id pub-id-type="doi">10.1007/s00234-019-02293-y</pub-id>
          <pub-id pub-id-type="medline">31598737</pub-id>
          <pub-id pub-id-type="pii">10.1007/s00234-019-02293-y</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref8">
        <label>8</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Devlin</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Chang</surname>
              <given-names>MW</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Toutanova</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>BERT: pre-training of deep bidirectional transformers for language understanding</article-title>
          <source>arXiv</source>
          <comment>Preprint posted online on October 11, 2018.</comment>
          <pub-id pub-id-type="doi">10.48550/arXiv.1810.04805</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref9">
        <label>9</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Soffer</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Glicksberg</surname>
              <given-names>BS</given-names>
            </name>
            <name name-style="western">
              <surname>Zimlichman</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Klang</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <article-title>BERT for the processing of radiological reports: an attention-based natural language processing algorithm</article-title>
          <source>Acad Radiol</source>
          <year>2022</year>
          <month>04</month>
          <volume>29</volume>
          <issue>4</issue>
          <fpage>634</fpage>
          <lpage>635</lpage>
          <pub-id pub-id-type="doi">10.1016/j.acra.2021.03.036</pub-id>
          <pub-id pub-id-type="medline">34362663</pub-id>
          <pub-id pub-id-type="pii">S1076-6332(21)00311-1</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref10">
        <label>10</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Alsentzer</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Murphy</surname>
              <given-names>JR</given-names>
            </name>
            <name name-style="western">
              <surname>Boag</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Weng</surname>
              <given-names>WH</given-names>
            </name>
            <name name-style="western">
              <surname>Jin</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Naumann</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>McDermott</surname>
              <given-names>MBA</given-names>
            </name>
          </person-group>
          <article-title>Publicly available clinical BERT embeddings</article-title>
          <source>arXiv</source>
          <comment>Preprint posted online on April 6, 2019.</comment>
          <pub-id pub-id-type="doi">10.48550/arXiv.1904.03323</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref11">
        <label>11</label>
        <nlm-citation citation-type="web">
          <article-title>Bio_ClinicalBert</article-title>
          <source>Hugging Face</source>
          <access-date>2024-08-15</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://huggingface.co/emilyalsentzer/Bio_ClinicalBERT">https://huggingface.co/emilyalsentzer/Bio_ClinicalBERT</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref12">
        <label>12</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Klang</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Levin</surname>
              <given-names>MA</given-names>
            </name>
            <name name-style="western">
              <surname>Soffer</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Zebrowski</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Glicksberg</surname>
              <given-names>BS</given-names>
            </name>
            <name name-style="western">
              <surname>Carr</surname>
              <given-names>BG</given-names>
            </name>
            <name name-style="western">
              <surname>Mcgreevy</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Reich</surname>
              <given-names>DL</given-names>
            </name>
            <name name-style="western">
              <surname>Freeman</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>A simple free-text-like method for extracting semi-structured data from electronic health records: exemplified in prediction of in-hospital mortality</article-title>
          <source>Big Data Cogn Comput</source>
          <year>2021</year>
          <month>08</month>
          <day>29</day>
          <volume>5</volume>
          <issue>3</issue>
          <fpage>40</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.3390/bdcc5030040[CrossRef"/>
          </comment>
          <pub-id pub-id-type="doi">10.3390/bdcc5030040</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref13">
        <label>13</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Khan</surname>
              <given-names>Jy</given-names>
            </name>
            <name name-style="western">
              <surname>Khondaker</surname>
              <given-names>MTI</given-names>
            </name>
            <name name-style="western">
              <surname>Afroz</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Uddin</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Iqbal</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>A benchmark study of machine learning models for online fake news detection</article-title>
          <source>Mach Learn Appl</source>
          <year>2021</year>
          <month>06</month>
          <day>15</day>
          <volume>4</volume>
          <fpage>100032</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1016/j.mlwa.2021.100032[CrossRef"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.mlwa.2021.100032</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref14">
        <label>14</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Yenkikar</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Babu</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>AirBERT: a fine-tuned language representation model for airlines tweet sentiment analysis</article-title>
          <source>Intelligent Decision Technologies</source>
          <year>2023</year>
          <month>05</month>
          <day>15</day>
          <volume>17</volume>
          <issue>2</issue>
          <fpage>435</fpage>
          <lpage>455</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://dl.acm.org/doi/abs/10.3233/IDT-220173"/>
          </comment>
          <pub-id pub-id-type="doi">10.3233/idt-220173</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref15">
        <label>15</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Lu</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Van</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Aerts</surname>
              <given-names>HJWL</given-names>
            </name>
            <name name-style="western">
              <surname>Savova</surname>
              <given-names>GK</given-names>
            </name>
            <name name-style="western">
              <surname>Bitterman</surname>
              <given-names>DS</given-names>
            </name>
          </person-group>
          <article-title>Evaluating the ChatGPT family of models for biomedical reasoning and classification</article-title>
          <source>J Am Med Inform Assoc</source>
          <year>2024</year>
          <month>04</month>
          <day>03</day>
          <volume>31</volume>
          <issue>4</issue>
          <fpage>940</fpage>
          <lpage>948</lpage>
          <pub-id pub-id-type="doi">10.1093/jamia/ocad256</pub-id>
          <pub-id pub-id-type="medline">38261400</pub-id>
          <pub-id pub-id-type="pii">7585396</pub-id>
          <pub-id pub-id-type="pmcid">PMC10990500</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref16">
        <label>16</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Xavier</surname>
              <given-names>BA</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>P</given-names>
            </name>
          </person-group>
          <article-title>Natural language processing for imaging protocol assignment: machine learning for multiclass classification of abdominal CT protocols using indication text data</article-title>
          <source>J Digit Imaging</source>
          <year>2022</year>
          <month>10</month>
          <volume>35</volume>
          <issue>5</issue>
          <fpage>1120</fpage>
          <lpage>1130</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/35654878"/>
          </comment>
          <pub-id pub-id-type="doi">10.1007/s10278-022-00633-8</pub-id>
          <pub-id pub-id-type="medline">35654878</pub-id>
          <pub-id pub-id-type="pii">10.1007/s10278-022-00633-8</pub-id>
          <pub-id pub-id-type="pmcid">PMC9582109</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref17">
        <label>17</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Dipnall</surname>
              <given-names>JF</given-names>
            </name>
            <name name-style="western">
              <surname>Lu</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Gabbe</surname>
              <given-names>BJ</given-names>
            </name>
            <name name-style="western">
              <surname>Cosic</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Edwards</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Page</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Du</surname>
              <given-names>L</given-names>
            </name>
          </person-group>
          <article-title>Comparison of state-of-the-art machine and deep learning algorithms to classify proximal humeral fractures using radiology text</article-title>
          <source>Eur J Radiol</source>
          <year>2022</year>
          <month>08</month>
          <volume>153</volume>
          <fpage>110366</fpage>
          <pub-id pub-id-type="doi">10.1016/j.ejrad.2022.110366</pub-id>
          <pub-id pub-id-type="medline">35623313</pub-id>
          <pub-id pub-id-type="pii">S0720-048X(22)00216-9</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref18">
        <label>18</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ji</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Hölttä</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Marttinen</surname>
              <given-names>P</given-names>
            </name>
          </person-group>
          <article-title>Does the magic of BERT apply to medical code assignment? a quantitative study</article-title>
          <source>Comput Biol Med</source>
          <year>2021</year>
          <month>12</month>
          <volume>139</volume>
          <fpage>104998</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S0010-4825(21)00792-7"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.compbiomed.2021.104998</pub-id>
          <pub-id pub-id-type="medline">34739971</pub-id>
          <pub-id pub-id-type="pii">S0010-4825(21)00792-7</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref19">
        <label>19</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Patzer</surname>
              <given-names>RE</given-names>
            </name>
            <name name-style="western">
              <surname>Pitts</surname>
              <given-names>SR</given-names>
            </name>
            <name name-style="western">
              <surname>Patzer</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Schrager</surname>
              <given-names>JD</given-names>
            </name>
          </person-group>
          <article-title>Prediction of emergency department hospital admission based on natural language processing and neural networks</article-title>
          <source>Methods Inf Med</source>
          <year>2017</year>
          <month>10</month>
          <day>26</day>
          <volume>56</volume>
          <issue>5</issue>
          <fpage>377</fpage>
          <lpage>389</lpage>
          <pub-id pub-id-type="doi">10.3414/ME17-01-0024</pub-id>
          <pub-id pub-id-type="medline">28816338</pub-id>
          <pub-id pub-id-type="pii">17-01-0024</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref20">
        <label>20</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Raita</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Goto</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Faridi</surname>
              <given-names>MK</given-names>
            </name>
            <name name-style="western">
              <surname>Brown</surname>
              <given-names>DFM</given-names>
            </name>
            <name name-style="western">
              <surname>Camargo</surname>
              <given-names>CA</given-names>
            </name>
            <name name-style="western">
              <surname>Hasegawa</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Emergency department triage prediction of clinical outcomes using machine learning models</article-title>
          <source>Crit Care</source>
          <year>2019</year>
          <month>02</month>
          <day>22</day>
          <volume>23</volume>
          <issue>1</issue>
          <fpage>64</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://ccforum.biomedcentral.com/articles/10.1186/s13054-019-2351-7"/>
          </comment>
          <pub-id pub-id-type="doi">10.1186/s13054-019-2351-7</pub-id>
          <pub-id pub-id-type="medline">30795786</pub-id>
          <pub-id pub-id-type="pii">10.1186/s13054-019-2351-7</pub-id>
          <pub-id pub-id-type="pmcid">PMC6387562</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref21">
        <label>21</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Klang</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Kummer</surname>
              <given-names>BR</given-names>
            </name>
            <name name-style="western">
              <surname>Dangayach</surname>
              <given-names>NS</given-names>
            </name>
            <name name-style="western">
              <surname>Zhong</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Kia</surname>
              <given-names>MA</given-names>
            </name>
            <name name-style="western">
              <surname>Timsina</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Cossentino</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Costa</surname>
              <given-names>AB</given-names>
            </name>
            <name name-style="western">
              <surname>Levin</surname>
              <given-names>MA</given-names>
            </name>
            <name name-style="western">
              <surname>Oermann</surname>
              <given-names>EK</given-names>
            </name>
          </person-group>
          <article-title>Predicting adult neuroscience intensive care unit admission from emergency department triage using a retrospective, tabular-free text machine learning approach</article-title>
          <source>Sci Rep</source>
          <year>2021</year>
          <month>01</month>
          <day>14</day>
          <volume>11</volume>
          <issue>1</issue>
          <fpage>1381</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/s41598-021-80985-3"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41598-021-80985-3</pub-id>
          <pub-id pub-id-type="medline">33446890</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41598-021-80985-3</pub-id>
          <pub-id pub-id-type="pmcid">PMC7809037</pub-id>
        </nlm-citation>
      </ref>
    </ref-list>
  </back>
</article>
