<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "http://dtd.nlm.nih.gov/publishing/2.0/journalpublishing.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article" dtd-version="2.0">
  <front>
    <journal-meta>
      <journal-id journal-id-type="publisher-id">JMIR</journal-id>
      <journal-id journal-id-type="nlm-ta">JMIR AI</journal-id>
      <journal-title>JMIR AI</journal-title>
      <issn pub-type="epub">2817-1705</issn>
      <publisher>
        <publisher-name>JMIR Publications</publisher-name>
        <publisher-loc>Toronto, Canada</publisher-loc>
      </publisher>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="publisher-id">v2i1e45770</article-id>
      <article-id pub-id-type="pmid">38875563</article-id>
      <article-id pub-id-type="doi">10.2196/45770</article-id>
      <article-categories>
        <subj-group subj-group-type="heading">
          <subject>Original Paper</subject>
        </subj-group>
        <subj-group subj-group-type="article-type">
          <subject>Original Paper</subject>
        </subj-group>
      </article-categories>
      <title-group>
        <article-title>The Evolution of Artificial Intelligence in Biomedicine: Bibliometric Analysis</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="editor">
          <name>
            <surname>El Emam</surname>
            <given-names>Khaled</given-names>
          </name>
        </contrib>
        <contrib contrib-type="editor">
          <name>
            <surname>Malin</surname>
            <given-names>Bradley</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Huang</surname>
            <given-names>Li</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Benítez-Andrades</surname>
            <given-names>José Alberto</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Kohen</surname>
            <given-names>David</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib id="contrib1" contrib-type="author" equal-contrib="yes">
          <name name-style="western">
            <surname>Gu</surname>
            <given-names>Jiasheng</given-names>
          </name>
          <degrees>MS</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0009-0004-2678-1607</ext-link>
        </contrib>
        <contrib id="contrib2" contrib-type="author">
          <name name-style="western">
            <surname>Gao</surname>
            <given-names>Chongyang</given-names>
          </name>
          <degrees>MS</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-2358-4710</ext-link>
        </contrib>
        <contrib id="contrib3" contrib-type="author" corresp="yes" equal-contrib="yes">
          <name name-style="western">
            <surname>Wang</surname>
            <given-names>Lili</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff3" ref-type="aff">3</xref>
          <address>
            <institution>Department of Computer Science</institution>
            <institution>Dartmouth College</institution>
            <addr-line>15 Thayer Dive</addr-line>
            <addr-line>Hanover, NH, 03755</addr-line>
            <country>United States</country>
            <phone>1 516 888 6691</phone>
            <email>lili.wang.gr@dartmouth.edu</email>
          </address>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-2864-7984</ext-link>
        </contrib>
      </contrib-group>
      <aff id="aff1">
        <label>1</label>
        <institution>Department of Computer Science</institution>
        <institution>University of Southern California</institution>
        <addr-line>Los Angeles, CA</addr-line>
        <country>United States</country>
      </aff>
      <aff id="aff2">
        <label>2</label>
        <institution>Department of Computer Science</institution>
        <institution>Northwestern University</institution>
        <addr-line>Evanston, IL</addr-line>
        <country>United States</country>
      </aff>
      <aff id="aff3">
        <label>3</label>
        <institution>Department of Computer Science</institution>
        <institution>Dartmouth College</institution>
        <addr-line>Hanover, NH</addr-line>
        <country>United States</country>
      </aff>
      <author-notes>
        <corresp>Corresponding Author: Lili Wang <email>lili.wang.gr@dartmouth.edu</email></corresp>
      </author-notes>
      <pub-date pub-type="collection">
        <year>2023</year>
      </pub-date>
      <pub-date pub-type="epub">
        <day>19</day>
        <month>12</month>
        <year>2023</year>
      </pub-date>
      <volume>2</volume>
      <elocation-id>e45770</elocation-id>
      <history>
        <date date-type="received">
          <day>16</day>
          <month>1</month>
          <year>2023</year>
        </date>
        <date date-type="rev-request">
          <day>19</day>
          <month>5</month>
          <year>2023</year>
        </date>
        <date date-type="rev-recd">
          <day>11</day>
          <month>6</month>
          <year>2023</year>
        </date>
        <date date-type="accepted">
          <day>29</day>
          <month>10</month>
          <year>2023</year>
        </date>
      </history>
      <copyright-statement>©Jiasheng Gu, Chongyang Gao, Lili Wang. Originally published in JMIR AI (https://ai.jmir.org), 19.12.2023.</copyright-statement>
      <copyright-year>2023</copyright-year>
      <license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/">
        <p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR AI, is properly cited. The complete bibliographic information, a link to the original publication on https://www.ai.jmir.org/, as well as this copyright and license information must be included.</p>
      </license>
      <self-uri xlink:href="https://ai.jmir.org/2023/1/e45770" xlink:type="simple"/>
      <abstract>
        <sec sec-type="background">
          <title>Background</title>
          <p>The utilization of artificial intelligence (AI) technologies in the biomedical field has attracted increasing attention in recent decades. Studying how past AI technologies have found their way into medicine over time can help to predict which current (and future) AI technologies have the potential to be utilized in medicine in the coming years, thereby providing a helpful reference for future research directions.</p>
        </sec>
        <sec sec-type="objective">
          <title>Objective</title>
          <p>The aim of this study was to predict the future trend of AI technologies used in different biomedical domains based on past trends of related technologies and biomedical domains.</p>
        </sec>
        <sec sec-type="methods">
          <title>Methods</title>
          <p>We collected a large corpus of articles from the PubMed database pertaining to the intersection of AI and biomedicine. Initially, we attempted to use regression on the extracted keywords alone; however, we found that this approach did not provide sufficient information. Therefore, we propose a method called “background-enhanced prediction” to expand the knowledge utilized by the regression algorithm by incorporating both the keywords and their surrounding context. This method of data construction resulted in improved performance across the six regression models evaluated. Our findings were confirmed through experiments on recurrent prediction and forecasting.</p>
        </sec>
        <sec sec-type="results">
          <title>Results</title>
          <p>In our analysis using background information for prediction, we found that a window size of 3 yielded the best results, outperforming the use of keywords alone. Furthermore, utilizing data only prior to 2017, our regression projections for the period of 2017-2021 exhibited a high coefficient of determination (R<sup>2</sup>), which reached up to 0.78, demonstrating the effectiveness of our method in predicting long-term trends. Based on the prediction, studies related to proteins and tumors will be pushed out of the top 20 and become replaced by early diagnostics, tomography, and other detection technologies. These are certain areas that are well-suited to incorporate AI technology. Deep learning, machine learning, and neural networks continue to be the dominant AI technologies in biomedical applications. Generative adversarial networks represent an emerging technology with a strong growth trend.</p>
        </sec>
        <sec sec-type="conclusions">
          <title>Conclusions</title>
          <p>In this study, we explored AI trends in the biomedical field and developed a predictive model to forecast future trends. Our findings were confirmed through experiments on current trends.</p>
        </sec>
      </abstract>
      <kwd-group>
        <kwd>bibliometrics</kwd>
        <kwd>trend forecasting</kwd>
        <kwd>AI in medicine</kwd>
        <kwd>Word2Vec</kwd>
        <kwd>regression models</kwd>
        <kwd>agglomerative clustering</kwd>
        <kwd>usage</kwd>
        <kwd>artificial intelligence</kwd>
        <kwd>utilization</kwd>
        <kwd>biomedical</kwd>
        <kwd>effectiveness</kwd>
        <kwd>AI trends</kwd>
        <kwd>predictive model</kwd>
        <kwd>development</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec sec-type="introduction">
      <title>Introduction</title>
      <sec>
        <title>Artificial Intelligence in Biomedicine</title>
        <p>Medicine has long been recognized as a prime area for applying artificial intelligence (AI) [<xref ref-type="bibr" rid="ref1">1</xref>], with biomedicine being a vibrant and promising field. Advances in technology and science have led to the use of various methods to obtain biomedical data, such as clinical analyses, biological parameters, and medical imaging. However, the diversity and complexity of these data, along with the need for more information on certain atypical diseases result in unbalanced and nonsmooth biomedical data. In this scenario, machine learning can improve medical big data analysis, reduce the risk of medical errors, and generate a more unified diagnostic and prognostic protocol.</p>
        <p>Recent AI research has leveraged machine learning methods to identify patterns and complex interactions from data, which require large amounts of data as support. Artificial neural networks and deep learning are currently among the most popular machine learning technologies. These methods are used in biomedicine across all medical dimensions, from genomic applications such as gene expression to public health care management such as for predicting population information or infectious disease outbreaks [<xref ref-type="bibr" rid="ref2">2</xref>]. AI has also significantly impacted biomedical processors such as electrocardiogram, electroencephalogram, and electromyography classification processors and hearing aid processors [<xref ref-type="bibr" rid="ref3">3</xref>].</p>
        <p>AI is increasingly being utilized in a variety of applications in the biomedical field. Notable examples include IBM Watson-Oncology, which selects drugs for cancer treatment with equal or superior efficiency compared to human experts; Microsoft’s Hanover project at Oregon, which personalizes cancer treatment plans through analysis of medical research; and the UK National Health Service utilizing Google’s DeepMind platform to detect health risks by analyzing mobile app data and medical images from patients. Additionally, algorithms developed at Stanford University have been shown to detect pneumonia more accurately than human radiologists; in the diabetic retinopathy challenge, the computer was as effective as an ophthalmologist in making referral decisions [<xref ref-type="bibr" rid="ref4">4</xref>]. Therefore, it is essential to analyze the trends in the integration of these AI-related technologies with the biomedical field to understand which technologies have played an important role in the past, predict the current and emerging technologies that are more likely to be important in the future, and determine which original technologies are regaining importance in a particular biomedical field.</p>
        <p>Language models offer an effective means to analyze texts and have become the basis for many applications, including machine translation and text classification. In all text-related fields, language models can bring new improvements and opportunities to a greater or lesser extent and assist in literature research.</p>
      </sec>
      <sec>
        <title>Co-word Analysis</title>
        <p>Recently, increased attention has been paid to the management of references and expansion of the research scope. Bibliometric analysis summarizes the structure of a field by analyzing the social and structural relationships between different research components such as authors, countries, institutions, and topics. Additionally, bibliometric analysis significantly impacts reorienting research and identifying popular issues. Thus, bibliometric analysis enables discovery of how research in a given field is distributed and changing. The data collected and the conclusions drawn from a bibliometric analysis can be used to track popular topics, predict promising technologies, and assist scientists in redirecting their research. There has been substantial research and application of bibliometric analysis in academia and industry, and extracting keywords to analyze texts is a very common strategy in such studies. Although it is intuitive to use the whole text as an object of analysis, this requires extensive computational resources. Moreover, many texts are not of high quality, some of them are repetitive or have no actual content, and a lot of noise can make the model learn the wrong knowledge. Therefore, keyword-focused analysis is often a better choice. Co-word analysis is one such technique that focuses on keywords and analyzes the content itself [<xref ref-type="bibr" rid="ref5">5</xref>]. This analysis aims to uncover the intrinsic connections of articles and discover trends within them with applications in many fields, including medicine and business.</p>
        <p>Co-word analysis was first proposed by French bibliometricians in the late 1970s [<xref ref-type="bibr" rid="ref6">6</xref>] as a technique for studying keywords in the content of publications. Words in the co-word analysis are typically derived from the article title, abstract, and full text. These words may be specifically extracted from certain parts of each component, depending on the goal of the analysis. Co-word analysis assumes that words that frequently occur together have thematic relationships with each other. Based on this assumption, co-word analysis can be used to predict future research in a field. Analysis of the keywords of published articles in a given field has the potential to predict keywords for future research in the field, which in turn portrays the future of the research field accordingly. Co-word analysis uses several methods based on covariate matrices, such as factor, cluster, multivariate, and social network analyses. These methods help researchers to obtain an overview of a field. Thus, co-word analysis is a method to analyze papers in a field and make valid judgments.</p>
      </sec>
      <sec>
        <title>Text Similarity</title>
        <p>Text similarity measurement is fundamental to natural language processing tasks and is essential in information retrieval, question answering, machine translation, and dialogue systems, among other applications. In recent years, various techniques for measuring semantic similarity have been proposed. Text similarity techniques can be divided into two main categories: text distance and text representation [<xref ref-type="bibr" rid="ref7">7</xref>].</p>
        <p>Text distance describes the semantic similarity of two text words from the perspective of distance. Length-based and distribution-based distance are the two main types of text distance. Traditionally, text similarity is evaluated by measuring the length distance, which uses the numerical properties of the text to calculate the text vector distance length, such as the Euclidean distance, cosine distance, or Manhattan distance [<xref ref-type="bibr" rid="ref8">8</xref>]. However, the text similarity should not be symmetric and the length distance does not consider the statistical characteristics of the data. The distribution distance determines the similarity between documents based on the similarity of their distribution, such as Jensen-Shannon divergence [<xref ref-type="bibr" rid="ref9">9</xref>], Kullback-Leibler divergence [<xref ref-type="bibr" rid="ref10">10</xref>], and Wasserstein distance [<xref ref-type="bibr" rid="ref11">11</xref>], among others.</p>
        <p>Text representation methods convert text to a numerical feature vector. These methods are mainly divided into a string-based method, corpus-based method, semantic text matching, and graph structure–based method. String-based methods operate on string sequences and character compositions to measure the similarity or dissimilarity (distance) between two text strings for approximate string matching or comparison. The advantage of such methods is that they are simple to compute. Representative string-based methods include longest common subsequence [<xref ref-type="bibr" rid="ref12">12</xref>], Edit distance [<xref ref-type="bibr" rid="ref13">13</xref>], Jaro similarity [<xref ref-type="bibr" rid="ref14">14</xref>], Dice [<xref ref-type="bibr" rid="ref15">15</xref>], and Jaccard [<xref ref-type="bibr" rid="ref16">16</xref>]. The corpus-based methods use information from the corpus to compute text similarity; this information can be either text features or co-occurrence probabilities. In recent studies, corpus-based approaches include three different measures: the bag-of-words model, distributed representation, and matrix decomposition method. The corpus-based methods mainly include bag-of-words [<xref ref-type="bibr" rid="ref17">17</xref>], text frequency-inverse document frequency [<xref ref-type="bibr" rid="ref18">18</xref>], Word2Vec [<xref ref-type="bibr" rid="ref19">19</xref>], latent semantic analysis [<xref ref-type="bibr" rid="ref20">20</xref>], and others. Semantic similarity determines the similarity between text and documents based on their meaning rather than character-by-character matching. Deep-structured semantic models [<xref ref-type="bibr" rid="ref21">21</xref>] are typical models in this regard. Graph-based text similarities are mainly based on a knowledge-graph representation and a graph neural network representation. The graph structure better enables determining the similarity between nodes. Knowledge graphs [<xref ref-type="bibr" rid="ref22">22</xref>] and graph neural networks [<xref ref-type="bibr" rid="ref23">23</xref>] are the main methods for exploiting a graph structure.</p>
      </sec>
      <sec>
        <title>Predicting the Future of AI in Health Care</title>
        <p>Some previous works have also discussed the application of AI in medicine and possible future directions. One is integrative analysis [<xref ref-type="bibr" rid="ref24">24</xref>], where data from different modalities can describe various aspects of a health problem. By mining these heterogeneous data in an integrated way, holistic and comprehensive insight into health can be obtained. In recent years, there has been a growing number of studies and initiatives related to AI in health, integrating different aspects of clinical data and linking drug development to clinical data. AI for precision medicine [<xref ref-type="bibr" rid="ref25">25</xref>] represents another promising combination of AI and medicine, which assists in solving the most complex problems in personalized care. For example, AI in microscopic diagnostics [<xref ref-type="bibr" rid="ref26">26</xref>] can improve the work of pathologists and may even gradually replace their work.</p>
        <p>In this study, we used language models to measure the relationship between keywords, which can subsequently assist in building aggregation models and using adjacent keywords. Specifically, we propose a background-enhanced prediction method for constructing data for prediction using adjacent keywords, which refer to matrices adjacent to a 2D correlation matrix constructed using a clustering algorithm. This approach allows regression models to learn better and more accurately predict the relationships between keywords. We applied this approach to predict the future trend of AI technologies used in different biomedical domains based on past trends of related technologies and biomedical domains. We further compared the prediction results to the patterns of current trends to evaluate the reliability of the prediction.</p>
      </sec>
    </sec>
    <sec sec-type="methods">
      <title>Methods</title>
      <sec>
        <title>Data Sets</title>
        <p>The data sets used in this study were obtained from the National Institutes of Health PubMed and PMC collections, with measures taken to avoid duplication by utilizing unique identifiers.</p>
        <p>The corpus utilized in this study consists of three parts: (1) 114,266 abstracts and 49,126 full texts from PubMed and PMC obtained by searching keywords such as “machine learning,” “data mining,” “artificial intelligence,” “deep learning,” and “classifier” in the Title/Abstract field; (2) 61,382 full-text papers from PMC obtained by searching keywords such as “machine learning,” “data mining,” “artificial intelligence,” and “deep learning” in all fields, serving as a complement to the previous part; and (3) 2,507,391 full-text papers retrieved from the PubMed Central Open Access section with no keyword filtering to capture a comprehensive understanding of the biomedical field.</p>
        <p>Due to permission restrictions, full-text access was limited for some papers. The full text of the papers primarily served for training, with the core of our experiments lying in the analysis and model prediction based on the abstracts.</p>
      </sec>
      <sec>
        <title>Language Model</title>
        <p>We utilized the word-embedding model Word2Vec as our language model owing to its advantages of efficiency and robustness among other available options [<xref ref-type="bibr" rid="ref27">27</xref>].</p>
        <p>Word embedding is a method of transforming a single word into a digital representation that captures various features of the word within a text, such as semantic relationships, definitions, and contexts. These digital representations can be used to identify similarities or dissimilarities between words.</p>
        <p>To feed text data into a machine learning model, the text must be converted into an embedding. A simple method to achieve this involves “hot-coding” the text data, where each vector is mapped to a category. However, such simple embeddings have limitations as they do not capture the features of the words and can be large depending on the corpus size.</p>
        <p>The effectiveness of Word2Vec is derived from its ability to combine vectors of similar words, leading to reliable estimates of word meaning based on their frequency in the corpus. This results in associations with other words, such as similar embedding vectors of “king” and “queen.” Algebraic operations on word embeddings can also provide approximations of word similarity, such as obtaining the vector for “queen” by subtracting the vector for “man” from the vector for “king” and adding the vector for “woman.” The cosine similarity measure is used to compare the similarity of two words, which is calculated according to the following formula:</p>
        <disp-formula><italic>cos</italic>(<italic>x</italic>,<italic>y</italic>)=<italic>x</italic>·<italic>y</italic>/∣<italic>x</italic>∣×∣<italic>y</italic>∣</disp-formula>
        <p>To improve the suitability of the original corpus for our language model, we performed extensive preprocessing to address any noise that may impact the model’s effectiveness. This included removing all numeric and nonalphabetic characters, except for the special character “-,” which is often used to link multiple words and create unique phrases. Additionally, to enhance the word vectors of biomedical- and AI-related keywords, we transformed multiword keywords in 114,266 abstracts into single tokens by merging them; for example, “machine learning” was merged into “machine+learning.”</p>
        <p>The selection of hyperparameters was based on the available computational resources and the training corpus size. Our Word2Vec model had 300 dimensions and a window size of 5. Our computational device is a cluster with 384 GB of memory and 16 CPU cores. The Word2Vec model was trained sequentially on the three data sets, with the entire training process taking approximately 72 hours.</p>
      </sec>
      <sec>
        <title>Background-Enhanced Prediction</title>
        <p>Technology tends to be heavily studied in similar areas of research. Conversely, technology and its similar variants may be very popular in the same field. For example, techniques used for one type of cancer may also be relevant to other types of cancer, and various artificial neural models can all be applied in the field of medical image recognition. Our model was developed to predict future research trends based on direct relationships between technologies and fields and related technologies and fields. More specifically, we extracted the top 500 most frequent AI terms and the top 1000 most frequent biomedical fields from the 114,266 abstracts. To distinguish AI terms from biomedical terms, we adopted a simple classifier. We obtained approximately 47,000 biomedical phrases from Medical Subject Headings and approximately 700 AI algorithms from Wikipedia. We used the average cosine similarity of each keyword and all terms in the two-word sets to predict whether the keyword should belong to the biomedical or AI domain. Next, Word2vec was used to obtain embeddings from each word. After converting all words into embeddings using Word2vec, we applied agglomerative clustering [<xref ref-type="bibr" rid="ref28">28</xref>] to classify all the keywords according to their embeddings. Agglomerative clustering is a bottom-up clustering process. Initially, each input object forms its cluster. In each subsequent step, the two “closest” clusters are merged until only one remains. In our case, words with similar meanings will be grouped. Such a hierarchy is useful in many applications, and we provide the resulting tree diagram next to the corresponding heat map to best visualize the relationships between the surrounding categories.</p>
        <p><xref rid="figure1" ref-type="fig">Figure 1</xref> depicts the co-occurrence frequency of biomedical and AI keywords. For regression prediction, we utilized not only the data from the orange part (information held by the keyword) but also from the green part (information held by the word neighboring the keyword). This inclusion provides a richer context, offering models that include more relevant information to learn from. The number 4 in the orange cell indicates the number of co-occurrences of “neural network” and “cancer,” which we not only used as input to predict the number of future co-occurrences of the terms “neural network” and “cancer” but also added the number of co-occurrences in the green section, 5+3+5+3+4+7+5+4, to obtain a more comprehensive prediction using the neighboring information.</p>
        <fig id="figure1" position="float">
          <label>Figure 1</label>
          <caption>
            <p>Co-occurrence frequency table of biomedical- and artificial intelligence–related keywords. Each number represents the number of co-occurrences of a given artificial intelligence model and biomedical term. The orange part represents the information held by the keyword and the green part represents the information held by the keyword's neighbors. CNN, convolutional neural network; LSTM: long short-term memory; MLP, multilayer perceptron; NN: neural network; RNN, recurrent neural network.</p>
          </caption>
          <graphic xlink:href="ai_v2i1e45770_fig1.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
      <sec>
        <title>Regression Model</title>
        <p>The inputs and outputs of the regression model represent the co-occurrence frequency of biomedical and AI keywords in previous years and the co-occurrence frequency of future biomedical and AI keywords obtained by prediction. Due to the limited number of AI-related papers from 1970 to 2000, we used semiannual statistics for January 2000 to December 2021 in our analysis. We incorporated each semiannual data set into a training and testing prediction model. Our model uses a small window of the heatmap for the past 6 months, which was constructed using specific technology and domain pairs as features, and the model was trained on all prior-year samples to predict the current year’s heat level. We employed six different regression algorithms: support vector regression, lasso regression, ridge regression, elastic net [<xref ref-type="bibr" rid="ref29">29</xref>], orthogonal matching pursuit [<xref ref-type="bibr" rid="ref30">30</xref>], and passive aggressive regressor [<xref ref-type="bibr" rid="ref31">31</xref>], using Scikit-learn [<xref ref-type="bibr" rid="ref32">32</xref>]. We set the parameters to random_state=0 for lasso, ridge, elastic net, and passive aggressive regressor; normalize=True for lasso and ridge; and left the other parameters as default values.</p>
        <p>The data from 2016 to 2021 were used as a validation set and the data from 2002 to 2021 were used to predict trends from 2021 to 2026.</p>
      </sec>
    </sec>
    <sec sec-type="results">
      <title>Results</title>
      <sec>
        <title>Visualization</title>
        <p><xref rid="figure2" ref-type="fig">Figure 2</xref> presents a heatmap that illustrates the distribution of publications from 1970 to 2021. To improve the visualization, we limited the analysis to the top 100 frequently occurring AI terms and the top 200 frequently occurring biomedical terms. However, in subsequent experiments, we expanded the analysis to include the top 500 AI terms and the top 1000 biomedical terms. The heatmap plots the intersection of computer technology and biomedical fields, with the heat representing the logarithm of the number of papers published between 1970 and 2021 that mention both areas in the abstract. This map demonstrates that neural network–based methods are the most popular AI tools for application in the medical field.</p>
        <fig id="figure2" position="float">
          <label>Figure 2</label>
          <caption>
            <p>Heatmap of the publications related to certain artificial intelligence (AI) technologies and biomedical fields from 1970 to 2021. The horizontal axis is the keywords in the biomedical field and the vertical axis is the keywords of AI technology. A higher resolution version of this figure is available in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>.</p>
          </caption>
          <graphic xlink:href="ai_v2i1e45770_fig2.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <p>After encoding words using Word2Vec, each word becomes a corresponding embedding. To evaluate the quality of the generated embeddings, we employed t-distributed stochastic neighbor embedding (t-SNE) [<xref ref-type="bibr" rid="ref33">33</xref>], a technique for visualizing high-dimensional data by projecting it onto a 2D map. The t-SNE plots in <xref rid="figure3" ref-type="fig">Figures 3</xref> and <xref rid="figure4" ref-type="fig">4</xref> reveal that the word embeddings obtained by Word2Vec do allow words with similar meanings to be close together in the embedding space. <xref rid="figure3" ref-type="fig">Figure 3</xref> highlights the vector positions of cancer-related keywords in 2D space, while <xref rid="figure4" ref-type="fig">Figure 4</xref> shows the positions of classifier-related keywords.</p>
        <fig id="figure3" position="float">
          <label>Figure 3</label>
          <caption>
            <p>Biomedical keywords in a t-distributed stochastic neighbor embedding plot.</p>
          </caption>
          <graphic xlink:href="ai_v2i1e45770_fig3.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <fig id="figure4" position="float">
          <label>Figure 4</label>
          <caption>
            <p>Artificial intelligence keywords in the t-distributed stochastic neighbor embedding plot.</p>
          </caption>
          <graphic xlink:href="ai_v2i1e45770_fig4.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
      <sec>
        <title>Future Trend Prediction</title>
        <p><xref rid="figure5" ref-type="fig">Figure 5</xref> illustrates the average R<sup>2</sup> values of all predicted and actual results from July 2002 to December 2021, with different window sizes of 1, 3, 5, 7, and 9. From <xref rid="figure5" ref-type="fig">Figure 5</xref>, we can also see that the elastic net model provided the best results when the window size was equal to 9, whereas some other models worked best when the window size was equal to 3. </p>
        <p>Since our model relies on the previous year’s heatmap as a feature, to predict a longer time horizon, we iteratively ran our model using the predicted heatmap of cycle x to predict the heatmap of cycle x+1. As shown in <xref rid="figure6" ref-type="fig">Figure 6</xref>, although the R<sup>2</sup> value decreased during the 5-year prediction, it was still relatively high. We also provide a 100×200 demonstration to visualize the prediction results in <xref rid="figure7" ref-type="fig">Figures 7</xref>-<xref rid="figure10" ref-type="fig">10</xref>. These heatmaps, like those in <xref rid="figure2" ref-type="fig">Figure 2</xref>, are also used to show the frequency of co-occurrence between the keywords of AI technology and biomedicine. <xref rid="figure8" ref-type="fig">Figure 8</xref> depicts the original publications that were recorded between July and December 2021, while <xref rid="figure9" ref-type="fig">Figure 9</xref> represents the predicted publications for the same time period. To effectively showcase the disparity between the actual and projected outcomes, a heatmap was generated using both the original and predicted heatmaps. This comparison is visually presented in <xref rid="figure10" ref-type="fig">Figure 10</xref>, allowing for a clear and easily understandable differentiation between the two sets of data.</p>
        <fig id="figure5" position="float">
          <label>Figure 5</label>
          <caption>
            <p>Mean R-square values obtained by forecasting in half-yearly intervals from July 2002 to December 2021 under different window sizes for different methods. SVR: support vector regression.</p>
          </caption>
          <graphic xlink:href="ai_v2i1e45770_fig5.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <fig id="figure6" position="float">
          <label>Figure 6</label>
          <caption>
            <p>Line graph of the forecast results for each half year from 2002 to 2021. The model used was elastic net with a 9×9 window size, as this resulted in the best prediction (R-square value).</p>
          </caption>
          <graphic xlink:href="ai_v2i1e45770_fig6.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <fig id="figure7" position="float">
          <label>Figure 7</label>
          <caption>
            <p>The predictions are iterated in half-year increments from July 2014 to December 2021, and the data obtained from the predictions are used as the data set for the subsequent prediction models for training. The horizontal axis is time and the vertical axis is the R-square value. A higher resolution version of this figure is available in <xref ref-type="supplementary-material" rid="app2">Multimedia Appendix 2</xref>.</p>
          </caption>
          <graphic xlink:href="ai_v2i1e45770_fig7.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <fig id="figure8" position="float">
          <label>Figure 8</label>
          <caption>
            <p>Heatmap from July to December 2021 for the actual intersection of artificial intelligence (AI) technology and biomedical field applications. The horizontal axis is the keywords in the medical field and the vertical axis is the keywords in AI technology. A higher resolution version of this figure is available in <xref ref-type="supplementary-material" rid="app3">Multimedia Appendix 3</xref>.</p>
          </caption>
          <graphic xlink:href="ai_v2i1e45770_fig8.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <fig id="figure9" position="float">
          <label>Figure 9</label>
          <caption>
            <p>Predicted heat map of the intersection of artificial intelligence (AI) technology and biomedical field applications from July to December 2021. The horizontal axis is the keywords in the medical field and the vertical axis is the keywords in AI technology. A higher resolution version of this figure is available in <xref ref-type="supplementary-material" rid="app4">Multimedia Appendix 4</xref>.</p>
          </caption>
          <graphic xlink:href="ai_v2i1e45770_fig9.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <fig id="figure10" position="float">
          <label>Figure 10</label>
          <caption>
            <p>Heatmap drawn from the difference between the predicted and actual heatmaps for July to December 2021 (Figures 9 and 8, respectively) representing the intersection of artificial intelligence (AI) technology and biomedical field applications. The horizontal axis is the keywords in the medical field and the vertical axis is the keywords in AI technology. A higher resolution version of this figure is available in <xref ref-type="supplementary-material" rid="app5">Multimedia Appendix 5</xref>.</p>
          </caption>
          <graphic xlink:href="ai_v2i1e45770_fig10.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
      <sec>
        <title>Co-occurrence Trend Analysis</title>
        <p>The data obtained through statistical analysis indicated that the number of papers combining AI with biomedicine is increasing in spurts. From <xref ref-type="table" rid="table1">Table 1</xref>, we can see which combinations between AI and biomedicine are the most popular. The field of genetics shows many combinations with various AI technologies, occupying 13 of the top 20 positions. Numerous papers on this topic highlight its popularity [<xref ref-type="bibr" rid="ref34">34</xref>-<xref ref-type="bibr" rid="ref36">36</xref>]. The combination of AI and protein ranked fourth, demonstrating that protein analysis is a very suitable field for the use of machines. Cancer and tumors are currently the main challenges in biomedicine, and their combination with AI is also a popular topic at present. In these biomedical fields, machine learning is the AI technology with the highest number of applications. Although deep learning and neural networks are trendy, traditional methods such as vector automata and random forests are still the main choices in biomedical fields. Many fundamental concepts of AI are also included in this ranking, such as classification, regression, cross-validation, feature extraction, receiver operating characteristic, and others. Overall, this analysis shows that AI has become a key technology in the biomedical field and requires the proficiency of biomedical scientists.</p>
        <table-wrap position="float" id="table1">
          <label>Table 1</label>
          <caption>
            <p>The top 20 combinations of artificial intelligence (AI) technologies and biomedical fields that have appeared in the literature in the last 5 years.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="70"/>
            <col width="630"/>
            <col width="300"/>
            <thead>
              <tr valign="top">
                <td>Rank</td>
                <td>AI technology, biomedical field</td>
                <td>Proportion of publications, %</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>1</td>
                <td>machine learning, gene</td>
                <td>1.650</td>
              </tr>
              <tr valign="top">
                <td>2</td>
                <td>classification, gene</td>
                <td>1.038</td>
              </tr>
              <tr valign="top">
                <td>3</td>
                <td>neural network, gene</td>
                <td>0.634</td>
              </tr>
              <tr valign="top">
                <td>4</td>
                <td>deep learning, gene</td>
                <td>0.453</td>
              </tr>
              <tr valign="top">
                <td>5</td>
                <td>support vector machine, gene</td>
                <td>0.447</td>
              </tr>
              <tr valign="top">
                <td>6</td>
                <td>machine learning, protein</td>
                <td>0.404</td>
              </tr>
              <tr valign="top">
                <td>7</td>
                <td>regression, gene</td>
                <td>0.402</td>
              </tr>
              <tr valign="top">
                <td>8</td>
                <td>learning algorithm, gene</td>
                <td>0.385</td>
              </tr>
              <tr valign="top">
                <td>9</td>
                <td>machine learning, cancer</td>
                <td>0.380</td>
              </tr>
              <tr valign="top">
                <td>10</td>
                <td>classification, cancer</td>
                <td>0.351</td>
              </tr>
              <tr valign="top">
                <td>11</td>
                <td>random forest, gene</td>
                <td>0.331</td>
              </tr>
              <tr valign="top">
                <td>12</td>
                <td>artificial intelligence, gene</td>
                <td>0.249</td>
              </tr>
              <tr valign="top">
                <td>13</td>
                <td>convolution neural network, gene</td>
                <td>0.241</td>
              </tr>
              <tr valign="top">
                <td>14</td>
                <td>cross-validation, gene</td>
                <td>0.219</td>
              </tr>
              <tr valign="top">
                <td>15</td>
                <td>feature selection, gene</td>
                <td>0.191</td>
              </tr>
              <tr valign="top">
                <td>16</td>
                <td>neural network, cancer</td>
                <td>0.176</td>
              </tr>
              <tr valign="top">
                <td>17</td>
                <td>classification, tumor</td>
                <td>0.172</td>
              </tr>
              <tr valign="top">
                <td>18</td>
                <td>supervised learning, gene</td>
                <td>0.171</td>
              </tr>
              <tr valign="top">
                <td>19</td>
                <td>machine learning, tumor</td>
                <td>0.171</td>
              </tr>
              <tr valign="top">
                <td>20</td>
                <td>receiver operating characteristic, gene</td>
                <td>0.169</td>
              </tr>
            </tbody>
          </table>
        </table-wrap>
        <p>Since many combinations between AI and biomedicine have a very small contribution or are nonexistent, some are not meaningful; therefore, we set a reasonable threshold to filter such combinations, avoiding the situation where the original minimal combination grows by a considerable percentage with little growth so that the table showing the trend of changes is more meaningful. From <xref ref-type="table" rid="table2">Table 2</xref>, we can see the very rapid growth of cases combining AI and biomedicine in the last 5 years. This is because genes, proteins, oncology, and many other fields are growing rapidly, and core medical testing technology such as magnetic resonance imaging is compatible with AI.</p>
        <p>We used the best model from our proposed methodology to forecast the trends in AI technology and biomedicine over the next 5 years. The prediction results for the contributions of each combination and their growth are shown in <xref ref-type="table" rid="table3">Table 3</xref> and <xref ref-type="table" rid="table4">Table 4</xref>, respectively. The regression results were rounded for brevity of presentation in the tables. We can use these predicted results to provide an outlook on the future development of AI in biomedicine. From the point of view of AI technologies, standard techniques such as deep learning, machine learning, and neural networks still dominate. Traditional machine learning methods such as random forest and support vector machine are outside the top 20 prediction results. Deep learning will gradually become the mainstream AI technology combined with biomedicine [<xref ref-type="bibr" rid="ref37">37</xref>]. From a biomedical perspective, genetics will continue to dominate. At the same time, studies focusing on proteins and tumors will leave the top 20 and be replaced by early diagnostics, tomography, and other detection technologies. These are certain areas that are well suited to incorporate AI technology.</p>
        <table-wrap position="float" id="table2">
          <label>Table 2</label>
          <caption>
            <p>The 20 most rapidly growing combinations of artificial intelligence (AI) technologies and biomedical fields in the last 5 years.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="70"/>
            <col width="630"/>
            <col width="300"/>
            <thead>
              <tr valign="top">
                <td>Rank</td>
                <td>AI technology, biomedical field</td>
                <td>Growth, %</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>1</td>
                <td>electronic, health records</td>
                <td>1054.545</td>
              </tr>
              <tr valign="top">
                <td>2</td>
                <td>electronic health records, electronic health</td>
                <td>1054.545</td>
              </tr>
              <tr valign="top">
                <td>3</td>
                <td>machine learning, electronic health record</td>
                <td>1033.333</td>
              </tr>
              <tr valign="top">
                <td>4</td>
                <td>machine learning, health care</td>
                <td>820.000</td>
              </tr>
              <tr valign="top">
                <td>5</td>
                <td>machine learning, risk factor</td>
                <td>816.667</td>
              </tr>
              <tr valign="top">
                <td>6</td>
                <td>machine learning, public health</td>
                <td>735.000</td>
              </tr>
              <tr valign="top">
                <td>7</td>
                <td>neural network, gene</td>
                <td>700.483</td>
              </tr>
              <tr valign="top">
                <td>8</td>
                <td>neural network, cancer</td>
                <td>647.059</td>
              </tr>
              <tr valign="top">
                <td>9</td>
                <td>machine learning, tumor</td>
                <td>619.697</td>
              </tr>
              <tr valign="top">
                <td>10</td>
                <td>image analysis, gene</td>
                <td>613.333</td>
              </tr>
              <tr valign="top">
                <td>11</td>
                <td>machine learning, clinical trial</td>
                <td>572.414</td>
              </tr>
              <tr valign="top">
                <td>12</td>
                <td>machine learning, clinical practice</td>
                <td>566.667</td>
              </tr>
              <tr valign="top">
                <td>13</td>
                <td>decision making, gene</td>
                <td>547.619</td>
              </tr>
              <tr valign="top">
                <td>14</td>
                <td>artificial intelligence, gene</td>
                <td>511.111</td>
              </tr>
              <tr valign="top">
                <td>15</td>
                <td>random forest, cancer</td>
                <td>493.617</td>
              </tr>
              <tr valign="top">
                <td>16</td>
                <td>machine learning, clinical data</td>
                <td>487.179</td>
              </tr>
              <tr valign="top">
                <td>17</td>
                <td>electronic medical record, medical records</td>
                <td>480.000</td>
              </tr>
              <tr valign="top">
                <td>18</td>
                <td>next generation sequencing, gene</td>
                <td>467.647</td>
              </tr>
              <tr valign="top">
                <td>19</td>
                <td>random forest, tumor</td>
                <td>466.667</td>
              </tr>
              <tr valign="top">
                <td>20</td>
                <td>machine learning, magnetic resonance</td>
                <td>456.579</td>
              </tr>
            </tbody>
          </table>
        </table-wrap>
        <table-wrap position="float" id="table3">
          <label>Table 3</label>
          <caption>
            <p>The top 20 combinations of artificial intelligence (AI) technologies and biomedical fields that will emerge in the next 5 years.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="70"/>
            <col width="630"/>
            <col width="300"/>
            <thead>
              <tr valign="top">
                <td>Rank</td>
                <td>AI technology, biomedical field</td>
                <td>Predicted proportion of publications, %</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>1</td>
                <td>machine learning, gene</td>
                <td>2.331</td>
              </tr>
              <tr valign="top">
                <td>2</td>
                <td>artificial intelligence, early diagnosis</td>
                <td>2.289</td>
              </tr>
              <tr valign="top">
                <td>3</td>
                <td>artificial intelligence, early detection</td>
                <td>1.901</td>
              </tr>
              <tr valign="top">
                <td>4</td>
                <td>artificial intelligence, gene</td>
                <td>1.487</td>
              </tr>
              <tr valign="top">
                <td>5</td>
                <td>neural network, gene</td>
                <td>1.392</td>
              </tr>
              <tr valign="top">
                <td>6</td>
                <td>deep learning, computed tomography</td>
                <td>1.288</td>
              </tr>
              <tr valign="top">
                <td>7</td>
                <td>artificial intelligence, systematic reviews</td>
                <td>1.239</td>
              </tr>
              <tr valign="top">
                <td>8</td>
                <td>classification, gene</td>
                <td>1.197</td>
              </tr>
              <tr valign="top">
                <td>9</td>
                <td>supervised learning, gene</td>
                <td>1.188</td>
              </tr>
              <tr valign="top">
                <td>10</td>
                <td>generative adversarial network, gene</td>
                <td>1.040</td>
              </tr>
              <tr valign="top">
                <td>11</td>
                <td>artificial intelligence, personalized treatment</td>
                <td>0.881</td>
              </tr>
              <tr valign="top">
                <td>12</td>
                <td>machine learning, risk factors</td>
                <td>0.659</td>
              </tr>
              <tr valign="top">
                <td>13</td>
                <td>deep learning, gene</td>
                <td>0.633</td>
              </tr>
              <tr valign="top">
                <td>14</td>
                <td>artificial intelligence, systematic review</td>
                <td>0.617</td>
              </tr>
              <tr valign="top">
                <td>15</td>
                <td>convolution neural network, gene</td>
                <td>0.604</td>
              </tr>
              <tr valign="top">
                <td>16</td>
                <td>learning algorithm, gene</td>
                <td>0.593</td>
              </tr>
              <tr valign="top">
                <td>17</td>
                <td>receiver operating characteristic, computed tomography scans</td>
                <td>0.581</td>
              </tr>
              <tr valign="top">
                <td>18</td>
                <td>machine learning, medical records</td>
                <td>0.578</td>
              </tr>
              <tr valign="top">
                <td>19</td>
                <td>machine learning, blood pressure</td>
                <td>0.569</td>
              </tr>
              <tr valign="top">
                <td>20</td>
                <td>artificial intelligence, imaging modalities</td>
                <td>0.554</td>
              </tr>
            </tbody>
          </table>
        </table-wrap>
        <table-wrap position="float" id="table4">
          <label>Table 4</label>
          <caption>
            <p>The top 20 rapidly growing combinations of artificial intelligence (AI) technology and biomedical fields in the next 5 years.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="70"/>
            <col width="630"/>
            <col width="300"/>
            <thead>
              <tr valign="top">
                <td>Rank</td>
                <td>AI technology, biomedical field</td>
                <td>Predicted growth, %</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>1</td>
                <td>artificial intelligence, gene</td>
                <td>2253.521</td>
              </tr>
              <tr valign="top">
                <td>2</td>
                <td>machine learning, risk factor</td>
                <td>2184.491</td>
              </tr>
              <tr valign="top">
                <td>3</td>
                <td>cross-validation, gene</td>
                <td>2164.150</td>
              </tr>
              <tr valign="top">
                <td>4</td>
                <td>receiver operating characteristic, gene</td>
                <td>1504.581</td>
              </tr>
              <tr valign="top">
                <td>5</td>
                <td>learning algorithm, gene</td>
                <td>1421.751</td>
              </tr>
              <tr valign="top">
                <td>6</td>
                <td>neural network, gene</td>
                <td>1340.880</td>
              </tr>
              <tr valign="top">
                <td>7</td>
                <td>convolution neural network, gene</td>
                <td>1296.067</td>
              </tr>
              <tr valign="top">
                <td>8</td>
                <td>classification, gene</td>
                <td>1280.985</td>
              </tr>
              <tr valign="top">
                <td>9</td>
                <td>machine learning, gene</td>
                <td>1261.342</td>
              </tr>
              <tr valign="top">
                <td>10</td>
                <td>classification, cancer</td>
                <td>888.106</td>
              </tr>
              <tr valign="top">
                <td>11</td>
                <td>support vector machine, gene</td>
                <td>791.807</td>
              </tr>
              <tr valign="top">
                <td>12</td>
                <td>neural network, cancer</td>
                <td>665.430</td>
              </tr>
              <tr valign="top">
                <td>13</td>
                <td>artificial intelligence, cancer</td>
                <td>621.627</td>
              </tr>
              <tr valign="top">
                <td>14</td>
                <td>deep learning, gene</td>
                <td>502.318</td>
              </tr>
              <tr valign="top">
                <td>15</td>
                <td>classification, tumor</td>
                <td>415.298</td>
              </tr>
              <tr valign="top">
                <td>16</td>
                <td>regression, gene</td>
                <td>377.864</td>
              </tr>
              <tr valign="top">
                <td>17</td>
                <td>machine learning, protein</td>
                <td>333.778</td>
              </tr>
              <tr valign="top">
                <td>18</td>
                <td>random forest, gene</td>
                <td>322.787</td>
              </tr>
              <tr valign="top">
                <td>19</td>
                <td>deep learning, cancer</td>
                <td>200.080</td>
              </tr>
              <tr valign="top">
                <td>20</td>
                <td>natural language processing, natural language</td>
                <td>192.518</td>
              </tr>
            </tbody>
          </table>
        </table-wrap>
      </sec>
    </sec>
    <sec sec-type="discussion">
      <title>Discussion</title>
      <sec>
        <title>Principal Findings</title>
        <sec>
          <title>AI Technology Trends in Biomedicine</title>
          <p>Our findings confirm that standard AI techniques, including deep learning, machine learning, and neural networks, continue to be the primary driving forces behind the integration of AI into biomedicine. However, it is noteworthy that generative adversarial networks (GANs) [<xref ref-type="bibr" rid="ref38">38</xref>] are gaining prominence, particularly in the genetics field. GANs hold immense potential for applications in medical imaging and drug discovery owing to their ability to generate synthetic images across various modalities.</p>
        </sec>
        <sec>
          <title>Evolution of Biomedical Research</title>
          <p>The data also highlight the shifting landscape of biomedical research. While genetics remains dominant, areas such as proteins and tumors are gradually giving way to early diagnostics, tomography, and other detection technologies. These developments align with the suitability of these fields for AI integration, resulting in promising advancements in health care analysis and diagnostics.</p>
        </sec>
        <sec>
          <title>Impact of AI on Health Care</title>
          <p>As suggested by previous research [<xref ref-type="bibr" rid="ref24">24</xref>], the future of AI in health care is promising. AI has the potential to enhance the accuracy of cancer diagnosis and prognosis beyond that of average statistical experts [<xref ref-type="bibr" rid="ref39">39</xref>,<xref ref-type="bibr" rid="ref40">40</xref>]. Furthermore, as AI technology continues to advance, it will enable the resolution of more complex and specialized health care problems, further transforming the biomedical landscape.</p>
        </sec>
      </sec>
      <sec>
        <title>Future Work</title>
        <p>By utilizing keywords to filter medical papers that have applied AI techniques, we identified key connections and trends among them. The approach of using keywords aggregated based on text similarity performed well in the regression model. This approach is intuitive and leads to improved co-word analysis for trend prediction.</p>
        <p>Fundamentally, incorporating peripheral information led to higher regression accuracy and more accurate predictions of future trends. Additionally, this approach also takes into account internal relationships within a class compared to previous methods. However, this also raises the question of how to best measure the degree of keyword association.</p>
        <p>We made some simple assumptions that words with similar meanings would complement the information of the others. Specifically, considering only their own meanings tends to make the predictions one-sided, while having more reference information naturally makes the predictions more robust. This can be seen as a type of data augmentation. There are still many directions to explore regarding this approach. In future research, it may be possible to use different text similarity methods such as convolutional neural network, bidirectional encoder representations from transformers, and various regression models, where the reliability of text similarity determines whether the information obtained from the surrounding context is valid. Additionally, different time spans for the prediction can be studied. Although this study focused on AI techniques in the biomedical field, the applicability of the proposed approach extends to any study involving co-word analysis.</p>
      </sec>
      <sec>
        <title>Limitations</title>
        <p>While our study provides valuable insights into the trends of AI technologies in the biomedical domain based on a comprehensive data set from PubMed, there are several limitations to consider. First, there is a limitation of the data source, since our study solely relies on PubMed as the primary source of articles, which might introduce a selection bias. There are numerous other databases and grey literature sources that were not considered, and their inclusion might have offered a more comprehensive view. Second, our study lacks external validity. Our findings, although significant in the context of our data set, require validation with real-world applications and events to check their external validity.</p>
      </sec>
      <sec>
        <title>Conclusions</title>
        <p>In this study, we aimed to explore the analysis and prediction of trends at the intersection of biomedical and AI research. To accomplish this, we collected a large corpus of articles from PubMed on the intersection of AI and biomedicine. Initially, we attempted to use regression on the extracted keywords alone. However, we found that this approach was lacking in information. Therefore, we proposed a method called background-enhanced prediction to expand the knowledge utilized by the regression algorithm by incorporating both the keywords and their surrounding context. This data construction method improved the performance of our forecasting models. Our findings were validated through comparisons with current trends. In particular, the integration of electronic medical record big data with AI, laboratory data, clinical trials, and imaging diagnostic tools has emerged as a prominent approach.</p>
      </sec>
    </sec>
  </body>
  <back>
    <app-group>
      <supplementary-material id="app1">
        <label>Multimedia Appendix 1</label>
        <p>Higher resolution version of <xref rid="figure2" ref-type="fig">Figure 2</xref>.</p>
        <media xlink:href="ai_v2i1e45770_app1.pdf" xlink:title="PDF File  (Adobe PDF File), 427 KB"/>
      </supplementary-material>
      <supplementary-material id="app2">
        <label>Multimedia Appendix 2</label>
        <p>Higher resolution version of <xref rid="figure7" ref-type="fig">Figure 7</xref>.</p>
        <media xlink:href="ai_v2i1e45770_app2.pdf" xlink:title="PDF File  (Adobe PDF File), 209 KB"/>
      </supplementary-material>
      <supplementary-material id="app3">
        <label>Multimedia Appendix 3</label>
        <p>Higher resolution version of <xref rid="figure8" ref-type="fig">Figure 8</xref>.</p>
        <media xlink:href="ai_v2i1e45770_app3.pdf" xlink:title="PDF File  (Adobe PDF File), 346 KB"/>
      </supplementary-material>
      <supplementary-material id="app4">
        <label>Multimedia Appendix 4</label>
        <p>Higher resolution version of <xref rid="figure9" ref-type="fig">Figure 9</xref>.</p>
        <media xlink:href="ai_v2i1e45770_app4.pdf" xlink:title="PDF File  (Adobe PDF File), 477 KB"/>
      </supplementary-material>
      <supplementary-material id="app5">
        <label>Multimedia Appendix 5</label>
        <p>Higher resolution version of <xref rid="figure10" ref-type="fig">Figure 10</xref>.</p>
        <media xlink:href="ai_v2i1e45770_app5.pdf" xlink:title="PDF File  (Adobe PDF File), 421 KB"/>
      </supplementary-material>
    </app-group>
    <glossary>
      <title>Abbreviations</title>
      <def-list>
        <def-item>
          <term id="abb1">AI</term>
          <def>
            <p>artificial intelligence</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb2">GAN</term>
          <def>
            <p>generative adversarial network</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb3">t-SNE</term>
          <def>
            <p>t-distributed stochastic neighbor embedding</p>
          </def>
        </def-item>
      </def-list>
    </glossary>
    <ack>
      <p>We are sincerely grateful to our advisor, Professor Soroush Vosoughi, for the invaluable guidance and support provided during the development of this paper. We would also like to thank our colleagues and the staff at Minds, Machines, and Society Lab for their helpful insights and assistance.</p>
    </ack>
    <notes>
      <sec>
        <title>Data Availability</title>
        <p>The code is available at GitHub [<xref ref-type="bibr" rid="ref41">41</xref>].</p>
      </sec>
    </notes>
    <fn-group>
      <fn fn-type="conflict">
        <p>None declared.</p>
      </fn>
    </fn-group>
    <ref-list>
      <ref id="ref1">
        <label>1</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Yu</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Beam</surname>
              <given-names>AL</given-names>
            </name>
            <name name-style="western">
              <surname>Kohane</surname>
              <given-names>IS</given-names>
            </name>
          </person-group>
          <article-title>Artificial intelligence in healthcare</article-title>
          <source>Nat Biomed Eng</source>
          <year>2018</year>
          <month>10</month>
          <volume>2</volume>
          <issue>10</issue>
          <fpage>719</fpage>
          <lpage>731</lpage>
          <pub-id pub-id-type="doi">10.1038/s41551-018-0305-z</pub-id>
          <pub-id pub-id-type="medline">31015651</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41551-018-0305-z</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref2">
        <label>2</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zemouri</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Zerhouni</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Racoceanu</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>Deep learning in the biomedical applications: recent and future status</article-title>
          <source>Appl Sci</source>
          <year>2019</year>
          <month>04</month>
          <day>12</day>
          <volume>9</volume>
          <issue>8</issue>
          <fpage>1526</fpage>
          <pub-id pub-id-type="doi">10.3390/app9081526</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref3">
        <label>3</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wei</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Zhou</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Luo</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Ren</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Huang</surname>
              <given-names>L</given-names>
            </name>
          </person-group>
          <source>IEEE Trans Biomed Circuits Syst</source>
          <year>2020</year>
          <month>04</month>
          <volume>14</volume>
          <issue>2</issue>
          <fpage>145</fpage>
          <lpage>163</lpage>
          <pub-id pub-id-type="doi">10.1109/TBCAS.2020.2974154</pub-id>
          <pub-id pub-id-type="medline">32078560</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref4">
        <label>4</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Bali</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Garg</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Bali</surname>
              <given-names>RT</given-names>
            </name>
          </person-group>
          <article-title>Artificial intelligence (AI) in healthcare and biomedical research: Why a strong computational/AI bioethics framework is required?</article-title>
          <source>Indian J Ophthalmol</source>
          <year>2019</year>
          <month>01</month>
          <volume>67</volume>
          <issue>1</issue>
          <fpage>3</fpage>
          <lpage>6</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/30574881"/>
          </comment>
          <pub-id pub-id-type="doi">10.4103/ijo.IJO_1292_18</pub-id>
          <pub-id pub-id-type="medline">30574881</pub-id>
          <pub-id pub-id-type="pii">IndianJOphthalmol_2019_67_1_3_248118</pub-id>
          <pub-id pub-id-type="pmcid">PMC6324122</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref5">
        <label>5</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Donthu</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Kumar</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Mukherjee</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Pandey</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Lim</surname>
              <given-names>WM</given-names>
            </name>
          </person-group>
          <article-title>How to conduct a bibliometric analysis: an overview and guidelines</article-title>
          <source>J Bus Res</source>
          <year>2021</year>
          <month>09</month>
          <volume>133</volume>
          <fpage>285</fpage>
          <lpage>296</lpage>
          <pub-id pub-id-type="doi">10.1016/j.jbusres.2021.04.070</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref6">
        <label>6</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>He</surname>
              <given-names>Q</given-names>
            </name>
          </person-group>
          <article-title>Knowledge discovery through co-word analysis</article-title>
          <source>Library Trends</source>
          <year>1999</year>
          <volume>48</volume>
          <issue>1</issue>
          <fpage>133</fpage>
          <lpage>159</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.ideals.illinois.edu/items/8226"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref7">
        <label>7</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Dong</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Measurement of text similarity: a survey</article-title>
          <source>Information</source>
          <year>2020</year>
          <month>08</month>
          <day>31</day>
          <volume>11</volume>
          <issue>9</issue>
          <fpage>421</fpage>
          <pub-id pub-id-type="doi">10.3390/info11090421</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref8">
        <label>8</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Deza</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Deza</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <source>Encyclopedia of distances</source>
          <year>2009</year>
          <publisher-loc>Berlin, Heidelberg</publisher-loc>
          <publisher-name>Springer</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref9">
        <label>9</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Manning</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Schütze</surname>
              <given-names>H</given-names>
            </name>
          </person-group>
          <source>Foundations of statistical natural language processing</source>
          <year>1999</year>
          <publisher-loc>Cambridge, MA</publisher-loc>
          <publisher-name>MIT Press</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref10">
        <label>10</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kullback</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Leibler</surname>
              <given-names>RA</given-names>
            </name>
          </person-group>
          <article-title>On information and sufficiency</article-title>
          <source>Ann Math Statist</source>
          <year>1951</year>
          <month>03</month>
          <volume>22</volume>
          <issue>1</issue>
          <fpage>79</fpage>
          <lpage>86</lpage>
          <pub-id pub-id-type="doi">10.1214/aoms/1177729694</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref11">
        <label>11</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Weng</surname>
              <given-names>L</given-names>
            </name>
          </person-group>
          <article-title>From GAN to WGAN</article-title>
          <source>arXiv</source>
          <year>2019</year>
          <access-date>2023-11-26</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://arxiv.org/abs/1904.08994">http://arxiv.org/abs/1904.08994</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref12">
        <label>12</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Irving</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Fraser</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <person-group person-group-type="editor">
            <name name-style="western">
              <surname>Apostolico</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Crochemore</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Galil</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Manber</surname>
              <given-names>U</given-names>
            </name>
          </person-group>
          <article-title>Two algorithms for the longest common subsequence of three (or more) strings</article-title>
          <source>Combinatorial Pattern Matching Combinatorial Pattern Matching. CPM 1992. Lecture Notes in Computer Science, vol 644</source>
          <year>1992</year>
          <publisher-loc>Berlin, Heidelberg</publisher-loc>
          <publisher-name>Springer</publisher-name>
          <fpage>214</fpage>
          <lpage>229</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref13">
        <label>13</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Damerau</surname>
              <given-names>FJ</given-names>
            </name>
          </person-group>
          <article-title>A technique for computer detection and correction of spelling errors</article-title>
          <source>Commun ACM</source>
          <year>1964</year>
          <month>03</month>
          <volume>7</volume>
          <issue>3</issue>
          <fpage>171</fpage>
          <lpage>176</lpage>
          <pub-id pub-id-type="doi">10.1145/363958.363994</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref14">
        <label>14</label>
        <nlm-citation citation-type="web">
          <source>ERIC: String Comparator Metrics and Enhanced Decision Rules in the Fellegi-Sunter Model of Record Linkage</source>
          <access-date>2023-11-26</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://eric.ed.gov/?id=ED325505">https://eric.ed.gov/?id=ED325505</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref15">
        <label>15</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Dice</surname>
              <given-names>LR</given-names>
            </name>
          </person-group>
          <article-title>Measures of the amount of ecologic association between species</article-title>
          <source>Ecology</source>
          <year>1945</year>
          <month>07</month>
          <volume>26</volume>
          <issue>3</issue>
          <fpage>297</fpage>
          <lpage>302</lpage>
          <pub-id pub-id-type="doi">10.2307/1932409</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref16">
        <label>16</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Jaccard</surname>
              <given-names>P</given-names>
            </name>
          </person-group>
          <article-title>The distribution of the flora in the alpine zone</article-title>
          <source>New Phytologist</source>
          <year>1912</year>
          <volume>11</volume>
          <issue>2</issue>
          <fpage>37</fpage>
          <lpage>50</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.jstor.org/stable/2427226"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref17">
        <label>17</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Salton</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Buckley</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>Term weighting approaches in automatic text retrieval</article-title>
          <source>Inf Process Manag</source>
          <year>1988</year>
          <volume>24</volume>
          <issue>5</issue>
          <fpage>513</fpage>
          <lpage>523</lpage>
          <pub-id pub-id-type="doi">10.1016/0306-4573(88)90021-0</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref18">
        <label>18</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Robertson</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Walker</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Some simple effective approximations to the 2-Poisson model for probabilistic weighted retrieval</article-title>
          <year>1994</year>
          <conf-name>Seventeenth Annual International ACM-SIGIR Conference on Research and Development in Information Retrieval (SIGR '94)</conf-name>
          <conf-date>July 3-6, 1994</conf-date>
          <conf-loc>Dublin, Ireland</conf-loc>
          <pub-id pub-id-type="doi">10.1007/978-1-4471-2099-5_24</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref19">
        <label>19</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Rong</surname>
              <given-names>X</given-names>
            </name>
          </person-group>
          <article-title>word2vec parameter learning explained</article-title>
          <source>arXiv</source>
          <year>2016</year>
          <access-date>2023-11-26</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://arxiv.org/abs/1411.2738">http://arxiv.org/abs/1411.2738</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref20">
        <label>20</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Deerwester</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Dumais</surname>
              <given-names>ST</given-names>
            </name>
            <name name-style="western">
              <surname>Furnas</surname>
              <given-names>GW</given-names>
            </name>
            <name name-style="western">
              <surname>Landauer</surname>
              <given-names>TK</given-names>
            </name>
            <name name-style="western">
              <surname>Harshman</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Indexing by latent semantic analysis</article-title>
          <source>J Am Soc Inf Sci</source>
          <year>1990</year>
          <month>09</month>
          <volume>41</volume>
          <issue>6</issue>
          <fpage>391</fpage>
          <lpage>407</lpage>
          <pub-id pub-id-type="doi">10.1002/(sici)1097-4571(199009)41:6&#60;391::aid-asi1&#62;3.0.co;2-9</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref21">
        <label>21</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Shen</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>He</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Gao</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Deng</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Mesnil</surname>
              <given-names>G</given-names>
            </name>
          </person-group>
          <article-title>A latent semantic model with convolutional-pooling structure for information retrieval</article-title>
          <year>2014</year>
          <conf-name>23rd ACM International Conference on Conference on Information Knowledge Management</conf-name>
          <conf-date>November 3-7, 2014</conf-date>
          <conf-loc>Shanghai, China</conf-loc>
          <pub-id pub-id-type="doi">10.1145/2661829.2661935</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref22">
        <label>22</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Jia</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Xiang</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>A review: knowledge reasoning over knowledge graph</article-title>
          <source>Expert Syst Appl</source>
          <year>2020</year>
          <month>03</month>
          <volume>141</volume>
          <fpage>112948</fpage>
          <pub-id pub-id-type="doi">10.1016/j.eswa.2019.112948</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref23">
        <label>23</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zhou</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Cui</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Hu</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Graph neural networks: a review of methods and applications</article-title>
          <source>AI Open</source>
          <year>2020</year>
          <volume>1</volume>
          <fpage>57</fpage>
          <lpage>81</lpage>
          <pub-id pub-id-type="doi">10.1016/j.aiopen.2021.01.001</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref24">
        <label>24</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Preininger</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>AI in health: state of the art, challenges, and future directions</article-title>
          <source>Yearb Med Inform</source>
          <year>2019</year>
          <month>08</month>
          <volume>28</volume>
          <issue>1</issue>
          <fpage>16</fpage>
          <lpage>26</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://www.thieme-connect.com/DOI/DOI?10.1055/s-0039-1677908"/>
          </comment>
          <pub-id pub-id-type="doi">10.1055/s-0039-1677908</pub-id>
          <pub-id pub-id-type="medline">31419814</pub-id>
          <pub-id pub-id-type="pmcid">PMC6697503</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref25">
        <label>25</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Johnson</surname>
              <given-names>KB</given-names>
            </name>
            <name name-style="western">
              <surname>Wei</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Weeraratne</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Frisse</surname>
              <given-names>ME</given-names>
            </name>
            <name name-style="western">
              <surname>Misulis</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Rhee</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Zhao</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Snowdon</surname>
              <given-names>JL</given-names>
            </name>
          </person-group>
          <article-title>Precision medicine, AI, and the future of personalized health care</article-title>
          <source>Clin Transl Sci</source>
          <year>2021</year>
          <month>01</month>
          <volume>14</volume>
          <issue>1</issue>
          <fpage>86</fpage>
          <lpage>93</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/32961010"/>
          </comment>
          <pub-id pub-id-type="doi">10.1111/cts.12884</pub-id>
          <pub-id pub-id-type="medline">32961010</pub-id>
          <pub-id pub-id-type="pmcid">PMC7877825</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref26">
        <label>26</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ahmad</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Rahim</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Zubair</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Abdul-Ghafar</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Artificial intelligence (AI) in medicine, current applications and future role with special emphasis on its potential and promise in pathology: present and future impact, obstacles including costs and acceptance among pathologists, practical and philosophical considerations. A comprehensive review</article-title>
          <source>Diagn Pathol</source>
          <year>2021</year>
          <month>03</month>
          <day>17</day>
          <volume>16</volume>
          <issue>1</issue>
          <fpage>24</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://diagnosticpathology.biomedcentral.com/articles/10.1186/s13000-021-01085-4"/>
          </comment>
          <pub-id pub-id-type="doi">10.1186/s13000-021-01085-4</pub-id>
          <pub-id pub-id-type="medline">33731170</pub-id>
          <pub-id pub-id-type="pii">10.1186/s13000-021-01085-4</pub-id>
          <pub-id pub-id-type="pmcid">PMC7971952</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref27">
        <label>27</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hu</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Luo</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Qi</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Mao</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Fu</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Zheng</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Guo</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>Q</given-names>
            </name>
          </person-group>
          <article-title>Understanding the topic evolution of scientific literatures like an evolving city: using Google Word2Vec model and spatial autocorrelation analysis</article-title>
          <source>Inf Processing Manag</source>
          <year>2019</year>
          <month>07</month>
          <volume>56</volume>
          <issue>4</issue>
          <fpage>1185</fpage>
          <lpage>1203</lpage>
          <pub-id pub-id-type="doi">10.1016/j.ipm.2019.02.014</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref28">
        <label>28</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ackermann</surname>
              <given-names>MR</given-names>
            </name>
            <name name-style="western">
              <surname>Blömer</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Kuntze</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Sohler</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>Analysis of agglomerative clustering</article-title>
          <source>Algorithmica</source>
          <year>2012</year>
          <month>12</month>
          <day>12</day>
          <volume>69</volume>
          <issue>1</issue>
          <fpage>184</fpage>
          <lpage>215</lpage>
          <pub-id pub-id-type="doi">10.1007/s00453-012-9717-4</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref29">
        <label>29</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Xiao</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Cui</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Qiao</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Zheng</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Early diagnosis model of Alzheimer’s Disease based on sparse logistic regression</article-title>
          <source>Multimed Tools Appl</source>
          <year>2020</year>
          <month>09</month>
          <day>25</day>
          <volume>80</volume>
          <issue>3</issue>
          <fpage>3969</fpage>
          <lpage>3980</lpage>
          <pub-id pub-id-type="doi">10.1007/s11042-020-09738-0</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref30">
        <label>30</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zarei</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Asl</surname>
              <given-names>BM</given-names>
            </name>
          </person-group>
          <article-title>Automatic seizure detection using orthogonal matching pursuit, discrete wavelet transform, and entropy based features of EEG signals</article-title>
          <source>Comput Biol Med</source>
          <year>2021</year>
          <month>04</month>
          <volume>131</volume>
          <fpage>104250</fpage>
          <pub-id pub-id-type="doi">10.1016/j.compbiomed.2021.104250</pub-id>
          <pub-id pub-id-type="medline">33578071</pub-id>
          <pub-id pub-id-type="pii">S0010-4825(21)00044-5</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref31">
        <label>31</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Malki</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Atlam</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Hassanien</surname>
              <given-names>AE</given-names>
            </name>
            <name name-style="western">
              <surname>Dagnew</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Elhosseini</surname>
              <given-names>MA</given-names>
            </name>
            <name name-style="western">
              <surname>Gad</surname>
              <given-names>I</given-names>
            </name>
          </person-group>
          <article-title>Association between weather data and COVID-19 pandemic predicting mortality rate: machine learning approaches</article-title>
          <source>Chaos Solitons Fractals</source>
          <year>2020</year>
          <month>09</month>
          <volume>138</volume>
          <fpage>110137</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/32834583"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.chaos.2020.110137</pub-id>
          <pub-id pub-id-type="medline">32834583</pub-id>
          <pub-id pub-id-type="pii">S0960-0779(20)30533-6</pub-id>
          <pub-id pub-id-type="pmcid">PMC7367008</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref32">
        <label>32</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Pedregosa</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Varoquaux</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Gramfort</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Michel</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Thirion</surname>
              <given-names>B</given-names>
            </name>
          </person-group>
          <article-title>Scikit-learn: machine learning in Python</article-title>
          <source>J Machine Learn Res</source>
          <year>2011</year>
          <volume>12</volume>
          <fpage>2825</fpage>
          <lpage>2830</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.jmlr.org/papers/volume12/pedregosa11a/pedregosa11a.pdf"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref33">
        <label>33</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>van der Maaten</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Hinton</surname>
              <given-names>G</given-names>
            </name>
          </person-group>
          <article-title>Visualizing data using t-SNE</article-title>
          <source>J Mach Learn Res</source>
          <year>2008</year>
          <volume>9</volume>
          <fpage>2579</fpage>
          <lpage>2605</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.jmlr.org/papers/volume9/vandermaaten08a/vandermaaten08a.pdf?fbcl"/>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref34">
        <label>34</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Abdulqader</surname>
              <given-names>DM</given-names>
            </name>
            <name name-style="western">
              <surname>Abdulazeez</surname>
              <given-names>AM</given-names>
            </name>
            <name name-style="western">
              <surname>Zeebaree</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>Machine learning supervised algorithms of gene selection: a review</article-title>
          <source>Technology Reports of Kansai University</source>
          <year>2020</year>
          <volume>62</volume>
          <issue>3</issue>
          <fpage>233</fpage>
          <lpage>244</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.researchgate.net/profile/Adnan-Mohsin-Abdulazeez/publication/341119469_Machine_Learning_Supervised_Algorithms_of_Gene_Selection_A_Review/links/60b372e692851cd0d985f0cf/Machine-Learning-Supervised-Algorithms-of-Gene-Selection-A-Review.pdf"/>
          </comment>
          <pub-id pub-id-type="doi">10.1109/icset53708.2021.9612526</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref35">
        <label>35</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Mahood</surname>
              <given-names>EH</given-names>
            </name>
            <name name-style="western">
              <surname>Kruse</surname>
              <given-names>LH</given-names>
            </name>
            <name name-style="western">
              <surname>Moghe</surname>
              <given-names>GD</given-names>
            </name>
          </person-group>
          <article-title>Machine learning: a powerful tool for gene function prediction in plants</article-title>
          <source>Appl Plant Sci</source>
          <year>2020</year>
          <month>07</month>
          <volume>8</volume>
          <issue>7</issue>
          <fpage>e11376</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/32765975"/>
          </comment>
          <pub-id pub-id-type="doi">10.1002/aps3.11376</pub-id>
          <pub-id pub-id-type="medline">32765975</pub-id>
          <pub-id pub-id-type="pii">APS311376</pub-id>
          <pub-id pub-id-type="pmcid">PMC7394712</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref36">
        <label>36</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Mochida</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Koda</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Inoue</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Nishii</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Statistical and machine learning approaches to predict gene regulatory networks from transcriptome datasets</article-title>
          <source>Front Plant Sci</source>
          <year>2018</year>
          <volume>9</volume>
          <fpage>1770</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/30555503"/>
          </comment>
          <pub-id pub-id-type="doi">10.3389/fpls.2018.01770</pub-id>
          <pub-id pub-id-type="medline">30555503</pub-id>
          <pub-id pub-id-type="pmcid">PMC6281826</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref37">
        <label>37</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Tan</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Han</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>H</given-names>
            </name>
          </person-group>
          <article-title>From machine learning to deep learning: progress in machine intelligence for rational drug discovery</article-title>
          <source>Drug Discov Today</source>
          <year>2017</year>
          <month>11</month>
          <volume>22</volume>
          <issue>11</issue>
          <fpage>1680</fpage>
          <lpage>1685</lpage>
          <pub-id pub-id-type="doi">10.1016/j.drudis.2017.08.010</pub-id>
          <pub-id pub-id-type="medline">28881183</pub-id>
          <pub-id pub-id-type="pii">S1359-6446(16)30436-6</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref38">
        <label>38</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Goodfellow</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Pouget-Abadie</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Mirza</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Warde-Farley</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Ozair</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Courville</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Bengio</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Generative adversarial networks</article-title>
          <source>Commun ACM</source>
          <year>2020</year>
          <month>10</month>
          <day>22</day>
          <volume>63</volume>
          <issue>11</issue>
          <fpage>139</fpage>
          <lpage>144</lpage>
          <pub-id pub-id-type="doi">10.1145/3422622</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref39">
        <label>39</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Huang</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Fong</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Zhao</surname>
              <given-names>Q</given-names>
            </name>
          </person-group>
          <article-title>Artificial intelligence in cancer diagnosis and prognosis: opportunities and challenges</article-title>
          <source>Cancer Lett</source>
          <year>2020</year>
          <month>02</month>
          <day>28</day>
          <volume>471</volume>
          <fpage>61</fpage>
          <lpage>71</lpage>
          <pub-id pub-id-type="doi">10.1016/j.canlet.2019.12.007</pub-id>
          <pub-id pub-id-type="medline">31830558</pub-id>
          <pub-id pub-id-type="pii">S0304-3835(19)30613-5</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref40">
        <label>40</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Belić</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Bobić</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Badža</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Šolaja</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Đurić-Jovičić</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Kostić</surname>
              <given-names>VS</given-names>
            </name>
          </person-group>
          <article-title>Artificial intelligence for assisting diagnostics and assessment of Parkinson's disease-A review</article-title>
          <source>Clin Neurol Neurosurg</source>
          <year>2019</year>
          <month>09</month>
          <volume>184</volume>
          <fpage>105442</fpage>
          <pub-id pub-id-type="doi">10.1016/j.clineuro.2019.105442</pub-id>
          <pub-id pub-id-type="medline">31351213</pub-id>
          <pub-id pub-id-type="pii">S0303-8467(19)30238-0</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref41">
        <label>41</label>
        <nlm-citation citation-type="web">
          <article-title>Code used in this study</article-title>
          <source>GitHub</source>
          <access-date>2023-11-26</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://github.com/jiashenggu/ai_in_bio">https://github.com/jiashenggu/ai_in_bio</ext-link>
          </comment>
        </nlm-citation>
      </ref>
    </ref-list>
  </back>
</article>
