Artificial intelligence in pediatric healthcare: current applications, potential, and implementation considerations
Article information
Abstract
Artificial intelligence (AI) has transformed pediatric healthcare by supporting diagnostics, personalized treatment strategies, and prognosis predictions. Although it offers significant promise in these areas, its application in pediatric settings poses distinct challenges compared with that in adults due to variable developmental status, the limited availability of pediatric data, and ethical concerns regarding bias and transparency. This narrative review summarizes the key concepts of AI and its clinical applications across clinical fields in the treatment of children and adolescents. Here we highlight the emerging role of large language models in performing administrative tasks and clinical documentation and supporting decision-making. We also address the evolving impact of AI integration in surgical care as an example while exploring ongoing concerns regarding reliability and diagnostic safety. Furthermore, we survey AI-enabled medical devices and discuss the current regulatory frameworks relevant to pediatric care. This review provides a balanced overview of opportunities and challenges from a pediatrician's standpoint and aims to facilitate effective alignment and collaboration with key stakeholders in pediatric healthcare. Pediatricians must implement AI solutions cautiously and accountably to avoid unintended harm and realize their potential.
Key message
Artificial intelligence (AI) offers potential benefits in pediatric care, but its real-world adoption requires clinician literacy, ethical and legal safeguards, and cautious implementation. Large language models are emerging across healthcare, but their use in pediatric clinical practice remains premature. Thus, the cautious and accountable implementation of AI is crucial to preventing unintended harm and realizing its potential.
Introduction
Data-driven innovations that enhance disease prevention, diagnostics, treatment, outcomes, and the prediction of clinical courses are transforming pediatric healthcare. Moreover, artificial intelligence (AI), fueled by increasing computing power and extensive data availability, is rapidly advancing, with innovative approaches and algorithms continuously emerging. Nevertheless, pediatric AI applications face distinct challenges. For instance, variability in children’s developmental stages and the limited availability of high-quality data pose significant obstacles to developing robust AI models [1].
This review provides contextual background, clarifies key terminology, explores current clinical applications, and discusses the challenges of implementing AI in pediatrics. While here we do not attempt to offer a comprehensive or fully up-to-date catalog, we incorporate diverse perspectives from all key stakeholders in the healthcare ecosystem, including patients, parents or guardians, physicians, hospitals, insurance providers, and industrial partners developing AI-based medical devices. By bridging the gap between technological innovation and clinical practice, we provide an unbiased assessment of the opportunities and challenges surrounding AI-enabled pediatric healthcare, which ultimately aims to enhance patient outcomes and overall well-being.
Understanding AI: history and fundamentals
AI is a field of interdisciplinary research that spans computer science, linguistics, philosophy, and other areas [2]. It focuses on developing methods that enable machines to perceive their environment, learn from data, store knowledge, and make decisions to achieve defined goals [3].
Modern AI research can be traced back to 1956, when John McCarthy used the term “artificial intelligence” during the Dartmouth Summer Research Project. Early successes, such as the Logic Theorist [4], a program capable of proving mathematical theorems, generated optimism about its potential. During the 1980s, AI experienced a resurgence with the development of expert systems, which applied expert knowledge to complex decision-making tasks. A notable example is MYCIN, developed at Stanford University that recommend effective antibiotics for bacterial infections [5]. However, despite their initial promise and successful implementations, these expert systems struggled to adapt to new situations and proved costly to maintain and update their knowledge bases [6].
The field advanced rapidly with the rise of machine learning (ML), a branch of AI focused on training algorithms to recognize patterns and make decisions accordingly. ML enables computers to solve problems without requiring explicit programming. ML methods fall into three broad categories: supervised learning, unsupervised learning, and reinforcement learning (see Table 1 for brief descriptions).
The current advent of AI stems from deep learning (DL) [7], pioneered by Hinton et al. [8,9] through deep neural networks. DL is a subset of ML methods that utilize artificial neural networks (Table 1), typically featuring large layered networks of processing units with millions to billions of adjustable parameters. DL models, such as large language models (LLMs) like the generative pretrained transformer (GPT) behind ChatGPT [10,11], are trained on vast amounts of text data (Fig. 1 left). These LLMs are examples of so-called foundation models: large, general-purpose models that can be adapted to various domains. While most current foundation models are based on the transformer architecture with LLMs as their most popular examples, the concept extends to other architecture and data modalities such as images, audio, and multimodal data. Because foundation models were initially trained for general purposes, they often undergo fine-tuning using smaller, domain-specific datasets to specialize them for specific tasks (Fig. 1 right). For example, ChatGPT is a fine-tuned model built upon the GPT foundation model and optimized for conversation.

Basic concepts in modern artificial intelligence (AI). (Left) The hierarchy of AI, machine learning, deep learning, and foundation models. These concepts are independent of data types such as text, image, or audio. (Right) An example showing how large language models (LLMs; foundation models for text data) are trained on vast amounts of text data (hundreds of billions of words). These models are then fine-tuned for specific tasks, such as generating responses to users’ prompts. The figure shows only a subset of popular applications. GPT, generative pretrained transformer; LLaMA, large language model meta AI.
The combination of enhanced computing power and large volumes of multimodal data has driven AI’s recent exponential growth, thereby impacting various sectors including healthcare [6,12]. These technologies have demonstrated success performing diagnostic tasks [12], predicting patient outcomes [13], and assisting with surgical procedures [14]. Understanding the fundamental concepts of AI will help healthcare professionals navigate emerging technologies and leverage these tools to deliver better patient care [2,15].
Applications of AI in pediatric healthcare
AI technologies are increasingly integrated into various aspects of pediatric care, thereby offering potential benefits while also presenting unique challenges.
1. AI applications in pediatric subspecialities: transforming diagnosis and management
AI and ML are transforming pediatric subspecialties by augmenting specific clinical tasks such as early detection, precision diagnostics, and outcome predictions. AI-driven early warning systems are particularly valuable in time-sensitive conditions such as pediatric sepsis. These models leverage biosignal data to predict deterioration and guide timely interventions, potentially improving survival rates [1,16,17]. Furthermore, in pediatric allergy and immunology, ML models incorporating environmental data have shown promise in forecasting allergen exposure, thereby helping prevent acute exacerbations [18,19]. In pediatric neurology and oncology, DL models trained on multimodal inputs including genetic, neuroimaging, and clinical data have enhanced diagnostic precision for complex conditions such as pediatric brain tumors [20,21]. Similarly, in pediatric cardiology, DL models applied to echocardiograms and electrocardiograms have demonstrated strong ability to detect biventricular dysfunction and ventricular dilation, which are often challenging to assess using traditional methods [22].
Beyond diagnostics, AI models also contribute to postoperative risk assessments. For instance, an interpretable ML model like optimal classification trees has been successfully used to predict the clinical outcomes of patients with a history of undergoing congenital heart surgery despite non-linear data relationships [23,24]. These tools can help clinicians tailor follow-up care and surveillance strategies based on individual risk profiles. These examples demonstrate how AI models both enhance accuracy and efficiency in pediatric care and enable more proactive and personalized clinical management.
2. Large language models
LLMs may help streamline administrative tasks, enhance doctor–patient communication, and support clinical decisions. LLMs can be utilized in tasks that requires their natural language processing capacity, such as generating clinical notes from conversations. Moreover, they can be integrated into electronic medical records systems to assist with patient summaries, discharge instructions, and personalized education efforts [25]. LLM tools can rapidly produce medical documentation of comparable quality to that of physician-written notes, significantly reducing physicians' administrative burdens [26]. LLMs can also aid communication with caregivers by generating contextually relevant and patient-specific responses to common parental questions in the pediatric intensive care unit [27].
In terms of clinical decision support, LLMs have shown potential for diagnostic assistance, treatment planning, and information retrieval [28]. They demonstrated high accuracy and significantly faster responses at pediatric drug dosage calculations [29]. LLMs showed diagnostic performance comparable to that of pediatricians in supporting the differential diagnosis of common presentations in primary care [30]. As LLMs evolve to handle more complex interactions, such as comprehensive history-taking and open-ended questioning, their use might improve diagnostic accuracy [31].
Despite growing enthusiasm for their clinical use, LLMs have primarily been evaluated through medical examination questions, with limited assessments of their real-world clinical performance [32]. In recent stress tests using clinically based prompts, GPT-4o demonstrated a 16% inappropriate response rate. Notably, 21.5% of responses previously considered acceptable with GPT-3.5 were subsequently deemed inappropriate when reassessed with newer models [33]. These findings highlight the need for continuous and rigorous model assessments in clinical applications, especially in pediatric populations, where limited age-specific data may exacerbate the risk of biased or underperforming LLMs.
Other significant risks include hallucinations and biases, which raise serious questions about these models’ accuracy and reliability because they can produce misinformation, distort context, and strengthen bias. Data privacy is also a major concern. Uploading patient information into external LLMs (like public versions of Chat GPT) carries considerable risk, as input data might be used for model training that potentially exposes identifiable health information.
These risks and concerns require that pediatricians and caregivers exercise caution when using these tools.
3. AI/ML-based medical devices and regulation
AI/ML technologies embedded in medical devices are increasingly used in direct clinical interventions, necessitating a closer examination of their approval processes, safety, and pediatric-specific challenges. While the U.S. Food and Drug Administration (FDA) authorized the first AI-enabled device in 1995, the number of AI/ML-based medical devices, including software, has risen sharply only over the past decade [34].
Table 2 summarizes representative cases based on their specific patient care step they aim to address: prevention, diagnosis, treatment, or prediction. These devices can analyze vast amounts of medical data and detect subtle changes in patterns that may be overlooked by clinicians, facilitating faster and more accurate diagnoses. Predictive analytics powered by AI/ML can estimate a patient's disease risk and likelihood of hospital readmission [35]. Patient treatment can be personalized using these devices, such as smart insulin pumps, which autonomously adjust insulin delivery in response to real-time patient data [36]. Wearable devices and sensors integrated with AI/ML enable continuous monitoring, allowing for timely interventions [37]. Furthermore, AI-enhanced robotic surgery systems improve operative precision and safety [38].

Summary of selected U.S. Food and Drug Administration/Conformité Européenne approved artificial intelligence-based medical devices (pediatric applications preferred)
Most currently approved AI/ML medical devices are categorized as class I or II, indicating low and moderate risk, respectively, and authorized through the 510(k) pathway, which requires demonstrating substantial equivalence to a previously approved device [39]. This regulatory approach often reflects incremental improvements to existing technologies rather than breakthrough innovations, with approval processes that may not require new clinical trials but instead rely on comparisons to predicate devices. Developers may attempt to bypass the complexities of full regulatory approval by initially launching their products as wellness apps to collect patient data for a future regulatory approval application, potentially compromising safety and data sharing [40,41].
Unlike static (“locked”) algorithms, AI/ML systems evolve using new data. While this allows ongoing optimization, it raises concerns about postmarket safety and reliability. These changes may introduce unexpected risks, particularly in pediatric populations, if not supported by robust validations. Recognizing these challenges, the FDA has proposed a lifecycle-based regulatory framework that includes pre-market reviews, algorithm change protocols, and postmarket surveillance to allow ongoing updates while ensuring sustained safety and effectiveness [42,43].
AI/ML medical devices for pediatric healthcare face unique development and implementation challenges. Medical technology for children has traditionally lagged behind that for adults [44]. AI/ML technologies are no exception to this trend. This systematic delay often places pediatric clinicians in the position of using adult devices for children without adequate evidence of safety or efficacy in off-label applications. The unique characteristics of the pediatric population (i.e., children are not "small adults") make the extension of algorithms developed for adult populations to children risky due to fundamental differences in physiology, disease presentation, and care requirements.
Moreover, children are frequently underrepresented in test cohorts used to validate these technologies. The limited financial incentive due to the smaller market size further delays developing pediatric-specific AI/ML applications. Despite children comprising approximately 23% of the U.S. population, they account for only 10% of total healthcare expenditures [45]. Consequently, only a small fraction (17%; 149 devices in 1995–2024) of FDA-approved AI/ML devices are labeled for pediatric use, primariliy in the radiology, cardiology, and neurology [46].
Therefore, when considering the use of these technologies for vulnerable pediatric populations, pediatricians and related specialists must balance the potential benefits of AI/ML medical devices against their current limitations and regulatory uncertainties.
AI Integration in Pediatric Surgery
Minimally invasive surgery is a surgical care area that actively adopted AI early on [47,48]. Due to the complex and delicate anatomical structures of pediatric patients, conventional surgical approaches require high technical expertise. However, the integration of AI technologies like ML, LLMs, and computer vision may contribute to improved surgical outcomes in peditric patients by enhancing diagnostic accuracy, surgical planning, intraoperative guidance, and postoperative monitoring [48,49]. The real-time identification of critical anatomical structures assisted by AI-based image analysis can minimize unnecessary tissue damage and potentially accelerating postoperative recovery [50]. An automated analysis of surgical procedures enables the generation of surgical documentation without manual input [51]. In robotic surgery, AI improves outcomes by stabilizing instruments, enhancing visualization, identifying critical anatomical structures, providing early warnings of intraoperative risks, and enabling partial automation of robotic control [52,53]. Despite this potential, challenges include a lack of pediatric-sized robotic instruments, high implementation costs, and limited or biased pediatric datasets trained primarily on adult populations that hinder model generalizability and reduce clinical reliability in children [49].
Of note, digital twin technology is increasingly being explored within surgical fields. Digital twin-assisted surgery (DTAS), enhanced by AI and DL, creates patient-specific virtual replicas that dynamically reflect intraoperative conditions, offering real-time guidance. By integrating live physiological data with predictive modeling, DTAS can improve surgical precision, for example by accurately delineating tumor margins and identifying critical anatomical structures to minimize collateral damage and ensure complete resection. Clinically, this approach has been associated with reduced operative times, fewer complications, and improved patient outcomes [54]. Moreover, DTAS serves as a powerful educational tool by allowing trainees to practice with highly realistic simulations of complex procedures within a safe and controlled environment [54]. While DTAS contributes to enhanced surgical precision, its performance is highly dependent on the quality and accuracy of the real-world anatomical data used to construct the virtual models. Given the rarity and low volume of pediatric surgical cases, collaborative multicenter efforts are essential to the development and curation of training datasets.
Like with all AI tools in healthcare, those used in pediatric surgery must undergo rigorous validation for safety and transparency, during which processes pediatric surgeons should play a pivotal role to ensure their clinical reliability [55].
Key stakeholder perspectives on AI applications
Implementing AI in healthcare places healthcare providers at the center of a network of interactions with insurers/payers, patients/caregivers, and the healthcare industry (Fig. 2).

Key areas for artificial intelligence in healthcare, highlighting interactions between stakeholders. Healthcare providers gain diagnostic accuracy and streamlined workflows. Insurers improve efficiency and detect fraud but face investment challenges. Patients benefit from early detection and better care, but privacy concerns persist. The industry sees market differentiation but faces data and integration issues. Balancing these aspects is crucial for effective artificial intelligence implementation. PHI, protected health information.
Healthcare providers benefit from AI through enhanced diagnostic accuracy and speed, streamlined workflows, and access to up-to-date medical information, leading to better patient care and outcomes. For example, AI can assist with the diagnosis of pediatric conditions, thereby enabling timely interventions. However, providers face increased responsibility in validating AI outputs. Conflicts between AI recommendations and clinical judgment can create legal and ethical uncertainties [56].
In their interactions with insurers and payers, healthcare providers leverage AI to improve operational efficiency and cost optimization. By accurately detecting fraudulent claims, providers help insurers reduce financial losses, leading to more affordable healthcare plans. However, both parties must address challenges such as the initial investment required for AI implementation and ensuring that the algorithms are free from bias to maintain fairness in patient care [57,58].
Healthcare providers also play a crucial role in addressing patient and caregiver concerns. While AI can aid the early detection of health issues, reduce medical errors, and enhance overall care quality, providers must ensure that patient data are protected during such usage to address privacy concerns and maintain trust. They need to manage the cost of AI technologies to avoid burdening patients and caregivers with increased expenses. Providers must also work to overcome resistance to adopting AI by educating and reassuring their patients about its benefits and safety.
In the broader industry, healthcare providers contribute to market differentiation and utilization of the transformative potential of AI in healthcare. They may ensure regulatory compliance by integrating only those AI tools that meet industry standards. However, providers must navigate challenges such as the scarcity of pediatric data for training AI models and the complexity of integrating AI with existing systems while addressing ethical considerations related to AI use, such as ensuring patient consent and transparency in AI’s decision-making. Incentive structures surrounding reimbursement policies vary across countries and influence AI adoption in clinical care [59].
By effectively managing these interactions and challenges, healthcare providers can leverage AI to improve healthcare outcomes, enhance patient experiences, and drive healthcare innovations. Balancing the benefits and challenges of AI implementation is crucial to maximizing its potential and ensuring responsible use across all stakeholders.
Outstanding challenges for AI applications in pediatrics
Integrating AI into pediatric practice presents substantial challenges that necessitate careful consideration of ethical implications and robust policy oversight. One primary challenge is the "black box" problem. Many AI systems operate without transparency, making it difficult for clinicians to understand and trust their recommendations [60,61]. This lack of transparency hinders their adoption in pediatric healthcare, in which clinical decisions require explicit justification and caregiver confidence. Without sufficient explainability, clinicians may hesitate to adopt AI-driven advice, particularly in sensitive pediatric contexts. Thus, advancing explainable AI (XAI; Table 1) methods is essential for building clinician trust and AI reliability [62]. Nonetheless, current XAI technologies remain in early development and require substantial research and validation.
Bias in healthcare AI models can lead to disparities in diagnostics, treatment, and resource allocation, disproportionately affecting underserved populations. Such bias stems from data bias and algorithmic bias. Data bias occurs when training data are unrepresentative or reflect existing inequalities. For example, AI dermatology tools trained mostly on light-skinned patients performed poorly for those with darker skin tones [63]. On the other hand, algorithmic bias emerges from the model’s design and assumptions regardless of data quality. Models can unintentionally discriminate despite data balance. For instance, using health costs to predict medical needs can underestimate needs of groups with limited access, such as Black patients, who spend less in healthcare [64].
Another significant challenge involves false-positive alerts from AI-based diagnostic and monitoring tools [65]. Frequent false alarms can cause alert fatigue in clinicians, diminishing their responsiveness to genuine clinical emergencies. In pediatric care, in which timely decisions and interventions are crucial, reduced clinician vigilance can have particularly serious consequences [66]. Moreover, false-positive alerts trigger unnecessary medical interventions, including unwarranted diagnostic tests and treatments, increasing healthcare costs and excessive resource utilization. Such unnecessary interventions also raise ethical concerns, as pediatric patients are exposed to avoidable stress and risks from procedures, while their families experience unnecessary anxiety. To address this, stakeholders should prioritize the development of AI algorithms that balance sensitivity with specificity. Thus, the rigorous validation of AI systems tailored explicitly for pediatric populations is essential. Furthermore, healthcare institutions and regulatory bodies must establish standards that ensure the clinical relevance of AI-generated alerts to ultimately enhance patient safety, reduce costs, and improve overall efficiency.
Financial incentives also pose an additional significant challenge preventing AI integration in pediatric healthcare. Economic factors can subtly influence diagnostic thresholds—known as “biomarkup”—leading AI systems to detect minor or borderline conditions that prompt unnecessary but billable interventions [57,67]. Such subtle bias toward profitability over clinical necessity risks compromising patient-centered care. To mitigate this risk, stakeholders, including AI developers, healthcare institutions, and insurers, must commit to observing transparent practices and aligning AI deployment strictly for patient benefit. Robust oversight, clear regulatory guidelines, and independent audits are required to ensure that AI technologies serve pediatric populations equitably, ethically, and effectively and prioritize clinical benefit over financial gain.
Liability issues arising from AI integration in clinical care present complex challenges beyond those encountered in traditional practice. When AI-driven decisions cause patient harm, evaluating responsibility across clinicians, developers, and healthcare institutions raises intricate legal and ethical questions [57]. Additionally, while reliance on incomplete or biased datasets can produce inaccurate outputs, proving the connection between AI errors and patient harm with sufficient or acceptable evidence may be difficult due to issues including AI explainability and its evolving nature [56].
Addressing these challenges requires collaboration among pediatric clinicians, data scientists, ethicists, patient advocates, regulatory authorities, and lawyers. Rigorous oversight and ethical governance of AI applications will allow the medical community to safely integrate AI into pediatric care. Such an approach will enable clinicians to leverage the transformative potential of AI, ultimately improving patient outcomes while preserving patient welfare and trust.
Conclusion
Persistent concerns about data quality, algorithmic transparency, limited explainability, and ethical considerations including bias and financial incentives in AI integration emphasize the need for sustained research, cross-disciplinary collaboration, and robust governance. As AI increasingly integrates into pediatric clinical practice, clinicians must develop core competencies with these technologies to effectively deliver their evolving functions. However, the current limitations of LLMs, including concerns regarding reliability, hallucinations, data bias, and insufficient validation within pediatric contexts, suggest that their widespread clinical application may still be premature. Pediatricians are well-positioned to guide the thoughtful adoption of AI to improve the health outcomes of children and adolescents. Therefore, cautious and accountable implementation is crucial to preventing unintended harm induced by AI and realizing its potential.
Notes
Conflicts of interest
No potential conflict of interest relevant to this article was reported.
Funding
This work was supported by the Gyeongsang National University Fund for Professors on Sabbatical Leave 2024 (TP) and by the NIH R01NS129188 and U54HG01 2513 (SWK).
Acknowledgments
We are grateful to Yunxi Li (Harvard Medical School) for her insightful feedback and constructive comments on the manuscript.
Author Contribution
Conceptualization: SWK, TP; Investigation: TP, IHL, SWL; Resources: SWK; Visualization: IHL; Supervision: SWK, IHL; Writing–original draft:TP, IHL, SWK; Writing– review & editing: IHL, SWK