ArticlePDF Available

Clinical deployment of machine learning models in craniofacial surgery: considerations for adoption and implementation

OAE Publishing Inc.
Artificial Intelligence Surgery
Authors:

Abstract

The volume and complexity of clinical data are growing rapidly. The potential for artificial intelligence (AI) and machine learning (ML) to significantly impact plastic and craniofacial surgery is immense. This manuscript reviews the overall landscape of AI in craniofacial surgery, highlighting the scarcity of prospective and clinically translated models. It examines the numerous clinical promises and challenges associated with AI, such as the lack of robust legislation and structured frameworks for its integration into medicine. Clinical translation considerations are discussed, including the importance of ensuring clinical utility for real-world use. Finally, this commentary brings forward how clinicians can build trust and sustainability toward model-driven clinical care.
Roy et al. Art Int Surg 2024;4:427-34
DOI: 10.20517/ais.2024.69 Artificial
Intelligence Surgery
© The Author(s) 2024. Open Access This article is licensed under a Creative Commons Attribution 4.0
International License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, sharing,
adaptation, distribution and reproduction in any medium or format, for any purpose, even commercially, as
long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and
indicate if changes were made.
www.oaepublish.com/ais
Open AccessCommentary
Clinical deployment of machine learning models in
craniofacial surgery: considerations for adoption
and implementation
Mélissa Roy1, Russell R. Reid2, Senthujan Senkaiahliyan3, Andrea S. Doria4,5, John H. Phillips3, Michael
Brudno6,7,8,9, Devin Singh10
1Division of Plastic, Department of Surgery, McMaster University, Hamilton L8S 4L8, Canada.
2Section of Plastic Surgery, Department of Surgery, University of Chicago, Chicago, IL 60637, USA.
3Division of Plastic, Reconstructive and Aesthetic Surgery, The Hospital for Sick Children, Toronto M5G 1X8, Canada.
4Department of Medical Imaging, University of Toronto, Toronto M5T 1W7, Canada.
5Department of Diagnostic Imaging, Research Institute, The Hospital for Sick Children, Toronto M5G 1E8, Canada.
6Department of Computer Science, University of Toronto, Toronto M5S 2E4, Canada.
7Vector Institute, Toronto M5G 1M1, Canada.
8Digital Team and Techna Institute, University Health Network, Toronto M5G 2C4, Canada.
9Genetics and Genome Biology, Research Institute, The Hospital for Sick Children, Toronto M5G 0A4, Canada.
10Department of Paediatrics, The Hospital for Sick Children, Toronto M5G 1X8, Canada.
Correspondence to: Dr. Devin Singh, Department of Paediatrics, The Hospital for Sick Children, 555 University Avenue, Toronto
M5G 1X8, Canada. E-mail: devin.singh@sickkids.ca
How to cite this article: Roy M, Reid RR, Senkaiahliyan S, Doria AS, Phillips JH, Brudno M, Singh D. Clinical deployment of
machine learning models in craniofacial surgery: considerations for adoption and implementation. Art Int Surg 2024;4:427-34.
https://dx.doi.org/10.20517/ais.2024.69
Received: 16 Aug 2024 First Decision: 24 Sep 2024 Revised: 11 Nov 2024 Accepted: 14 Nov 2024 Published: 13 Dec 2024
Academic Editors: Ernest S. Chiu, Andrew A. Gumbs Copy Editor: Pei-Yun Wang Production Editor: Pei-Yun Wang
Abstract
The volume and complexity of clinical data are growing rapidly. The potential for artificial intelligence (AI) and
machine learning (ML) to significantly impact plastic and craniofacial surgery is immense. This manuscript reviews
the overall landscape of AI in craniofacial surgery, highlighting the scarcity of prospective and clinically translated
models. It examines the numerous clinical promises and challenges associated with AI, such as the lack of robust
legislation and structured frameworks for its integration into medicine. Clinical translation considerations are
discussed, including the importance of ensuring clinical utility for real-world use. Finally, this commentary brings
forward how clinicians can build trust and sustainability toward model-driven clinical care.
Keywords: Artificial intelligence, machine learning, craniofacial surgery, clinical translation
CLINICAL DEPLOYMENT OF MACHINE LEARNING MODELS IN CRANIOFACIAL
Page 428 Roy et al. Art Int Surg 2024;4:427-34 https://dx.doi.org/10.20517/ais.2024.69
SURGERY: CONSIDERATIONS FOR ADOPTION AND IMPLEMENTATION
The volume and complexity of clinical data are growing rapidly across all fields of medicine. In parallel,
computational power is expanding and becoming more accessible, while human resources continue to
remain relatively stagnant and limited in healthcare[1]. It has been predicted that artificial intelligence (AI)
and machine learning (ML) will be ubiquitous in future clinical care[1]. Significant areas of interest that
could revolutionize care include processing of electronic health record data, image classification, and
identification of medical errors[2,3]. The promising appeal to incorporate AI into clinical practice must be
contextualized, and the intrinsic limitations to the use of ML algorithms acknowledged. Clinical translation
of ML tools and prospective validation articles are scarce to date and a new set of considerations for
adoption and implementation have been unveiled. This article reviews the current ML landscape in
craniofacial surgery and highlights promises, challenges, and considerations for successful clinical
translation.
Current landscape of ML in craniofacial surgery
The role of ML and AI in craniofacial surgery has previously been thoroughly reviewed[4-9]. In a scoping
review by Mak et al. (2020), the authors identified numerous craniofacial-based studies developing ML
models[4]. ML is particularly relevant to craniofacial surgery as it is a specialty that: (1) relies on imaging for
diagnostic purposes; (2) uses standardized and universal anatomical landmarks (soft tissue and bony); (3)
benefits from three-dimensional planning and surgical navigation; and (4) has variable outcomes and
benefits from risk prediction[9]. Thus far, published craniofacial surgery studies using AI have been
experimental and theoretical in nature, mostly relying on retrospective datasets with very limited sample
sizes and generally single-centered.
Cleft surgery
The potential benefit of integrating AI into cleft care spans numerous facets of clinical care due to the varied
presentations (from cleft lip to velopharyngeal insufficiency and dentoalveolar discrepancies) and evolution
over patients’ growth and development. A scoping review identified previously explored areas for
implementation of AI in cleft care[10]: prediction of risk of developing cleft lip or palate, diagnosis (prenatal
cleft presence), severity of morphological deformities of nose[11], speech evaluation (presence of
hypernasality, assessment of intelligibility)[12], surgical planning (estimation of volumetric defect of alveolar
cleft), prediction of need for orthognathic surgery, and more.
Orthognathic surgery
Orthognathic surgery lands itself well to be augmented by AI, although few studies have been published
thus far. A review of possible applications demonstrated areas of interest for future studies, including[13]:
complex diagnoses (superimposing numerous diagnostic tools for measurement of upper airways and
management of obstructive sleep apnea), common diagnoses (lateral cephalogram review), treatment
planning (taking into account the symbiotic relationship with orthodontic changes), creation of custom
dental appliances, and much more. An early study (2019) within the field of orthognathic surgery proposed
a proof of concept that AI (specifically convolutional neural networks) can be used in order to score facial
attractiveness and apparent age in orthognathic surgery patients[14]. Assistance in diagnosis and surgical
planning has also been proposed using CT scan images and comparing discrepancies in cephalometric
measures between the AI-generated plan and the postoperative images[15].
Craniosynostosis and head shape difference
Head shape differences are ubiquitous, and the rarity of craniosynostosis and the clinical ramifications of
later diagnosis generate huge diagnostic importance, which can be aided by AI. A review from 2023
explored the existing literature on the topic, highlighting that most studies thus far have used two-
Roy et al. Art Int Surg 2024;4:427-34 https://dx.doi.org/10.20517/ais.2024.69 Page 429
dimensional photographs for model building in which image orientation had a large impact on outcome[16].
A team at the Hospital for Sick Children, Toronto, is also developing a model compatible with three-
dimensional photographic analysis for diagnostic purposes and is validating it into a mobile capture for
widespread use and patient empowerment[17].
Craniofacial trauma
The detection of facial trauma on imaging has been an area of interest for the introduction of AI[18]. Overall
performance of models being developed, such as the DeepCT, has been excellent with high reported
sensitivity (89%) and specificity (95%) while using two-dimensional models with CT scan images[19].
Detection of facial fractures using three-dimensional images is also being explored within the biotechnology
literature[20]. Various models have been used, including CNNs and deep learning systems using a one-stage
detection called you only look once (YOLO)[18,21,22].
Facial reanimation
Thus far, AI represents a great promise of revolutionizing outcomes assessment for facial reanimation.
Traditionally, researchers and surgeons have used static images of smile commissure excursion using
various tools to provide an outcome assessment[23]. With the use of ML technology, facial expression
tracking and, therefore, dynamic facial assessment are becoming possible. Video analysis of cross-facial
nerve graft (CFNG) and free gracilis muscle patients took place in a proof of concept study by Boonipat
et al.[24]. This methodology enabled the authors to analyze symmetry, excursion, and overall facial
movements via the review of more than 500 facial landmarks. Another aspect of facial reanimation surgery
that has remained difficult to capture with existing outcome measures includes smile spontaneity.
Dusseldorp et al. compared controls with patients having undergone CFNG, masseter nerve coaptation, and
dual innervation to assess the feasibility of using such a tool to review spontaneity[25]. Although promising
innovative outcomes analyses are being developed, the surgical reconstruction of facial palsy focuses on the
lower face and smile, whereas AI technology considers the face as a whole.
Clinical promises
To many clinicians, AI represents a new technological advancement that can revolutionize their practice,
yet also a poorly understood and intimidating one. Such paradoxical perceptions of AI can be explained by
the power that is inherent to it and the often lack of transparency behind its use on our electronic devices,
social media, and more. Transparency, safety, and methodological rigor are central to evidence-based
medicine and can be enabled with established reporting standards tools. A new initiative, the MINimum
Information for Medical AI Reporting (MINIMAR), proposes such reporting standards[2]. MINIMAR
highlights four fundamental areas of transparency and reporting: (1) study population and setting; (2)
patient demographic characteristics; (3) model architecture; and (4) model evaluation. This concept is
further explored by research done by Sendak et al. with the “Model Facts” label[26]. To maximize the effective
and positive implementation of an AI model, adherence to such standards of reporting is highly
recommended.
Some of the highly anticipated promises of AI are centered around the optimization of patient-centered care
and outcomes. The hope is that detailed and precise ML algorithms can help enhance a clinician’s
diagnostic, management, and prognostic capabilities, as well as monitor and decrease medical errors. AI
may also allow patients to have a sense of improved ownership of their data and empower them with the
ability to interpret their health information[1]. Such optimization of care would have a system-level impact
and could allow for more efficient workflow, better resource allocation, and improved clinical outcomes, to
name a few[1].
Page 430 Roy et al. Art Int Surg 2024;4:427-34 https://dx.doi.org/10.20517/ais.2024.69
Clinical challenges
Some of the most feared consequences of AI are the potential for patient harm and the increased
accountability for clinicians. At baseline, introducing new technology into the medical field requires close
monitoring and auditing. In the case of AI, a flawed algorithm can lead to the dissemination of iatrogenic
harm, medical errors, and malpractice[1]. Incorrect model outputs leading to adverse patient outcomes also
increase the potential liability for physicians. This can deter healthcare professionals from utilizing AI as it is
currently not essential to the delivery of care. Beyond an inaccurate algorithm also exists a black box one, or
one for which the operations and output cannot be explained and is at the center of much controversy[27,28].
In the current climate, careful model validation, prospective auditing, and algorithm enhancements are
therefore necessary, as well as human support and oversight upon deployment of any AI models[1].
Transparency regarding model development and data sources used, algorithms and overall methodological
disclosure according to the MINIMAR reporting guidelines can mitigate some of these challenges[2].
The Gender Shades project further underscores the importance of reporting guidelines to ensure rigorous
bias assessments are completed[29]. The study highlights significant racial and gender bias in commercial
facial recognition technologies. Their research revealed that AI systems from companies like IBM,
Microsoft, and Face++ were less accurate at identifying gender in darker-skinned individuals, particularly
women, with error rates of up to 34.7% for darker-skinned females, compared to less than 1% for lighter-
skinned males. This disparity is linked to the underrepresentation of diverse phenotypes in the datasets used
to train these models, which overwhelmingly consist of lighter-skinned individuals.
These findings are particularly relevant for photo-based AI applications in craniofacial surgery. As these
technologies become integrated into surgical planning and diagnostics, biases could disproportionately
affect individuals with darker skin tones, potentially leading to misdiagnoses or improper treatment
recommendations. It underscores the necessity of using diverse and balanced datasets in the development of
AI models in craniofacial surgery as well as conducting detailed subgroup bias assessments on gender, age,
race, ethnicity, and in some instances, skin tone (for image/photography-based applications). Of note,
transparency to patients on how a model is expected to work specifically for them, based on these subgroup
bias assessments, is important to facilitate genuine informed consent. It is expected that variations in
performance will occur, and with transparency in understanding the limitations of a model, risk can be
mitigated to ensure care delivery remains equitable even when a model may not perform equally across
various populations.
While large data pools are required to create reliable and generalizable ML models, the acquisition of such
information may lead to concerns regarding health data ownership, privacy, and security[30,31]. Possible
malicious uses of AI technology have been reviewed and could include: breaches of data security and
privacy, hacking of algorithms with the intent to harm, manipulation of data, and much more[32].
Governance bodies should seek to define best practices, to mitigate security and safety threats, and to have
action plans in case a malicious event occurs[32]. Guidance navigating expectations and consequences of the
use of ML models in healthcare should be encouraged at all levels (patient, physician, institutions, and
governing bodies). Excessive or absolute dependence on experimental models should be avoided until
robust foundations and infrastructures are in place to mitigate risks associated with AI.
Clinical translation: real-world introduction
Watson et al. conducted semi-structured interviews with American Academic Medical Centers regarding
their use of predictive modeling and ML techniques within clinical care[33]. The team identified specific
barriers to the adoption and implementation of such models that encompassed several themes: culture and
personnel, clinical utility, financing, technology, and data. Overall, multidisciplinary development teams
Roy et al. Art Int Surg 2024;4:427-34 https://dx.doi.org/10.20517/ais.2024.69 Page 431
were found to be essential to ensure the integration of ML tools into the clinical workflow. A well-defined
clinical utility with clinically relevant parameters and actionable alerts ensured the usefulness of the ML
tools. Securing funding was seen as a significant challenge to overcome to support all phases of ML model
development and deployment. Partnerships with vendors could be considered to help overcome challenges
associated with translation and the long-term sustainability of model deployment[33].
The generalizability of ML models to the real-world clinical realm can be limited despite rigorous internal
and external validation studies. It has been shown that the real-world introduction of ML models sometimes
leads to lower accuracy and higher false positives[34]. This discrepancy between experimentation and reality
can be partially due to the datasets used. Research datasets have been shown to be constrained by stringent
inclusion and exclusion criteria[35,36]. Clinical deployment, therefore, requires close model and output
monitoring, followed by adjustments. Another aspect of validation can be the challenges in data sharing. In
that regard, “federated learning” could enable the use of large multi-institution datasets by decentralizing
data analysis and sharing computational models rather than data[37].
A disconnect between developers and users may sometimes occur. The technical expert team developing
craniofacial surgery ML models may not be versed in the clinical needs and settings in which the technology
will be deployed. An in-depth understanding of the clinical environments is key for both the development
and translation of the ML tools to the bedside. Are there support team members available to perform data
entry? Is the information obtained novel or more accurate than the one recorded from conventional clinical
assessment? Can the outputs be easily interpreted? Are the output results clinically relevant and helpful in
guiding the management of patients?[35]. The clinical utility of ML models needs to be properly estimated
and clinical needs should therefore guide model development and tool creation.
Fostering clinical trust
Beyond weighing the recognized benefits and risks of the introduction of ML and AI in their practices,
surgeons may experience a distrust toward AI systems and their outputs. This skepticism may come from a
lack of transparency or understanding of the processes. The explainability factor is important for users in
the context of clinical decision support systems[38,39]. Explainable AI (XAI) is an emerging field bridging
humans with machines by enabling a better understanding of how a model’s decisions and predictions were
made[40,41]. For clinicians to trust AI and ML models, such bidirectional dialogue and reasoning is crucial.
Using XAI involves significant trade-offs, such as the cost of its incorporation. Costs mostly come from the
computation required to create a dialogue and learning capabilities between the model and clinician.
Another identified trade-off lies between performance and interpretability. It appears that the models
offering the best performance metrics are also often the least explainable[42]. As medicine strives for the best
clinical performance and outcomes, deployment of explainable yet less performant models may be
questionable. Ultimately, surgeons will have to justify clinical decisions with models that they trust and can
understand in order to provide optimal machine-augmented care. Explainability can help answer some
ethical, societal, and regulatory apprehensions and requirements if paired with rigorous model validation
techniques and bias assessments[38]. To sustainably translate ML models into clinical practice, XAI appears
to be a fundamental investment that requires further attention and development. However, it is important
to note that model explainability is not a substitute for model evaluation following evidence-based medicine
best practices and robust statistical validation.
Sustainability after clinical translation
The long-term sustainability of ML in practice requires financial support for data quality and access,
governance, security, continuous model validation, and operational deployment. The implementation of AI
models in clinical practice may also require the creation of new roles to facilitate the adoption and
Page 432 Roy et al. Art Int Surg 2024;4:427-34 https://dx.doi.org/10.20517/ais.2024.69
maintenance of AI algorithms, such as data scientists and machine learning operations (MLOps) engineers.
For example, MLOps engineers will help create systems that continuously monitor models to ensure a
decline in model performance does not occur without clinical and operational teams being aware. This is
critical as models can decline in performance based on data drift and other changes in the clinical real-
world environment. To support this, business models and associated governance structures should be
created[43]. They may vary in size and range from innovation clusters, combining local expertise in AI,
translational research, digital health, statistics, and more, to centers of excellence in large organizations[44].
Final thoughts
As we enter an age of increased intersections between society, data, and technology, we will notice the rapid
proliferation of ML models migrating from development labs into real-world surgical settings. We
recommend craniofacial surgeons be open and enthusiastic about upcoming ML models and tools but also
aware of their numerous limitations. Clinical deployment of such models is arduous yet promising for the
advancement of surgical care at the patient, team and system levels. Successful and safe integration of these
models into practice requires input from surgeons. Although clinical expertise cannot be replaced at this
time, it can be augmented by ML models, and surgeons should not be afraid of being innovators or early
adopters if they are equipped with knowledge and awareness.
The greatest enemy of knowledge is not ignorance, it is the illusion of knowledge - Stephen Hawking
DECLARATIONS
Authors’ contributions
Made substantial contributions to the conception, literature search, writing and review of the study: Roy M,
Reid RR, Senkaiahliyan S, Doria AS, Phillips JH, Brudno M, Singh D
Availability of data and materials
Not applicable.
Financial support and sponsorship
None.
Conflicts of interest
Singh D is the co-founder and CEO of a Canadian healthcare technology start-up company called Hero AI,
while the other authors have declared that they have no conflicts of interest.
Ethical approval and consent to participate
Not applicable.
Consent for publication
Not applicable.
Copyright
© The Author(s) 2024.
REFERENCES
Topol EJ. High-performance medicine: the convergence of human and artificial intelligence. Nat Med 2019;25:44-56. DOI PubMed1.
Hernandez-Boussard T, Bozkurt S, Ioannidis JPA, Shah NH. MINIMAR (MINimum Information for Medical AI Reporting):
developing reporting standards for artificial intelligence in health care. J Am Med Inform Assoc 2020;27:2011-5. DOI PubMed PMC
2.
Schiff GD, Bates DW. Can electronic clinical documentation help prevent diagnostic errors? N Engl J Med 2010;362:1066-9. DOI 3.
Roy et al. Art Int Surg 2024;4:427-34 https://dx.doi.org/10.20517/ais.2024.69 Page 433
PubMed
Mak ML, Al-Shaqsi SZ, Phillips J. Prevalence of machine learning in craniofacial surgery. J Craniofac Surg 2020;31:898-903. DOI
PubMed
4.
Jarvis T, Thornburg D, Rebecca AM, Teven CM. Artificial intelligence in plastic surgery: current applications, future directions, and
ethical implications. Plast Reconstr Surg Glob Open 2020;8:e3200. DOI PubMed PMC
5.
Kanevsky J, Corban J, Gaster R, Kanevsky A, Lin S, Gilardino M. Big data and machine learning in plastic surgery: a new frontier in
surgical innovation. Plast Reconstr Surg 2016;137:890e-7e. DOI PubMed
6.
Zhu VZ, Tuggle CT, Au AF. Promise and limitations of big data research in plastic surgery. Ann Plast Surg 2016;76:453-8. DOI
PubMed
7.
Kim YJ, Kelley BP, Nasser JS, Chung KC. Implementing precision medicine and artificial intelligence in plastic surgery: concepts and
future prospects. Plast Reconstr Surg Glob Open 2019;7:e2113. DOI PubMed PMC
8.
Murphy DC, Saleh DB. Artificial intelligence in plastic surgery: what is it? Where are we now? What is on the horizon? Ann R Coll
Surg Engl 2020;102:577-80. DOI PubMed PMC
9.
Dhillon H, Chaudhari PK, Dhingra K, et al. Current applications of artificial intelligence in cleft care: a scoping review. Front Med
2021;8:676490. DOI PubMed PMC
10.
Wu J, Tse R, Shapiro LG. Learning to rank the severity of unrepaired cleft lip nasal deformity on 3D mesh data. Proc IAPR Int Conf
Pattern Recogn 2014;2014:460-4. DOI PubMed PMC
11.
Maier A, Hönig F, Bocklet T, et al. Automatic detection of articulation disorders in children with cleft lip and palate. J Acoust Soc Am
2009;126:2589-602. DOI PubMed
12.
Bouletreau P, Makaremi M, Ibrahim B, Louvrier A, Sigaux N. Artificial intelligence: applications in orthognathic surgery. J Stomatol
Oral Maxillofac Surg 2019;120:347-54. DOI PubMed
13.
Patcas R, Bernini DAJ, Volokitin A, Agustsson E, Rothe R, Timofte R. Applying artificial intelligence to assess the impact of
orthognathic treatment on facial attractiveness and estimated age. Int J Oral Maxillofac Surg 2019;48:77-83. DOI PubMed
14.
Du W, Bi W, Liu Y, Zhu Z, Tai Y, Luo E. Machine learning-based decision support system for orthognathic diagnosis and treatment
planning. BMC Oral Health 2024;24:286. DOI PubMed PMC
15.
Qamar A, Bangi SF, Barve R. Artificial intelligence applications in diagnosing and managing non-syndromic craniosynostosis: a
comprehensive review. Cureus 2023;15:e45318. DOI PubMed PMC
16.
Mashouri P, Skreta M, Phillips J, et al. 3D photography based neural network craniosynostosis triaging system. In: Proceedings of the
Machine Learning for Health NeurIPS Workshop PMLR; 2020. pp.226-37. Available from: https://proceedings.mlr.press/v136/
mashouri20a.html. [Last accessed on 5 Dec 2024].
17.
Pham TD, Holmes SB, Coulthard P. A review on artificial intelligence for the diagnosis of fractures in facial trauma imaging. Front
Artif Intell 2023;6:1278529. DOI PubMed PMC
18.
Wang HC, Wang SC, Yan JL, Ko LW. Artificial intelligence model trained with sparse data to detect facial and cranial bone fractures
from head CT. J Digit Imaging 2023;36:1408-18. DOI PubMed PMC
19.
Moon G, Lee D, Kim WJ, Kim Y, Sung KY, Choi HS. Very fast, high-resolution aggregation 3D detection CAM to quickly and
accurately find facial fracture areas. Comput Methods Programs Biomed 2024;256:108379. DOI
20.
Moon G, Kim S, Kim W, Kim Y, Jeong Y, Choi H. Computer aided facial bone fracture diagnosis (CA-FBFD) system based on object
detection model. IEEE Access 2022;10:79061-70. DOI
21.
Wang X, Xu Z, Tong Y, et al. Detection and classification of mandibular fracture on CT scan using deep convolutional neural network.
Clin Oral Investig 2022;26:4593-601. DOI
22.
Roy M, Corkum JP, Shah PS, et al. Effectiveness and safety of the use of gracilis muscle for dynamic smile restoration in facial
paralysis: a systematic review and meta-analysis. J Plast Reconstr Aesthet Surg 2019;72:1254-64. DOI
23.
Boonipat T, Asaad M, Lin J, Glass GE, Mardini S, Stotland M. Using artificial intelligence to measure facial expression following
facial reanimation surgery. Plast Reconstr Surg 2020;146:1147-50. DOI PubMed
24.
Dusseldorp JR, Guarin DL, van Veen MM, Miller M, Jowett N, Hadlock TA. Automated spontaneity assessment after smile
reanimation: a machine learning approach. Plast Reconstr Surg 2022;149:1393-402. DOI PubMed
25.
Sendak MP, Gao M, Brajer N, Balu S. Presenting machine learning model information to clinical end users with model facts labels.
NPJ Digit Med 2020;3:41. DOI PubMed PMC
26.
Castelvecchi D. Can we open the black box of AI? Nature 2016;538:20-3. DOI PubMed27.
Knight W. The dark secret at the heart of AI. Available from: https://www.technologyreview.com/2017/04/11/5113/the-dark-secret-at-
the-heart-of-ai/. [Last accessed on 5 Dec 2024].
28.
Buolamwini J, Gebru T. Gender shades: intersectional accuracy disparities in commercial gender classification. In: Proceedings of the
1st Conference on Fairness, Accountability and Transparency. PMLR; 2018. pp. 77-91. Available from: https://proceedings.mlr.press/
v81/buolamwini18a.html?mod=article_inline&ref=akusion-ci-shi-dai-bizinesumedeia. [Last accessed on 5 Dec 2024].
29.
Kish LJ, Topol EJ. Unpatients-why patients should own their medical data. Nat Biotechnol 2015;33:921-4. DOI PubMed30.
Murdoch B. Privacy and artificial intelligence: challenges for protecting health information in a new era. BMC Med Ethics
2021;22:122. DOI PubMed PMC
31.
Brundage M, Avin S, Clark J, et al. The malicious use of artificial intelligence: forecasting, prevention, and mitigation. ArXiv.
[Preprint.] Dec 1, 2024 [accessed 2024 Dec 5].Available from: https://doi.org/10.48550/arXiv.1802.07228.
32.
Page 434 Roy et al. Art Int Surg 2024;4:427-34 https://dx.doi.org/10.20517/ais.2024.69
Watson J, Hutyra CA, Clancy SM, et al. Overcoming barriers to the adoption and implementation of predictive modeling and machine
learning in clinical care: what can we learn from US academic medical centers? JAMIA Open 2020;3:167-72. DOI PubMed PMC
33.
Salehinejad H, Kitamura J, Ditkofsky N, et al. A real-world demonstration of machine learning generalizability in the detection of
intracranial hemorrhage on head computerized tomography. Sci Rep 2021;11:17051. DOI PubMed PMC
34.
Mechelli A, Vieira S. From models to tools: clinical translation of machine learning studies in psychosis. NPJ Schizophr 2020;6:4.
DOI PubMed PMC
35.
Patel R, Oduola S, Callard F, et al. What proportion of patients with psychosis is willing to take part in research? A mental health
electronic case register analysis. BMJ Open 2017;7:e013113. DOI PubMed PMC
36.
Collaborative learning without sharing data. Nat Mach Intell 2021;3:459. DOI37.
Antoniadi AM, Du Y, Guendouz Y, et al. Current challenges and future opportunities for XAI in machine learning-based clinical
decision support systems: a systematic review. Appl Sci 2021;11:5088. DOI
38.
Bussone A, Stumpf S, O’Sullivan D. The role of explanations on trust and reliance in clinical decision support systems. In: 2015
International Conference on Healthcare Informatics; 2015 Oct 21-23; Dallas, USA. IEEE; 2015. pp. 160-9. DOI
39.
Amann J, Blasimme A, Vayena E, Frey D, Madai VI; Precise4Q consortium. Explainability for artificial intelligence in healthcare: a
multidisciplinary perspective. BMC Med Inform Decis Mak 2020;20:310. DOI PubMed PMC
40.
Barredo Arrieta A, Díaz-Rodríguez N, Del Ser J, et al. Explainable artificial intelligence (XAI): concepts, taxonomies, opportunities
and challenges toward responsible AI. Inform Fusion 2020;58:82-115. DOI
41.
Linardatos P, Papastefanopoulos V, Kotsiantis S. Explainable AI: a review of machine learning interpretability methods. Entropy
2020;23:18. DOI PubMed PMC
42.
Accuracy. Artificial Intelligence Episode 1 - overview of leading artificial intelligence clusters around the globe. Available from:
https://www.accuracy.com/perspectives/overview-leading-artificial-intelligence-clusters-around-globe. [Last accessed on 5 Dec 2024].
43.
McKinsey and Company. Transforming healthcare with AI: the impact on the workforce and organizations. Available from: https://
www.mckinsey.com/industries/healthcare-systems-and-services/our-insights/transforming-healthcare-with-ai. [Last accessed on 5 Dec
2024].
44.
ResearchGate has not been able to resolve any citations for this publication.
Article
Full-text available
Background Dento-maxillofacial deformities are common problems. Orthodontic–orthognathic surgery is the primary treatment but accurate diagnosis and careful surgical planning are essential for optimum outcomes. This study aimed to establish and verify a machine learning–based decision support system for treatment of dento-maxillofacial malformations. Methods Patients (n = 574) with dento-maxillofacial deformities undergoing spiral CT during January 2015 to August 2020 were enrolled to train diagnostic models based on five different machine learning algorithms; the diagnostic performances were compared with expert diagnoses. Accuracy, sensitivity, specificity, and area under the curve (AUC) were calculated. The adaptive artificial bee colony algorithm was employed to formulate the orthognathic surgical plan, and subsequently evaluated by maxillofacial surgeons in a cohort of 50 patients. The objective evaluation included the difference in bone position between the artificial intelligence (AI) generated and actual surgical plans for the patient, along with discrepancies in postoperative cephalometric analysis outcomes. Results The binary relevance extreme gradient boosting model performed best, with diagnostic success rates > 90% for six different kinds of dento-maxillofacial deformities; the exception was maxillary overdevelopment (89.27%). AUC was > 0.88 for all diagnostic types. Median score for the surgical plans was 9, and was improved after human–computer interaction. There was no statistically significant difference between the actual and AI- groups. Conclusions Machine learning algorithms are effective for diagnosis and surgical planning of dento-maxillofacial deformities and help improve diagnostic efficiency, especially in lower medical centers.
Article
Full-text available
Patients with facial trauma may suffer from injuries such as broken bones, bleeding, swelling, bruising, lacerations, burns, and deformity in the face. Common causes of facial-bone fractures are the results of road accidents, violence, and sports injuries. Surgery is needed if the trauma patient would be deprived of normal functioning or subject to facial deformity based on findings from radiology. Although the image reading by radiologists is useful for evaluating suspected facial fractures, there are certain challenges in human-based diagnostics. Artificial intelligence (AI) is making a quantum leap in radiology, producing significant improvements of reports and workflows. Here, an updated literature review is presented on the impact of AI in facial trauma with a special reference to fracture detection in radiology. The purpose is to gain insights into the current development and demand for future research in facial trauma. This review also discusses limitations to be overcome and current important issues for investigation in order to make AI applications to the trauma more effective and realistic in practical settings. The publications selected for review were based on their clinical significance, journal metrics, and journal indexing.
Article
Full-text available
Craniosynostosis is characterised by the premature fusion of one or more cranial sutures, resulting in an abnormal head shape. The management of craniosynostosis requires early diagnosis, surgical intervention, and long-term monitoring. With the advancements in artificial intelligence (AI) technologies, there is great potential for AI to assist in various aspects of managing craniosynostosis. The main aim of this article is to review available literature describing the current uses of AI in craniosynostosis. The main applications highlighted include diagnosis, surgical planning, and outcome prediction. Many studies have demonstrated the accuracy of AI in differentiating subtypes of craniosynostosis using machine learning (ML) algorithms to classify craniosynostosis based on simple photographs. This demonstrates its potential to be used as a screening tool and may allow patients to monitor disease progression reducing the need for CT scanning. ML algorithms can also analyse CT scans to aid in the accurate and efficient diagnosis of craniosynostosis, particularly when training junior surgeons. However, the lack of sufficient data currently limits this clinical application. Virtual surgical planning for cranial vault remodelling using prefabricated cutting guides has been shown to allow more precise reconstruction by minimising the subjectivity of the clinicians’ assessment. This was particularly beneficial in reducing operating length and preventing the need for blood transfusions. Despite the potential benefits, there are numerous challenges associated with implementing AI in craniosynostosis. The integration of AI in craniosynostosis holds significant promise for improving the management of craniosynostosis. Further collaboration between clinicians, researchers, and AI experts is necessary to harness its full potential.
Article
Full-text available
Facial bone fractures must be diagnosed and treated as early as possible to avoid complications and sequelae. CT images need to be analyzed to detect fractures, but the analysis is time-consuming, and enough specialists are not available to analyze them. Many classification and object detection studies are being conducted to address these issues. The ability of classification-based studies to pinpoint the exact location of fractures is limited. Object detection-based research, by contrast, is problematic because the shape of a fracture is ambiguous. We propose a computer-aided facial bone fracture diagnosis (CA-FBFD) system to address the aforementioned challenges. This system adopts the object detection model YoloX-S, which is trained using only IoU Loss for box prediction, along with CT image Mixup data augmentation. For training, we used only nasal bone fracture data, whereas for testing, we used several other facial fracture data. During evaluation, the CA-FBFD system achieved an average precision of 69.8% for facial fractures, which is better than the baseline YoloX-S model by a large margin of 10.2%. In addition, the CA-FBFD system achieved a sensitivity/person of 100% for facial fractures, which is considerably better than that exhibited by the baseline YoloX-S model by a margin of 66.7%. Therefore, the CA-FBFD system can effectively minimize the labor of doctors who need to determine facial bone fractures in facial CT.
Article
Full-text available
Objectives This study aimed to evaluate the accuracy and reliability of convolutional neural networks (CNNs) for the detection and classification of mandibular fracture on spiral computed tomography (CT). Materials and methods Between January 2013 and July 2020, 686 patients with mandibular fractures who underwent CT scan were classified and annotated by three experienced maxillofacial surgeons serving as the ground truth. An algorithm including two convolutional neural networks (U-Net and ResNet) was trained, validated, and tested using 222, 56, and 408 CT scans, respectively. The diagnostic performance of the algorithm was compared with the ground truth and evaluated by DICE, accuracy, sensitivity, specificity, and area under the ROC curve (AUC). Results One thousand five hundred six mandibular fractures in nine subregions of 686 patients were diagnosed. The DICE of mandible segmentation using U-Net was 0.943. The accuracies of nine subregions were all above 90%, with a mean AUC of 0.956. Conclusions CNNs showed comparable reliability and accuracy in detecting and classifying mandibular fractures on CT. Clinical relevance The algorithm for automatic detection and classification of mandibular fractures will help improve diagnostic efficiency and provide expertise to areas with lower medical levels.
Article
Full-text available
Background Advances in healthcare artificial intelligence (AI) are occurring rapidly and there is a growing discussion about managing its development. Many AI technologies end up owned and controlled by private entities. The nature of the implementation of AI could mean such corporations, clinics and public bodies will have a greater than typical role in obtaining, utilizing and protecting patient health information. This raises privacy issues relating to implementation and data security. Main body The first set of concerns includes access, use and control of patient data in private hands. Some recent public–private partnerships for implementing AI have resulted in poor protection of privacy. As such, there have been calls for greater systemic oversight of big data health research. Appropriate safeguards must be in place to maintain privacy and patient agency. Private custodians of data can be impacted by competing goals and should be structurally encouraged to ensure data protection and to deter alternative use thereof. Another set of concerns relates to the external risk of privacy breaches through AI-driven methods. The ability to deidentify or anonymize patient health data may be compromised or even nullified in light of new algorithms that have successfully reidentified such data. This could increase the risk to patient data under private custodianship. Conclusions We are currently in a familiar situation in which regulation and oversight risk falling behind the technologies they govern. Regulation should emphasize patient agency and consent, and should encourage increasingly sophisticated methods of data anonymization and protection.
Article
The presence of cranial and facial bone fractures is an important finding on non-enhanced head computed tomography (CT) scans from patients who have sustained head trauma. Some prior studies have proposed automatic cranial fracture detections, but studies on facial fractures are lacking. We propose a deep learning system to automatically detect both cranial and facial bone fractures. Our system incorporated models consisting of YOLOv4 for one-stage fracture detection and improved ResUNet (ResUNet++) for the segmentation of cranial and facial bones. The results from the two models mapped together provided the location of the fracture and the name of the fractured bone as the final output. The training data for the detection model were the soft tissue algorithm images from a total of 1,447 head CT studies (a total of 16,985 images), and the training data for the segmentation model included 1,538 selected head CT images. The trained models were tested on a test dataset consisting of 192 head CT studies (a total of 5,890 images). The overall performance achieved a sensitivity of 88.66%, a precision of 94.51%, and an F1 score of 0.9149. Specifically, the cranial and facial regions were evaluated and resulted in a sensitivity of 84.78% and 80.77%, a precision of 92.86% and 87.50%, and F1 scores of 0.8864 and 0.8400, respectively. The average accuracy for the segmentation labels concerning all predicted fracture bounding boxes was 80.90%. Our deep learning system could accurately detect cranial and facial bone fractures and identify the fractured bone region simultaneously.
Article
Craniosynostosis (synostosis) is a serious disease where the sutures of a newborn’s skull fuse prematurely leading to debilitating head shape deformities. Due to the seriousness of this condition many normal infants and those with benign head shape abnormalities are referred to pediatric craniofacial plastic surgeons, leading to a high referral burden and delays in diagnosis for patients. A diagnostic delay beyond 4 months of age excludes patients from being treated with minimally invasive endoscopic procedures, leading to higher risk open surgeries. Machine learning (ML) image classifiers can enhance the triaging process of these referrals through the use of 3D images taken by a multicamera & angle setup during patient visits. In doing so, children with synostosis can be identified earlier, qualifying them for less invasive endoscopic surgical intervention. After training a variety of convolutional neural network (CNN) models on 3D images supplemented with synthetic images using generative adversarial networks (GANs), the best-performing model was found to be a novel approach developed in our study called a multi-view collapsed 3D CNN, which achieved area under the receiver operating curves (AUROC) between 90.00-97.00% for detecting various sub-types of synostosis. These results demonstrate the ability for ML models to potentially streamline the detection of children with synostosis and help overcome challenges associated with high referral burdens for these patients.
Article
Background: Recreation of a spontaneous, emotional smile remains a paramount goal of smile reanimation surgery. However, optimal techniques to reliably restore spontaneity remain unknown. Dual automated machine-learning tools were used to develop an objective tool to analyze spontaneous smiling. The feasibility of this tool was tested in a sample of functional free muscle transfers. Methods: Validated humorous videos were used to elicit spontaneous smiles. Automated facial landmark recognition (Emotrics) and emotion detection software (Affdex) were used to analyze video clips of spontaneous smiling in nine normal subjects and 39 facial reanimation cases. Emotionality quotient was used to quantify the ability of spontaneous smiles to express joy. Results: The software could analyze spontaneous smiling in all subjects. Spontaneous smiles of normal subjects exhibited median 100 percent joy and 0 percent negative emotion (emotional quotient score, +100/0). Spontaneous smiles of facial palsy patients after smile reanimation, using cross-facial nerve graft, masseteric nerve, and dual innervation, yielded median emotional quotient scores of +82/0, 0/-48, and +10/-24 respectively (joy, p = 0.006; negative emotion, p = 0.034). Conclusions: Computer vision software can objectively quantify spontaneous smiling outcomes. Of the retrospective sample of cases reviewed in this study, cross-facial nerve graft-innervated gracilis functional free muscle transfer achieved a greater degree of emotionality during spontaneous smiling than masseteric or dually innervated transfer. Quantification of spontaneous smiling from standard video clips could facilitate future, blinded, multicenter trials with sufficient long-term follow-up to definitively establish the rates of spontaneity from a range of reanimation procedures. Clinical question/level of evidence: Diagnostic, IV.