Skip to main content

MINI REVIEW

J. Abdom. Wall Surg., 06 September 2024
This article is part of the Special Issue Minimally Invasive Repair of Ventral and Incisional Hernias (ipom, e-TEP, v-TAPP, MILOS, etc) View all 3 articles

Artificial Intelligence—What to Expect From Machine Learning and Deep Learning in Hernia Surgery

Robert Vogel
Robert Vogel*Bjrn Mück
Björn Mück*
  • Klinikum Kempten - Klinikverbund Allgäu, Kempten, Germany

This mini-review explores the integration of Artificial Intelligence (AI) within hernia surgery, highlighting the role of Machine Learning (ML) and Deep Learning (DL). The term AI incorporates various technologies including ML, Neural Networks (NN), and DL. Classical ML algorithms depend on structured, labeled data for predictions, requiring significant human oversight. In contrast, DL, a subset of ML, generally leverages unlabeled, raw data such as images and videos to autonomously identify patterns and make intricate deductions. This process is enabled by neural networks used in DL, where hidden layers between the input and output capture complex data patterns. These layers’ configuration and weighting are pivotal in developing effective models for various applications, such as image and speech recognition, natural language processing, and more specifically, surgical procedures and outcomes in hernia surgery. Significant advancements have been achieved with DL models in surgical settings, particularly in predicting the complexity of abdominal wall reconstruction (AWR) and other postoperative outcomes, which are elaborated in detail within the context of this mini-review. The review method involved analyzing relevant literature from databases such as PubMed and Google Scholar, focusing on studies related to preoperative planning, intraoperative techniques, and postoperative management within hernia surgery. Only recent, peer-reviewed publications in English that directly relate to the topic were included, highlighting the latest advancements in the field to depict potential benefits and current limitations of AI technologies in hernia surgery, advocating for further research and application in this evolving field.

Introduction

“A computer would deserve to be called intelligent if it could deceive a human into believing that it was human” [1] a phrase once coined by Alan Mathison Turing, who is widely regarded as a pioneer in computer science and artificial intelligence. Turing’s early substantial contributions to AI in the mid-20th century, notably his role in deciphering the Enigma code during World War II, laid the foundation for the field [2].

Adopting Turing’s quote to contemporary times, one could identify any device capable of producing realistic videos, photos, and conversations as intelligent since artificial Intelligence (AI) has reached a level of sophistication where it can convincingly mimic human behavior. As a result, the term AI is now used ubiquitously as it is entangled within a vast area in modern computer science. Consequently, related terms are often incorrectly used interchangeably [3]. For simplicity reasons, AI can be considered an umbrella term that includes concepts such as machine learning (ML), neural networks (NN), and deep learning (DL) (Figure 1).

Figure 1
www.frontiersin.org

Figure 1. Schematic depicting the relationship between AI, artificial intelligence; ML, machine learning; DL, deep learning; NN, neural networks.

Many of these technologies have naturally also found applications in medicine, including hernia surgery. Machine learning algorithms (MLA) for example can, in simple terms, learn from existing data, generalize them, and extrapolate future, yet unknown data, allowing them to perform tasks without explicit instructions [4]. For instance, Hassan et al [5] trained an MLA on preoperative clinical data to accurately predict complications in abdominal wall reconstruction (AWR) surgery.

Classical machine learning algorithms typically rely on structured, labeled data (i.e., data organized according to a set hierarchy) to make predictions, making them generally more dependent on human intervention for learning. In contrast, deep learning (DL), being a subset of machine learning, does not necessarily require labeled datasets. Deep Learning algorithms can be regarded both as a refined as well as mathematically complex evolution of machine learning algorithms. To put it illustratively, just as a human brain uses its neural pathways to process information and reach decisions, deep learning employs artificial neural networks that simulate this process. These networks are capable of learning from data in an incremental manner, which enables them to make complex deductions as more information becomes available. It can utilize raw data, such as text, images, and videos, to identify features and distinguish explicit patterns, thus discovering data groupings without human interference [68].

The distinction between ML and DL is defined by the number of layers in a neural network also known as hidden layers (Figure 2) [9]. If more than three of these layers are present, the algorithm is considered a DL model. In summary, hidden layers act as intermediary stages between the input and output of a neural network. They play a crucial role in capturing complex patterns in data, which makes neural networks highly effective for diverse applications such as image and speech recognition and natural language processing. The configuration and weighting of these hidden layers are essential for creating effective neural network models. In particular, DLMs (deep learning models) involved in computer-aided diagnosis have been successfully applied in cranial [10, 11] trauma [12, 13] and oncologic [1416] computed tomography (CT) analysis. In this context, Elhage et al. [17] have demonstrated that an 8-layered convolutional neural network (CNN), a type of DL architecture, can effectively predict surgical complexity in AWR procedures, competing with a panel of expert surgeons. Furthermore, DLMs have proven to be effective in predicting surgical outcomes and post-operative complications [1820].

Figure 2
www.frontiersin.org

Figure 2. Schematic visualization of a deep learning model with input layer, n + 1 hidden layers and output layer).

The aim of this mini-review is to explore the advancements and applications of AI, particularly ML and DL, as well as other notable advancements in computing science in the field of hernia surgery, highlighting their potential benefits and current limitations.

Methodology

In conducting this mini review, we utilized literature from several databases including, PubMed and Google Scholar. The key terms used in our search were: “hernia,” “artificial intelligence, “deep learning,” “machine learning,” “abdominal wall reconstruction,” “hernia surgery,” and “augmented reality.”

Inclusion criteria for this review involved publications directly relevant to the study topic. Specifically, we included studies that focused on preoperative planning, intraoperative imaging and techniques, as well as postoperative management and follow-up in the context of abdominal wall reconstruction. We considered articles published in peer-reviewed journals, studies published in English, and publications from the last 10 years to ensure the inclusion of the most recent advancements.

Publications were excluded if they were not related to hernia surgery, studies not involving AI, ML, DL, or augmented reality in a surgical context, and non-peer-reviewed articles such as opinion pieces, editorials, or non-scientific reports. We also excluded articles focusing on animals or in vitro models rather than human subjects, and duplicate studies or publications with overlapping data sets.

Discussion

The potential impact of computing sciences on medicine was already being anticipated in the 1970s. W B Schwartz [21] predicted that “computing science would likely have significant effects by enhancing and possibly even replacing certain intellectual functions traditionally performed by physicians.” The author noted that the integration of computers in medical practice could profoundly influence physician manpower and the quality of healthcare. Schwartz’s foresight highlighted the transformative potential of computing science in medicine. 52 years later in 2022, a protocol was developed by Saeidi et al. [22] that enables a robot to autonomously perform a small bowel anastomosis with minimal human interference.

The rapid advancements in AI hence raise the question of not whether AI will shape our surgical future, but rather how it will do so.

In a quality improvement study conducted by Elhage et al. [17], the researchers aimed to assess the potential of image-based deep learning in predicting the complexity of AWR surgeries, specifically the need for component separation, as well as predicting pulmonary and wound complications. To achieve this, they developed an 8-layer CNN capable of analysing image characteristics.

The study design involved a comparison of the CNN-based surgical complexity model with a validation set of CT-images. The validation set was evaluated by a panel of 6 expert AWR surgeons, who were blinded to the surgical complexity DLM. The dataset utilised for analysis comprised 369 patients and 9303 CT images.

In summary, Elhage et al. found that the CNN-based DLM was more accurate than expert surgeon judgment in predicting the surgical complexity of AWR procedures [81.3% compared to the surgeons’ accuracy of 65.0% (p < 0.001)]. Furthermore, they observed that an additional DLM accurately predicted the occurrence of surgical site infections [AUC 0.898 (p < 0.001)]. These findings highlight the potential of image-based DLMs as valuable tools in forecasting surgical outcomes and improving decision-making in AWR surgery.

In another article published by Hassan et al. [5], the authors displayed the effectiveness of machine learning models (MLMs) in predicting hernia recurrence (HR), surgical site occurrences (SSOs), and 30-day readmission. Their study reported that MLMs achieved mean accuracy rates of 85% (95% CI 80%–90%) for HR prediction, 72% (95% CI 64%–80%) for SSOs, and 84% (95% CI 77%–90%) for 30-day readmission. These ML algorithms, trained on readily available preoperative clinical data, proved to be highly accurate in forecasting complications associated with AWR-surgery. The authors concluded to support the integration of MLMs into the preoperative evaluation process for patients undergoing AWR.

To facilitate the advancements of AI in hernia surgery, it is in our view crucial to expand patient databases on an international scale. Increasing the diversity and size of these databases will provide a broader range of data for training AI models and allow for more comprehensive and accurate predictions.

Since the effectiveness of AI predictions relies heavily on both the accuracy and the thoroughness of the input data, access to diverse patient populations will help address potential biases and ensure the reliability of AI algorithms across different demographics [23]. Expanding patient databases internationally can also help identify patterns and trends that may not be apparent in smaller or more localised datasets.

Moreover, the collection and analysis of video data in surgical procedures hold great potential for advancing the field of hernia surgery. By capturing surgical procedures through video recordings, it becomes possible to create detailed anatomical maps and explore the integration of AR with robotics in surgical interventions [24]. By integrating DLMs on video data, it appears to be possible to extract relevant information, such as anatomical landmarks, tissue characteristics, and procedural phase recognition, which can contribute to the development of more precise surgical interventions [25]. Additionally, the combination of video data with augmented reality has the potential to enhance surgical visualisation and navigation as well as robotic surgery training-simulators. By possibly overlaying real-time anatomical information and guidance onto the surgeon’s view, AR could provide valuable assistance during complex hernia surgeries. Surgeons can benefit from visual cues, real-time feedback, and enhanced precision, ultimately leading to improved surgical outcomes. For example, Cui et al. [26] used a CNN model comprised out of surgical videos from 35 patients in laparoscopic hernia repair in order to detect the vas deferens and their results suggested that the CNN promptly identifies and visualises vas deferens images.

Another promising approach combines wide-field, planar, near-infrared fluorescence imaging with AI for automated real-time guidance during surgery as highlighted by Gioux et al. [27]. This technology could help identify and avoid hidden tissues, such as nerves or blood vessels covered by fatty or connective tissue, by highlighting their location within the surgical field. Like following a breadcrumb trail, the AI system could guide surgeons in the direction of dissection, enhancing precision and hence reducing the risk of complications.

An essential pillar of AI is characterized through the concept of phase recognition. This involves the utilization of MLAs and computer vision techniques to automatically discern and categorize distinct phases or stages within a process.

Phase recognition AI utilizes algorithms, frequently rooted in DL, to scrutinize patterns, features, and temporal sequences within video data. This capability allows the system to differentiate between diverse stages of the surgical process [28, 29].

In a recently published study by Takeuchi et al. [30], the primary objective was to develop a DL-based automated phase-recognition system for identifying surgical phases in Transabdominal Preperitoneal (TAPP) procedures (i.e. preparation, peritoneal flap incision, peritoneal flap dissection, hernia dissection, mesh deployment, mesh fixation, peritoneal flap closure, and additional closure). A secondary aim was to explore the correlation between surgical skills and the duration of each phase. An AI model (AIM) was trained to automatically recognize surgical phases from videos, and the study assessed the relationship between phase duration and surgical skills. A fourfold cross-validation was used to evaluate the AIMs performance, achieving accuracies of 88.81% and 85.82% for unilateral and bilateral cases, respectively.

Ortenzi et al. [31] reported on an AI-based computer vision algorithm designed to automatically recognize surgical steps in videos of totally extraperitoneal (TEP) inguinal hernia repair. The algorithm achieved an overall accuracy of 88.8% in recognizing the complete procedure. The per-step accuracy was highest for the hernia sac reduction step at 94.3% and lowest for the preperitoneal dissection step at 72.2%. The authors concluded that this novel AIM could provide fully automated video analysis with a high level of accuracy. High-accuracy AIMs that enable automation of surgical video analysis allow for the identification and evaluation of surgical performance.

However, the integration of video data and DLMs in hernia surgery research and practice requires the accumulation of diverse and high-quality video datasets. It seems obvious, that these datasets should encompass various surgical techniques, patient characteristics, and procedural variations to ensure the robustness and generalisability of AIMs [32]. To optimize this process, Hashimoto et al. [33] propose extensive collaborations among surgeons as well as data scientists. These efforts are crucial to facilitate the sharing and pooling of video data, which will accelerate the development and refinement of AI-driven approaches in hernia surgery.

In a scoping review published by Taha et al. [34], the authors provided a comprehensive summary of the current objectives related to the integration of AI in the field of hernia surgery. They highlighted the potential applications and benefits of AI in areas such as medical imaging and surgical training. However, the authors also acknowledged the limited number of publications available on this specific topic, indicating a gap in the existing literature.

Based on this observation, Taha et al. emphasised the need for further research and the publication of original articles to explore and investigate the ways in which AI can effectively assist in medical imaging and support the training of surgeons in the context of hernia surgery.

Furthermore, ever since AI has emerged, ethical considerations have played a significant role in discussions surrounding the technology. While the potential benefits are undeniable, questions arise regarding accountability if complications occur. O’Sullivan et al. [35] categorise responsibility into accountability, liability, and culpability. Drawing a parallel to self-driving cars, they compare the surgeon overseeing the hypothetical autonomous robot to the driver of a car, making the surgeon ultimately responsible for the robot’s actions. Undeniably, a robust legal framework is paramount prior to implementing semi-autonomous programs or machines into any medical field.

In conclusion, further research on the applications of AI is undoubtedly of crucial importance. Moreover, we believe that promoting international collaboration in expanding patient and video databases are indispensable in facilitating this process.

Author Contributions

RV and BM contributed equally to this work; RV and BM drafted the manuscript; RV and BM are responsible for the interpretation of results and the critical revision of the manuscript. All authors contributed to the article and approved the submitted version.

Funding

The author(s) declare that no financial support was received for the research, authorship, and/or publication of this article.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. School of Mathematics and Statistics. Quotations by Alan Turing. United Kingdom: University of St Andrews (2024). Available from: https://mathshistory.st-andrews.ac.uk/Biographies/Turing/quotations/ (Accessed July 02, 2024).

Google Scholar

2. Encyclopædia Britannica, Inc. Artificial Intelligence. Chicago, IL, United States: Encyclopædia Britannica (2024). Available from: https://www.britannica.com/technology/artificial-intelligence/Alan-Turing-and-the-beginning-of-AI (Accessed July 02, 2024).

Google Scholar

3. Graziani, M, Dutkiewicz, L, Calvaresi, D, Amorim, JP, Yordanova, K, Vered, M, et al. A Global Taxonomy of Interpretable AI: Unifying the Terminology for the Technical and Social Sciences. Artif intelligence Rev (2023) 56(4):3473–504. doi:10.1007/s10462-022-10256-8

CrossRef Full Text | Google Scholar

4. Bell, J. What Is Machine Learning? In: Machine Learning and the City: Applications in Architecture and Urban Design (2022). p. 207–16.

CrossRef Full Text | Google Scholar

5. Hassan, AM, Lu, SC, Asaad, M, Liu, J, Offodile, AC, Sidey-Gibbons, C, et al. Novel Machine Learning Approach for the Prediction of Hernia Recurrence, Surgical Complication, and 30-Day Readmission After Abdominal Wall Reconstruction. J Am Coll Surg (2022) 234(5):918–27. doi:10.1097/XCS.0000000000000141

PubMed Abstract | CrossRef Full Text | Google Scholar

6. Choi, RY, Coyner, AS, Kalpathy-Cramer, J, Chiang, MF, and Campbell, JP. Introduction to Machine Learning, Neural Networks, and Deep Learning. Translational Vis Sci Technol (2020) 9(2):14. doi:10.1167/tvst.9.2.14

PubMed Abstract | CrossRef Full Text | Google Scholar

7. LeCun, Y, Bengio, Y, and Hinton, G. Deep Learning. Nature (2015) 521(7553):436–44. doi:10.1038/nature14539

PubMed Abstract | CrossRef Full Text | Google Scholar

8. Ghahramani, Z. Unsupervised Learning. In: Advanced Lectures on Machine Learning. Berlin, Heidelberg: Springer (2004). p. 72–112.

CrossRef Full Text | Google Scholar

9. IBM. AI vs. Machine Learning vs. Deep Learning vs. Neural Networks. Armonk, NY, United States: IBM (2023). Available from: https://www.ibm.com/think/topics/ai-vs-machine-learning-vs-deep-learning-vs-neural-networks (Accessed July 02, 2024).

Google Scholar

10. Gao, XW, Hui, R, and Tian, Z. Classification of CT Brain Images Based on Deep Learning Networks. Comput Methods Programs Biomed (2017) 138:49–56. doi:10.1016/j.cmpb.2016.10.007

PubMed Abstract | CrossRef Full Text | Google Scholar

11. Chilamkurthy, S, Ghosh, R, Tanamala, S, Biviji, M, Campeau, NG, Venugopal, VK, et al. Deep Learning Algorithms for Detection of Critical Findings in Head CT Scans: A Retrospective Study. Lancet (2018) 392(10162):2388–96. doi:10.1016/S0140-6736(18)31645-3

PubMed Abstract | CrossRef Full Text | Google Scholar

12. Cheng, CT, Lin, HH, Hsu, CP, Chen, HW, Huang, JF, Hsieh, CH, et al. Deep Learning for Automated Detection and Localization of Traumatic Abdominal Solid Organ Injuries on CT Scans. J Imaging Inform Med (2024) 37:1113–23. doi:10.1007/s10278-024-01038-5

PubMed Abstract | CrossRef Full Text | Google Scholar

13. Wang, J, Wood, A, Gao, C, Najarian, K, and Gryak, J. Automated Spleen Injury Detection Using 3D Active Contours and Machine Learning. Entropy (2021) 23(4):382. doi:10.3390/e23040382

PubMed Abstract | CrossRef Full Text | Google Scholar

14. Cheng, JZ, Ni, D, Chou, YH, Qin, J, Tiu, CM, Chang, YC, et al. Computer-Aided Diagnosis With Deep Learning Architecture: Applications to Breast Lesions in US Images and Pulmonary Nodules in CT Scans. Sci Rep (2016) 6:24454. doi:10.1038/srep24454

PubMed Abstract | CrossRef Full Text | Google Scholar

15. Gao, Y, Geras, KJ, Lewin, AA, and Moy, L. New Frontiers: An Update on Computer-Aided Diagnosis for Breast Imaging in the Age of Artificial Intelligence. Am J Roentgenology (2019) 212(2):300–7. doi:10.2214/AJR.18.20392

PubMed Abstract | CrossRef Full Text | Google Scholar

16. Hawkins, S, Wang, H, Liu, Y, Garcia, A, Stringfield, O, Krewer, H, et al. Predicting Malignant Nodules From Screening CT Scans. J Thorac Oncol (2016) 11(12):2120–8. doi:10.1016/j.jtho.2016.07.002

PubMed Abstract | CrossRef Full Text | Google Scholar

17. Elhage, SA, Deerenberg, EB, Ayuso, SA, Murphy, KJ, Shao, JM, Kercher, KW, et al. Development and Validation of Image-Based Deep Learning Models to Predict Surgical Complexity and Complications in Abdominal Wall Reconstruction. JAMA Surg (2021) 156(10):933–40. doi:10.1001/jamasurg.2021.3012

PubMed Abstract | CrossRef Full Text | Google Scholar

18. Xue, B, Li, D, Lu, C, King, CR, Wildes, T, Avidan, MS, et al. Use of Machine Learning to Develop and Evaluate Models Using Preoperative and Intraoperative Data to Identify Risks of Postoperative Complications. JAMA Netw open (2021) 4(3):e212240. doi:10.1001/jamanetworkopen.2021.2240

PubMed Abstract | CrossRef Full Text | Google Scholar

19. Bonde, A, Varadarajan, KM, Bonde, N, Troelsen, A, Muratoglu, OK, Malchau, H, et al. Assessing the Utility of Deep Neural Networks in Predicting Postoperative Surgical Complications: A Retrospective Study. The Lancet Digital Health (2021) 3(8):e471–e485. doi:10.1016/S2589-7500(21)00084-4

PubMed Abstract | CrossRef Full Text | Google Scholar

20. Ayuso, SA, Elhage, SA, Zhang, Y, Aladegbami, BG, Gersin, KS, Fischer, JP, et al. Predicting Rare Outcomes in Abdominal Wall Reconstruction Using Image-Based Deep Learning Models. Surgery (2023) 173(3):748–55. doi:10.1016/j.surg.2022.06.048

PubMed Abstract | CrossRef Full Text | Google Scholar

21. Schwartz, WB. Medicine and the Computer. The Promise and Problems of Change. New Engl J Med (1970) 283(23):1257–64. doi:10.1056/NEJM197012032832305

PubMed Abstract | CrossRef Full Text | Google Scholar

22. Saeidi, H, Opfermann, JD, Kam, M, Wei, S, Leonard, S, Hsieh, MH, et al. Autonomous Robotic Laparoscopic Surgery for Intestinal Anastomosis. Sci Robot (2022) 7(62):eabj2908. doi:10.1126/scirobotics.abj2908

PubMed Abstract | CrossRef Full Text | Google Scholar

23. Murthy, VH, Krumholz, HM, and Gross, CP. Participation in Cancer Clinical Trials: Race-Sex-And Age-Based Disparities. Jama (2004) 291(22):2720–6. doi:10.1001/jama.291.22.2720

PubMed Abstract | CrossRef Full Text | Google Scholar

24. Seetohul, J, Shafiee, M, and Sirlantzis, K. Augmented Reality (AR) for Surgical Robotic and Autonomous Systems: State of the Art, Challenges, and Solutions. Sensors (Basel) (2023) 23(13):6202. doi:10.3390/s23136202

PubMed Abstract | CrossRef Full Text | Google Scholar

25. Bar, O, Neimark, D, Zohar, M, Hager, GD, Girshick, R, Fried, GM, et al. Impact of Data on Generalization of AI for Surgical Intelligence Applications. Sci Rep (2020) 10:22208. doi:10.1038/s41598-020-79173-6

PubMed Abstract | CrossRef Full Text | Google Scholar

26. Cui, P, Zhao, S, and Chen, W. Identification of the Vas Deferens in Laparoscopic Inguinal Hernia Repair Surgery Using the Convolutional Neural Network. J Healthc Eng (2021) 2021:5578089. doi:10.1155/2021/5578089

PubMed Abstract | CrossRef Full Text | Google Scholar

27. Gioux, S, Choi, HS, and Frangioni, JV. Image-Guided Surgery Using Invisible Near-Infrared Light: Fundamentals of Clinical Translation. Mol Imaging (2010) 9(5):237–55. doi:10.2310/7290.2010.00034

PubMed Abstract | CrossRef Full Text | Google Scholar

28. Meeuwsen, FC, van Luyn, F, Blikkendaal, MD, Jansen, FW, and van den Dobbelsteen, JJ. Surgical Phase Modelling in Minimal Invasive Surgery. Surg Endosc (2019) 33(5):1426–32. doi:10.1007/s00464-018-6417-4

PubMed Abstract | CrossRef Full Text | Google Scholar

29. Ramesh, S, Dall’Alba, D, Gonzalez, C, Yu, T, Mascagni, P, Mutter, D, et al. Multi-Task Temporal Convolutional Networks for Joint Recognition of Surgical Phases and Steps in Gastric Bypass Procedures. Int J Comput Assist Radiol Surg (2021) 16:1111–9. doi:10.1007/s11548-021-02388-z

PubMed Abstract | CrossRef Full Text | Google Scholar

30. Takeuchi, M, Collins, T, Ndagijimana, A, Kawakubo, H, Kitagawa, Y, Marescaux, J, et al. Automatic Surgical Phase Recognition in Laparoscopic Inguinal Hernia Repair With Artificial Intelligence. Hernia (2022) 26(6):1669–78. doi:10.1007/s10029-022-02621-x

PubMed Abstract | CrossRef Full Text | Google Scholar

31. Ortenzi, M, Rapoport Ferman, J, Antolin, A, Bar, O, Zohar, M, Perry, O, et al. A Novel High Accuracy Model for Automatic Surgical Workflow Recognition Using Artificial Intelligence in Laparoscopic Totally Extraperitoneal Inguinal Hernia Repair (TEP). Surg Endosc (2023) 37(11):8818–28. doi:10.1007/s00464-023-10375-5

PubMed Abstract | CrossRef Full Text | Google Scholar

32. Gianfrancesco, MA, Tamang, S, Yazdany, J, and Schmajuk, G. Potential Biases in Machine Learning Algorithms Using Electronic Health Record Data. JAMA Intern Med (2018) 178(11):1544–7. doi:10.1001/jamainternmed.2018.3763

PubMed Abstract | CrossRef Full Text | Google Scholar

33. Hashimoto, DA, Rosman, G, Rus, D, and Meireles, OR. Artificial Intelligence in Surgery: Promises and Perils. Ann Surg (2018) 268(1):70–6. doi:10.1097/SLA.0000000000002693

PubMed Abstract | CrossRef Full Text | Google Scholar

34. Taha, A, Enodien, B, Frey, DM, and Taha-Mehlitz, S. The Development of Artificial Intelligence in Hernia Surgery: A Scoping Review. Front Surg (2022) 9:908014. doi:10.3389/fsurg.2022.908014

PubMed Abstract | CrossRef Full Text | Google Scholar

35. O'Sullivan, S, Nevejans, N, Allen, C, Blyth, A, Leonard, S, Pagallo, U, et al. Legal, Regulatory, and Ethical Frameworks for Development of Standards in Artificial Intelligence (AI) and Autonomous Robotic Surgery. Int J Med Robot (2019) 15(1):e1968. doi:10.1002/rcs.1968

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: AI, machine learning, deep learning, hernia surgery, neural network

Citation: Vogel R and Mück B (2024) Artificial Intelligence—What to Expect From Machine Learning and Deep Learning in Hernia Surgery. J. Abdom. Wall Surg. 3:13059. doi: 10.3389/jaws.2024.13059

Received: 29 March 2024; Accepted: 26 July 2024;
Published: 06 September 2024.

Copyright © 2024 Vogel and Mück. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Robert Vogel, robert.vogel@klinikverbund-allgaeu.de; Björn Mück, bjoern.mueck@klinikverbund-allgaeu.de

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.