Catalogue Search | MBRL
Search Results Heading
Explore the vast range of titles available.
MBRLSearchResults
-
DisciplineDiscipline
-
Is Peer ReviewedIs Peer Reviewed
-
Item TypeItem Type
-
SubjectSubject
-
YearFrom:-To:
-
More FiltersMore FiltersSourceLanguage
Done
Filters
Reset
42
result(s) for
"Barash, Yiftach"
Sort by:
Deep learning for pulmonary embolism detection on computed tomography pulmonary angiogram: a systematic review and meta-analysis
2021
Computed tomographic pulmonary angiography (CTPA) is the gold standard for pulmonary embolism (PE) diagnosis. However, this diagnosis is susceptible to misdiagnosis. In this study, we aimed to perform a systematic review of current literature applying deep learning for the diagnosis of PE on CTPA. MEDLINE/PUBMED were searched for studies that reported on the accuracy of deep learning algorithms for PE on CTPA. The risk of bias was evaluated using the QUADAS-2 tool. Pooled sensitivity and specificity were calculated. Summary receiver operating characteristic curves were plotted. Seven studies met our inclusion criteria. A total of 36,847 CTPA studies were analyzed. All studies were retrospective. Five studies provided enough data to calculate summary estimates. The pooled sensitivity and specificity for PE detection were 0.88 (95% CI 0.803–0.927) and 0.86 (95% CI 0.756–0.924), respectively. Most studies had a high risk of bias. Our study suggests that deep learning models can detect PE on CTPA with satisfactory sensitivity and an acceptable number of false positive cases. Yet, these are only preliminary retrospective works, indicating the need for future research to determine the clinical impact of automated PE detection on patient care. Deep learning models are gradually being implemented in hospital systems, and it is important to understand the strengths and limitations of these algorithms.
Journal Article
Deep learning visual analysis in laparoscopic surgery: a systematic review and diagnostic test accuracy meta-analysis
by
Horesh Nir
,
Rosin, Danny
,
Barash Yiftach
in
Artificial intelligence
,
Deep learning
,
Laparoscopy
2021
BackgroundIn the past decade, deep learning has revolutionized medical image processing. This technique may advance laparoscopic surgery. Study objective was to evaluate whether deep learning networks accurately analyze videos of laparoscopic procedures.MethodsMedline, Embase, IEEE Xplore, and the Web of science databases were searched from January 2012 to May 5, 2020. Selected studies tested a deep learning model, specifically convolutional neural networks, for video analysis of laparoscopic surgery. Study characteristics including the dataset source, type of operation, number of videos, and prediction application were compared. A random effects model was used for estimating pooled sensitivity and specificity of the computer algorithms. Summary receiver operating characteristic curves were calculated by the bivariate model of Reitsma.ResultsThirty-two out of 508 studies identified met inclusion criteria. Applications included instrument recognition and detection (45%), phase recognition (20%), anatomy recognition and detection (15%), action recognition (13%), surgery time prediction (5%), and gauze recognition (3%). The most common tested procedures were cholecystectomy (51%) and gynecological—mainly hysterectomy and myomectomy (26%). A total of 3004 videos were analyzed. Publications in clinical journals increased in 2020 compared to bio-computational ones. Four studies provided enough data to construct 8 contingency tables, enabling calculation of test accuracy with a pooled sensitivity of 0.93 (95% CI 0.85–0.97) and specificity of 0.96 (95% CI 0.84–0.99). Yet, the majority of papers had a high risk of bias.ConclusionsDeep learning research holds potential in laparoscopic surgery, but is limited in methodologies. Clinicians may advance AI in surgery, specifically by offering standardized visual databases and reporting.
Journal Article
A Gradient Boosting Machine Learning Model for Predicting Early Mortality in the Emergency Department Triage: Devising a Nine-Point Triage Score
by
Zimlichman Eyal
,
Barash Yiftach
,
Resheff, Yehezkel S
in
Algorithms
,
Artificial intelligence
,
Demographics
2020
BackgroundEmergency departments (ED) are becoming increasingly overwhelmed, increasing poor outcomes. Triage scores aim to optimize the waiting time and prioritize the resource usage. Artificial intelligence (AI) algorithms offer advantages for creating predictive clinical applications.ObjectiveEvaluate a state-of-the-art machine learning model for predicting mortality at the triage level and, by validating this automatic tool, improve the categorization of patients in the ED.DesignAn institutional review board (IRB) approval was granted for this retrospective study. Information of consecutive adult patients (ages 18–100) admitted at the emergency department (ED) of one hospital were retrieved (January 1, 2012–December 31, 2018). Features included the following: demographics, admission date, arrival mode, referral code, chief complaint, previous ED visits, previous hospitalizations, comorbidities, home medications, vital signs, and Emergency Severity Index (ESI). The following outcomes were evaluated: early mortality (up to 2 days post ED registration) and short-term mortality (2–30 days post ED registration). A gradient boosting model was trained on data from years 2012–2017 and examined on data from the final year (2018). The area under the curve (AUC) for mortality prediction was used as an outcome metric. Single-variable analysis was conducted to develop a nine-point triage score for early mortality.Key ResultsOverall, 799,522 ED visits were available for analysis. The early and short-term mortality rates were 0.6% and 2.5%, respectively. Models trained on the full set of features yielded an AUC of 0.962 for early mortality and 0.923 for short-term mortality. A model that utilized the nine features with the highest single-variable AUC scores (age, arrival mode, chief complaint, five primary vital signs, and ESI) yielded an AUC of 0.962 for early mortality.ConclusionThe gradient boosting model shows high predictive ability for screening patients at risk of early mortality utilizing data available at the time of triage in the ED.
Journal Article
Towards AI-Augmented Clinical Decision-Making: An Examination of ChatGPT's Utility in Acute Ulcerative Colitis Presentations
by
Levartovsky, Asaf
,
Barash, Yiftach
,
Ben-Horin, Shomron
in
Artificial Intelligence
,
Chatbots
,
Classification
2023
This study explores the potential of OpenAI's ChatGPT as a decision support tool for acute ulcerative colitis presentations in the setting of an emergency department. We assessed ChatGPT's performance in determining disease severity using TrueLove and Witts criteria and the necessity of hospitalization for patients with ulcerative colitis, comparing results with those of expert gastroenterologists. Of 20 cases, ChatGPT's assessments were found to be 80% consistent with gastroenterologist evaluations and indicated a high degree of reliability. This suggests that ChatGPT could provide as a clinical decision support tool in assessing acute ulcerative colitis, serving as an adjunct to clinical judgment.
Journal Article
Multimodal fusion models for pulmonary embolism mortality prediction
by
Greenspan, Hayit
,
Burshtein, Evyatar
,
Marom, Edith M.
in
631/114/1305
,
631/114/1564
,
631/114/2413
2023
Pulmonary embolism (PE) is a common, life threatening cardiovascular emergency. Risk stratification is one of the core principles of acute PE management and determines the choice of diagnostic and therapeutic strategies. In routine clinical practice, clinicians rely on the patient’s electronic health record (EHR) to provide a context for their medical imaging interpretation. Most deep learning models for radiology applications only consider pixel-value information without the clinical context. Only a few integrate both clinical and imaging data. In this work, we develop and compare multimodal fusion models that can utilize multimodal data by combining both volumetric pixel data and clinical patient data for automatic risk stratification of PE. Our best performing model is an intermediate fusion model that incorporates both bilinear attention and TabNet, and can be trained in an end-to-end manner. The results show that multimodality boosts performance by up to 14% with an area under the curve (AUC) of 0.96 for assessing PE severity, with a sensitivity of 90% and specificity of 94%, thus pointing to the value of using multimodal data to automatically assess PE severity.
Journal Article
Large language model (ChatGPT) as a support tool for breast tumor board
by
Sklair-Levy, Miri
,
Balint Lahat, Nora
,
Zippel, Douglas B
in
Artificial intelligence
,
Breast cancer
,
Chatbots
2023
Large language models (LLM) such as ChatGPT have gained public and scientific attention. The aim of this study is to evaluate ChatGPT as a support tool for breast tumor board decisions making. We inserted into ChatGPT-3.5 clinical information of ten consecutive patients presented in a breast tumor board in our institution. We asked the chatbot to recommend management. The results generated by ChatGPT were compared to the final recommendations of the tumor board. They were also graded independently by two senior radiologists. Grading scores were between 1–5 (1 = completely disagree, 5 = completely agree), and in three different categories: summarization, recommendation, and explanation. The mean age was 49.4, 8/10 (80%) of patients had invasive ductal carcinoma, one patient (1/10, 10%) had a ductal carcinoma in-situ and one patient (1/10, 10%) had a phyllodes tumor with atypia. In seven out of ten cases (70%), ChatGPT’s recommendations were similar to the tumor board’s decisions. Mean scores while grading the chatbot’s summarization, recommendation and explanation by the first reviewer were 3.7, 4.3, and 4.6 respectively. Mean values for the second reviewer were 4.3, 4.0, and 4.3, respectively. In this proof-of-concept study, we present initial results on the use of an LLM as a decision support tool in a breast tumor board. Given the significant advancements, it is warranted for clinicians to be familiar with the potential benefits and harms of the technology.
Journal Article
Findings on emergent magnetic resonance imaging in pregnant patients with suspected appendicitis: A single center perspective
by
Barash, Yiftach
,
Mashiach, Roy
,
Bufman, Hila
in
Appendicitis
,
Diagnosis
,
Magnetic resonance imaging
2024
This study’s aim is to describe the imaging findings in pregnant patients undergoing emergent MRI for suspected acute appendicitis, and the various alternative diagnoses seen on those MRI scans. This is a single center retrospective analysis in which we assessed the imaging, clinical and pathological data for all consecutive pregnant patients who underwent emergent MRI for suspected acute appendicitis between April 2013 and June 2021. Out of 167 patients, 35 patients (20.9%) were diagnosed with acute appendicitis on MRI. Thirty patients (18%) were diagnosed with an alternative diagnosis on MRI: 17/30 (56.7%) patients had a gynecological source of abdominal pain (e.g. ectopic pregnancy, red degeneration of a leiomyoma); 8 patients (26.7%) had urological findings such as pyelonephritis; and 6 patients (20%) had gastrointestinal diagnoses (e.g. abdominal wall hernia or inflammatory bowel disease). Our conclusions are that MRI is a good diagnostic tool in the pregnant patient, not only in diagnosing acute appendicitis, but also in providing information on alternative diagnoses to acute abdominal pain. Our findings show the various differential diagnoses on emergent MRI in pregnant patients with suspected acute appendicitis, which may assist clinicians and radiologists is patient assessment and imaging utilization.
Journal Article
Integrated visual and text-based analysis of ophthalmology clinical cases using a large language model
2025
Recent advancements in generative artificial intelligence have enabled analysis of text with visual data, which could have important implications in healthcare. Diagnosis in ophthalmology is often based on a combination of ocular examination, and clinical context. The aim of this study was to evaluate the performance of multimodal GPT-4 (GPT-4 V) in an integrated analysis of ocular images and clinical text. This retrospective study included 40 patients seen in our institution with images of their ocular examinations. Cases were selected by a board-certified ophthalmologist, to represent various pathologies. We provided the model with each patient image, without and then with the clinical context. We also asked two non-ophthalmology physicians to write diagnoses for each image, without and then with the clinical context. Answers for both GPT-4 V and the non-ophthalmologists were evaluated by two board-certified ophthalmologists. Performance accuracies were calculated and compared. GPT-4 V provided the correct diagnosis in 19/40 (47.5%) cases based on images without clinical context, and in 27/40 (67.5%) cases when clinical context was provided. Non-ophthalmologist physicians provided the correct diagnoses in 24/40 (60.0%), and 23/40 (57.5%) of cases without clinical context, and in 29/40 (72.5%) and 27/40 (67.5%) with clinical context. For all study participants adding context improved accuracy (
p
= 0.033). GPT-4 V is currently able to simultaneously analyze and integrate visual and textual data, and arrive at accurate clinical diagnoses in the majority of cases. Multimodal large language models like GPT-4 V have significant potential to advance both patient care and research in ophthalmology.
Journal Article
Diagnostic angiography for identification and management of late vascular injuries in war-related traumatic peripheral vascular injuries: A retrospective cohort study
2025
One of the feared complications of war-related peripheral vascular injury is the development of delayed hemorrhage. This study describes our experience with an innovative protocol of surveillance diagnostic angiography to detect occult late vascular complications in an effort to prevent delayed hemorrhage.
This retrospective cohort study was conducted at a single level one trauma center, reviewing patients with war-related peripheral vascular injuries caused by penetrating trauma from October 7th, 2023, to January 21st, 2024. Data collected included patient demographics, primary injury characteristics, associated complications, incidence of late vascular injuries (either symptomatic or occult), means of diagnosis, treatment strategies and outcomes.
The cohort included 41 patients with war-related peripheral vascular injuries affecting 51 limbs. All patients were male (100%) with a median age of 25 years, the majority being soldiers (85%). 24 occurrences of late vascular injuries were observed in 22 (43%) out of 51 limbs (100%). Half were symptomatic, with delayed hemorrhage occurring in 5 limbs in total (10%), and half were asymptomatic. A total of 17 surveillance diagnostic angiographies were performed with the sole indication of identifying occult late vascular injuries in asymptomatic patients, of which 4 (24%) were positive for findings. Five additional diagnostic angiographies were performed to assess late injuries discovered incidentally on imaging studies that were performed for other indications, and all were positive for late vascular injuries. Of all late vascular injuries, a total of 83% required subsequent treatment.
Late vascular injuries are a potentially lethal complication of war-related peripheral vascular injury. Aggressive surveillance with diagnostic angiography prior to discharge from a high intensity care unit can detect asymptomatic late vascular injuries, the treatment of which may prevent life-threatening hemorrhage.
Journal Article
AI in the ED: Assessing the efficacy of GPT models vs. physicians in medical score calculation
2024
Artificial Intelligence (AI) models like GPT-3.5 and GPT-4 have shown promise across various domains but remain underexplored in healthcare. Emergency Departments (ED) rely on established scoring systems, such as NIHSS and HEART score, to guide clinical decision-making. This study aims to evaluate the proficiency of GPT-3.5 and GPT-4 against experienced ED physicians in calculating five commonly used medical scores.
This retrospective study analyzed data from 150 patients who visited the ED over one week. Both AI models and two human physicians were tasked with calculating scores for NIH Stroke Scale, Canadian Syncope Risk Score, Alvarado Score for Acute Appendicitis, Canadian CT Head Rule, and HEART Score. Cohen's Kappa statistic and AUC values were used to assess inter-rater agreement and predictive performance, respectively.
The highest level of agreement was observed between the human physicians (Kappa = 0.681), while GPT-4 also showed moderate to substantial agreement with them (Kappa values of 0.473 and 0.576). GPT-3.5 had the lowest agreement with human scorers. These results highlight the superior predictive performance of human expertise over the currently available automated systems for this specific medical outcome. Human physicians achieved a higher ROC-AUC on 3 of the 5 scores, but none of the differences were statistically significant.
While AI models demonstrated some level of concordance with human expertise, they fell short in emulating the complex clinical judgments that physicians make. The study suggests that current AI models may serve as supplementary tools but are not ready to replace human expertise in high-stakes settings like the ED. Further research is needed to explore the capabilities and limitations of AI in emergency medicine.
Journal Article