Institute Affiliations


All Publications


  • Timing of Antiobesity Medications and Adolescent Metabolic and Bariatric Surgery. JAMA surgery Chinn, J. O., Shacker, M., Brennan, K. A., Esquivel, M. M., Pratt, J. S. 2025

    View details for DOI 10.1001/jamasurg.2025.4430

    View details for PubMedID 41123888

    View details for PubMedCentralID PMC12547670

  • Does the MBSAQIP Bariatric Surgical Risk/Benefit Calculator Accurately Predict Weight Loss in Adolescents? Obesity surgery Kochis, M. A., Chinn, J. O., Nzenwa, I. C., Brennan, K. A., Pratt, J. S., Griggs, C. L. 2025

    Abstract

    BACKGROUND: The Metabolic and Bariatric Surgery Accreditation and Quality Improvement Program (MBSAQIP) online calculator incorporates individual patient data to predict weight loss up to 1year after MBS, but it was derived from an adult database and has not been validated in younger cohorts. This study evaluates the accuracy of this calculator for adolescent MBS patients and explores patient factors which may be associated with prediction inaccuracy.METHODS: We include patients age≤21 who underwent laparoscopic sleeve gastrectomy at two major academic institutions from 2013 to 2023. Data were stratified between patients age<18 and 18-21. The calculator's predictions were compared to actual weight loss values at 1year. Relationships between various preoperative variables and the difference between predicted and actual weight loss were assessed using correlation, regression, and t-tests.RESULTS: There were 265 patients, with 176 age<18. The correlation coefficients for predicted and actual weight loss were 0.48 for patients age<18 and 0.38 for patients 18-21. On average, the proportion of predicted weight loss actually attained at 1year was 0.73. There were no statistically significant associations between calculator inaccuracy and patient age, sex, preoperative body mass index, or area deprivation index (all p>0.05).CONCLUSIONS: The MBASQIP calculator predictions show weak to moderate correlation with actual weight loss at 1year and should be used with caution when counseling pediatric patients considering MBS. This project underscores the importance of building multi-institutional collaborations and databases specific to the pediatric MBS context.

    View details for DOI 10.1007/s11695-025-08295-5

    View details for PubMedID 41068350

  • Comparing Phenotypes for Acute and Long-Term Response to Atrial Fibrillation Ablation Using Machine Learning. Circulation. Arrhythmia and electrophysiology Ganesan, P., Pedron, M., Feng, R., Rogers, A. J., Deb, B., Chang, H. J., Ruiperez-Campillo, S., Srivastava, V., Brennan, K. A., Giles, W., Baykaner, T., Clopton, P., Wang, P. J., Schotten, U., Krummen, D. E., Narayan, S. M. 2025: e012860

    Abstract

    BACKGROUND: It is difficult to identify patients with atrial fibrillation (AF) most likely to respond to ablation. While any arrhythmia patient may recur after acutely successful ablation, AF is unusual in that patients may have long-term arrhythmia freedom despite a lack of acute success. We hypothesized that acute and chronic AF ablation outcomes may reflect distinct physiology and used machine learning of multimodal data to identify their phenotypes.METHODS: We studied 561 consecutive patients in the Stanford AF ablation registry (66±10 years, 28% women, 67% nonparoxysmal), from whom we extracted 72 data features of electrograms, electrocardiogram, cardiac structure, lifestyle, and clinical variables. We compared 6 machine learning models to predict acute and long-term end points after ablation and used Shapley explainability analysis to contrast phenotypes. We validated our results in an independent external population of n=77 patients with AF.RESULTS: The 1-year success rate was 69.5%, and the acute termination rate was 49.6%, which correlated poorly on a patient-by-patient basis (phi coefficient=0.08). The best model for acute termination (area under the curve=0.86, Random Forest) was more predictive than for long-term outcomes (area under the curve=0.67, logistic regression; P<0.001). Phenotypes for long-term success reflected clinical and lifestyle features, while phenotypes for AF termination reflected electrical features. The need for AF induction predicted both phenotypes. The external validation cohort showed similar results (area under the curve=0.81 and 0.64, respectively) with similar phenotypes.CONCLUSIONS: Long-term and acute responses to AF ablation reflect distinct clinical and electrical physiology, respectively. This de-linking of phenotypes raises the question of whether long-term success operates through factors such as attenuated AF progression. There remains an urgent need to develop procedural predictors of long-term AF ablation success.

    View details for DOI 10.1161/CIRCEP.124.012860

    View details for PubMedID 39925268

  • Engineering of Generative Artificial Intelligence and Natural Language Processing Models to Accurately Identify Arrhythmia Recurrence. Circulation. Arrhythmia and electrophysiology Feng, R., Brennan, K. A., Azizi, Z., Goyal, J., Deb, B., Chang, H. J., Ganesan, P., Clopton, P., Pedron, M., Ruipe Rez-Campillo, S., Desai, Y., De Larochellière, H., Baykaner, T., Perez, M., Rodrigo, M., Rogers, A. J., Narayan, S. M. 2024: e013023

    Abstract

    Large language models (LLMs), such as ChatGPT, excel at interpreting unstructured data from public sources, yet are limited when responding to queries on private repositories, such as electronic health records (EHRs). We hypothesized that prompt engineering could enhance the accuracy of LLMs for interpreting EHR data without requiring domain knowledge, thus expanding their utility for patients and personalized diagnostics.We designed and systematically tested prompt engineering techniques to improve the ability of LLMs to interpret EHRs for nuanced diagnostic questions, referenced to a panel of medical experts. In 490 full-text EHR notes from 125 patients with prior life-threatening heart rhythm disorders, we asked GPT-4-turbo to identify recurrent arrhythmias distinct from prior events and tested 220 563 queries. To provide context, results were compared with rule-based natural language processing and BERT-based language models. Experiments were repeated for 2 additional LLMs.In an independent hold-out set of 389 notes, GPT-4-turbo had a balanced accuracy of 64.3%±4.7% out-of-the-box at baseline. This increased when asking GPT-4-turbo to provide a rationale for its answers, requiring a structured data output, and providing in-context exemplars, rose to a balanced accuracy of 91.4%±3.8% (P<0.05). This surpassed the traditional logic-based natural language processing and BERT-based models (P<0.05). Results were consistent for GPT-3.5-turbo and Jurassic-2 LLMs.The use of prompt engineering strategies enables LLMs to identify clinical end points from EHRs with an accuracy that surpassed natural language processing and approximated experts, yet without the need for expert knowledge. These approaches could be applied to LLM queries for other domains, to facilitate automated analysis of nuanced data sets with high accuracy by nonexperts.

    View details for DOI 10.1161/CIRCEP.124.013023

    View details for PubMedID 39676642