﻿<?xml version="1.0" encoding="UTF-8"?>
<ArticleSet>
  <Article>
    <Journal>
      <PublisherName>Academy of Medical Sciences of I.R. Iran</PublisherName>
      <JournalTitle>Archives of Iranian Medicine</JournalTitle>
      <Issn>1029-2977</Issn>
      <Volume>28</Volume>
      <Issue>12</Issue>
      <PubDate PubStatus="ppublish">
        <Year>2025</Year>
        <Month>12</Month>
        <DAY>01</DAY>
      </PubDate>
    </Journal>
    <ArticleTitle>Performance of ChatGPT and Gemini Compared with Emergency Physicians in NSTEMI Cases: A Prospective Cross-sectional Study</ArticleTitle>
    <FirstPage>696</FirstPage>
    <LastPage>702</LastPage>
    <ELocationID EIdType="doi">10.34172/aim.35274</ELocationID>
    <Language>EN</Language>
    <AuthorList>
      <Author>
        <FirstName>Mustafa</FirstName>
        <LastName>Yorgancıoğlu</LastName>
        <Identifier Source="ORCID">https://orcid.org/0000-0002-1101-7284</Identifier>
      </Author>
      <Author>
        <FirstName>Ekim</FirstName>
        <LastName>Saglam Gurmen</LastName>
      </Author>
    </AuthorList>
    <PublicationType>Journal Article</PublicationType>
    <ArticleIdList>
      <ArticleId IdType="doi">10.34172/aim.35274</ArticleId>
    </ArticleIdList>
    <History>
      <PubDate PubStatus="received">
        <Year>2024</Year>
        <Month>10</Month>
        <Day>12</Day>
      </PubDate>
      <PubDate PubStatus="accepted">
        <Year>2025</Year>
        <Month>11</Month>
        <Day>12</Day>
      </PubDate>
    </History>
    <Abstract>Background: Diagnosing non-ST elevation myocardial infarction (NSTEMI) in busy emergency departments is challenging. Artificial intelligence (AI) systems, particularly large language models (LLMs), offer potential as clinical decision support tools. This study aimed to evaluate the reliability of ChatGPT and Gemini in NSTEMI cases by comparing their responses to multiple-choice questions with those of emergency physicians. Methods: This prospective, cross-sectional study was conducted via an online survey among 1,106 emergency physicians in Turkey. The survey included ten NSTEMI-related multiple-choice questions based on the 2023 European Society of Cardiology guidelines. The same questions were presented to ChatGPT 4.0 and Gemini 2.5, queried using identical standardized prompts (temperature=0, no web access) on April 20, 2025. Statistical analyses were performed using SPSS 26.0. Results: AI models significantly outperformed physicians, correctly answering nine of ten questions versus the physicians’ mean of 7.62±1.32 (P&lt;0.001). Effect sizes indicated a very large difference for less experienced physicians and a moderate difference for specialists. Performance improved with experience, yet AI exceeded even the most experienced physicians. Participants from training and research hospitals scored higher than those from state hospitals. Conclusion: ChatGPT and Gemini demonstrated superior performance over emergency physicians in NSTEMI clinical questions, highlighting AI’s potential to enhance medical education, clinical decision support, and patient care. These findings, however, are limited by the non-proctored online setting and absence of real clinical context. Future research should focus on optimizing AI-clinician collaboration for safe and effective integration.  </Abstract>
    <ObjectList>
      <Object Type="keyword">
        <Param Name="value">Artificial intelligence</Param>
      </Object>
      <Object Type="keyword">
        <Param Name="value">Clinical decision support</Param>
      </Object>
      <Object Type="keyword">
        <Param Name="value">Emergency medicine</Param>
      </Object>
      <Object Type="keyword">
        <Param Name="value">Large language models</Param>
      </Object>
      <Object Type="keyword">
        <Param Name="value">Non-ST elevation myocardial infarction</Param>
      </Object>
    </ObjectList>
  </Article>
</ArticleSet>