Assessment of Large Language Models (LLMs) in decision-making support for gynecologic oncology
This study investigated the ability of Large Language Models (LLMs) to provide accurate and consistent answers by focusing on their performance in complex gynecologic cancer cases. LLMs are advancing rapidly and require a thorough evaluation to ensure that they can be safely and effectively used in...
Saved in:
Published in | Computational and structural biotechnology journal Vol. 23; pp. 4019 - 4026 |
---|---|
Main Authors | , , , , , , , , , , , , , , , , , , , , , , , , |
Format | Journal Article |
Language | English |
Published |
Netherlands
Elsevier B.V
01.12.2024
Research Network of Computational and Structural Biotechnology Elsevier |
Subjects | |
Online Access | Get full text |
Cover
Loading…
Abstract | This study investigated the ability of Large Language Models (LLMs) to provide accurate and consistent answers by focusing on their performance in complex gynecologic cancer cases.
LLMs are advancing rapidly and require a thorough evaluation to ensure that they can be safely and effectively used in clinical decision-making. Such evaluations are essential for confirming LLM reliability and accuracy in supporting medical professionals in casework.
We assessed three prominent LLMs—ChatGPT-4 (CG-4), Gemini Advanced (GemAdv), and Copilot—evaluating their accuracy, consistency, and overall performance. Fifteen clinical vignettes of varying difficulty and five open-ended questions based on real patient cases were used. The responses were coded, randomized, and evaluated blindly by six expert gynecologic oncologists using a 5-point Likert scale for relevance, clarity, depth, focus, and coherence.
GemAdv demonstrated superior accuracy (81.87 %) compared to both CG-4 (61.60 %) and Copilot (70.67 %) across all difficulty levels. GemAdv consistently provided correct answers more frequently (>60 % every day during the testing period). Although CG-4 showed a slight advantage in adhering to the National Comprehensive Cancer Network (NCCN) treatment guidelines, GemAdv excelled in the depth and focus of the answers provided, which are crucial aspects of clinical decision-making.
LLMs, especially GemAdv, show potential in supporting clinical practice by providing accurate, consistent, and relevant information for gynecologic cancer. However, further refinement is needed for more complex scenarios. This study highlights the promise of LLMs in gynecologic oncology, emphasizing the need for ongoing development and rigorous evaluation to maximize their clinical utility and reliability.
[Display omitted]
•Large Language Models (LLMs) are valuable tools in clinical practice, aiding healthcare professionals in making evidence-based decisions and improving patient care.•Gemini Advanced achieved 81.87 % accuracy in clinical decision-making.•Gemini Advanced consistently provided correct answers > 60 % every day during the testing period.•ChatGPT-4 and Gemini Advanced outperformed Copilot in treatment recommendations.•Further improvements are necessary to ensure accurate and relevant responses across clinical scenarios. |
---|---|
AbstractList | This study investigated the ability of Large Language Models (LLMs) to provide accurate and consistent answers by focusing on their performance in complex gynecologic cancer cases. LLMs are advancing rapidly and require a thorough evaluation to ensure that they can be safely and effectively used in clinical decision-making. Such evaluations are essential for confirming LLM reliability and accuracy in supporting medical professionals in casework. We assessed three prominent LLMs—ChatGPT-4 (CG-4), Gemini Advanced (GemAdv), and Copilot—evaluating their accuracy, consistency, and overall performance. Fifteen clinical vignettes of varying difficulty and five open-ended questions based on real patient cases were used. The responses were coded, randomized, and evaluated blindly by six expert gynecologic oncologists using a 5-point Likert scale for relevance, clarity, depth, focus, and coherence. GemAdv demonstrated superior accuracy (81.87 %) compared to both CG-4 (61.60 %) and Copilot (70.67 %) across all difficulty levels. GemAdv consistently provided correct answers more frequently (>60 % every day during the testing period). Although CG-4 showed a slight advantage in adhering to the National Comprehensive Cancer Network (NCCN) treatment guidelines, GemAdv excelled in the depth and focus of the answers provided, which are crucial aspects of clinical decision-making. LLMs, especially GemAdv, show potential in supporting clinical practice by providing accurate, consistent, and relevant information for gynecologic cancer. However, further refinement is needed for more complex scenarios. This study highlights the promise of LLMs in gynecologic oncology, emphasizing the need for ongoing development and rigorous evaluation to maximize their clinical utility and reliability. ga1 • Large Language Models (LLMs) are valuable tools in clinical practice, aiding healthcare professionals in making evidence-based decisions and improving patient care. • Gemini Advanced achieved 81.87 % accuracy in clinical decision-making. • Gemini Advanced consistently provided correct answers > 60 % every day during the testing period. • ChatGPT-4 and Gemini Advanced outperformed Copilot in treatment recommendations. • Further improvements are necessary to ensure accurate and relevant responses across clinical scenarios. This study investigated the ability of Large Language Models (LLMs) to provide accurate and consistent answers by focusing on their performance in complex gynecologic cancer cases. LLMs are advancing rapidly and require a thorough evaluation to ensure that they can be safely and effectively used in clinical decision-making. Such evaluations are essential for confirming LLM reliability and accuracy in supporting medical professionals in casework. We assessed three prominent LLMs—ChatGPT-4 (CG-4), Gemini Advanced (GemAdv), and Copilot—evaluating their accuracy, consistency, and overall performance. Fifteen clinical vignettes of varying difficulty and five open-ended questions based on real patient cases were used. The responses were coded, randomized, and evaluated blindly by six expert gynecologic oncologists using a 5-point Likert scale for relevance, clarity, depth, focus, and coherence. GemAdv demonstrated superior accuracy (81.87 %) compared to both CG-4 (61.60 %) and Copilot (70.67 %) across all difficulty levels. GemAdv consistently provided correct answers more frequently (>60 % every day during the testing period). Although CG-4 showed a slight advantage in adhering to the National Comprehensive Cancer Network (NCCN) treatment guidelines, GemAdv excelled in the depth and focus of the answers provided, which are crucial aspects of clinical decision-making. LLMs, especially GemAdv, show potential in supporting clinical practice by providing accurate, consistent, and relevant information for gynecologic cancer. However, further refinement is needed for more complex scenarios. This study highlights the promise of LLMs in gynecologic oncology, emphasizing the need for ongoing development and rigorous evaluation to maximize their clinical utility and reliability. [Display omitted] •Large Language Models (LLMs) are valuable tools in clinical practice, aiding healthcare professionals in making evidence-based decisions and improving patient care.•Gemini Advanced achieved 81.87 % accuracy in clinical decision-making.•Gemini Advanced consistently provided correct answers > 60 % every day during the testing period.•ChatGPT-4 and Gemini Advanced outperformed Copilot in treatment recommendations.•Further improvements are necessary to ensure accurate and relevant responses across clinical scenarios. This study investigated the ability of Large Language Models (LLMs) to provide accurate and consistent answers by focusing on their performance in complex gynecologic cancer cases. LLMs are advancing rapidly and require a thorough evaluation to ensure that they can be safely and effectively used in clinical decision-making. Such evaluations are essential for confirming LLM reliability and accuracy in supporting medical professionals in casework. We assessed three prominent LLMs-ChatGPT-4 (CG-4), Gemini Advanced (GemAdv), and Copilot-evaluating their accuracy, consistency, and overall performance. Fifteen clinical vignettes of varying difficulty and five open-ended questions based on real patient cases were used. The responses were coded, randomized, and evaluated blindly by six expert gynecologic oncologists using a 5-point Likert scale for relevance, clarity, depth, focus, and coherence. GemAdv demonstrated superior accuracy (81.87 %) compared to both CG-4 (61.60 %) and Copilot (70.67 %) across all difficulty levels. GemAdv consistently provided correct answers more frequently (>60 % every day during the testing period). Although CG-4 showed a slight advantage in adhering to the National Comprehensive Cancer Network (NCCN) treatment guidelines, GemAdv excelled in the depth and focus of the answers provided, which are crucial aspects of clinical decision-making. LLMs, especially GemAdv, show potential in supporting clinical practice by providing accurate, consistent, and relevant information for gynecologic cancer. However, further refinement is needed for more complex scenarios. This study highlights the promise of LLMs in gynecologic oncology, emphasizing the need for ongoing development and rigorous evaluation to maximize their clinical utility and reliability. Objective: This study investigated the ability of Large Language Models (LLMs) to provide accurate and consistent answers by focusing on their performance in complex gynecologic cancer cases. Background: LLMs are advancing rapidly and require a thorough evaluation to ensure that they can be safely and effectively used in clinical decision-making. Such evaluations are essential for confirming LLM reliability and accuracy in supporting medical professionals in casework. Study design: We assessed three prominent LLMs—ChatGPT-4 (CG-4), Gemini Advanced (GemAdv), and Copilot—evaluating their accuracy, consistency, and overall performance. Fifteen clinical vignettes of varying difficulty and five open-ended questions based on real patient cases were used. The responses were coded, randomized, and evaluated blindly by six expert gynecologic oncologists using a 5-point Likert scale for relevance, clarity, depth, focus, and coherence. Results: GemAdv demonstrated superior accuracy (81.87 %) compared to both CG-4 (61.60 %) and Copilot (70.67 %) across all difficulty levels. GemAdv consistently provided correct answers more frequently (>60 % every day during the testing period). Although CG-4 showed a slight advantage in adhering to the National Comprehensive Cancer Network (NCCN) treatment guidelines, GemAdv excelled in the depth and focus of the answers provided, which are crucial aspects of clinical decision-making. Conclusion: LLMs, especially GemAdv, show potential in supporting clinical practice by providing accurate, consistent, and relevant information for gynecologic cancer. However, further refinement is needed for more complex scenarios. This study highlights the promise of LLMs in gynecologic oncology, emphasizing the need for ongoing development and rigorous evaluation to maximize their clinical utility and reliability. This study investigated the ability of Large Language Models (LLMs) to provide accurate and consistent answers by focusing on their performance in complex gynecologic cancer cases.ObjectiveThis study investigated the ability of Large Language Models (LLMs) to provide accurate and consistent answers by focusing on their performance in complex gynecologic cancer cases.LLMs are advancing rapidly and require a thorough evaluation to ensure that they can be safely and effectively used in clinical decision-making. Such evaluations are essential for confirming LLM reliability and accuracy in supporting medical professionals in casework.BackgroundLLMs are advancing rapidly and require a thorough evaluation to ensure that they can be safely and effectively used in clinical decision-making. Such evaluations are essential for confirming LLM reliability and accuracy in supporting medical professionals in casework.We assessed three prominent LLMs-ChatGPT-4 (CG-4), Gemini Advanced (GemAdv), and Copilot-evaluating their accuracy, consistency, and overall performance. Fifteen clinical vignettes of varying difficulty and five open-ended questions based on real patient cases were used. The responses were coded, randomized, and evaluated blindly by six expert gynecologic oncologists using a 5-point Likert scale for relevance, clarity, depth, focus, and coherence.Study designWe assessed three prominent LLMs-ChatGPT-4 (CG-4), Gemini Advanced (GemAdv), and Copilot-evaluating their accuracy, consistency, and overall performance. Fifteen clinical vignettes of varying difficulty and five open-ended questions based on real patient cases were used. The responses were coded, randomized, and evaluated blindly by six expert gynecologic oncologists using a 5-point Likert scale for relevance, clarity, depth, focus, and coherence.GemAdv demonstrated superior accuracy (81.87 %) compared to both CG-4 (61.60 %) and Copilot (70.67 %) across all difficulty levels. GemAdv consistently provided correct answers more frequently (>60 % every day during the testing period). Although CG-4 showed a slight advantage in adhering to the National Comprehensive Cancer Network (NCCN) treatment guidelines, GemAdv excelled in the depth and focus of the answers provided, which are crucial aspects of clinical decision-making.ResultsGemAdv demonstrated superior accuracy (81.87 %) compared to both CG-4 (61.60 %) and Copilot (70.67 %) across all difficulty levels. GemAdv consistently provided correct answers more frequently (>60 % every day during the testing period). Although CG-4 showed a slight advantage in adhering to the National Comprehensive Cancer Network (NCCN) treatment guidelines, GemAdv excelled in the depth and focus of the answers provided, which are crucial aspects of clinical decision-making.LLMs, especially GemAdv, show potential in supporting clinical practice by providing accurate, consistent, and relevant information for gynecologic cancer. However, further refinement is needed for more complex scenarios. This study highlights the promise of LLMs in gynecologic oncology, emphasizing the need for ongoing development and rigorous evaluation to maximize their clinical utility and reliability.ConclusionLLMs, especially GemAdv, show potential in supporting clinical practice by providing accurate, consistent, and relevant information for gynecologic cancer. However, further refinement is needed for more complex scenarios. This study highlights the promise of LLMs in gynecologic oncology, emphasizing the need for ongoing development and rigorous evaluation to maximize their clinical utility and reliability. |
Author | Hsu, Yu-Cheng Hedianto, Tri Huang, Jingshan Irawan, Budi Wibowo, Bagus M. Brahmantara, Bagus Ngurah Yu, Zih-Ying Liao, Li-Na Indraprasta, Birama R. Ibrahim, Ibrahim H. Lu, Chien-Hsing Tjokroprawiro, Brahmana A. Tan, Ming Herlambang, Aditya Nugroho, Hari Bustomi, Ahmad Fadhli Pramuditya, Herlangga Mulawardhana, Pungky Yang, Jer-Yen Faridzi, Ach Salman Putra, Very Great E. Rahestyningtyas, Eccita Gumilar, Khanisyah Erza Tambunan, Zulkarnain Li, Dongqi |
Author_xml | – sequence: 1 givenname: Khanisyah Erza surname: Gumilar fullname: Gumilar, Khanisyah Erza email: khanisyah@fk.unair.ac.id organization: Graduate Institute of Biomedical Science, China Medical University, Taichung, Taiwan – sequence: 2 givenname: Birama R. surname: Indraprasta fullname: Indraprasta, Birama R. organization: Department of Obstetrics and Gynecology, Dr. Soetomo General Hospital - Faculty of Medicine, Universitas Airlangga, Surabaya, Indonesia – sequence: 3 givenname: Ach Salman surname: Faridzi fullname: Faridzi, Ach Salman organization: Department of Obstetrics and Gynecology, Dr. Soetomo General Hospital - Faculty of Medicine, Universitas Airlangga, Surabaya, Indonesia – sequence: 4 givenname: Bagus M. surname: Wibowo fullname: Wibowo, Bagus M. organization: Department of Obstetrics and Gynecology, Dr. Soetomo General Hospital - Faculty of Medicine, Universitas Airlangga, Surabaya, Indonesia – sequence: 5 givenname: Aditya surname: Herlambang fullname: Herlambang, Aditya organization: Department of Obstetrics and Gynecology, Dr. Soetomo General Hospital - Faculty of Medicine, Universitas Airlangga, Surabaya, Indonesia – sequence: 6 givenname: Eccita surname: Rahestyningtyas fullname: Rahestyningtyas, Eccita organization: Department of Obstetrics and Gynecology, Hospital of Universitas Airlangga - Faculty of Medicine, Universitas Airlangga, Surabaya, Indonesia – sequence: 7 givenname: Budi surname: Irawan fullname: Irawan, Budi organization: Department of Obstetrics and Gynecology, Dr. Soetomo General Hospital - Faculty of Medicine, Universitas Airlangga, Surabaya, Indonesia – sequence: 8 givenname: Zulkarnain surname: Tambunan fullname: Tambunan, Zulkarnain organization: Department of Obstetrics and Gynecology, Dr. Soetomo General Hospital - Faculty of Medicine, Universitas Airlangga, Surabaya, Indonesia – sequence: 9 givenname: Ahmad Fadhli surname: Bustomi fullname: Bustomi, Ahmad Fadhli organization: Department of Obstetrics and Gynecology, Dr. Soetomo General Hospital - Faculty of Medicine, Universitas Airlangga, Surabaya, Indonesia – sequence: 10 givenname: Bagus Ngurah surname: Brahmantara fullname: Brahmantara, Bagus Ngurah organization: Department of Obstetrics and Gynecology, Dr. Soetomo General Hospital - Faculty of Medicine, Universitas Airlangga, Surabaya, Indonesia – sequence: 11 givenname: Zih-Ying surname: Yu fullname: Yu, Zih-Ying organization: Department of Public Health, China Medical University, Taichung, Taiwan – sequence: 12 givenname: Yu-Cheng surname: Hsu fullname: Hsu, Yu-Cheng organization: Department of Public Health, China Medical University, Taichung, Taiwan – sequence: 13 givenname: Herlangga surname: Pramuditya fullname: Pramuditya, Herlangga organization: Department of Obstetrics and Gynecology, Dr. Ramelan Naval Hospital, Surabaya, Indonesia – sequence: 14 givenname: Very Great E. surname: Putra fullname: Putra, Very Great E. organization: Department of Obstetrics and Gynecology, Dr. Kariadi Central General Hospital, Semarang, Indonesia – sequence: 15 givenname: Hari surname: Nugroho fullname: Nugroho, Hari organization: Department of Obstetrics and Gynecology, Dr. Soetomo General Hospital - Faculty of Medicine, Universitas Airlangga, Surabaya, Indonesia – sequence: 16 givenname: Pungky surname: Mulawardhana fullname: Mulawardhana, Pungky organization: Department of Obstetrics and Gynecology, Hospital of Universitas Airlangga - Faculty of Medicine, Universitas Airlangga, Surabaya, Indonesia – sequence: 17 givenname: Brahmana A. surname: Tjokroprawiro fullname: Tjokroprawiro, Brahmana A. organization: Department of Obstetrics and Gynecology, Dr. Soetomo General Hospital - Faculty of Medicine, Universitas Airlangga, Surabaya, Indonesia – sequence: 18 givenname: Tri surname: Hedianto fullname: Hedianto, Tri organization: Faculty of Medicine and Health, Institut Teknologi Sepuluh Nopember, Surabaya, Indonesia – sequence: 19 givenname: Ibrahim H. surname: Ibrahim fullname: Ibrahim, Ibrahim H. organization: Graduate Institute of Biomedical Science, China Medical University, Taichung, Taiwan – sequence: 20 givenname: Jingshan surname: Huang fullname: Huang, Jingshan organization: School of Computing, College of Medicine, University of South Alabama, Mobile, AL, USA – sequence: 21 givenname: Dongqi surname: Li fullname: Li, Dongqi organization: School of Information and Computer Sciences, School of Social and Behavioral Sciences, University of California, Irvine, CA, USA – sequence: 22 givenname: Chien-Hsing surname: Lu fullname: Lu, Chien-Hsing organization: Department of Obstetrics and Gynecology, Taichung Veteran General Hospital, Taichung, Taiwan – sequence: 23 givenname: Jer-Yen surname: Yang fullname: Yang, Jer-Yen email: jyyang@cmu.edu.tw organization: Graduate Institute of Biomedical Science, China Medical University, Taichung, Taiwan – sequence: 24 givenname: Li-Na surname: Liao fullname: Liao, Li-Na email: linaliao@mail.cmu.edu.tw organization: Department of Public Health, China Medical University, Taichung, Taiwan – sequence: 25 givenname: Ming surname: Tan fullname: Tan, Ming email: mingtan@mail.cmu.edu.tw organization: Graduate Institute of Biomedical Science, China Medical University, Taichung, Taiwan |
BackLink | https://www.ncbi.nlm.nih.gov/pubmed/39610903$$D View this record in MEDLINE/PubMed |
BookMark | eNqFUk1v1DAQjVARLaV_gAPKsRx2GccfSSQkVFV8VErFBa5YzmQcvGTtxc5W2n-Pt1uqlkPxwR6N33sez5uXxZEPnoriNYMlA6berZaY-tWygkrkxBIkPCtOKgC2AF7D0YP4uDhLaQV5NUy1HF4Ux7xVDFrgJ8WPi5QopTX5uQy27EwcKe9-3JocXIeBplSed911els6Xw6ELrngF2vzy_mxTNvNJsS5tCGW484ThimMDsvgb6Pdq-K5NVOis7vztPj-6eO3yy-L7uvnq8uLboFSynkhGlMRUkVS9IKjMJYUNMArqZhQRJykbaUBsIb1Ei0BEtpWqcGSbXHgp8XVQXcIZqU30a1N3OlgnL5NhDhqE2eHE2mqYZC1lTU2TNS87a2o8wNKtWiqgYus9eGgtdn2axowtyaa6ZHo4xvvfuox3GjGFHCANiuc3ynE8HtLadZrl5CmyXgK26R59oZnxxr-fyjjAlQLAjL0zcO67gv6a2YGVAcAxpBSJHsPYaD3Q6NXej80ej80-1wuIZOaf0joZjNnj_Pf3PQ09f2BmmeEbhxFndCRRxpcJJxz691T9D9LHN1J |
CitedBy_id | crossref_primary_10_3390_jcm14030875 crossref_primary_10_1016_j_csbj_2025_03_026 crossref_primary_10_1016_j_csbj_2024_12_013 |
Cites_doi | 10.1109/ICAAIC56838.2023.10140214 10.1016/j.ajog.2023.04.020 10.1007/s00403-019-01957-2 10.2196/51873 10.1111/iej.13985 10.2196/45312 10.5468/ogs.23231 10.1115/1.4062773 10.1093/annonc/mdq417 10.1371/journal.pdig.0000198 10.1109/JAS.2023.123618 10.1186/s13000-024-01464-7 10.2214/AJR.21.27195 10.1093/asjof/ojad084 10.6004/jnccn.2023.0006 10.1097/CCO.0000000000000916 10.1007/s10459-023-10257-4 10.1148/radiol.230922 |
ContentType | Journal Article |
Copyright | 2024 The Authors 2024 The Authors. 2024 The Authors 2024 |
Copyright_xml | – notice: 2024 The Authors – notice: 2024 The Authors. – notice: 2024 The Authors 2024 |
DBID | 6I. AAFTH AAYXX CITATION NPM 7X8 7S9 L.6 5PM DOA |
DOI | 10.1016/j.csbj.2024.10.050 |
DatabaseName | ScienceDirect Open Access Titles Elsevier:ScienceDirect:Open Access CrossRef PubMed MEDLINE - Academic AGRICOLA AGRICOLA - Academic PubMed Central (Full Participant titles) DOAJ Directory of Open Access Journals |
DatabaseTitle | CrossRef PubMed MEDLINE - Academic AGRICOLA AGRICOLA - Academic |
DatabaseTitleList | AGRICOLA PubMed MEDLINE - Academic |
Database_xml | – sequence: 1 dbid: DOA name: DOAJ Directory of Open Access Journals url: https://www.doaj.org/ sourceTypes: Open Website – sequence: 2 dbid: NPM name: PubMed url: https://proxy.k.utb.cz/login?url=http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?db=PubMed sourceTypes: Index Database |
DeliveryMethod | fulltext_linktorsrc |
Discipline | Engineering |
EISSN | 2001-0370 |
EndPage | 4026 |
ExternalDocumentID | oai_doaj_org_article_e70d57f57c814739bf47e3e669ca2d34 PMC11603009 39610903 10_1016_j_csbj_2024_10_050 S2001037024003702 |
Genre | Journal Article |
GroupedDBID | 0R~ 457 53G 5VS 6I. AACTN AAEDT AAEDW AAFTH AAHBH AAIKJ AALRI AAXUO ABMAC ACGFS ADBBV ADEZE ADRAZ ADVLN AEXQZ AFTJW AGHFR AITUG AKRWK ALMA_UNASSIGNED_HOLDINGS AMRAJ AOIJS BAWUL BCNDV DIK EBS EJD FDB GROUPED_DOAJ HYE IPNFZ KQ8 M41 M48 M~E O9- OK1 RIG ROL RPM SSZ AAYWO AAYXX ACVFH ADCNI AEUPX AFPUW AIGII AKBMS AKYEP CITATION 0SF NCXOZ NPM 7X8 7S9 L.6 5PM |
ID | FETCH-LOGICAL-c555t-48a2ece2e54b43c4afe60803256146ee3e5f95a00fa1b5cfe0cecf966dfef9cd3 |
IEDL.DBID | M48 |
ISSN | 2001-0370 |
IngestDate | Wed Aug 27 01:11:20 EDT 2025 Thu Aug 21 18:36:28 EDT 2025 Fri Aug 22 20:40:24 EDT 2025 Fri Jul 11 14:26:16 EDT 2025 Thu Jan 02 22:25:24 EST 2025 Thu Apr 24 23:08:24 EDT 2025 Tue Jul 01 05:22:45 EDT 2025 Sat Mar 08 15:48:10 EST 2025 |
IsDoiOpenAccess | true |
IsOpenAccess | true |
IsPeerReviewed | true |
IsScholarly | true |
Keywords | Accuracy Large Language Models Artificial intelligence Gynecologic cancer Consistency |
Language | English |
License | This is an open access article under the CC BY-NC-ND license. 2024 The Authors. This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/). |
LinkModel | DirectLink |
MergedId | FETCHMERGED-LOGICAL-c555t-48a2ece2e54b43c4afe60803256146ee3e5f95a00fa1b5cfe0cecf966dfef9cd3 |
Notes | ObjectType-Article-1 SourceType-Scholarly Journals-1 ObjectType-Feature-2 content type line 23 |
OpenAccessLink | http://journals.scholarsportal.info/openUrl.xqy?doi=10.1016/j.csbj.2024.10.050 |
PMID | 39610903 |
PQID | 3134069040 |
PQPubID | 23479 |
PageCount | 8 |
ParticipantIDs | doaj_primary_oai_doaj_org_article_e70d57f57c814739bf47e3e669ca2d34 pubmedcentral_primary_oai_pubmedcentral_nih_gov_11603009 proquest_miscellaneous_3200305083 proquest_miscellaneous_3134069040 pubmed_primary_39610903 crossref_primary_10_1016_j_csbj_2024_10_050 crossref_citationtrail_10_1016_j_csbj_2024_10_050 elsevier_sciencedirect_doi_10_1016_j_csbj_2024_10_050 |
ProviderPackageCode | CITATION AAYXX |
PublicationCentury | 2000 |
PublicationDate | 2024-12-01 |
PublicationDateYYYYMMDD | 2024-12-01 |
PublicationDate_xml | – month: 12 year: 2024 text: 2024-12-01 day: 01 |
PublicationDecade | 2020 |
PublicationPlace | Netherlands |
PublicationPlace_xml | – name: Netherlands |
PublicationTitle | Computational and structural biotechnology journal |
PublicationTitleAlternate | Comput Struct Biotechnol J |
PublicationYear | 2024 |
Publisher | Elsevier B.V Research Network of Computational and Structural Biotechnology Elsevier |
Publisher_xml | – name: Elsevier B.V – name: Research Network of Computational and Structural Biotechnology – name: Elsevier |
References | Seth, Lim, Xie, Cevik, Rozen, Ross (bib17) 2023; 5 Kapoor, Haj-Mirzaian, Yan, Wickner, Giess, Eappen (bib14) 2022; 219 Voigt, Trautwein (bib24) 2023; 35 Kung, Cheatham, Medenilla, Sillos, De Leon, Elepano (bib21) 2023; 2 Ullah, Parwani, Baig, Singh (bib23) 2024; 19 Bhardwaz S. ,Kumar J., An Extensive Comparative Analysis of Chatbot Technologies - ChatGPT, Google BARD and Microsoft Bing, in 2023 2nd International Conference on Applied Artificial Intelligence and Computing (ICAAIC). 2023. p. 673–679. Kamo, Dandapani, Miksad, Houlihan, Kaplan, Regan (bib16) 2011; 22 Lee, Kim (bib4) 2024; 67 Gilson, Safranek, Huang, Socrates, Chi, Taylor (bib20) 2023; 9 Ellaway, Tolsgaard (bib3) 2023; 28 Sikander, Baker, Deveci, Lund, Rosenberg (bib11) 2023; 15 What’s the most popular LLM? 2024 [cited 2024 June 10]; Available from Daniel, Ma, Singh, Beatrice Bloom, Nilda Adair, William Chen (bib13) 2023; 8 Li, Kemp, Logan, Dimri, Singh, Mattar (bib19) 2023; 229 Wu, He, Liu, Sun, Liu, Han (bib8) 2023; 10 Gumilar, Indraprasta, Hsu, Yiou, Hong, Irawan (bib10) 2024 Brodnik, Carton, Muir, Ghosh, Downey, Echlin (bib2) 2023; 90 Suarez, Diaz-Flores Garcia, Algar, Gomez Sanchez, Llorente de Pedro, Freire (bib22) 2024; 57 . Vaidya, Mori, Dusza, Rossi, Nehal, Lee (bib15) 2019; 311 Rahsepar, Tavakoli, Kim, Hassani, Abtin, Bedayat (bib7) 2023; 307 Veras, Dyer, Rooney, Barros Silva, Rutherford, Kairy (bib12) 2023; 12 Abu-Rustum, Yashar, Arend, Barber, Bradley, Brooks (bib5) 2023; 21 Gordon, Towbin, Wingrove, Shafique, Haas, Kitts (bib6) 2023 Zuniga Salazar, Zuniga, Vindel, Yoong, Hincapie, Zuniga (bib18) 2023; 15 Voigt (10.1016/j.csbj.2024.10.050_bib24) 2023; 35 Ellaway (10.1016/j.csbj.2024.10.050_bib3) 2023; 28 Abu-Rustum (10.1016/j.csbj.2024.10.050_bib5) 2023; 21 10.1016/j.csbj.2024.10.050_bib9 Zuniga Salazar (10.1016/j.csbj.2024.10.050_bib18) 2023; 15 Veras (10.1016/j.csbj.2024.10.050_bib12) 2023; 12 Lee (10.1016/j.csbj.2024.10.050_bib4) 2024; 67 10.1016/j.csbj.2024.10.050_bib1 Li (10.1016/j.csbj.2024.10.050_bib19) 2023; 229 Kapoor (10.1016/j.csbj.2024.10.050_bib14) 2022; 219 Vaidya (10.1016/j.csbj.2024.10.050_bib15) 2019; 311 Gilson (10.1016/j.csbj.2024.10.050_bib20) 2023; 9 Kung (10.1016/j.csbj.2024.10.050_bib21) 2023; 2 Suarez (10.1016/j.csbj.2024.10.050_bib22) 2024; 57 Brodnik (10.1016/j.csbj.2024.10.050_bib2) 2023; 90 Gordon (10.1016/j.csbj.2024.10.050_bib6) 2023 Wu (10.1016/j.csbj.2024.10.050_bib8) 2023; 10 Sikander (10.1016/j.csbj.2024.10.050_bib11) 2023; 15 Kamo (10.1016/j.csbj.2024.10.050_bib16) 2011; 22 Gumilar (10.1016/j.csbj.2024.10.050_bib10) 2024 Ullah (10.1016/j.csbj.2024.10.050_bib23) 2024; 19 Rahsepar (10.1016/j.csbj.2024.10.050_bib7) 2023; 307 Daniel (10.1016/j.csbj.2024.10.050_bib13) 2023; 8 Seth (10.1016/j.csbj.2024.10.050_bib17) 2023; 5 |
References_xml | – reference: Bhardwaz S. ,Kumar J., An Extensive Comparative Analysis of Chatbot Technologies - ChatGPT, Google BARD and Microsoft Bing, in 2023 2nd International Conference on Applied Artificial Intelligence and Computing (ICAAIC). 2023. p. 673–679. – volume: 311 start-page: 691 year: 2019 end-page: 696 ident: bib15 article-title: Appearance-related psychosocial distress following facial skin cancer surgery using the FACE-Q Skin Cancer publication-title: Arch Dermatol Res – volume: 28 start-page: 659 year: 2023 end-page: 664 ident: bib3 article-title: Artificial scholarship: LLMs in health professions education research publication-title: Adv Health Sci Educ Theory Pr – volume: 229 year: 2023 ident: bib19 article-title: ChatGPT outscored human candidates in a virtual objective structured clinical examination in obstetrics and gynecology publication-title: Am J Obstet Gynecol – volume: 35 start-page: 68 year: 2023 end-page: 77 ident: bib24 article-title: Improved guideline adherence in oncology through clinical decision-support systems: still hindered by current health IT infrastructures? publication-title: Curr Opin Oncol – volume: 10 start-page: 1122 year: 2023 end-page: 1136 ident: bib8 article-title: A brief overview of ChatGPT: the history, status quo and potential future development publication-title: IEEE/CAA J Autom Sin – volume: 57 start-page: 108 year: 2024 end-page: 113 ident: bib22 article-title: Unveiling the ChatGPT phenomenon: evaluating the consistency and accuracy of endodontic question answers publication-title: Int Endod J – volume: 67 start-page: 153 year: 2024 end-page: 159 ident: bib4 article-title: Potential applications of ChatGPT in obstetrics and gynecology in Korea: a review article publication-title: Obstet Gynecol Sci – volume: 90 year: 2023 ident: bib2 article-title: Perspective: large language models in applied mechanics publication-title: J Appl Mech – volume: 219 start-page: 338 year: 2022 end-page: 345 ident: bib14 article-title: Patient experience scores for radiologists: comparison with nonradiologist physicians and changes after public posting in an institutional online provider directory publication-title: Am J Roentgenol – volume: 8 start-page: 5 year: 2023 ident: bib13 article-title: Patient experience performance at a primary cancer center versus affiliated community facilities publication-title: Adv Radiat Oncol – reference: What’s the most popular LLM? 2024 [cited 2024 June 10]; Available from: – volume: 5 start-page: ojad084 year: 2023 ident: bib17 article-title: Comparing the efficacy of large language models ChatGPT, BARD, and Bing AI in providing information on rhinoplasty: an observational study publication-title: Aesthet Surg J Open Forum – volume: 15 year: 2023 ident: bib18 article-title: Efficacy of AI chats to determine an emergency: a comparison between open AI's ChatGPT, Google Bard, and Microsoft Bing AI Chat publication-title: Cureus – volume: 15 year: 2023 ident: bib11 article-title: ChatGPT-4 and human researchers are equal in writing scientific introduction sections: a blinded, randomized, non-inferiority controlled study publication-title: Cureus – volume: 19 start-page: 43 year: 2024 ident: bib23 article-title: Challenges and barriers of using large language models (LLM) such as ChatGPT for diagnostic medicine with a focus on digital pathology - a recent scoping review publication-title: Diagn Pathol – year: 2023 ident: bib6 article-title: Enhancing patient communication with Chat-GPT in radiology: evaluating the efficacy and readability of answers to common imaging-related questions publication-title: J Am Coll Radio – reference: . – year: 2024 ident: bib10 article-title: Disparities in medical recommendations from AIbased chatbots across different countries/regions publication-title: Res Sq – volume: 12 year: 2023 ident: bib12 article-title: Usability and efficacy of artificial intelligence chatbots (ChatGPT) for health sciences students: protocol for a crossover randomized controlled trial publication-title: JMIR Res Protoc – volume: 9 year: 2023 ident: bib20 article-title: How Does ChatGPT perform on the United States medical licensing examination (USMLE)? The implications of large language models for medical education and knowledge assessment publication-title: JMIR Med Educ – volume: 2 year: 2023 ident: bib21 article-title: Performance of ChatGPT on USMLE: potential for AI-assisted medical education using large language models publication-title: PLOS Digit Health – volume: 307 year: 2023 ident: bib7 article-title: How AI responds to common lung cancer questions: ChatGPT vs Google Bard publication-title: Radiology – volume: 21 start-page: 181 year: 2023 end-page: 209 ident: bib5 article-title: Uterine neoplasms, version 1.2023, NCCN clinical practice guidelines in oncology publication-title: J Natl Compr Canc Netw – volume: 22 start-page: 723 year: 2011 end-page: 729 ident: bib16 article-title: Evaluation of the SCA instrument for measuring patient satisfaction with cancer care administered via paper or via the Internet publication-title: Ann Oncol – ident: 10.1016/j.csbj.2024.10.050_bib9 doi: 10.1109/ICAAIC56838.2023.10140214 – volume: 229 issue: 2 year: 2023 ident: 10.1016/j.csbj.2024.10.050_bib19 article-title: ChatGPT outscored human candidates in a virtual objective structured clinical examination in obstetrics and gynecology publication-title: Am J Obstet Gynecol doi: 10.1016/j.ajog.2023.04.020 – year: 2023 ident: 10.1016/j.csbj.2024.10.050_bib6 article-title: Enhancing patient communication with Chat-GPT in radiology: evaluating the efficacy and readability of answers to common imaging-related questions publication-title: J Am Coll Radio – volume: 311 start-page: 691 issue: 9 year: 2019 ident: 10.1016/j.csbj.2024.10.050_bib15 article-title: Appearance-related psychosocial distress following facial skin cancer surgery using the FACE-Q Skin Cancer publication-title: Arch Dermatol Res doi: 10.1007/s00403-019-01957-2 – volume: 15 issue: 9 year: 2023 ident: 10.1016/j.csbj.2024.10.050_bib18 article-title: Efficacy of AI chats to determine an emergency: a comparison between open AI's ChatGPT, Google Bard, and Microsoft Bing AI Chat publication-title: Cureus – year: 2024 ident: 10.1016/j.csbj.2024.10.050_bib10 article-title: Disparities in medical recommendations from AIbased chatbots across different countries/regions publication-title: Res Sq – volume: 12 year: 2023 ident: 10.1016/j.csbj.2024.10.050_bib12 article-title: Usability and efficacy of artificial intelligence chatbots (ChatGPT) for health sciences students: protocol for a crossover randomized controlled trial publication-title: JMIR Res Protoc doi: 10.2196/51873 – volume: 57 start-page: 108 issue: 1 year: 2024 ident: 10.1016/j.csbj.2024.10.050_bib22 article-title: Unveiling the ChatGPT phenomenon: evaluating the consistency and accuracy of endodontic question answers publication-title: Int Endod J doi: 10.1111/iej.13985 – volume: 15 issue: 11 year: 2023 ident: 10.1016/j.csbj.2024.10.050_bib11 article-title: ChatGPT-4 and human researchers are equal in writing scientific introduction sections: a blinded, randomized, non-inferiority controlled study publication-title: Cureus – volume: 9 year: 2023 ident: 10.1016/j.csbj.2024.10.050_bib20 article-title: How Does ChatGPT perform on the United States medical licensing examination (USMLE)? The implications of large language models for medical education and knowledge assessment publication-title: JMIR Med Educ doi: 10.2196/45312 – volume: 67 start-page: 153 issue: 2 year: 2024 ident: 10.1016/j.csbj.2024.10.050_bib4 article-title: Potential applications of ChatGPT in obstetrics and gynecology in Korea: a review article publication-title: Obstet Gynecol Sci doi: 10.5468/ogs.23231 – volume: 90 issue: 10 year: 2023 ident: 10.1016/j.csbj.2024.10.050_bib2 article-title: Perspective: large language models in applied mechanics publication-title: J Appl Mech doi: 10.1115/1.4062773 – ident: 10.1016/j.csbj.2024.10.050_bib1 – volume: 22 start-page: 723 issue: 3 year: 2011 ident: 10.1016/j.csbj.2024.10.050_bib16 article-title: Evaluation of the SCA instrument for measuring patient satisfaction with cancer care administered via paper or via the Internet publication-title: Ann Oncol doi: 10.1093/annonc/mdq417 – volume: 2 issue: 2 year: 2023 ident: 10.1016/j.csbj.2024.10.050_bib21 article-title: Performance of ChatGPT on USMLE: potential for AI-assisted medical education using large language models publication-title: PLOS Digit Health doi: 10.1371/journal.pdig.0000198 – volume: 10 start-page: 1122 issue: 5 year: 2023 ident: 10.1016/j.csbj.2024.10.050_bib8 article-title: A brief overview of ChatGPT: the history, status quo and potential future development publication-title: IEEE/CAA J Autom Sin doi: 10.1109/JAS.2023.123618 – volume: 19 start-page: 43 issue: 1 year: 2024 ident: 10.1016/j.csbj.2024.10.050_bib23 article-title: Challenges and barriers of using large language models (LLM) such as ChatGPT for diagnostic medicine with a focus on digital pathology - a recent scoping review publication-title: Diagn Pathol doi: 10.1186/s13000-024-01464-7 – volume: 219 start-page: 338 issue: 2 year: 2022 ident: 10.1016/j.csbj.2024.10.050_bib14 article-title: Patient experience scores for radiologists: comparison with nonradiologist physicians and changes after public posting in an institutional online provider directory publication-title: Am J Roentgenol doi: 10.2214/AJR.21.27195 – volume: 5 start-page: ojad084 year: 2023 ident: 10.1016/j.csbj.2024.10.050_bib17 article-title: Comparing the efficacy of large language models ChatGPT, BARD, and Bing AI in providing information on rhinoplasty: an observational study publication-title: Aesthet Surg J Open Forum doi: 10.1093/asjof/ojad084 – volume: 8 start-page: 5 year: 2023 ident: 10.1016/j.csbj.2024.10.050_bib13 article-title: Patient experience performance at a primary cancer center versus affiliated community facilities publication-title: Adv Radiat Oncol – volume: 21 start-page: 181 issue: 2 year: 2023 ident: 10.1016/j.csbj.2024.10.050_bib5 article-title: Uterine neoplasms, version 1.2023, NCCN clinical practice guidelines in oncology publication-title: J Natl Compr Canc Netw doi: 10.6004/jnccn.2023.0006 – volume: 35 start-page: 68 issue: 1 year: 2023 ident: 10.1016/j.csbj.2024.10.050_bib24 article-title: Improved guideline adherence in oncology through clinical decision-support systems: still hindered by current health IT infrastructures? publication-title: Curr Opin Oncol doi: 10.1097/CCO.0000000000000916 – volume: 28 start-page: 659 issue: 3 year: 2023 ident: 10.1016/j.csbj.2024.10.050_bib3 article-title: Artificial scholarship: LLMs in health professions education research publication-title: Adv Health Sci Educ Theory Pr doi: 10.1007/s10459-023-10257-4 – volume: 307 issue: 5 year: 2023 ident: 10.1016/j.csbj.2024.10.050_bib7 article-title: How AI responds to common lung cancer questions: ChatGPT vs Google Bard publication-title: Radiology doi: 10.1148/radiol.230922 |
SSID | ssj0000816930 |
Score | 2.3461647 |
Snippet | This study investigated the ability of Large Language Models (LLMs) to provide accurate and consistent answers by focusing on their performance in complex... ga1 • Large Language Models (LLMs) are valuable tools in clinical practice, aiding healthcare professionals in making evidence-based decisions and improving... Objective: This study investigated the ability of Large Language Models (LLMs) to provide accurate and consistent answers by focusing on their performance in... |
SourceID | doaj pubmedcentral proquest pubmed crossref elsevier |
SourceType | Open Website Open Access Repository Aggregation Database Index Database Enrichment Source Publisher |
StartPage | 4019 |
SubjectTerms | Accuracy Artificial intelligence biotechnology Consistency decision making Gynecologic cancer Large Language Models patients |
SummonAdditionalLinks | – databaseName: DOAJ Directory of Open Access Journals dbid: DOA link: http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwrV1Lb9QwELZQT3BAlOfykpE4gFAgiR9xjgVRVWjLiUo9YcX2GLYq2YrsHvrvO2NnV1mQlguXKLKdhz1jzzfJ-BvGXjsXlIc6FpUToZAqmMJ02hShc0obtMhtSRucT7_qkzP55VydT1J9UUxYpgfOA_cBmjKoJqrGm0o2onVRNiBA69Z3dRCJCRRt3sSZSmuwIZKRctwlkwO6_OAu0CGs5XsK56KN9hNLlAj7dwzS34Dzz7jJiSE6vsfujgiSH-U3P2S3oL_P7kx4BR-w70dbvk2-jHxO0d54zF8mOaU_uxz4m_n8dHjLFz0PY56d4ldKTcWH9RWhco54lv-47sHnBZIv-3R2_ZCdHX_-9umkGBMpFF4ptSqk6WpAiYCSTgovuwgakaKoiQZUAw6niq3qyjJ2lVM-QunBR3SEQoTY-iAesYN-2cMTxqkEjDetkV7qsnUI4KR2KqAn5modZqzaDKr1I8s4Jbu4tJtwsgtLgrAkCCpDQczYu-01V5ljY2_rjySrbUvix04FqDV21Br7L62ZMbWRtB2hRoYQeKvF3oe_2qiFxXlIP1e6HpbrwYpK0CZiXBP3tKmTy4aod8YeZ1XadkO0RHxfYo3ZUbKdfu7W9IufiQ-8olThiJWf_o-RecZuU39zxM5zdrD6vYYXiLtW7mWaYjcP3S0a priority: 102 providerName: Directory of Open Access Journals |
Title | Assessment of Large Language Models (LLMs) in decision-making support for gynecologic oncology |
URI | https://dx.doi.org/10.1016/j.csbj.2024.10.050 https://www.ncbi.nlm.nih.gov/pubmed/39610903 https://www.proquest.com/docview/3134069040 https://www.proquest.com/docview/3200305083 https://pubmed.ncbi.nlm.nih.gov/PMC11603009 https://doaj.org/article/e70d57f57c814739bf47e3e669ca2d34 |
Volume | 23 |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV1bb9MwFLam7QUeEHfKoDISDyCUKYkvcR4QKohpYi0PiIo9EcW30akkW9NK9N9zTuKUFaaKl6hynET2sX2-4x5_HyEvtbbCuNRHiWY24sKqSJVSRbbUQirwyHmMB5wnn-XJlH86E2d7pJc7Ch3Y3BjaoZ7UdDE_-nW1fgcT_u2fXC3T6AuI9VJ-hJlaGMIfgGfKUNFgEuB-uzIrpB7BbZeQSZTF4RzNza_Z8lUtpf-Wy_oXkv6dWXnNVR3fJXcCxqSjblDcI3uuuk9uX2MefEC-jzaMnLT2dIz54HDt9i4pCqTNG_pqPJ40r-msojYo8UQ_W_Eq2qwuscsoIF56vq6c6ZZQWlftr_VDMj3--PXDSRSkFiIjhFhGXJWpA5s5wTVnhpfeScCSLEWiUOkcc8LnooxjXyZaGO9i44yHUMl653Nj2SOyX9WVe0IoljhlVK644TLONUA8LrWwEKvpVNoBSfpOLUzgIUc5jHnRJ5xdFGiIAg2BZWCIAXmzeeayY-HYWfs92mpTExm024J6cV6ECVm4LLYi8yIzKuEZy7XnGTRTytyUqWV8QERv6SKAkQ5kwKtmOz_-oh8WBcxU_PulrFy9agqWMDxmDKvmjjppG9QBLh6Qx91Q2jSD5UiNH8MdtTXIttq5faea_WgZwxMUEwc0_fQ_PnxIbmFzupSdZ2R_uVi55wC8lnpIDkanX76dDtuNi2E7t34DWtgu8A |
linkProvider | Scholars Portal |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Assessment+of+Large+Language+Models+%28LLMs%29+in+decision-making+support+for+gynecologic+oncology&rft.jtitle=Computational+and+structural+biotechnology+journal&rft.au=Gumilar%2C+Khanisyah+Erza&rft.au=Indraprasta%2C+Birama+R&rft.au=Faridzi%2C+Ach+Salman&rft.au=Wibowo%2C+Bagus+M&rft.date=2024-12-01&rft.issn=2001-0370&rft.eissn=2001-0370&rft.volume=23+p.4019-4026&rft.spage=4019&rft.epage=4026&rft_id=info:doi/10.1016%2Fj.csbj.2024.10.050&rft.externalDBID=NO_FULL_TEXT |
thumbnail_l | http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/lc.gif&issn=2001-0370&client=summon |
thumbnail_m | http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/mc.gif&issn=2001-0370&client=summon |
thumbnail_s | http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/sc.gif&issn=2001-0370&client=summon |