Assessment of Large Language Models (LLMs) in decision-making support for gynecologic oncology

This study investigated the ability of Large Language Models (LLMs) to provide accurate and consistent answers by focusing on their performance in complex gynecologic cancer cases. LLMs are advancing rapidly and require a thorough evaluation to ensure that they can be safely and effectively used in...

Full description

Saved in:
Bibliographic Details
Published inComputational and structural biotechnology journal Vol. 23; pp. 4019 - 4026
Main Authors Gumilar, Khanisyah Erza, Indraprasta, Birama R., Faridzi, Ach Salman, Wibowo, Bagus M., Herlambang, Aditya, Rahestyningtyas, Eccita, Irawan, Budi, Tambunan, Zulkarnain, Bustomi, Ahmad Fadhli, Brahmantara, Bagus Ngurah, Yu, Zih-Ying, Hsu, Yu-Cheng, Pramuditya, Herlangga, Putra, Very Great E., Nugroho, Hari, Mulawardhana, Pungky, Tjokroprawiro, Brahmana A., Hedianto, Tri, Ibrahim, Ibrahim H., Huang, Jingshan, Li, Dongqi, Lu, Chien-Hsing, Yang, Jer-Yen, Liao, Li-Na, Tan, Ming
Format Journal Article
LanguageEnglish
Published Netherlands Elsevier B.V 01.12.2024
Research Network of Computational and Structural Biotechnology
Elsevier
Subjects
Online AccessGet full text

Cover

Loading…
Abstract This study investigated the ability of Large Language Models (LLMs) to provide accurate and consistent answers by focusing on their performance in complex gynecologic cancer cases. LLMs are advancing rapidly and require a thorough evaluation to ensure that they can be safely and effectively used in clinical decision-making. Such evaluations are essential for confirming LLM reliability and accuracy in supporting medical professionals in casework. We assessed three prominent LLMs—ChatGPT-4 (CG-4), Gemini Advanced (GemAdv), and Copilot—evaluating their accuracy, consistency, and overall performance. Fifteen clinical vignettes of varying difficulty and five open-ended questions based on real patient cases were used. The responses were coded, randomized, and evaluated blindly by six expert gynecologic oncologists using a 5-point Likert scale for relevance, clarity, depth, focus, and coherence. GemAdv demonstrated superior accuracy (81.87 %) compared to both CG-4 (61.60 %) and Copilot (70.67 %) across all difficulty levels. GemAdv consistently provided correct answers more frequently (>60 % every day during the testing period). Although CG-4 showed a slight advantage in adhering to the National Comprehensive Cancer Network (NCCN) treatment guidelines, GemAdv excelled in the depth and focus of the answers provided, which are crucial aspects of clinical decision-making. LLMs, especially GemAdv, show potential in supporting clinical practice by providing accurate, consistent, and relevant information for gynecologic cancer. However, further refinement is needed for more complex scenarios. This study highlights the promise of LLMs in gynecologic oncology, emphasizing the need for ongoing development and rigorous evaluation to maximize their clinical utility and reliability. [Display omitted] •Large Language Models (LLMs) are valuable tools in clinical practice, aiding healthcare professionals in making evidence-based decisions and improving patient care.•Gemini Advanced achieved 81.87 % accuracy in clinical decision-making.•Gemini Advanced consistently provided correct answers > 60 % every day during the testing period.•ChatGPT-4 and Gemini Advanced outperformed Copilot in treatment recommendations.•Further improvements are necessary to ensure accurate and relevant responses across clinical scenarios.
AbstractList This study investigated the ability of Large Language Models (LLMs) to provide accurate and consistent answers by focusing on their performance in complex gynecologic cancer cases. LLMs are advancing rapidly and require a thorough evaluation to ensure that they can be safely and effectively used in clinical decision-making. Such evaluations are essential for confirming LLM reliability and accuracy in supporting medical professionals in casework. We assessed three prominent LLMs—ChatGPT-4 (CG-4), Gemini Advanced (GemAdv), and Copilot—evaluating their accuracy, consistency, and overall performance. Fifteen clinical vignettes of varying difficulty and five open-ended questions based on real patient cases were used. The responses were coded, randomized, and evaluated blindly by six expert gynecologic oncologists using a 5-point Likert scale for relevance, clarity, depth, focus, and coherence. GemAdv demonstrated superior accuracy (81.87 %) compared to both CG-4 (61.60 %) and Copilot (70.67 %) across all difficulty levels. GemAdv consistently provided correct answers more frequently (>60 % every day during the testing period). Although CG-4 showed a slight advantage in adhering to the National Comprehensive Cancer Network (NCCN) treatment guidelines, GemAdv excelled in the depth and focus of the answers provided, which are crucial aspects of clinical decision-making. LLMs, especially GemAdv, show potential in supporting clinical practice by providing accurate, consistent, and relevant information for gynecologic cancer. However, further refinement is needed for more complex scenarios. This study highlights the promise of LLMs in gynecologic oncology, emphasizing the need for ongoing development and rigorous evaluation to maximize their clinical utility and reliability.
ga1 • Large Language Models (LLMs) are valuable tools in clinical practice, aiding healthcare professionals in making evidence-based decisions and improving patient care. • Gemini Advanced achieved 81.87 % accuracy in clinical decision-making. • Gemini Advanced consistently provided correct answers > 60 % every day during the testing period. • ChatGPT-4 and Gemini Advanced outperformed Copilot in treatment recommendations. • Further improvements are necessary to ensure accurate and relevant responses across clinical scenarios.
This study investigated the ability of Large Language Models (LLMs) to provide accurate and consistent answers by focusing on their performance in complex gynecologic cancer cases. LLMs are advancing rapidly and require a thorough evaluation to ensure that they can be safely and effectively used in clinical decision-making. Such evaluations are essential for confirming LLM reliability and accuracy in supporting medical professionals in casework. We assessed three prominent LLMs—ChatGPT-4 (CG-4), Gemini Advanced (GemAdv), and Copilot—evaluating their accuracy, consistency, and overall performance. Fifteen clinical vignettes of varying difficulty and five open-ended questions based on real patient cases were used. The responses were coded, randomized, and evaluated blindly by six expert gynecologic oncologists using a 5-point Likert scale for relevance, clarity, depth, focus, and coherence. GemAdv demonstrated superior accuracy (81.87 %) compared to both CG-4 (61.60 %) and Copilot (70.67 %) across all difficulty levels. GemAdv consistently provided correct answers more frequently (>60 % every day during the testing period). Although CG-4 showed a slight advantage in adhering to the National Comprehensive Cancer Network (NCCN) treatment guidelines, GemAdv excelled in the depth and focus of the answers provided, which are crucial aspects of clinical decision-making. LLMs, especially GemAdv, show potential in supporting clinical practice by providing accurate, consistent, and relevant information for gynecologic cancer. However, further refinement is needed for more complex scenarios. This study highlights the promise of LLMs in gynecologic oncology, emphasizing the need for ongoing development and rigorous evaluation to maximize their clinical utility and reliability. [Display omitted] •Large Language Models (LLMs) are valuable tools in clinical practice, aiding healthcare professionals in making evidence-based decisions and improving patient care.•Gemini Advanced achieved 81.87 % accuracy in clinical decision-making.•Gemini Advanced consistently provided correct answers > 60 % every day during the testing period.•ChatGPT-4 and Gemini Advanced outperformed Copilot in treatment recommendations.•Further improvements are necessary to ensure accurate and relevant responses across clinical scenarios.
This study investigated the ability of Large Language Models (LLMs) to provide accurate and consistent answers by focusing on their performance in complex gynecologic cancer cases. LLMs are advancing rapidly and require a thorough evaluation to ensure that they can be safely and effectively used in clinical decision-making. Such evaluations are essential for confirming LLM reliability and accuracy in supporting medical professionals in casework. We assessed three prominent LLMs-ChatGPT-4 (CG-4), Gemini Advanced (GemAdv), and Copilot-evaluating their accuracy, consistency, and overall performance. Fifteen clinical vignettes of varying difficulty and five open-ended questions based on real patient cases were used. The responses were coded, randomized, and evaluated blindly by six expert gynecologic oncologists using a 5-point Likert scale for relevance, clarity, depth, focus, and coherence. GemAdv demonstrated superior accuracy (81.87 %) compared to both CG-4 (61.60 %) and Copilot (70.67 %) across all difficulty levels. GemAdv consistently provided correct answers more frequently (>60 % every day during the testing period). Although CG-4 showed a slight advantage in adhering to the National Comprehensive Cancer Network (NCCN) treatment guidelines, GemAdv excelled in the depth and focus of the answers provided, which are crucial aspects of clinical decision-making. LLMs, especially GemAdv, show potential in supporting clinical practice by providing accurate, consistent, and relevant information for gynecologic cancer. However, further refinement is needed for more complex scenarios. This study highlights the promise of LLMs in gynecologic oncology, emphasizing the need for ongoing development and rigorous evaluation to maximize their clinical utility and reliability.
Objective: This study investigated the ability of Large Language Models (LLMs) to provide accurate and consistent answers by focusing on their performance in complex gynecologic cancer cases. Background: LLMs are advancing rapidly and require a thorough evaluation to ensure that they can be safely and effectively used in clinical decision-making. Such evaluations are essential for confirming LLM reliability and accuracy in supporting medical professionals in casework. Study design: We assessed three prominent LLMs—ChatGPT-4 (CG-4), Gemini Advanced (GemAdv), and Copilot—evaluating their accuracy, consistency, and overall performance. Fifteen clinical vignettes of varying difficulty and five open-ended questions based on real patient cases were used. The responses were coded, randomized, and evaluated blindly by six expert gynecologic oncologists using a 5-point Likert scale for relevance, clarity, depth, focus, and coherence. Results: GemAdv demonstrated superior accuracy (81.87 %) compared to both CG-4 (61.60 %) and Copilot (70.67 %) across all difficulty levels. GemAdv consistently provided correct answers more frequently (>60 % every day during the testing period). Although CG-4 showed a slight advantage in adhering to the National Comprehensive Cancer Network (NCCN) treatment guidelines, GemAdv excelled in the depth and focus of the answers provided, which are crucial aspects of clinical decision-making. Conclusion: LLMs, especially GemAdv, show potential in supporting clinical practice by providing accurate, consistent, and relevant information for gynecologic cancer. However, further refinement is needed for more complex scenarios. This study highlights the promise of LLMs in gynecologic oncology, emphasizing the need for ongoing development and rigorous evaluation to maximize their clinical utility and reliability.
This study investigated the ability of Large Language Models (LLMs) to provide accurate and consistent answers by focusing on their performance in complex gynecologic cancer cases.ObjectiveThis study investigated the ability of Large Language Models (LLMs) to provide accurate and consistent answers by focusing on their performance in complex gynecologic cancer cases.LLMs are advancing rapidly and require a thorough evaluation to ensure that they can be safely and effectively used in clinical decision-making. Such evaluations are essential for confirming LLM reliability and accuracy in supporting medical professionals in casework.BackgroundLLMs are advancing rapidly and require a thorough evaluation to ensure that they can be safely and effectively used in clinical decision-making. Such evaluations are essential for confirming LLM reliability and accuracy in supporting medical professionals in casework.We assessed three prominent LLMs-ChatGPT-4 (CG-4), Gemini Advanced (GemAdv), and Copilot-evaluating their accuracy, consistency, and overall performance. Fifteen clinical vignettes of varying difficulty and five open-ended questions based on real patient cases were used. The responses were coded, randomized, and evaluated blindly by six expert gynecologic oncologists using a 5-point Likert scale for relevance, clarity, depth, focus, and coherence.Study designWe assessed three prominent LLMs-ChatGPT-4 (CG-4), Gemini Advanced (GemAdv), and Copilot-evaluating their accuracy, consistency, and overall performance. Fifteen clinical vignettes of varying difficulty and five open-ended questions based on real patient cases were used. The responses were coded, randomized, and evaluated blindly by six expert gynecologic oncologists using a 5-point Likert scale for relevance, clarity, depth, focus, and coherence.GemAdv demonstrated superior accuracy (81.87 %) compared to both CG-4 (61.60 %) and Copilot (70.67 %) across all difficulty levels. GemAdv consistently provided correct answers more frequently (>60 % every day during the testing period). Although CG-4 showed a slight advantage in adhering to the National Comprehensive Cancer Network (NCCN) treatment guidelines, GemAdv excelled in the depth and focus of the answers provided, which are crucial aspects of clinical decision-making.ResultsGemAdv demonstrated superior accuracy (81.87 %) compared to both CG-4 (61.60 %) and Copilot (70.67 %) across all difficulty levels. GemAdv consistently provided correct answers more frequently (>60 % every day during the testing period). Although CG-4 showed a slight advantage in adhering to the National Comprehensive Cancer Network (NCCN) treatment guidelines, GemAdv excelled in the depth and focus of the answers provided, which are crucial aspects of clinical decision-making.LLMs, especially GemAdv, show potential in supporting clinical practice by providing accurate, consistent, and relevant information for gynecologic cancer. However, further refinement is needed for more complex scenarios. This study highlights the promise of LLMs in gynecologic oncology, emphasizing the need for ongoing development and rigorous evaluation to maximize their clinical utility and reliability.ConclusionLLMs, especially GemAdv, show potential in supporting clinical practice by providing accurate, consistent, and relevant information for gynecologic cancer. However, further refinement is needed for more complex scenarios. This study highlights the promise of LLMs in gynecologic oncology, emphasizing the need for ongoing development and rigorous evaluation to maximize their clinical utility and reliability.
Author Hsu, Yu-Cheng
Hedianto, Tri
Huang, Jingshan
Irawan, Budi
Wibowo, Bagus M.
Brahmantara, Bagus Ngurah
Yu, Zih-Ying
Liao, Li-Na
Indraprasta, Birama R.
Ibrahim, Ibrahim H.
Lu, Chien-Hsing
Tjokroprawiro, Brahmana A.
Tan, Ming
Herlambang, Aditya
Nugroho, Hari
Bustomi, Ahmad Fadhli
Pramuditya, Herlangga
Mulawardhana, Pungky
Yang, Jer-Yen
Faridzi, Ach Salman
Putra, Very Great E.
Rahestyningtyas, Eccita
Gumilar, Khanisyah Erza
Tambunan, Zulkarnain
Li, Dongqi
Author_xml – sequence: 1
  givenname: Khanisyah Erza
  surname: Gumilar
  fullname: Gumilar, Khanisyah Erza
  email: khanisyah@fk.unair.ac.id
  organization: Graduate Institute of Biomedical Science, China Medical University, Taichung, Taiwan
– sequence: 2
  givenname: Birama R.
  surname: Indraprasta
  fullname: Indraprasta, Birama R.
  organization: Department of Obstetrics and Gynecology, Dr. Soetomo General Hospital - Faculty of Medicine, Universitas Airlangga, Surabaya, Indonesia
– sequence: 3
  givenname: Ach Salman
  surname: Faridzi
  fullname: Faridzi, Ach Salman
  organization: Department of Obstetrics and Gynecology, Dr. Soetomo General Hospital - Faculty of Medicine, Universitas Airlangga, Surabaya, Indonesia
– sequence: 4
  givenname: Bagus M.
  surname: Wibowo
  fullname: Wibowo, Bagus M.
  organization: Department of Obstetrics and Gynecology, Dr. Soetomo General Hospital - Faculty of Medicine, Universitas Airlangga, Surabaya, Indonesia
– sequence: 5
  givenname: Aditya
  surname: Herlambang
  fullname: Herlambang, Aditya
  organization: Department of Obstetrics and Gynecology, Dr. Soetomo General Hospital - Faculty of Medicine, Universitas Airlangga, Surabaya, Indonesia
– sequence: 6
  givenname: Eccita
  surname: Rahestyningtyas
  fullname: Rahestyningtyas, Eccita
  organization: Department of Obstetrics and Gynecology, Hospital of Universitas Airlangga - Faculty of Medicine, Universitas Airlangga, Surabaya, Indonesia
– sequence: 7
  givenname: Budi
  surname: Irawan
  fullname: Irawan, Budi
  organization: Department of Obstetrics and Gynecology, Dr. Soetomo General Hospital - Faculty of Medicine, Universitas Airlangga, Surabaya, Indonesia
– sequence: 8
  givenname: Zulkarnain
  surname: Tambunan
  fullname: Tambunan, Zulkarnain
  organization: Department of Obstetrics and Gynecology, Dr. Soetomo General Hospital - Faculty of Medicine, Universitas Airlangga, Surabaya, Indonesia
– sequence: 9
  givenname: Ahmad Fadhli
  surname: Bustomi
  fullname: Bustomi, Ahmad Fadhli
  organization: Department of Obstetrics and Gynecology, Dr. Soetomo General Hospital - Faculty of Medicine, Universitas Airlangga, Surabaya, Indonesia
– sequence: 10
  givenname: Bagus Ngurah
  surname: Brahmantara
  fullname: Brahmantara, Bagus Ngurah
  organization: Department of Obstetrics and Gynecology, Dr. Soetomo General Hospital - Faculty of Medicine, Universitas Airlangga, Surabaya, Indonesia
– sequence: 11
  givenname: Zih-Ying
  surname: Yu
  fullname: Yu, Zih-Ying
  organization: Department of Public Health, China Medical University, Taichung, Taiwan
– sequence: 12
  givenname: Yu-Cheng
  surname: Hsu
  fullname: Hsu, Yu-Cheng
  organization: Department of Public Health, China Medical University, Taichung, Taiwan
– sequence: 13
  givenname: Herlangga
  surname: Pramuditya
  fullname: Pramuditya, Herlangga
  organization: Department of Obstetrics and Gynecology, Dr. Ramelan Naval Hospital, Surabaya, Indonesia
– sequence: 14
  givenname: Very Great E.
  surname: Putra
  fullname: Putra, Very Great E.
  organization: Department of Obstetrics and Gynecology, Dr. Kariadi Central General Hospital, Semarang, Indonesia
– sequence: 15
  givenname: Hari
  surname: Nugroho
  fullname: Nugroho, Hari
  organization: Department of Obstetrics and Gynecology, Dr. Soetomo General Hospital - Faculty of Medicine, Universitas Airlangga, Surabaya, Indonesia
– sequence: 16
  givenname: Pungky
  surname: Mulawardhana
  fullname: Mulawardhana, Pungky
  organization: Department of Obstetrics and Gynecology, Hospital of Universitas Airlangga - Faculty of Medicine, Universitas Airlangga, Surabaya, Indonesia
– sequence: 17
  givenname: Brahmana A.
  surname: Tjokroprawiro
  fullname: Tjokroprawiro, Brahmana A.
  organization: Department of Obstetrics and Gynecology, Dr. Soetomo General Hospital - Faculty of Medicine, Universitas Airlangga, Surabaya, Indonesia
– sequence: 18
  givenname: Tri
  surname: Hedianto
  fullname: Hedianto, Tri
  organization: Faculty of Medicine and Health, Institut Teknologi Sepuluh Nopember, Surabaya, Indonesia
– sequence: 19
  givenname: Ibrahim H.
  surname: Ibrahim
  fullname: Ibrahim, Ibrahim H.
  organization: Graduate Institute of Biomedical Science, China Medical University, Taichung, Taiwan
– sequence: 20
  givenname: Jingshan
  surname: Huang
  fullname: Huang, Jingshan
  organization: School of Computing, College of Medicine, University of South Alabama, Mobile, AL, USA
– sequence: 21
  givenname: Dongqi
  surname: Li
  fullname: Li, Dongqi
  organization: School of Information and Computer Sciences, School of Social and Behavioral Sciences, University of California, Irvine, CA, USA
– sequence: 22
  givenname: Chien-Hsing
  surname: Lu
  fullname: Lu, Chien-Hsing
  organization: Department of Obstetrics and Gynecology, Taichung Veteran General Hospital, Taichung, Taiwan
– sequence: 23
  givenname: Jer-Yen
  surname: Yang
  fullname: Yang, Jer-Yen
  email: jyyang@cmu.edu.tw
  organization: Graduate Institute of Biomedical Science, China Medical University, Taichung, Taiwan
– sequence: 24
  givenname: Li-Na
  surname: Liao
  fullname: Liao, Li-Na
  email: linaliao@mail.cmu.edu.tw
  organization: Department of Public Health, China Medical University, Taichung, Taiwan
– sequence: 25
  givenname: Ming
  surname: Tan
  fullname: Tan, Ming
  email: mingtan@mail.cmu.edu.tw
  organization: Graduate Institute of Biomedical Science, China Medical University, Taichung, Taiwan
BackLink https://www.ncbi.nlm.nih.gov/pubmed/39610903$$D View this record in MEDLINE/PubMed
BookMark eNqFUk1v1DAQjVARLaV_gAPKsRx2GccfSSQkVFV8VErFBa5YzmQcvGTtxc5W2n-Pt1uqlkPxwR6N33sez5uXxZEPnoriNYMlA6berZaY-tWygkrkxBIkPCtOKgC2AF7D0YP4uDhLaQV5NUy1HF4Ux7xVDFrgJ8WPi5QopTX5uQy27EwcKe9-3JocXIeBplSed911els6Xw6ELrngF2vzy_mxTNvNJsS5tCGW484ThimMDsvgb6Pdq-K5NVOis7vztPj-6eO3yy-L7uvnq8uLboFSynkhGlMRUkVS9IKjMJYUNMArqZhQRJykbaUBsIb1Ei0BEtpWqcGSbXHgp8XVQXcIZqU30a1N3OlgnL5NhDhqE2eHE2mqYZC1lTU2TNS87a2o8wNKtWiqgYus9eGgtdn2axowtyaa6ZHo4xvvfuox3GjGFHCANiuc3ynE8HtLadZrl5CmyXgK26R59oZnxxr-fyjjAlQLAjL0zcO67gv6a2YGVAcAxpBSJHsPYaD3Q6NXej80ej80-1wuIZOaf0joZjNnj_Pf3PQ09f2BmmeEbhxFndCRRxpcJJxz691T9D9LHN1J
CitedBy_id crossref_primary_10_3390_jcm14030875
crossref_primary_10_1016_j_csbj_2025_03_026
crossref_primary_10_1016_j_csbj_2024_12_013
Cites_doi 10.1109/ICAAIC56838.2023.10140214
10.1016/j.ajog.2023.04.020
10.1007/s00403-019-01957-2
10.2196/51873
10.1111/iej.13985
10.2196/45312
10.5468/ogs.23231
10.1115/1.4062773
10.1093/annonc/mdq417
10.1371/journal.pdig.0000198
10.1109/JAS.2023.123618
10.1186/s13000-024-01464-7
10.2214/AJR.21.27195
10.1093/asjof/ojad084
10.6004/jnccn.2023.0006
10.1097/CCO.0000000000000916
10.1007/s10459-023-10257-4
10.1148/radiol.230922
ContentType Journal Article
Copyright 2024 The Authors
2024 The Authors.
2024 The Authors 2024
Copyright_xml – notice: 2024 The Authors
– notice: 2024 The Authors.
– notice: 2024 The Authors 2024
DBID 6I.
AAFTH
AAYXX
CITATION
NPM
7X8
7S9
L.6
5PM
DOA
DOI 10.1016/j.csbj.2024.10.050
DatabaseName ScienceDirect Open Access Titles
Elsevier:ScienceDirect:Open Access
CrossRef
PubMed
MEDLINE - Academic
AGRICOLA
AGRICOLA - Academic
PubMed Central (Full Participant titles)
DOAJ Directory of Open Access Journals
DatabaseTitle CrossRef
PubMed
MEDLINE - Academic
AGRICOLA
AGRICOLA - Academic
DatabaseTitleList AGRICOLA


PubMed

MEDLINE - Academic
Database_xml – sequence: 1
  dbid: DOA
  name: DOAJ Directory of Open Access Journals
  url: https://www.doaj.org/
  sourceTypes: Open Website
– sequence: 2
  dbid: NPM
  name: PubMed
  url: https://proxy.k.utb.cz/login?url=http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?db=PubMed
  sourceTypes: Index Database
DeliveryMethod fulltext_linktorsrc
Discipline Engineering
EISSN 2001-0370
EndPage 4026
ExternalDocumentID oai_doaj_org_article_e70d57f57c814739bf47e3e669ca2d34
PMC11603009
39610903
10_1016_j_csbj_2024_10_050
S2001037024003702
Genre Journal Article
GroupedDBID 0R~
457
53G
5VS
6I.
AACTN
AAEDT
AAEDW
AAFTH
AAHBH
AAIKJ
AALRI
AAXUO
ABMAC
ACGFS
ADBBV
ADEZE
ADRAZ
ADVLN
AEXQZ
AFTJW
AGHFR
AITUG
AKRWK
ALMA_UNASSIGNED_HOLDINGS
AMRAJ
AOIJS
BAWUL
BCNDV
DIK
EBS
EJD
FDB
GROUPED_DOAJ
HYE
IPNFZ
KQ8
M41
M48
M~E
O9-
OK1
RIG
ROL
RPM
SSZ
AAYWO
AAYXX
ACVFH
ADCNI
AEUPX
AFPUW
AIGII
AKBMS
AKYEP
CITATION
0SF
NCXOZ
NPM
7X8
7S9
L.6
5PM
ID FETCH-LOGICAL-c555t-48a2ece2e54b43c4afe60803256146ee3e5f95a00fa1b5cfe0cecf966dfef9cd3
IEDL.DBID M48
ISSN 2001-0370
IngestDate Wed Aug 27 01:11:20 EDT 2025
Thu Aug 21 18:36:28 EDT 2025
Fri Aug 22 20:40:24 EDT 2025
Fri Jul 11 14:26:16 EDT 2025
Thu Jan 02 22:25:24 EST 2025
Thu Apr 24 23:08:24 EDT 2025
Tue Jul 01 05:22:45 EDT 2025
Sat Mar 08 15:48:10 EST 2025
IsDoiOpenAccess true
IsOpenAccess true
IsPeerReviewed true
IsScholarly true
Keywords Accuracy
Large Language Models
Artificial intelligence
Gynecologic cancer
Consistency
Language English
License This is an open access article under the CC BY-NC-ND license.
2024 The Authors.
This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/).
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-c555t-48a2ece2e54b43c4afe60803256146ee3e5f95a00fa1b5cfe0cecf966dfef9cd3
Notes ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 23
OpenAccessLink http://journals.scholarsportal.info/openUrl.xqy?doi=10.1016/j.csbj.2024.10.050
PMID 39610903
PQID 3134069040
PQPubID 23479
PageCount 8
ParticipantIDs doaj_primary_oai_doaj_org_article_e70d57f57c814739bf47e3e669ca2d34
pubmedcentral_primary_oai_pubmedcentral_nih_gov_11603009
proquest_miscellaneous_3200305083
proquest_miscellaneous_3134069040
pubmed_primary_39610903
crossref_primary_10_1016_j_csbj_2024_10_050
crossref_citationtrail_10_1016_j_csbj_2024_10_050
elsevier_sciencedirect_doi_10_1016_j_csbj_2024_10_050
ProviderPackageCode CITATION
AAYXX
PublicationCentury 2000
PublicationDate 2024-12-01
PublicationDateYYYYMMDD 2024-12-01
PublicationDate_xml – month: 12
  year: 2024
  text: 2024-12-01
  day: 01
PublicationDecade 2020
PublicationPlace Netherlands
PublicationPlace_xml – name: Netherlands
PublicationTitle Computational and structural biotechnology journal
PublicationTitleAlternate Comput Struct Biotechnol J
PublicationYear 2024
Publisher Elsevier B.V
Research Network of Computational and Structural Biotechnology
Elsevier
Publisher_xml – name: Elsevier B.V
– name: Research Network of Computational and Structural Biotechnology
– name: Elsevier
References Seth, Lim, Xie, Cevik, Rozen, Ross (bib17) 2023; 5
Kapoor, Haj-Mirzaian, Yan, Wickner, Giess, Eappen (bib14) 2022; 219
Voigt, Trautwein (bib24) 2023; 35
Kung, Cheatham, Medenilla, Sillos, De Leon, Elepano (bib21) 2023; 2
Ullah, Parwani, Baig, Singh (bib23) 2024; 19
Bhardwaz S. ,Kumar J., An Extensive Comparative Analysis of Chatbot Technologies - ChatGPT, Google BARD and Microsoft Bing, in 2023 2nd International Conference on Applied Artificial Intelligence and Computing (ICAAIC). 2023. p. 673–679.
Kamo, Dandapani, Miksad, Houlihan, Kaplan, Regan (bib16) 2011; 22
Lee, Kim (bib4) 2024; 67
Gilson, Safranek, Huang, Socrates, Chi, Taylor (bib20) 2023; 9
Ellaway, Tolsgaard (bib3) 2023; 28
Sikander, Baker, Deveci, Lund, Rosenberg (bib11) 2023; 15
What’s the most popular LLM? 2024 [cited 2024 June 10]; Available from
Daniel, Ma, Singh, Beatrice Bloom, Nilda Adair, William Chen (bib13) 2023; 8
Li, Kemp, Logan, Dimri, Singh, Mattar (bib19) 2023; 229
Wu, He, Liu, Sun, Liu, Han (bib8) 2023; 10
Gumilar, Indraprasta, Hsu, Yiou, Hong, Irawan (bib10) 2024
Brodnik, Carton, Muir, Ghosh, Downey, Echlin (bib2) 2023; 90
Suarez, Diaz-Flores Garcia, Algar, Gomez Sanchez, Llorente de Pedro, Freire (bib22) 2024; 57
.
Vaidya, Mori, Dusza, Rossi, Nehal, Lee (bib15) 2019; 311
Rahsepar, Tavakoli, Kim, Hassani, Abtin, Bedayat (bib7) 2023; 307
Veras, Dyer, Rooney, Barros Silva, Rutherford, Kairy (bib12) 2023; 12
Abu-Rustum, Yashar, Arend, Barber, Bradley, Brooks (bib5) 2023; 21
Gordon, Towbin, Wingrove, Shafique, Haas, Kitts (bib6) 2023
Zuniga Salazar, Zuniga, Vindel, Yoong, Hincapie, Zuniga (bib18) 2023; 15
Voigt (10.1016/j.csbj.2024.10.050_bib24) 2023; 35
Ellaway (10.1016/j.csbj.2024.10.050_bib3) 2023; 28
Abu-Rustum (10.1016/j.csbj.2024.10.050_bib5) 2023; 21
10.1016/j.csbj.2024.10.050_bib9
Zuniga Salazar (10.1016/j.csbj.2024.10.050_bib18) 2023; 15
Veras (10.1016/j.csbj.2024.10.050_bib12) 2023; 12
Lee (10.1016/j.csbj.2024.10.050_bib4) 2024; 67
10.1016/j.csbj.2024.10.050_bib1
Li (10.1016/j.csbj.2024.10.050_bib19) 2023; 229
Kapoor (10.1016/j.csbj.2024.10.050_bib14) 2022; 219
Vaidya (10.1016/j.csbj.2024.10.050_bib15) 2019; 311
Gilson (10.1016/j.csbj.2024.10.050_bib20) 2023; 9
Kung (10.1016/j.csbj.2024.10.050_bib21) 2023; 2
Suarez (10.1016/j.csbj.2024.10.050_bib22) 2024; 57
Brodnik (10.1016/j.csbj.2024.10.050_bib2) 2023; 90
Gordon (10.1016/j.csbj.2024.10.050_bib6) 2023
Wu (10.1016/j.csbj.2024.10.050_bib8) 2023; 10
Sikander (10.1016/j.csbj.2024.10.050_bib11) 2023; 15
Kamo (10.1016/j.csbj.2024.10.050_bib16) 2011; 22
Gumilar (10.1016/j.csbj.2024.10.050_bib10) 2024
Ullah (10.1016/j.csbj.2024.10.050_bib23) 2024; 19
Rahsepar (10.1016/j.csbj.2024.10.050_bib7) 2023; 307
Daniel (10.1016/j.csbj.2024.10.050_bib13) 2023; 8
Seth (10.1016/j.csbj.2024.10.050_bib17) 2023; 5
References_xml – reference: Bhardwaz S. ,Kumar J., An Extensive Comparative Analysis of Chatbot Technologies - ChatGPT, Google BARD and Microsoft Bing, in 2023 2nd International Conference on Applied Artificial Intelligence and Computing (ICAAIC). 2023. p. 673–679.
– volume: 311
  start-page: 691
  year: 2019
  end-page: 696
  ident: bib15
  article-title: Appearance-related psychosocial distress following facial skin cancer surgery using the FACE-Q Skin Cancer
  publication-title: Arch Dermatol Res
– volume: 28
  start-page: 659
  year: 2023
  end-page: 664
  ident: bib3
  article-title: Artificial scholarship: LLMs in health professions education research
  publication-title: Adv Health Sci Educ Theory Pr
– volume: 229
  year: 2023
  ident: bib19
  article-title: ChatGPT outscored human candidates in a virtual objective structured clinical examination in obstetrics and gynecology
  publication-title: Am J Obstet Gynecol
– volume: 35
  start-page: 68
  year: 2023
  end-page: 77
  ident: bib24
  article-title: Improved guideline adherence in oncology through clinical decision-support systems: still hindered by current health IT infrastructures?
  publication-title: Curr Opin Oncol
– volume: 10
  start-page: 1122
  year: 2023
  end-page: 1136
  ident: bib8
  article-title: A brief overview of ChatGPT: the history, status quo and potential future development
  publication-title: IEEE/CAA J Autom Sin
– volume: 57
  start-page: 108
  year: 2024
  end-page: 113
  ident: bib22
  article-title: Unveiling the ChatGPT phenomenon: evaluating the consistency and accuracy of endodontic question answers
  publication-title: Int Endod J
– volume: 67
  start-page: 153
  year: 2024
  end-page: 159
  ident: bib4
  article-title: Potential applications of ChatGPT in obstetrics and gynecology in Korea: a review article
  publication-title: Obstet Gynecol Sci
– volume: 90
  year: 2023
  ident: bib2
  article-title: Perspective: large language models in applied mechanics
  publication-title: J Appl Mech
– volume: 219
  start-page: 338
  year: 2022
  end-page: 345
  ident: bib14
  article-title: Patient experience scores for radiologists: comparison with nonradiologist physicians and changes after public posting in an institutional online provider directory
  publication-title: Am J Roentgenol
– volume: 8
  start-page: 5
  year: 2023
  ident: bib13
  article-title: Patient experience performance at a primary cancer center versus affiliated community facilities
  publication-title: Adv Radiat Oncol
– reference: What’s the most popular LLM? 2024 [cited 2024 June 10]; Available from:
– volume: 5
  start-page: ojad084
  year: 2023
  ident: bib17
  article-title: Comparing the efficacy of large language models ChatGPT, BARD, and Bing AI in providing information on rhinoplasty: an observational study
  publication-title: Aesthet Surg J Open Forum
– volume: 15
  year: 2023
  ident: bib18
  article-title: Efficacy of AI chats to determine an emergency: a comparison between open AI's ChatGPT, Google Bard, and Microsoft Bing AI Chat
  publication-title: Cureus
– volume: 15
  year: 2023
  ident: bib11
  article-title: ChatGPT-4 and human researchers are equal in writing scientific introduction sections: a blinded, randomized, non-inferiority controlled study
  publication-title: Cureus
– volume: 19
  start-page: 43
  year: 2024
  ident: bib23
  article-title: Challenges and barriers of using large language models (LLM) such as ChatGPT for diagnostic medicine with a focus on digital pathology - a recent scoping review
  publication-title: Diagn Pathol
– year: 2023
  ident: bib6
  article-title: Enhancing patient communication with Chat-GPT in radiology: evaluating the efficacy and readability of answers to common imaging-related questions
  publication-title: J Am Coll Radio
– reference: .
– year: 2024
  ident: bib10
  article-title: Disparities in medical recommendations from AIbased chatbots across different countries/regions
  publication-title: Res Sq
– volume: 12
  year: 2023
  ident: bib12
  article-title: Usability and efficacy of artificial intelligence chatbots (ChatGPT) for health sciences students: protocol for a crossover randomized controlled trial
  publication-title: JMIR Res Protoc
– volume: 9
  year: 2023
  ident: bib20
  article-title: How Does ChatGPT perform on the United States medical licensing examination (USMLE)? The implications of large language models for medical education and knowledge assessment
  publication-title: JMIR Med Educ
– volume: 2
  year: 2023
  ident: bib21
  article-title: Performance of ChatGPT on USMLE: potential for AI-assisted medical education using large language models
  publication-title: PLOS Digit Health
– volume: 307
  year: 2023
  ident: bib7
  article-title: How AI responds to common lung cancer questions: ChatGPT vs Google Bard
  publication-title: Radiology
– volume: 21
  start-page: 181
  year: 2023
  end-page: 209
  ident: bib5
  article-title: Uterine neoplasms, version 1.2023, NCCN clinical practice guidelines in oncology
  publication-title: J Natl Compr Canc Netw
– volume: 22
  start-page: 723
  year: 2011
  end-page: 729
  ident: bib16
  article-title: Evaluation of the SCA instrument for measuring patient satisfaction with cancer care administered via paper or via the Internet
  publication-title: Ann Oncol
– ident: 10.1016/j.csbj.2024.10.050_bib9
  doi: 10.1109/ICAAIC56838.2023.10140214
– volume: 229
  issue: 2
  year: 2023
  ident: 10.1016/j.csbj.2024.10.050_bib19
  article-title: ChatGPT outscored human candidates in a virtual objective structured clinical examination in obstetrics and gynecology
  publication-title: Am J Obstet Gynecol
  doi: 10.1016/j.ajog.2023.04.020
– year: 2023
  ident: 10.1016/j.csbj.2024.10.050_bib6
  article-title: Enhancing patient communication with Chat-GPT in radiology: evaluating the efficacy and readability of answers to common imaging-related questions
  publication-title: J Am Coll Radio
– volume: 311
  start-page: 691
  issue: 9
  year: 2019
  ident: 10.1016/j.csbj.2024.10.050_bib15
  article-title: Appearance-related psychosocial distress following facial skin cancer surgery using the FACE-Q Skin Cancer
  publication-title: Arch Dermatol Res
  doi: 10.1007/s00403-019-01957-2
– volume: 15
  issue: 9
  year: 2023
  ident: 10.1016/j.csbj.2024.10.050_bib18
  article-title: Efficacy of AI chats to determine an emergency: a comparison between open AI's ChatGPT, Google Bard, and Microsoft Bing AI Chat
  publication-title: Cureus
– year: 2024
  ident: 10.1016/j.csbj.2024.10.050_bib10
  article-title: Disparities in medical recommendations from AIbased chatbots across different countries/regions
  publication-title: Res Sq
– volume: 12
  year: 2023
  ident: 10.1016/j.csbj.2024.10.050_bib12
  article-title: Usability and efficacy of artificial intelligence chatbots (ChatGPT) for health sciences students: protocol for a crossover randomized controlled trial
  publication-title: JMIR Res Protoc
  doi: 10.2196/51873
– volume: 57
  start-page: 108
  issue: 1
  year: 2024
  ident: 10.1016/j.csbj.2024.10.050_bib22
  article-title: Unveiling the ChatGPT phenomenon: evaluating the consistency and accuracy of endodontic question answers
  publication-title: Int Endod J
  doi: 10.1111/iej.13985
– volume: 15
  issue: 11
  year: 2023
  ident: 10.1016/j.csbj.2024.10.050_bib11
  article-title: ChatGPT-4 and human researchers are equal in writing scientific introduction sections: a blinded, randomized, non-inferiority controlled study
  publication-title: Cureus
– volume: 9
  year: 2023
  ident: 10.1016/j.csbj.2024.10.050_bib20
  article-title: How Does ChatGPT perform on the United States medical licensing examination (USMLE)? The implications of large language models for medical education and knowledge assessment
  publication-title: JMIR Med Educ
  doi: 10.2196/45312
– volume: 67
  start-page: 153
  issue: 2
  year: 2024
  ident: 10.1016/j.csbj.2024.10.050_bib4
  article-title: Potential applications of ChatGPT in obstetrics and gynecology in Korea: a review article
  publication-title: Obstet Gynecol Sci
  doi: 10.5468/ogs.23231
– volume: 90
  issue: 10
  year: 2023
  ident: 10.1016/j.csbj.2024.10.050_bib2
  article-title: Perspective: large language models in applied mechanics
  publication-title: J Appl Mech
  doi: 10.1115/1.4062773
– ident: 10.1016/j.csbj.2024.10.050_bib1
– volume: 22
  start-page: 723
  issue: 3
  year: 2011
  ident: 10.1016/j.csbj.2024.10.050_bib16
  article-title: Evaluation of the SCA instrument for measuring patient satisfaction with cancer care administered via paper or via the Internet
  publication-title: Ann Oncol
  doi: 10.1093/annonc/mdq417
– volume: 2
  issue: 2
  year: 2023
  ident: 10.1016/j.csbj.2024.10.050_bib21
  article-title: Performance of ChatGPT on USMLE: potential for AI-assisted medical education using large language models
  publication-title: PLOS Digit Health
  doi: 10.1371/journal.pdig.0000198
– volume: 10
  start-page: 1122
  issue: 5
  year: 2023
  ident: 10.1016/j.csbj.2024.10.050_bib8
  article-title: A brief overview of ChatGPT: the history, status quo and potential future development
  publication-title: IEEE/CAA J Autom Sin
  doi: 10.1109/JAS.2023.123618
– volume: 19
  start-page: 43
  issue: 1
  year: 2024
  ident: 10.1016/j.csbj.2024.10.050_bib23
  article-title: Challenges and barriers of using large language models (LLM) such as ChatGPT for diagnostic medicine with a focus on digital pathology - a recent scoping review
  publication-title: Diagn Pathol
  doi: 10.1186/s13000-024-01464-7
– volume: 219
  start-page: 338
  issue: 2
  year: 2022
  ident: 10.1016/j.csbj.2024.10.050_bib14
  article-title: Patient experience scores for radiologists: comparison with nonradiologist physicians and changes after public posting in an institutional online provider directory
  publication-title: Am J Roentgenol
  doi: 10.2214/AJR.21.27195
– volume: 5
  start-page: ojad084
  year: 2023
  ident: 10.1016/j.csbj.2024.10.050_bib17
  article-title: Comparing the efficacy of large language models ChatGPT, BARD, and Bing AI in providing information on rhinoplasty: an observational study
  publication-title: Aesthet Surg J Open Forum
  doi: 10.1093/asjof/ojad084
– volume: 8
  start-page: 5
  year: 2023
  ident: 10.1016/j.csbj.2024.10.050_bib13
  article-title: Patient experience performance at a primary cancer center versus affiliated community facilities
  publication-title: Adv Radiat Oncol
– volume: 21
  start-page: 181
  issue: 2
  year: 2023
  ident: 10.1016/j.csbj.2024.10.050_bib5
  article-title: Uterine neoplasms, version 1.2023, NCCN clinical practice guidelines in oncology
  publication-title: J Natl Compr Canc Netw
  doi: 10.6004/jnccn.2023.0006
– volume: 35
  start-page: 68
  issue: 1
  year: 2023
  ident: 10.1016/j.csbj.2024.10.050_bib24
  article-title: Improved guideline adherence in oncology through clinical decision-support systems: still hindered by current health IT infrastructures?
  publication-title: Curr Opin Oncol
  doi: 10.1097/CCO.0000000000000916
– volume: 28
  start-page: 659
  issue: 3
  year: 2023
  ident: 10.1016/j.csbj.2024.10.050_bib3
  article-title: Artificial scholarship: LLMs in health professions education research
  publication-title: Adv Health Sci Educ Theory Pr
  doi: 10.1007/s10459-023-10257-4
– volume: 307
  issue: 5
  year: 2023
  ident: 10.1016/j.csbj.2024.10.050_bib7
  article-title: How AI responds to common lung cancer questions: ChatGPT vs Google Bard
  publication-title: Radiology
  doi: 10.1148/radiol.230922
SSID ssj0000816930
Score 2.3461647
Snippet This study investigated the ability of Large Language Models (LLMs) to provide accurate and consistent answers by focusing on their performance in complex...
ga1 • Large Language Models (LLMs) are valuable tools in clinical practice, aiding healthcare professionals in making evidence-based decisions and improving...
Objective: This study investigated the ability of Large Language Models (LLMs) to provide accurate and consistent answers by focusing on their performance in...
SourceID doaj
pubmedcentral
proquest
pubmed
crossref
elsevier
SourceType Open Website
Open Access Repository
Aggregation Database
Index Database
Enrichment Source
Publisher
StartPage 4019
SubjectTerms Accuracy
Artificial intelligence
biotechnology
Consistency
decision making
Gynecologic cancer
Large Language Models
patients
SummonAdditionalLinks – databaseName: DOAJ Directory of Open Access Journals
  dbid: DOA
  link: http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwrV1Lb9QwELZQT3BAlOfykpE4gFAgiR9xjgVRVWjLiUo9YcX2GLYq2YrsHvrvO2NnV1mQlguXKLKdhz1jzzfJ-BvGXjsXlIc6FpUToZAqmMJ02hShc0obtMhtSRucT7_qkzP55VydT1J9UUxYpgfOA_cBmjKoJqrGm0o2onVRNiBA69Z3dRCJCRRt3sSZSmuwIZKRctwlkwO6_OAu0CGs5XsK56KN9hNLlAj7dwzS34Dzz7jJiSE6vsfujgiSH-U3P2S3oL_P7kx4BR-w70dbvk2-jHxO0d54zF8mOaU_uxz4m_n8dHjLFz0PY56d4ldKTcWH9RWhco54lv-47sHnBZIv-3R2_ZCdHX_-9umkGBMpFF4ptSqk6WpAiYCSTgovuwgakaKoiQZUAw6niq3qyjJ2lVM-QunBR3SEQoTY-iAesYN-2cMTxqkEjDetkV7qsnUI4KR2KqAn5modZqzaDKr1I8s4Jbu4tJtwsgtLgrAkCCpDQczYu-01V5ljY2_rjySrbUvix04FqDV21Br7L62ZMbWRtB2hRoYQeKvF3oe_2qiFxXlIP1e6HpbrwYpK0CZiXBP3tKmTy4aod8YeZ1XadkO0RHxfYo3ZUbKdfu7W9IufiQ-8olThiJWf_o-RecZuU39zxM5zdrD6vYYXiLtW7mWaYjcP3S0a
  priority: 102
  providerName: Directory of Open Access Journals
Title Assessment of Large Language Models (LLMs) in decision-making support for gynecologic oncology
URI https://dx.doi.org/10.1016/j.csbj.2024.10.050
https://www.ncbi.nlm.nih.gov/pubmed/39610903
https://www.proquest.com/docview/3134069040
https://www.proquest.com/docview/3200305083
https://pubmed.ncbi.nlm.nih.gov/PMC11603009
https://doaj.org/article/e70d57f57c814739bf47e3e669ca2d34
Volume 23
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV1bb9MwFLam7QUeEHfKoDISDyCUKYkvcR4QKohpYi0PiIo9EcW30akkW9NK9N9zTuKUFaaKl6hynET2sX2-4x5_HyEvtbbCuNRHiWY24sKqSJVSRbbUQirwyHmMB5wnn-XJlH86E2d7pJc7Ch3Y3BjaoZ7UdDE_-nW1fgcT_u2fXC3T6AuI9VJ-hJlaGMIfgGfKUNFgEuB-uzIrpB7BbZeQSZTF4RzNza_Z8lUtpf-Wy_oXkv6dWXnNVR3fJXcCxqSjblDcI3uuuk9uX2MefEC-jzaMnLT2dIz54HDt9i4pCqTNG_pqPJ40r-msojYo8UQ_W_Eq2qwuscsoIF56vq6c6ZZQWlftr_VDMj3--PXDSRSkFiIjhFhGXJWpA5s5wTVnhpfeScCSLEWiUOkcc8LnooxjXyZaGO9i44yHUMl653Nj2SOyX9WVe0IoljhlVK644TLONUA8LrWwEKvpVNoBSfpOLUzgIUc5jHnRJ5xdFGiIAg2BZWCIAXmzeeayY-HYWfs92mpTExm024J6cV6ECVm4LLYi8yIzKuEZy7XnGTRTytyUqWV8QERv6SKAkQ5kwKtmOz_-oh8WBcxU_PulrFy9agqWMDxmDKvmjjppG9QBLh6Qx91Q2jSD5UiNH8MdtTXIttq5faea_WgZwxMUEwc0_fQ_PnxIbmFzupSdZ2R_uVi55wC8lnpIDkanX76dDtuNi2E7t34DWtgu8A
linkProvider Scholars Portal
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Assessment+of+Large+Language+Models+%28LLMs%29+in+decision-making+support+for+gynecologic+oncology&rft.jtitle=Computational+and+structural+biotechnology+journal&rft.au=Gumilar%2C+Khanisyah+Erza&rft.au=Indraprasta%2C+Birama+R&rft.au=Faridzi%2C+Ach+Salman&rft.au=Wibowo%2C+Bagus+M&rft.date=2024-12-01&rft.issn=2001-0370&rft.eissn=2001-0370&rft.volume=23+p.4019-4026&rft.spage=4019&rft.epage=4026&rft_id=info:doi/10.1016%2Fj.csbj.2024.10.050&rft.externalDBID=NO_FULL_TEXT
thumbnail_l http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/lc.gif&issn=2001-0370&client=summon
thumbnail_m http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/mc.gif&issn=2001-0370&client=summon
thumbnail_s http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/sc.gif&issn=2001-0370&client=summon