Classifying Unstructured Text in Electronic Health Records for Mental Health Prediction Models: Large Language Model Evaluation Study
Prediction models have demonstrated a range of applications across medicine, including using electronic health record (EHR) data to identify hospital readmission and mortality risk. Large language models (LLMs) can transform unstructured EHR text into structured features, which can then be integrate...
Saved in:
Published in | JMIR medical informatics Vol. 13; p. e65454 |
---|---|
Main Authors | , , , , , , , |
Format | Journal Article |
Language | English |
Published |
Canada
JMIR Publications
21.01.2025
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Abstract | Prediction models have demonstrated a range of applications across medicine, including using electronic health record (EHR) data to identify hospital readmission and mortality risk. Large language models (LLMs) can transform unstructured EHR text into structured features, which can then be integrated into statistical prediction models, ensuring that the results are both clinically meaningful and interpretable.
This study aims to compare the classification decisions made by clinical experts with those generated by a state-of-the-art LLM, using terms extracted from a large EHR data set of individuals with mental health disorders seen in emergency departments (EDs).
Using a dataset from the EHR systems of more than 50 health care provider organizations in the United States from 2016 to 2021, we extracted all clinical terms that appeared in at least 1000 records of individuals admitted to the ED for a mental health-related problem from a source population of over 6 million ED episodes. Two experienced mental health clinicians (one medically trained psychiatrist and one clinical psychologist) reached consensus on the classification of EHR terms and diagnostic codes into categories. We evaluated an LLM's agreement with clinical judgment across three classification tasks as follows: (1) classify terms into "mental health" or "physical health", (2) classify mental health terms into 1 of 42 prespecified categories, and (3) classify physical health terms into 1 of 19 prespecified broad categories.
There was high agreement between the LLM and clinical experts when categorizing 4553 terms as "mental health" or "physical health" (κ=0.77, 95% CI 0.75-0.80). However, there was still considerable variability in LLM-clinician agreement on the classification of mental health terms (κ=0.62, 95% CI 0.59-0.66) and physical health terms (κ=0.69, 95% CI 0.67-0.70).
The LLM displayed high agreement with clinical experts when classifying EHR terms into certain mental health or physical health term categories. However, agreement with clinical experts varied considerably within both sets of mental and physical health term categories. Importantly, the use of LLMs presents an alternative to manual human coding, presenting great potential to create interpretable features for prediction models. |
---|---|
AbstractList | Abstract BackgroundPrediction models have demonstrated a range of applications across medicine, including using electronic health record (EHR) data to identify hospital readmission and mortality risk. Large language models (LLMs) can transform unstructured EHR text into structured features, which can then be integrated into statistical prediction models, ensuring that the results are both clinically meaningful and interpretable. ObjectiveThis study aims to compare the classification decisions made by clinical experts with those generated by a state-of-the-art LLM, using terms extracted from a large EHR data set of individuals with mental health disorders seen in emergency departments (EDs). MethodsUsing a dataset from the EHR systems of more than 50 health care provider organizations in the United States from 2016 to 2021, we extracted all clinical terms that appeared in at least 1000 records of individuals admitted to the ED for a mental health–related problem from a source population of over 6 million ED episodes. Two experienced mental health clinicians (one medically trained psychiatrist and one clinical psychologist) reached consensus on the classification of EHR terms and diagnostic codes into categories. We evaluated an LLM’s agreement with clinical judgment across three classification tasks as follows: (1) classify terms into “mental health” or “physical health”, (2) classify mental health terms into 1 of 42 prespecified categories, and (3) classify physical health terms into 1 of 19 prespecified broad categories. ResultsThere was high agreement between the LLM and clinical experts when categorizing 4553 terms as “mental health” or “physical health” (κ=0.77, 95% CI 0.75-0.80). However, there was still considerable variability in LLM-clinician agreement on the classification of mental health terms (κ=0.62, 95% CI 0.59‐0.66) and physical health terms (κ=0.69, 95% CI 0.67‐0.70). ConclusionsThe LLM displayed high agreement with clinical experts when classifying EHR terms into certain mental health or physical health term categories. However, agreement with clinical experts varied considerably within both sets of mental and physical health term categories. Importantly, the use of LLMs presents an alternative to manual human coding, presenting great potential to create interpretable features for prediction models. Prediction models have demonstrated a range of applications across medicine, including using electronic health record (EHR) data to identify hospital readmission and mortality risk. Large language models (LLMs) can transform unstructured EHR text into structured features, which can then be integrated into statistical prediction models, ensuring that the results are both clinically meaningful and interpretable. This study aims to compare the classification decisions made by clinical experts with those generated by a state-of-the-art LLM, using terms extracted from a large EHR data set of individuals with mental health disorders seen in emergency departments (EDs). Using a dataset from the EHR systems of more than 50 health care provider organizations in the United States from 2016 to 2021, we extracted all clinical terms that appeared in at least 1000 records of individuals admitted to the ED for a mental health-related problem from a source population of over 6 million ED episodes. Two experienced mental health clinicians (one medically trained psychiatrist and one clinical psychologist) reached consensus on the classification of EHR terms and diagnostic codes into categories. We evaluated an LLM's agreement with clinical judgment across three classification tasks as follows: (1) classify terms into "mental health" or "physical health", (2) classify mental health terms into 1 of 42 prespecified categories, and (3) classify physical health terms into 1 of 19 prespecified broad categories. There was high agreement between the LLM and clinical experts when categorizing 4553 terms as "mental health" or "physical health" (κ=0.77, 95% CI 0.75-0.80). However, there was still considerable variability in LLM-clinician agreement on the classification of mental health terms (κ=0.62, 95% CI 0.59-0.66) and physical health terms (κ=0.69, 95% CI 0.67-0.70). The LLM displayed high agreement with clinical experts when classifying EHR terms into certain mental health or physical health term categories. However, agreement with clinical experts varied considerably within both sets of mental and physical health term categories. Importantly, the use of LLMs presents an alternative to manual human coding, presenting great potential to create interpretable features for prediction models. Background:Prediction models have demonstrated a range of applications across medicine, including using electronic health record (EHR) data to identify hospital readmission and mortality risk. Large language models (LLMs) can transform unstructured EHR text into structured features, which can then be integrated into statistical prediction models, ensuring that the results are both clinically meaningful and interpretable.Objective:This study aims to compare the classification decisions made by clinical experts with those generated by a state-of-the-art LLM, using terms extracted from a large EHR data set of individuals with mental health disorders seen in emergency departments (EDs).Methods:Using a dataset from the EHR systems of more than 50 health care provider organizations in the United States from 2016 to 2021, we extracted all clinical terms that appeared in at least 1000 records of individuals admitted to the ED for a mental health–related problem from a source population of over 6 million ED episodes. Two experienced mental health clinicians (one medically trained psychiatrist and one clinical psychologist) reached consensus on the classification of EHR terms and diagnostic codes into categories. We evaluated an LLM’s agreement with clinical judgment across three classification tasks as follows: (1) classify terms into “mental health” or “physical health”, (2) classify mental health terms into 1 of 42 prespecified categories, and (3) classify physical health terms into 1 of 19 prespecified broad categories.Results:There was high agreement between the LLM and clinical experts when categorizing 4553 terms as “mental health” or “physical health” (κ=0.77, 95% CI 0.75-0.80). However, there was still considerable variability in LLM-clinician agreement on the classification of mental health terms (κ=0.62, 95% CI 0.59‐0.66) and physical health terms (κ=0.69, 95% CI 0.67‐0.70).Conclusions:The LLM displayed high agreement with clinical experts when classifying EHR terms into certain mental health or physical health term categories. However, agreement with clinical experts varied considerably within both sets of mental and physical health term categories. Importantly, the use of LLMs presents an alternative to manual human coding, presenting great potential to create interpretable features for prediction models. Prediction models have demonstrated a range of applications across medicine, including using electronic health record (EHR) data to identify hospital readmission and mortality risk. Large language models (LLMs) can transform unstructured EHR text into structured features, which can then be integrated into statistical prediction models, ensuring that the results are both clinically meaningful and interpretable.BackgroundPrediction models have demonstrated a range of applications across medicine, including using electronic health record (EHR) data to identify hospital readmission and mortality risk. Large language models (LLMs) can transform unstructured EHR text into structured features, which can then be integrated into statistical prediction models, ensuring that the results are both clinically meaningful and interpretable.This study aims to compare the classification decisions made by clinical experts with those generated by a state-of-the-art LLM, using terms extracted from a large EHR data set of individuals with mental health disorders seen in emergency departments (EDs).ObjectiveThis study aims to compare the classification decisions made by clinical experts with those generated by a state-of-the-art LLM, using terms extracted from a large EHR data set of individuals with mental health disorders seen in emergency departments (EDs).Using a dataset from the EHR systems of more than 50 health care provider organizations in the United States from 2016 to 2021, we extracted all clinical terms that appeared in at least 1000 records of individuals admitted to the ED for a mental health-related problem from a source population of over 6 million ED episodes. Two experienced mental health clinicians (one medically trained psychiatrist and one clinical psychologist) reached consensus on the classification of EHR terms and diagnostic codes into categories. We evaluated an LLM's agreement with clinical judgment across three classification tasks as follows: (1) classify terms into "mental health" or "physical health", (2) classify mental health terms into 1 of 42 prespecified categories, and (3) classify physical health terms into 1 of 19 prespecified broad categories.MethodsUsing a dataset from the EHR systems of more than 50 health care provider organizations in the United States from 2016 to 2021, we extracted all clinical terms that appeared in at least 1000 records of individuals admitted to the ED for a mental health-related problem from a source population of over 6 million ED episodes. Two experienced mental health clinicians (one medically trained psychiatrist and one clinical psychologist) reached consensus on the classification of EHR terms and diagnostic codes into categories. We evaluated an LLM's agreement with clinical judgment across three classification tasks as follows: (1) classify terms into "mental health" or "physical health", (2) classify mental health terms into 1 of 42 prespecified categories, and (3) classify physical health terms into 1 of 19 prespecified broad categories.There was high agreement between the LLM and clinical experts when categorizing 4553 terms as "mental health" or "physical health" (κ=0.77, 95% CI 0.75-0.80). However, there was still considerable variability in LLM-clinician agreement on the classification of mental health terms (κ=0.62, 95% CI 0.59-0.66) and physical health terms (κ=0.69, 95% CI 0.67-0.70).ResultsThere was high agreement between the LLM and clinical experts when categorizing 4553 terms as "mental health" or "physical health" (κ=0.77, 95% CI 0.75-0.80). However, there was still considerable variability in LLM-clinician agreement on the classification of mental health terms (κ=0.62, 95% CI 0.59-0.66) and physical health terms (κ=0.69, 95% CI 0.67-0.70).The LLM displayed high agreement with clinical experts when classifying EHR terms into certain mental health or physical health term categories. However, agreement with clinical experts varied considerably within both sets of mental and physical health term categories. Importantly, the use of LLMs presents an alternative to manual human coding, presenting great potential to create interpretable features for prediction models.ConclusionsThe LLM displayed high agreement with clinical experts when classifying EHR terms into certain mental health or physical health term categories. However, agreement with clinical experts varied considerably within both sets of mental and physical health term categories. Importantly, the use of LLMs presents an alternative to manual human coding, presenting great potential to create interpretable features for prediction models. |
Author | Olfson, Mark Ungar, Lyle Cullen, Sara W Cardamone, Nicholas C Liu, Tony Schmutte, Timothy Williams, Nathaniel J Marcus, Steven C |
Author_xml | – sequence: 1 givenname: Nicholas C orcidid: 0000-0001-9854-8565 surname: Cardamone fullname: Cardamone, Nicholas C – sequence: 2 givenname: Mark orcidid: 0000-0002-3958-5662 surname: Olfson fullname: Olfson, Mark – sequence: 3 givenname: Timothy orcidid: 0000-0003-1711-1906 surname: Schmutte fullname: Schmutte, Timothy – sequence: 4 givenname: Lyle orcidid: 0000-0003-2047-1443 surname: Ungar fullname: Ungar, Lyle – sequence: 5 givenname: Tony orcidid: 0000-0002-3707-3989 surname: Liu fullname: Liu, Tony – sequence: 6 givenname: Sara W orcidid: 0000-0002-7846-5727 surname: Cullen fullname: Cullen, Sara W – sequence: 7 givenname: Nathaniel J orcidid: 0000-0002-3948-7480 surname: Williams fullname: Williams, Nathaniel J – sequence: 8 givenname: Steven C orcidid: 0000-0001-7819-3824 surname: Marcus fullname: Marcus, Steven C |
BackLink | https://www.ncbi.nlm.nih.gov/pubmed/39864953$$D View this record in MEDLINE/PubMed |
BookMark | eNpdklFrFDEQxxep2FrvK0hABF9Ok002m_gi5Tht4Yqi7XPIZme3OXJJTbLF-wB-b3N3bWl9mQwz__nxZzKvqyMfPFTVjOCPNZH8E29Yw15UJ3UtyVxyyY6e5MfVLKU1xpgwwjlvX1XHVArOZENPqr8Lp1Oyw9b6EV37lONk8hShR1fwJyPr0dKByTF4a9A5aJdv0E8wIfYJDSGiS_BZu4fOjzJoTbbBo8vQg0uf0UrHEUr046RLsi-j5Z12k97rfuWp376pXg7aJZjdv6fV9dfl1eJ8vvr-7WJxtpobVtd5zqRpTNv2ZoCm4yBqKYEZIQkF1pfQgTS6ExL3jdBaNJLCwHGnAZdJ1tX0tLo4cPug1-o22o2OWxW0VftCiKPSMVvjQHFaE6NbDZQwBoIILk3XYtNJUsuGmcL6cmDdTt0GelMWEbV7Bn3e8fZGjeFOESIEo60ohA_3hBh-T5Cy2thkwDntIUxJUcIxbgnlO-Pv_pOuwxR92VVRNaJpCOY74Nunlh69PHx3Ebw_CEwMKUUYHiUEq90lqf0l0X_pzLrE |
Cites_doi | 10.2196/55318 10.1038/s41746-022-00558-0 10.1101/2023.12.25.23300525 10.1001/jamanetworkopen.2018.5097 10.1109/TCBB.2019.2937862 10.48084/etasr.7200 10.21203/rs.3.rs-3914899/v1 10.3390/informatics7030025 10.2196/preprints.64088 10.1016/j.cct.2020.106075 10.1145/3636555.3636910 10.1007/978-3-319-33383-0_5 10.1073/pnas.2305016120 10.1016/j.jamda.2023.09.006 10.1136/oem.48.7.503 10.1136/bmj.m958 10.3233/SHTI190219 10.3389/fpsyt.2021.707916 10.18653/v1/2020.acl-main.468 10.2196/preprints.48659 10.3390/informatics6010004 10.1109/ICCECE58645.2024.10497313 10.1109/TNNLS.2013.2292894 10.1007/s10462-024-10896-y 10.1201/b13617 10.3390/math11102320 10.1145/3581754.3584136 |
ContentType | Journal Article |
Copyright | Nicholas C Cardamone, Mark Olfson, Timothy Schmutte, Lyle Ungar, Tony Liu, Sara W Cullen, Nathaniel J Williams, Steven C Marcus. Originally published in JMIR Medical Informatics (https://medinform.jmir.org). 2025. This work is licensed under https://creativecommons.org/licenses/by/4.0/" target="_blank">https://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License. Copyright © Nicholas C Cardamone, Mark Olfson, Timothy Schmutte, Lyle Ungar, Tony Liu, Sara W Cullen, Nathaniel J Williams, Steven C Marcus. Originally published in JMIR Medical Informatics (https://medinform.jmir.org) 2025 |
Copyright_xml | – notice: Nicholas C Cardamone, Mark Olfson, Timothy Schmutte, Lyle Ungar, Tony Liu, Sara W Cullen, Nathaniel J Williams, Steven C Marcus. Originally published in JMIR Medical Informatics (https://medinform.jmir.org). – notice: 2025. This work is licensed under https://creativecommons.org/licenses/by/4.0/" target="_blank">https://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License. – notice: Copyright © Nicholas C Cardamone, Mark Olfson, Timothy Schmutte, Lyle Ungar, Tony Liu, Sara W Cullen, Nathaniel J Williams, Steven C Marcus. Originally published in JMIR Medical Informatics (https://medinform.jmir.org) 2025 |
DBID | AAYXX CITATION CGR CUY CVF ECM EIF NPM 3V. 7X7 7XB 88C 8FI 8FJ 8FK ABUWG AFKRA AZQEC BENPR CCPQU DWQXO FYUFA GHDGH K9. M0S M0T PHGZM PHGZT PIMPY PJZUB PKEHL PPXIY PQEST PQQKQ PQUKI PRINS 7X8 5PM DOA |
DOI | 10.2196/65454 |
DatabaseName | CrossRef Medline MEDLINE MEDLINE (Ovid) MEDLINE MEDLINE PubMed ProQuest Central (Corporate) Health & Medical Collection ProQuest Central (purchase pre-March 2016) Healthcare Administration Database (Alumni) Hospital Premium Collection Hospital Premium Collection (Alumni Edition) ProQuest Central (Alumni) (purchase pre-March 2016) ProQuest Central (Alumni) ProQuest Central UK/Ireland ProQuest Central Essentials ProQuest Central ProQuest One Community College ProQuest Central Korea Health Research Premium Collection Health Research Premium Collection (Alumni) ProQuest Health & Medical Complete (Alumni) ProQuest Health & Medical Collection Health Management Database ProQuest Central Premium ProQuest One Academic Publicly Available Content Database ProQuest Health & Medical Research Collection ProQuest One Academic Middle East (New) ProQuest One Health & Nursing ProQuest One Academic Eastern Edition (DO NOT USE) ProQuest One Academic ProQuest One Academic UKI Edition ProQuest Central China MEDLINE - Academic PubMed Central (Full Participant titles) DOAJ Directory of Open Access Journals |
DatabaseTitle | CrossRef MEDLINE Medline Complete MEDLINE with Full Text PubMed MEDLINE (Ovid) Publicly Available Content Database ProQuest One Academic Middle East (New) ProQuest Central Essentials ProQuest Health & Medical Complete (Alumni) ProQuest Central (Alumni Edition) ProQuest One Community College ProQuest One Health & Nursing ProQuest Central China ProQuest Central ProQuest Health & Medical Research Collection Health Research Premium Collection Health and Medicine Complete (Alumni Edition) ProQuest Central Korea Health & Medical Research Collection ProQuest Central (New) ProQuest One Academic Eastern Edition ProQuest Health Management ProQuest Hospital Collection Health Research Premium Collection (Alumni) ProQuest Hospital Collection (Alumni) ProQuest Health & Medical Complete ProQuest One Academic UKI Edition ProQuest Health Management (Alumni Edition) ProQuest One Academic ProQuest One Academic (New) ProQuest Central (Alumni) MEDLINE - Academic |
DatabaseTitleList | MEDLINE Publicly Available Content Database MEDLINE - Academic |
Database_xml | – sequence: 1 dbid: DOA name: DOAJ Directory of Open Access Journals url: https://www.doaj.org/ sourceTypes: Open Website – sequence: 2 dbid: NPM name: PubMed url: https://proxy.k.utb.cz/login?url=http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?db=PubMed sourceTypes: Index Database – sequence: 3 dbid: EIF name: MEDLINE url: https://proxy.k.utb.cz/login?url=https://www.webofscience.com/wos/medline/basic-search sourceTypes: Index Database – sequence: 4 dbid: BENPR name: ProQuest Central url: https://www.proquest.com/central sourceTypes: Aggregation Database |
DeliveryMethod | fulltext_linktorsrc |
Discipline | Medicine |
EISSN | 2291-9694 |
EndPage | e65454 |
ExternalDocumentID | oai_doaj_org_article_6321ca7ae3144e81869cb70cb912954c PMC11884378 39864953 10_2196_65454 |
Genre | Journal Article |
GeographicLocations | United States |
GeographicLocations_xml | – name: United States |
GrantInformation_xml | – fundername: NIMH NIH HHS grantid: R01 MH126895 |
GroupedDBID | 53G 5VS 7X7 8FI 8FJ AAFWJ AAYXX ABUWG ADBBV AFKRA AFPKN ALIPV ALMA_UNASSIGNED_HOLDINGS AOIJS BAWUL BCNDV BENPR CCPQU CITATION DIK EMOBN FYUFA GROUPED_DOAJ HMCUK HYE KQ8 M0T M48 M~E OK1 PGMZT PHGZM PHGZT PIMPY RPM UKHRP CGR CUY CVF ECM EIF NPM 3V. 7XB 8FK AZQEC DWQXO K9. PJZUB PKEHL PPXIY PQEST PQQKQ PQUKI PRINS 7X8 5PM PUEGO |
ID | FETCH-LOGICAL-c422t-49c5c77dcfe5b6e8299e4c8913e4d13ebe9cab890d58aa8593ef60bae049c4b23 |
IEDL.DBID | 7X7 |
ISSN | 2291-9694 |
IngestDate | Wed Aug 27 01:30:38 EDT 2025 Thu Aug 21 18:33:45 EDT 2025 Tue Aug 05 11:15:25 EDT 2025 Fri Jul 25 22:25:15 EDT 2025 Tue May 06 01:31:28 EDT 2025 Tue Jul 01 05:19:22 EDT 2025 |
IsDoiOpenAccess | true |
IsOpenAccess | true |
IsPeerReviewed | true |
IsScholarly | true |
Keywords | natural language processing mental health AI mental health disorder LLM emergency department large language model machine learning EHR EHR system artificial intelligence NLP physical health ChatGPT predictive modeling health informatics text dataset electronic health record ML |
Language | English |
License | Nicholas C Cardamone, Mark Olfson, Timothy Schmutte, Lyle Ungar, Tony Liu, Sara W Cullen, Nathaniel J Williams, Steven C Marcus. Originally published in JMIR Medical Informatics (https://medinform.jmir.org). This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on https://medinform.jmir.org/, as well as this copyright and license information must be included. |
LinkModel | DirectLink |
MergedId | FETCHMERGED-LOGICAL-c422t-49c5c77dcfe5b6e8299e4c8913e4d13ebe9cab890d58aa8593ef60bae049c4b23 |
Notes | ObjectType-Article-1 SourceType-Scholarly Journals-1 ObjectType-Feature-2 content type line 14 content type line 23 None declared. |
ORCID | 0000-0001-9854-8565 0000-0002-3958-5662 0000-0003-1711-1906 0000-0002-7846-5727 0000-0002-3948-7480 0000-0003-2047-1443 0000-0001-7819-3824 0000-0002-3707-3989 |
OpenAccessLink | https://www.proquest.com/docview/3158551068?pq-origsite=%requestingapplication% |
PMID | 39864953 |
PQID | 3158551068 |
PQPubID | 4997117 |
ParticipantIDs | doaj_primary_oai_doaj_org_article_6321ca7ae3144e81869cb70cb912954c pubmedcentral_primary_oai_pubmedcentral_nih_gov_11884378 proquest_miscellaneous_3160071362 proquest_journals_3158551068 pubmed_primary_39864953 crossref_primary_10_2196_65454 |
ProviderPackageCode | CITATION AAYXX |
PublicationCentury | 2000 |
PublicationDate | 20250121 |
PublicationDateYYYYMMDD | 2025-01-21 |
PublicationDate_xml | – month: 1 year: 2025 text: 20250121 day: 21 |
PublicationDecade | 2020 |
PublicationPlace | Canada |
PublicationPlace_xml | – name: Canada – name: Toronto – name: Toronto, Canada |
PublicationTitle | JMIR medical informatics |
PublicationTitleAlternate | JMIR Med Inform |
PublicationYear | 2025 |
Publisher | JMIR Publications |
Publisher_xml | – name: JMIR Publications |
References | Li (R30) Hossain (R1); 18 Lee (R5); 7 Garrido-Merchan (R29); 2 Kumar (R20); 14 R21 R23 R22 R25 Bayramli (R6); 5 R24 Boudreaux (R3); 95 R26 Lin (R36); 57 Boudreaux (R4); 12 Mahajan (R7); 264 Fung (R27); 48 R8 Sivarajkumar (R34); 12 Akbilgic (R9); 6 R31 R12 Lee (R32); 11 Frenay (R37); 25 Mahmoudi (R2); 369 R11 R33 R14 R13 R35 R16 R15 R18 Scharp (R28); 25 R19 Marafino (R10); 1 Gilardi (R17); 120 |
References_xml | – volume: 12 ident: R34 article-title: An empirical evaluation of prompting strategies for large language models in zero-shot clinical natural language processing: algorithm development and validation study publication-title: JMIR Med Inform doi: 10.2196/55318 – ident: R13 – volume: 5 issue: 1 ident: R6 article-title: Predictive structured-unstructured interactions in EHR models: a case study of suicide prediction publication-title: NPJ Digit Med doi: 10.1038/s41746-022-00558-0 – ident: R21 doi: 10.1101/2023.12.25.23300525 – volume: 1 issue: 8 ident: R10 article-title: Validation of prediction models for critical care outcomes using natural language processing of electronic health record data publication-title: JAMA Netw Open doi: 10.1001/jamanetworkopen.2018.5097 – volume: 18 start-page: 745 issue: 2 ident: R1 article-title: Use of electronic health data for disease prediction: a comprehensive literature review publication-title: IEEE/ACM Trans Comput Biol Bioinform doi: 10.1109/TCBB.2019.2937862 – volume: 14 start-page: 14219 issue: 3 ident: R20 article-title: Towards optimal NLP solutions: analyzing GPT and LLaMA-2 models across model scale, dataset size, and task diversity publication-title: Eng Technol Appl Sci Res doi: 10.48084/etasr.7200 – ident: R31 doi: 10.21203/rs.3.rs-3914899/v1 – volume: 7 issue: 3 ident: R5 article-title: Clinical implementation of predictive models embedded within electronic health record systems: a systematic review publication-title: Informatics (MDPI) doi: 10.3390/informatics7030025 – ident: R19 doi: 10.2196/preprints.64088 – ident: R30 publication-title: arxiv – volume: 95 ident: R3 article-title: Emergency department safety assessment and follow-up evaluation 2: an implementation trial to improve suicide prevention publication-title: Contemp Clin Trials doi: 10.1016/j.cct.2020.106075 – ident: R25 – ident: R16 doi: 10.1145/3636555.3636910 – ident: R26 doi: 10.1007/978-3-319-33383-0_5 – volume: 120 issue: 30 ident: R17 article-title: ChatGPT outperforms crowd workers for text-annotation tasks publication-title: Proc Natl Acad Sci U S A doi: 10.1073/pnas.2305016120 – volume: 25 start-page: 69 issue: 1 ident: R28 article-title: Natural language processing applied to clinical documentation in post-acute care settings: a scoping review publication-title: J Am Med Dir Assoc doi: 10.1016/j.jamda.2023.09.006 – volume: 48 start-page: 503 issue: 7 ident: R27 article-title: Bootstrap estimate of the variance and confidence interval of kappa publication-title: Br J Ind Med doi: 10.1136/oem.48.7.503 – volume: 369 ident: R2 article-title: Use of electronic medical records in development and validation of risk prediction models of hospital readmission: systematic review publication-title: BMJ doi: 10.1136/bmj.m958 – volume: 264 ident: R7 article-title: Combining structured and unstructured data for predicting risk of readmission for heart failure patients publication-title: Stud Health Technol Inform doi: 10.3233/SHTI190219 – ident: R18 – volume: 12 ident: R4 article-title: Applying machine learning approaches to suicide prediction using healthcare data: overview and future directions publication-title: Front Psychiatry doi: 10.3389/fpsyt.2021.707916 – ident: R33 – ident: R14 – ident: R12 – ident: R35 doi: 10.18653/v1/2020.acl-main.468 – ident: R23 doi: 10.2196/preprints.48659 – volume: 6 start-page: 4 issue: 1 ident: R9 article-title: Unstructured text in EMR improves prediction of death after surgery in children publication-title: Informatics (MDPI) doi: 10.3390/informatics6010004 – ident: R11 doi: 10.1109/ICCECE58645.2024.10497313 – volume: 25 start-page: 845 issue: 5 ident: R37 article-title: Classification in the presence of label noise: a survey publication-title: IEEE Trans Neural Netw Learning Syst doi: 10.1109/TNNLS.2013.2292894 – volume: 57 start-page: 243 issue: 9 ident: R36 article-title: Towards trustworthy LLMs: a review on debiasing and dehallucinating in large language models publication-title: Artif Intell Rev doi: 10.1007/s10462-024-10896-y – ident: R22 – ident: R24 – ident: R8 doi: 10.1201/b13617 – volume: 11 start-page: 2320 issue: 10 ident: R32 article-title: A mathematical investigation of hallucination and creativity in GPT models publication-title: Mathematics doi: 10.3390/math11102320 – ident: R15 doi: 10.1145/3581754.3584136 – volume: 2 ident: R29 publication-title: J Comput Cogn Eng |
SSID | ssj0001416667 |
Score | 2.3075974 |
Snippet | Prediction models have demonstrated a range of applications across medicine, including using electronic health record (EHR) data to identify hospital... Background:Prediction models have demonstrated a range of applications across medicine, including using electronic health record (EHR) data to identify... Abstract BackgroundPrediction models have demonstrated a range of applications across medicine, including using electronic health record (EHR) data to identify... |
SourceID | doaj pubmedcentral proquest pubmed crossref |
SourceType | Open Website Open Access Repository Aggregation Database Index Database |
StartPage | e65454 |
SubjectTerms | AI Language Models in Health Care Anxiety Artificial intelligence Chronic fatigue syndrome Classification Data warehouses Datasets Decision Support for Health Professionals Diagnostic Tools in Mental Health Disease Eating disorders Electronic Health Records Electronic Health Records - statistics & numerical data Emergency Service, Hospital Emotional disorders Humans Impulsivity Large Language Models Machine Learning Medical coding Mental disorders Mental Disorders - diagnosis Mental Health Methods and New Tools in Mental Health Research Mood disorders Multimedia Natural Language Processing New Technologies Original Paper Personality disorders Psychosis Reconciliation Review boards Self destructive behavior United States |
SummonAdditionalLinks | – databaseName: DOAJ Directory of Open Access Journals dbid: DOA link: http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwrV1La9wwEB7KHkKhlKRPp0lQoVeTlfV0b23YEEJaetiFvRlJlulC8ZZscugPyP_OjOQ1u6WQSy8-SMbImpHmG83oG4BP3ujO8qBLyztZSkT0pdPBlyZwGXUgvpiUIPtdXy3k9VItd0p9UU5YpgfOE3euRcWDMy4KhP6R-Nfq4M00-JpTiCrQ7os2b8eZSqcrksJh5gBeUK4zatm5Rqgg94xP4uj_F7D8Oz9yx-BcHsLLASmyL3mER_As9q_g4NsQC38ND6me5SrdU2KLgQj2_ja2bI4bLlv1bDaWuGH5thHLzuaGIVJlmb1n2_Pjlj5MUmJUHu3X5jO7oSRxfOYDzdzMZiM7OKMUxD9vYHE5m19clUNRhTLIqrorZR1UMKYNXVReR4vmKMpAwcooW3z4WAfnbT1tlXWO2NBip6feRXQlgvSVeAuTft3H98A6JzrvnffCeGmV9VRcQSlVdbw1CEMLONvOdvM7c2c06HOQOJokjgK-kgzGTqK6Tg2oAM2gAM1TClDAyVaCzbD-No3g6AbhdqNtAR_Hblw5FA5xfVzf0zs6-ei6KuBdFvg4EkGs9bUSBdg9Vdgb6n5Pv_qZ2LnRY7NSGHv8P37uAzyvqODwlJcVP4EJqlI8RRR058-Swj8CUjUHqQ priority: 102 providerName: Directory of Open Access Journals – databaseName: Scholars Portal Journals: Open Access dbid: M48 link: http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwfV1La9wwEB7aFEIhlD4Tt0lQIVe3K-vpQglp2BBKU3LIQm9GkuVkIXjb3QSaH9D_3RnZa-ISetFBEkZ4RppvNKNvAA680Y3lQeeWNzKXiOhzp4PPTeAy6kB8MSlB9rs-ncmvP9S9bML-B64edO2ontRsef3h96-7Q9zwnymNGRXoo0YUIB_DEzRGhooYnPUIP12zSIqLmU3YGs0eWaFE1v8Qwvw3UfKe5Tl5Ds96yMiOOhm_gEexfQmbZ31Q_BX8SYUt5-nBEpv1jLC3y1izCzx52bxl06HWDeueHbHO61wxhKyso_FZj5wv6cMkLkZ10q5Xn9g3yhbHtrvZ7LrZdKAJZ5SLePcaZifTi-PTvK-ukAdZFDe5LIMKxtShicrraNEuRRkoahlljY2PZXDelpNaWeeIFi02euJdRJ8iSF-IN7DRLtq4A6xxovHeeS-Ml1ZZT1UWlFJFw2uDeDSD_fXfrn52JBoVOh8kjiqJI4MvJINhkDivU8dieVn1W6jSouDBGRcFOoGRmPjK4M0k-JJTsDJksLuWYLXWo0pw9Ifw3NE2g_fDMG4hiou4Ni5uaY5OzrouMtjuBD6sRBB9falEBnakCqOljkfa-VWi6UbXzUph7Nv_r-sdPC2opvCE5wXfhQ1UkriHQOfG7ydV_gvzQgA2 priority: 102 providerName: Scholars Portal |
Title | Classifying Unstructured Text in Electronic Health Records for Mental Health Prediction Models: Large Language Model Evaluation Study |
URI | https://www.ncbi.nlm.nih.gov/pubmed/39864953 https://www.proquest.com/docview/3158551068 https://www.proquest.com/docview/3160071362 https://pubmed.ncbi.nlm.nih.gov/PMC11884378 https://doaj.org/article/6321ca7ae3144e81869cb70cb912954c |
Volume | 13 |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwfV3daxQxEB-0hSKI-O1qPSL4GnrZZJOsL2LlShFbivTg3pYkm9UD2at37YN_gP-3M9nc1hPxJQ87yxJ2JslvPvIbgLfe6M6KoLkVneIKET13OnhuglBRB-KLSQWy5_p0rj4tqkUOuG1yWeV2T0wbdbsKFCM_kgKBLRqQtu-vfnDqGkXZ1dxC4y7sE3UZWbVZmNsYi6KkmDmA-1TxjLZ2pBEwqJ0jKDH1_wte_l0l-cexc_IQHmS8yD4MCn4Ed2L_GA7Ockb8CfxKXS2X6bYSm2c62Jt1bNklbrts2bPZ2OiGDXeO2OBybhjiVTZw-GwlF2v6MOmKUZO075t37DOViuM4hDWHx2w2coQzKkT8-RTmJ7PLj6c8t1bgQZXlNVd1qIIxbehi5XW0eChFFShlGVWLg491cN7W07ayzhEnWuz01LuIDkVQvpTPYK9f9fEFsM7JznvnvTRe2cp6arFQVVXZidYgGC1gsv3bzdXAoNGg50HqaJI6CjgmHYxCIrxOD1brr01eP42WpQjOuCjRA4xEw1cHb6bB14IylaGAw60Gm7wKN82tzRTwZhTj-qGkiOvj6obe0clT12UBzweFjzORxF1fV7IAu2MKO1PdlfTLb4mjG_02q6SxL_8_r1dwr6SGwlPBS3EIe2gk8TWinGs_SaY8gf3j2fnFl0mKFeB4puxvUNQDQg |
linkProvider | ProQuest |
linkToHtml | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwtV3bbtQwEB1VRSpICHEnUIqR4DFqnDi2g4QQl622dFvxsCvtW7AdB1ZC2bLbCvUD-B2-kRnnUhYh3vqSh3i1sjLH4xnP-ByAF1bJWnMnY81rEQuM6GMjnY2V48JLR3wxoUH2RI5n4uM8n2_Br_4uDLVV9j4xOOpq6eiMfD_jGNgigKR-c_o9JtUoqq72EhotLI78xQ9M2davDz-gfV-m6cFo-n4cd6oCsRNpehaLwuVOqcrVPrfSa_THXjiq1nlR4cP6whmri6TKtTFEB-ZrmVjjMZZ2whLRAbr8a7jxJpTsqbm6PNMRVIRTO3CTOqwR2_sSAxSxseUFZYB_hbN_d2X-sc0d3IZbXXzK3raAugNbvrkLO8ddBf4e_AwqmotwO4rNOvrZ85Wv2BTdPFs0bDQI67D2jhNrU9w1w_iYtZxB_cinFf0xYYORKNu39Ss2odZ0fLbHqO1rNho4yRk1Pl7ch9mVfPQHsN0sG_8IWG2y2lpjbaas0Lm2JOmQ53la80ph8BvBXv-1y9OWsaPETIfMUQZzRPCObDAMEsF2eLFcfSm79VrKLOXOKOMzzDg90f4VzqrE2YJTZdRFsNtbsOxW_bq8xGgEz4dhXK9UhDGNX57Tb2Q4GZBpBA9bgw8zyYgrv8izCPQGFDamujnSLL4GTnDME7XIlH78_3k9g-vj6fGknByeHD2BGymJGSc8TvkubCNg_FOMsM7sXoA1g89XvY5-A8_ZPig |
linkToPdf | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwtV1ta9swED5KCmEwxt7rres02D6axLIsyYNR2jWhXbsQRgP95kmyvAWG0yUtoz9gf6q_rnd-6zLGvvWLP1ghCN3p_Jzu9DwAb62ShY6cDHVUiFAgog-NdDZULhJeOuKLqRpkJ_JwJj6dJWcbcN3ehaG2yjYmVoE6Xzg6Ix_EEQJbdCCpB0XTFjE9GO-e_wxJQYoqra2cRu0ix_7qF6Zvqw9HB2jrd5yPR6cfD8NGYSB0gvOLUKQucUrlrvCJlV5jbPbCUeXOixwf1qfOWJ0O80QbQ9RgvpBDazziaicskR5g-N9UlBX1YHN_NJl-uT3hEVSSU324T_3W6OkDiXBFrH0AK52Af4Hbv3s0__jojR_Cgwatsr3avR7Bhi8fQ_9zU49_Ar8rTc15dVeKzRoy2sulz9kprhybl2zUyeyw-sYTqxPeFUO0zGoGoXZkuqQ_Jk9hJNH2Y_WenVCjOj7rQ9X6NRt1DOWM2iCvnsLsTpb9GfTKRem3gBUmLqw11sbKCp1oSwIPSZLwIsoVQuEAdtrVzs5r_o4M8x4yR1aZI4B9skE3SHTb1YvF8lvW7N5MxjxyRhkfY_7piQQwdVYNnU0jqpO6ALZbC2ZNDFhltx4bwJtuGHcvlWRM6ReX9BtZnRNIHsDz2uDdTGJizk-TOAC95gprU10fKeffK4ZwzBq1iJV-8f95vYY-7qHs5Ghy_BLucVI2HkYhj7ahh_7iXyHcurA7jV8z-HrXW-kGp4xDww |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Classifying+Unstructured+Text+in+Electronic+Health+Records+for+Mental+Health+Prediction+Models%3A+Large+Language+Model+Evaluation+Study&rft.jtitle=JMIR+medical+informatics&rft.au=Cardamone%2C+Nicholas+C&rft.au=Olfson%2C+Mark&rft.au=Schmutte%2C+Timothy&rft.au=Ungar%2C+Lyle&rft.date=2025-01-21&rft.pub=JMIR+Publications&rft.eissn=2291-9694&rft.volume=13&rft.spage=e65454&rft_id=info:doi/10.2196%2F65454&rft.externalDBID=HAS_PDF_LINK |
thumbnail_l | http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/lc.gif&issn=2291-9694&client=summon |
thumbnail_m | http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/mc.gif&issn=2291-9694&client=summon |
thumbnail_s | http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/sc.gif&issn=2291-9694&client=summon |