Classifying Unstructured Text in Electronic Health Records for Mental Health Prediction Models: Large Language Model Evaluation Study

Prediction models have demonstrated a range of applications across medicine, including using electronic health record (EHR) data to identify hospital readmission and mortality risk. Large language models (LLMs) can transform unstructured EHR text into structured features, which can then be integrate...

Full description

Saved in:
Bibliographic Details
Published inJMIR medical informatics Vol. 13; p. e65454
Main Authors Cardamone, Nicholas C, Olfson, Mark, Schmutte, Timothy, Ungar, Lyle, Liu, Tony, Cullen, Sara W, Williams, Nathaniel J, Marcus, Steven C
Format Journal Article
LanguageEnglish
Published Canada JMIR Publications 21.01.2025
Subjects
Online AccessGet full text

Cover

Loading…
Abstract Prediction models have demonstrated a range of applications across medicine, including using electronic health record (EHR) data to identify hospital readmission and mortality risk. Large language models (LLMs) can transform unstructured EHR text into structured features, which can then be integrated into statistical prediction models, ensuring that the results are both clinically meaningful and interpretable. This study aims to compare the classification decisions made by clinical experts with those generated by a state-of-the-art LLM, using terms extracted from a large EHR data set of individuals with mental health disorders seen in emergency departments (EDs). Using a dataset from the EHR systems of more than 50 health care provider organizations in the United States from 2016 to 2021, we extracted all clinical terms that appeared in at least 1000 records of individuals admitted to the ED for a mental health-related problem from a source population of over 6 million ED episodes. Two experienced mental health clinicians (one medically trained psychiatrist and one clinical psychologist) reached consensus on the classification of EHR terms and diagnostic codes into categories. We evaluated an LLM's agreement with clinical judgment across three classification tasks as follows: (1) classify terms into "mental health" or "physical health", (2) classify mental health terms into 1 of 42 prespecified categories, and (3) classify physical health terms into 1 of 19 prespecified broad categories. There was high agreement between the LLM and clinical experts when categorizing 4553 terms as "mental health" or "physical health" (κ=0.77, 95% CI 0.75-0.80). However, there was still considerable variability in LLM-clinician agreement on the classification of mental health terms (κ=0.62, 95% CI 0.59-0.66) and physical health terms (κ=0.69, 95% CI 0.67-0.70). The LLM displayed high agreement with clinical experts when classifying EHR terms into certain mental health or physical health term categories. However, agreement with clinical experts varied considerably within both sets of mental and physical health term categories. Importantly, the use of LLMs presents an alternative to manual human coding, presenting great potential to create interpretable features for prediction models.
AbstractList Abstract BackgroundPrediction models have demonstrated a range of applications across medicine, including using electronic health record (EHR) data to identify hospital readmission and mortality risk. Large language models (LLMs) can transform unstructured EHR text into structured features, which can then be integrated into statistical prediction models, ensuring that the results are both clinically meaningful and interpretable. ObjectiveThis study aims to compare the classification decisions made by clinical experts with those generated by a state-of-the-art LLM, using terms extracted from a large EHR data set of individuals with mental health disorders seen in emergency departments (EDs). MethodsUsing a dataset from the EHR systems of more than 50 health care provider organizations in the United States from 2016 to 2021, we extracted all clinical terms that appeared in at least 1000 records of individuals admitted to the ED for a mental health–related problem from a source population of over 6 million ED episodes. Two experienced mental health clinicians (one medically trained psychiatrist and one clinical psychologist) reached consensus on the classification of EHR terms and diagnostic codes into categories. We evaluated an LLM’s agreement with clinical judgment across three classification tasks as follows: (1) classify terms into “mental health” or “physical health”, (2) classify mental health terms into 1 of 42 prespecified categories, and (3) classify physical health terms into 1 of 19 prespecified broad categories. ResultsThere was high agreement between the LLM and clinical experts when categorizing 4553 terms as “mental health” or “physical health” (κ=0.77, 95% CI 0.75-0.80). However, there was still considerable variability in LLM-clinician agreement on the classification of mental health terms (κ=0.62, 95% CI 0.59‐0.66) and physical health terms (κ=0.69, 95% CI 0.67‐0.70). ConclusionsThe LLM displayed high agreement with clinical experts when classifying EHR terms into certain mental health or physical health term categories. However, agreement with clinical experts varied considerably within both sets of mental and physical health term categories. Importantly, the use of LLMs presents an alternative to manual human coding, presenting great potential to create interpretable features for prediction models.
Prediction models have demonstrated a range of applications across medicine, including using electronic health record (EHR) data to identify hospital readmission and mortality risk. Large language models (LLMs) can transform unstructured EHR text into structured features, which can then be integrated into statistical prediction models, ensuring that the results are both clinically meaningful and interpretable. This study aims to compare the classification decisions made by clinical experts with those generated by a state-of-the-art LLM, using terms extracted from a large EHR data set of individuals with mental health disorders seen in emergency departments (EDs). Using a dataset from the EHR systems of more than 50 health care provider organizations in the United States from 2016 to 2021, we extracted all clinical terms that appeared in at least 1000 records of individuals admitted to the ED for a mental health-related problem from a source population of over 6 million ED episodes. Two experienced mental health clinicians (one medically trained psychiatrist and one clinical psychologist) reached consensus on the classification of EHR terms and diagnostic codes into categories. We evaluated an LLM's agreement with clinical judgment across three classification tasks as follows: (1) classify terms into "mental health" or "physical health", (2) classify mental health terms into 1 of 42 prespecified categories, and (3) classify physical health terms into 1 of 19 prespecified broad categories. There was high agreement between the LLM and clinical experts when categorizing 4553 terms as "mental health" or "physical health" (κ=0.77, 95% CI 0.75-0.80). However, there was still considerable variability in LLM-clinician agreement on the classification of mental health terms (κ=0.62, 95% CI 0.59-0.66) and physical health terms (κ=0.69, 95% CI 0.67-0.70). The LLM displayed high agreement with clinical experts when classifying EHR terms into certain mental health or physical health term categories. However, agreement with clinical experts varied considerably within both sets of mental and physical health term categories. Importantly, the use of LLMs presents an alternative to manual human coding, presenting great potential to create interpretable features for prediction models.
Background:Prediction models have demonstrated a range of applications across medicine, including using electronic health record (EHR) data to identify hospital readmission and mortality risk. Large language models (LLMs) can transform unstructured EHR text into structured features, which can then be integrated into statistical prediction models, ensuring that the results are both clinically meaningful and interpretable.Objective:This study aims to compare the classification decisions made by clinical experts with those generated by a state-of-the-art LLM, using terms extracted from a large EHR data set of individuals with mental health disorders seen in emergency departments (EDs).Methods:Using a dataset from the EHR systems of more than 50 health care provider organizations in the United States from 2016 to 2021, we extracted all clinical terms that appeared in at least 1000 records of individuals admitted to the ED for a mental health–related problem from a source population of over 6 million ED episodes. Two experienced mental health clinicians (one medically trained psychiatrist and one clinical psychologist) reached consensus on the classification of EHR terms and diagnostic codes into categories. We evaluated an LLM’s agreement with clinical judgment across three classification tasks as follows: (1) classify terms into “mental health” or “physical health”, (2) classify mental health terms into 1 of 42 prespecified categories, and (3) classify physical health terms into 1 of 19 prespecified broad categories.Results:There was high agreement between the LLM and clinical experts when categorizing 4553 terms as “mental health” or “physical health” (κ=0.77, 95% CI 0.75-0.80). However, there was still considerable variability in LLM-clinician agreement on the classification of mental health terms (κ=0.62, 95% CI 0.59‐0.66) and physical health terms (κ=0.69, 95% CI 0.67‐0.70).Conclusions:The LLM displayed high agreement with clinical experts when classifying EHR terms into certain mental health or physical health term categories. However, agreement with clinical experts varied considerably within both sets of mental and physical health term categories. Importantly, the use of LLMs presents an alternative to manual human coding, presenting great potential to create interpretable features for prediction models.
Prediction models have demonstrated a range of applications across medicine, including using electronic health record (EHR) data to identify hospital readmission and mortality risk. Large language models (LLMs) can transform unstructured EHR text into structured features, which can then be integrated into statistical prediction models, ensuring that the results are both clinically meaningful and interpretable.BackgroundPrediction models have demonstrated a range of applications across medicine, including using electronic health record (EHR) data to identify hospital readmission and mortality risk. Large language models (LLMs) can transform unstructured EHR text into structured features, which can then be integrated into statistical prediction models, ensuring that the results are both clinically meaningful and interpretable.This study aims to compare the classification decisions made by clinical experts with those generated by a state-of-the-art LLM, using terms extracted from a large EHR data set of individuals with mental health disorders seen in emergency departments (EDs).ObjectiveThis study aims to compare the classification decisions made by clinical experts with those generated by a state-of-the-art LLM, using terms extracted from a large EHR data set of individuals with mental health disorders seen in emergency departments (EDs).Using a dataset from the EHR systems of more than 50 health care provider organizations in the United States from 2016 to 2021, we extracted all clinical terms that appeared in at least 1000 records of individuals admitted to the ED for a mental health-related problem from a source population of over 6 million ED episodes. Two experienced mental health clinicians (one medically trained psychiatrist and one clinical psychologist) reached consensus on the classification of EHR terms and diagnostic codes into categories. We evaluated an LLM's agreement with clinical judgment across three classification tasks as follows: (1) classify terms into "mental health" or "physical health", (2) classify mental health terms into 1 of 42 prespecified categories, and (3) classify physical health terms into 1 of 19 prespecified broad categories.MethodsUsing a dataset from the EHR systems of more than 50 health care provider organizations in the United States from 2016 to 2021, we extracted all clinical terms that appeared in at least 1000 records of individuals admitted to the ED for a mental health-related problem from a source population of over 6 million ED episodes. Two experienced mental health clinicians (one medically trained psychiatrist and one clinical psychologist) reached consensus on the classification of EHR terms and diagnostic codes into categories. We evaluated an LLM's agreement with clinical judgment across three classification tasks as follows: (1) classify terms into "mental health" or "physical health", (2) classify mental health terms into 1 of 42 prespecified categories, and (3) classify physical health terms into 1 of 19 prespecified broad categories.There was high agreement between the LLM and clinical experts when categorizing 4553 terms as "mental health" or "physical health" (κ=0.77, 95% CI 0.75-0.80). However, there was still considerable variability in LLM-clinician agreement on the classification of mental health terms (κ=0.62, 95% CI 0.59-0.66) and physical health terms (κ=0.69, 95% CI 0.67-0.70).ResultsThere was high agreement between the LLM and clinical experts when categorizing 4553 terms as "mental health" or "physical health" (κ=0.77, 95% CI 0.75-0.80). However, there was still considerable variability in LLM-clinician agreement on the classification of mental health terms (κ=0.62, 95% CI 0.59-0.66) and physical health terms (κ=0.69, 95% CI 0.67-0.70).The LLM displayed high agreement with clinical experts when classifying EHR terms into certain mental health or physical health term categories. However, agreement with clinical experts varied considerably within both sets of mental and physical health term categories. Importantly, the use of LLMs presents an alternative to manual human coding, presenting great potential to create interpretable features for prediction models.ConclusionsThe LLM displayed high agreement with clinical experts when classifying EHR terms into certain mental health or physical health term categories. However, agreement with clinical experts varied considerably within both sets of mental and physical health term categories. Importantly, the use of LLMs presents an alternative to manual human coding, presenting great potential to create interpretable features for prediction models.
Author Olfson, Mark
Ungar, Lyle
Cullen, Sara W
Cardamone, Nicholas C
Liu, Tony
Schmutte, Timothy
Williams, Nathaniel J
Marcus, Steven C
Author_xml – sequence: 1
  givenname: Nicholas C
  orcidid: 0000-0001-9854-8565
  surname: Cardamone
  fullname: Cardamone, Nicholas C
– sequence: 2
  givenname: Mark
  orcidid: 0000-0002-3958-5662
  surname: Olfson
  fullname: Olfson, Mark
– sequence: 3
  givenname: Timothy
  orcidid: 0000-0003-1711-1906
  surname: Schmutte
  fullname: Schmutte, Timothy
– sequence: 4
  givenname: Lyle
  orcidid: 0000-0003-2047-1443
  surname: Ungar
  fullname: Ungar, Lyle
– sequence: 5
  givenname: Tony
  orcidid: 0000-0002-3707-3989
  surname: Liu
  fullname: Liu, Tony
– sequence: 6
  givenname: Sara W
  orcidid: 0000-0002-7846-5727
  surname: Cullen
  fullname: Cullen, Sara W
– sequence: 7
  givenname: Nathaniel J
  orcidid: 0000-0002-3948-7480
  surname: Williams
  fullname: Williams, Nathaniel J
– sequence: 8
  givenname: Steven C
  orcidid: 0000-0001-7819-3824
  surname: Marcus
  fullname: Marcus, Steven C
BackLink https://www.ncbi.nlm.nih.gov/pubmed/39864953$$D View this record in MEDLINE/PubMed
BookMark eNpdklFrFDEQxxep2FrvK0hABF9Ok002m_gi5Tht4Yqi7XPIZme3OXJJTbLF-wB-b3N3bWl9mQwz__nxZzKvqyMfPFTVjOCPNZH8E29Yw15UJ3UtyVxyyY6e5MfVLKU1xpgwwjlvX1XHVArOZENPqr8Lp1Oyw9b6EV37lONk8hShR1fwJyPr0dKByTF4a9A5aJdv0E8wIfYJDSGiS_BZu4fOjzJoTbbBo8vQg0uf0UrHEUr046RLsi-j5Z12k97rfuWp376pXg7aJZjdv6fV9dfl1eJ8vvr-7WJxtpobVtd5zqRpTNv2ZoCm4yBqKYEZIQkF1pfQgTS6ExL3jdBaNJLCwHGnAZdJ1tX0tLo4cPug1-o22o2OWxW0VftCiKPSMVvjQHFaE6NbDZQwBoIILk3XYtNJUsuGmcL6cmDdTt0GelMWEbV7Bn3e8fZGjeFOESIEo60ohA_3hBh-T5Cy2thkwDntIUxJUcIxbgnlO-Pv_pOuwxR92VVRNaJpCOY74Nunlh69PHx3Ebw_CEwMKUUYHiUEq90lqf0l0X_pzLrE
Cites_doi 10.2196/55318
10.1038/s41746-022-00558-0
10.1101/2023.12.25.23300525
10.1001/jamanetworkopen.2018.5097
10.1109/TCBB.2019.2937862
10.48084/etasr.7200
10.21203/rs.3.rs-3914899/v1
10.3390/informatics7030025
10.2196/preprints.64088
10.1016/j.cct.2020.106075
10.1145/3636555.3636910
10.1007/978-3-319-33383-0_5
10.1073/pnas.2305016120
10.1016/j.jamda.2023.09.006
10.1136/oem.48.7.503
10.1136/bmj.m958
10.3233/SHTI190219
10.3389/fpsyt.2021.707916
10.18653/v1/2020.acl-main.468
10.2196/preprints.48659
10.3390/informatics6010004
10.1109/ICCECE58645.2024.10497313
10.1109/TNNLS.2013.2292894
10.1007/s10462-024-10896-y
10.1201/b13617
10.3390/math11102320
10.1145/3581754.3584136
ContentType Journal Article
Copyright Nicholas C Cardamone, Mark Olfson, Timothy Schmutte, Lyle Ungar, Tony Liu, Sara W Cullen, Nathaniel J Williams, Steven C Marcus. Originally published in JMIR Medical Informatics (https://medinform.jmir.org).
2025. This work is licensed under https://creativecommons.org/licenses/by/4.0/" target="_blank">https://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Copyright © Nicholas C Cardamone, Mark Olfson, Timothy Schmutte, Lyle Ungar, Tony Liu, Sara W Cullen, Nathaniel J Williams, Steven C Marcus. Originally published in JMIR Medical Informatics (https://medinform.jmir.org) 2025
Copyright_xml – notice: Nicholas C Cardamone, Mark Olfson, Timothy Schmutte, Lyle Ungar, Tony Liu, Sara W Cullen, Nathaniel J Williams, Steven C Marcus. Originally published in JMIR Medical Informatics (https://medinform.jmir.org).
– notice: 2025. This work is licensed under https://creativecommons.org/licenses/by/4.0/" target="_blank">https://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
– notice: Copyright © Nicholas C Cardamone, Mark Olfson, Timothy Schmutte, Lyle Ungar, Tony Liu, Sara W Cullen, Nathaniel J Williams, Steven C Marcus. Originally published in JMIR Medical Informatics (https://medinform.jmir.org) 2025
DBID AAYXX
CITATION
CGR
CUY
CVF
ECM
EIF
NPM
3V.
7X7
7XB
88C
8FI
8FJ
8FK
ABUWG
AFKRA
AZQEC
BENPR
CCPQU
DWQXO
FYUFA
GHDGH
K9.
M0S
M0T
PHGZM
PHGZT
PIMPY
PJZUB
PKEHL
PPXIY
PQEST
PQQKQ
PQUKI
PRINS
7X8
5PM
DOA
DOI 10.2196/65454
DatabaseName CrossRef
Medline
MEDLINE
MEDLINE (Ovid)
MEDLINE
MEDLINE
PubMed
ProQuest Central (Corporate)
Health & Medical Collection
ProQuest Central (purchase pre-March 2016)
Healthcare Administration Database (Alumni)
Hospital Premium Collection
Hospital Premium Collection (Alumni Edition)
ProQuest Central (Alumni) (purchase pre-March 2016)
ProQuest Central (Alumni)
ProQuest Central UK/Ireland
ProQuest Central Essentials
ProQuest Central
ProQuest One Community College
ProQuest Central Korea
Health Research Premium Collection
Health Research Premium Collection (Alumni)
ProQuest Health & Medical Complete (Alumni)
ProQuest Health & Medical Collection
Health Management Database
ProQuest Central Premium
ProQuest One Academic
Publicly Available Content Database
ProQuest Health & Medical Research Collection
ProQuest One Academic Middle East (New)
ProQuest One Health & Nursing
ProQuest One Academic Eastern Edition (DO NOT USE)
ProQuest One Academic
ProQuest One Academic UKI Edition
ProQuest Central China
MEDLINE - Academic
PubMed Central (Full Participant titles)
DOAJ Directory of Open Access Journals
DatabaseTitle CrossRef
MEDLINE
Medline Complete
MEDLINE with Full Text
PubMed
MEDLINE (Ovid)
Publicly Available Content Database
ProQuest One Academic Middle East (New)
ProQuest Central Essentials
ProQuest Health & Medical Complete (Alumni)
ProQuest Central (Alumni Edition)
ProQuest One Community College
ProQuest One Health & Nursing
ProQuest Central China
ProQuest Central
ProQuest Health & Medical Research Collection
Health Research Premium Collection
Health and Medicine Complete (Alumni Edition)
ProQuest Central Korea
Health & Medical Research Collection
ProQuest Central (New)
ProQuest One Academic Eastern Edition
ProQuest Health Management
ProQuest Hospital Collection
Health Research Premium Collection (Alumni)
ProQuest Hospital Collection (Alumni)
ProQuest Health & Medical Complete
ProQuest One Academic UKI Edition
ProQuest Health Management (Alumni Edition)
ProQuest One Academic
ProQuest One Academic (New)
ProQuest Central (Alumni)
MEDLINE - Academic
DatabaseTitleList
MEDLINE
Publicly Available Content Database
MEDLINE - Academic
Database_xml – sequence: 1
  dbid: DOA
  name: DOAJ Directory of Open Access Journals
  url: https://www.doaj.org/
  sourceTypes: Open Website
– sequence: 2
  dbid: NPM
  name: PubMed
  url: https://proxy.k.utb.cz/login?url=http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?db=PubMed
  sourceTypes: Index Database
– sequence: 3
  dbid: EIF
  name: MEDLINE
  url: https://proxy.k.utb.cz/login?url=https://www.webofscience.com/wos/medline/basic-search
  sourceTypes: Index Database
– sequence: 4
  dbid: BENPR
  name: ProQuest Central
  url: https://www.proquest.com/central
  sourceTypes: Aggregation Database
DeliveryMethod fulltext_linktorsrc
Discipline Medicine
EISSN 2291-9694
EndPage e65454
ExternalDocumentID oai_doaj_org_article_6321ca7ae3144e81869cb70cb912954c
PMC11884378
39864953
10_2196_65454
Genre Journal Article
GeographicLocations United States
GeographicLocations_xml – name: United States
GrantInformation_xml – fundername: NIMH NIH HHS
  grantid: R01 MH126895
GroupedDBID 53G
5VS
7X7
8FI
8FJ
AAFWJ
AAYXX
ABUWG
ADBBV
AFKRA
AFPKN
ALIPV
ALMA_UNASSIGNED_HOLDINGS
AOIJS
BAWUL
BCNDV
BENPR
CCPQU
CITATION
DIK
EMOBN
FYUFA
GROUPED_DOAJ
HMCUK
HYE
KQ8
M0T
M48
M~E
OK1
PGMZT
PHGZM
PHGZT
PIMPY
RPM
UKHRP
CGR
CUY
CVF
ECM
EIF
NPM
3V.
7XB
8FK
AZQEC
DWQXO
K9.
PJZUB
PKEHL
PPXIY
PQEST
PQQKQ
PQUKI
PRINS
7X8
5PM
PUEGO
ID FETCH-LOGICAL-c422t-49c5c77dcfe5b6e8299e4c8913e4d13ebe9cab890d58aa8593ef60bae049c4b23
IEDL.DBID 7X7
ISSN 2291-9694
IngestDate Wed Aug 27 01:30:38 EDT 2025
Thu Aug 21 18:33:45 EDT 2025
Tue Aug 05 11:15:25 EDT 2025
Fri Jul 25 22:25:15 EDT 2025
Tue May 06 01:31:28 EDT 2025
Tue Jul 01 05:19:22 EDT 2025
IsDoiOpenAccess true
IsOpenAccess true
IsPeerReviewed true
IsScholarly true
Keywords natural language processing
mental health
AI
mental health disorder
LLM
emergency department
large language model
machine learning
EHR
EHR system
artificial intelligence
NLP
physical health
ChatGPT
predictive modeling
health informatics
text
dataset
electronic health record
ML
Language English
License Nicholas C Cardamone, Mark Olfson, Timothy Schmutte, Lyle Ungar, Tony Liu, Sara W Cullen, Nathaniel J Williams, Steven C Marcus. Originally published in JMIR Medical Informatics (https://medinform.jmir.org).
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on https://medinform.jmir.org/, as well as this copyright and license information must be included.
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-c422t-49c5c77dcfe5b6e8299e4c8913e4d13ebe9cab890d58aa8593ef60bae049c4b23
Notes ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 14
content type line 23
None declared.
ORCID 0000-0001-9854-8565
0000-0002-3958-5662
0000-0003-1711-1906
0000-0002-7846-5727
0000-0002-3948-7480
0000-0003-2047-1443
0000-0001-7819-3824
0000-0002-3707-3989
OpenAccessLink https://www.proquest.com/docview/3158551068?pq-origsite=%requestingapplication%
PMID 39864953
PQID 3158551068
PQPubID 4997117
ParticipantIDs doaj_primary_oai_doaj_org_article_6321ca7ae3144e81869cb70cb912954c
pubmedcentral_primary_oai_pubmedcentral_nih_gov_11884378
proquest_miscellaneous_3160071362
proquest_journals_3158551068
pubmed_primary_39864953
crossref_primary_10_2196_65454
ProviderPackageCode CITATION
AAYXX
PublicationCentury 2000
PublicationDate 20250121
PublicationDateYYYYMMDD 2025-01-21
PublicationDate_xml – month: 1
  year: 2025
  text: 20250121
  day: 21
PublicationDecade 2020
PublicationPlace Canada
PublicationPlace_xml – name: Canada
– name: Toronto
– name: Toronto, Canada
PublicationTitle JMIR medical informatics
PublicationTitleAlternate JMIR Med Inform
PublicationYear 2025
Publisher JMIR Publications
Publisher_xml – name: JMIR Publications
References Li (R30)
Hossain (R1); 18
Lee (R5); 7
Garrido-Merchan (R29); 2
Kumar (R20); 14
R21
R23
R22
R25
Bayramli (R6); 5
R24
Boudreaux (R3); 95
R26
Lin (R36); 57
Boudreaux (R4); 12
Mahajan (R7); 264
Fung (R27); 48
R8
Sivarajkumar (R34); 12
Akbilgic (R9); 6
R31
R12
Lee (R32); 11
Frenay (R37); 25
Mahmoudi (R2); 369
R11
R33
R14
R13
R35
R16
R15
R18
Scharp (R28); 25
R19
Marafino (R10); 1
Gilardi (R17); 120
References_xml – volume: 12
  ident: R34
  article-title: An empirical evaluation of prompting strategies for large language models in zero-shot clinical natural language processing: algorithm development and validation study
  publication-title: JMIR Med Inform
  doi: 10.2196/55318
– ident: R13
– volume: 5
  issue: 1
  ident: R6
  article-title: Predictive structured-unstructured interactions in EHR models: a case study of suicide prediction
  publication-title: NPJ Digit Med
  doi: 10.1038/s41746-022-00558-0
– ident: R21
  doi: 10.1101/2023.12.25.23300525
– volume: 1
  issue: 8
  ident: R10
  article-title: Validation of prediction models for critical care outcomes using natural language processing of electronic health record data
  publication-title: JAMA Netw Open
  doi: 10.1001/jamanetworkopen.2018.5097
– volume: 18
  start-page: 745
  issue: 2
  ident: R1
  article-title: Use of electronic health data for disease prediction: a comprehensive literature review
  publication-title: IEEE/ACM Trans Comput Biol Bioinform
  doi: 10.1109/TCBB.2019.2937862
– volume: 14
  start-page: 14219
  issue: 3
  ident: R20
  article-title: Towards optimal NLP solutions: analyzing GPT and LLaMA-2 models across model scale, dataset size, and task diversity
  publication-title: Eng Technol Appl Sci Res
  doi: 10.48084/etasr.7200
– ident: R31
  doi: 10.21203/rs.3.rs-3914899/v1
– volume: 7
  issue: 3
  ident: R5
  article-title: Clinical implementation of predictive models embedded within electronic health record systems: a systematic review
  publication-title: Informatics (MDPI)
  doi: 10.3390/informatics7030025
– ident: R19
  doi: 10.2196/preprints.64088
– ident: R30
  publication-title: arxiv
– volume: 95
  ident: R3
  article-title: Emergency department safety assessment and follow-up evaluation 2: an implementation trial to improve suicide prevention
  publication-title: Contemp Clin Trials
  doi: 10.1016/j.cct.2020.106075
– ident: R25
– ident: R16
  doi: 10.1145/3636555.3636910
– ident: R26
  doi: 10.1007/978-3-319-33383-0_5
– volume: 120
  issue: 30
  ident: R17
  article-title: ChatGPT outperforms crowd workers for text-annotation tasks
  publication-title: Proc Natl Acad Sci U S A
  doi: 10.1073/pnas.2305016120
– volume: 25
  start-page: 69
  issue: 1
  ident: R28
  article-title: Natural language processing applied to clinical documentation in post-acute care settings: a scoping review
  publication-title: J Am Med Dir Assoc
  doi: 10.1016/j.jamda.2023.09.006
– volume: 48
  start-page: 503
  issue: 7
  ident: R27
  article-title: Bootstrap estimate of the variance and confidence interval of kappa
  publication-title: Br J Ind Med
  doi: 10.1136/oem.48.7.503
– volume: 369
  ident: R2
  article-title: Use of electronic medical records in development and validation of risk prediction models of hospital readmission: systematic review
  publication-title: BMJ
  doi: 10.1136/bmj.m958
– volume: 264
  ident: R7
  article-title: Combining structured and unstructured data for predicting risk of readmission for heart failure patients
  publication-title: Stud Health Technol Inform
  doi: 10.3233/SHTI190219
– ident: R18
– volume: 12
  ident: R4
  article-title: Applying machine learning approaches to suicide prediction using healthcare data: overview and future directions
  publication-title: Front Psychiatry
  doi: 10.3389/fpsyt.2021.707916
– ident: R33
– ident: R14
– ident: R12
– ident: R35
  doi: 10.18653/v1/2020.acl-main.468
– ident: R23
  doi: 10.2196/preprints.48659
– volume: 6
  start-page: 4
  issue: 1
  ident: R9
  article-title: Unstructured text in EMR improves prediction of death after surgery in children
  publication-title: Informatics (MDPI)
  doi: 10.3390/informatics6010004
– ident: R11
  doi: 10.1109/ICCECE58645.2024.10497313
– volume: 25
  start-page: 845
  issue: 5
  ident: R37
  article-title: Classification in the presence of label noise: a survey
  publication-title: IEEE Trans Neural Netw Learning Syst
  doi: 10.1109/TNNLS.2013.2292894
– volume: 57
  start-page: 243
  issue: 9
  ident: R36
  article-title: Towards trustworthy LLMs: a review on debiasing and dehallucinating in large language models
  publication-title: Artif Intell Rev
  doi: 10.1007/s10462-024-10896-y
– ident: R22
– ident: R24
– ident: R8
  doi: 10.1201/b13617
– volume: 11
  start-page: 2320
  issue: 10
  ident: R32
  article-title: A mathematical investigation of hallucination and creativity in GPT models
  publication-title: Mathematics
  doi: 10.3390/math11102320
– ident: R15
  doi: 10.1145/3581754.3584136
– volume: 2
  ident: R29
  publication-title: J Comput Cogn Eng
SSID ssj0001416667
Score 2.3075974
Snippet Prediction models have demonstrated a range of applications across medicine, including using electronic health record (EHR) data to identify hospital...
Background:Prediction models have demonstrated a range of applications across medicine, including using electronic health record (EHR) data to identify...
Abstract BackgroundPrediction models have demonstrated a range of applications across medicine, including using electronic health record (EHR) data to identify...
SourceID doaj
pubmedcentral
proquest
pubmed
crossref
SourceType Open Website
Open Access Repository
Aggregation Database
Index Database
StartPage e65454
SubjectTerms AI Language Models in Health Care
Anxiety
Artificial intelligence
Chronic fatigue syndrome
Classification
Data warehouses
Datasets
Decision Support for Health Professionals
Diagnostic Tools in Mental Health
Disease
Eating disorders
Electronic Health Records
Electronic Health Records - statistics & numerical data
Emergency Service, Hospital
Emotional disorders
Humans
Impulsivity
Large Language Models
Machine Learning
Medical coding
Mental disorders
Mental Disorders - diagnosis
Mental Health
Methods and New Tools in Mental Health Research
Mood disorders
Multimedia
Natural Language Processing
New Technologies
Original Paper
Personality disorders
Psychosis
Reconciliation
Review boards
Self destructive behavior
United States
SummonAdditionalLinks – databaseName: DOAJ Directory of Open Access Journals
  dbid: DOA
  link: http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwrV1La9wwEB7KHkKhlKRPp0lQoVeTlfV0b23YEEJaetiFvRlJlulC8ZZscugPyP_OjOQ1u6WQSy8-SMbImpHmG83oG4BP3ujO8qBLyztZSkT0pdPBlyZwGXUgvpiUIPtdXy3k9VItd0p9UU5YpgfOE3euRcWDMy4KhP6R-Nfq4M00-JpTiCrQ7os2b8eZSqcrksJh5gBeUK4zatm5Rqgg94xP4uj_F7D8Oz9yx-BcHsLLASmyL3mER_As9q_g4NsQC38ND6me5SrdU2KLgQj2_ja2bI4bLlv1bDaWuGH5thHLzuaGIVJlmb1n2_Pjlj5MUmJUHu3X5jO7oSRxfOYDzdzMZiM7OKMUxD9vYHE5m19clUNRhTLIqrorZR1UMKYNXVReR4vmKMpAwcooW3z4WAfnbT1tlXWO2NBip6feRXQlgvSVeAuTft3H98A6JzrvnffCeGmV9VRcQSlVdbw1CEMLONvOdvM7c2c06HOQOJokjgK-kgzGTqK6Tg2oAM2gAM1TClDAyVaCzbD-No3g6AbhdqNtAR_Hblw5FA5xfVzf0zs6-ei6KuBdFvg4EkGs9bUSBdg9Vdgb6n5Pv_qZ2LnRY7NSGHv8P37uAzyvqODwlJcVP4EJqlI8RRR058-Swj8CUjUHqQ
  priority: 102
  providerName: Directory of Open Access Journals
– databaseName: Scholars Portal Journals: Open Access
  dbid: M48
  link: http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwfV1La9wwEB7aFEIhlD4Tt0lQIVe3K-vpQglp2BBKU3LIQm9GkuVkIXjb3QSaH9D_3RnZa-ISetFBEkZ4RppvNKNvAA680Y3lQeeWNzKXiOhzp4PPTeAy6kB8MSlB9rs-ncmvP9S9bML-B64edO2ontRsef3h96-7Q9zwnymNGRXoo0YUIB_DEzRGhooYnPUIP12zSIqLmU3YGs0eWaFE1v8Qwvw3UfKe5Tl5Ds96yMiOOhm_gEexfQmbZ31Q_BX8SYUt5-nBEpv1jLC3y1izCzx52bxl06HWDeueHbHO61wxhKyso_FZj5wv6cMkLkZ10q5Xn9g3yhbHtrvZ7LrZdKAJZ5SLePcaZifTi-PTvK-ukAdZFDe5LIMKxtShicrraNEuRRkoahlljY2PZXDelpNaWeeIFi02euJdRJ8iSF-IN7DRLtq4A6xxovHeeS-Ml1ZZT1UWlFJFw2uDeDSD_fXfrn52JBoVOh8kjiqJI4MvJINhkDivU8dieVn1W6jSouDBGRcFOoGRmPjK4M0k-JJTsDJksLuWYLXWo0pw9Ifw3NE2g_fDMG4hiou4Ni5uaY5OzrouMtjuBD6sRBB9falEBnakCqOljkfa-VWi6UbXzUph7Nv_r-sdPC2opvCE5wXfhQ1UkriHQOfG7ydV_gvzQgA2
  priority: 102
  providerName: Scholars Portal
Title Classifying Unstructured Text in Electronic Health Records for Mental Health Prediction Models: Large Language Model Evaluation Study
URI https://www.ncbi.nlm.nih.gov/pubmed/39864953
https://www.proquest.com/docview/3158551068
https://www.proquest.com/docview/3160071362
https://pubmed.ncbi.nlm.nih.gov/PMC11884378
https://doaj.org/article/6321ca7ae3144e81869cb70cb912954c
Volume 13
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwfV3daxQxEB-0hSKI-O1qPSL4GnrZZJOsL2LlShFbivTg3pYkm9UD2at37YN_gP-3M9nc1hPxJQ87yxJ2JslvPvIbgLfe6M6KoLkVneIKET13OnhuglBRB-KLSQWy5_p0rj4tqkUOuG1yWeV2T0wbdbsKFCM_kgKBLRqQtu-vfnDqGkXZ1dxC4y7sE3UZWbVZmNsYi6KkmDmA-1TxjLZ2pBEwqJ0jKDH1_wte_l0l-cexc_IQHmS8yD4MCn4Ed2L_GA7Ockb8CfxKXS2X6bYSm2c62Jt1bNklbrts2bPZ2OiGDXeO2OBybhjiVTZw-GwlF2v6MOmKUZO075t37DOViuM4hDWHx2w2coQzKkT8-RTmJ7PLj6c8t1bgQZXlNVd1qIIxbehi5XW0eChFFShlGVWLg491cN7W07ayzhEnWuz01LuIDkVQvpTPYK9f9fEFsM7JznvnvTRe2cp6arFQVVXZidYgGC1gsv3bzdXAoNGg50HqaJI6CjgmHYxCIrxOD1brr01eP42WpQjOuCjRA4xEw1cHb6bB14IylaGAw60Gm7wKN82tzRTwZhTj-qGkiOvj6obe0clT12UBzweFjzORxF1fV7IAu2MKO1PdlfTLb4mjG_02q6SxL_8_r1dwr6SGwlPBS3EIe2gk8TWinGs_SaY8gf3j2fnFl0mKFeB4puxvUNQDQg
linkProvider ProQuest
linkToHtml http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwtV3bbtQwEB1VRSpICHEnUIqR4DFqnDi2g4QQl622dFvxsCvtW7AdB1ZC2bLbCvUD-B2-kRnnUhYh3vqSh3i1sjLH4xnP-ByAF1bJWnMnY81rEQuM6GMjnY2V48JLR3wxoUH2RI5n4uM8n2_Br_4uDLVV9j4xOOpq6eiMfD_jGNgigKR-c_o9JtUoqq72EhotLI78xQ9M2davDz-gfV-m6cFo-n4cd6oCsRNpehaLwuVOqcrVPrfSa_THXjiq1nlR4cP6whmri6TKtTFEB-ZrmVjjMZZ2whLRAbr8a7jxJpTsqbm6PNMRVIRTO3CTOqwR2_sSAxSxseUFZYB_hbN_d2X-sc0d3IZbXXzK3raAugNbvrkLO8ddBf4e_AwqmotwO4rNOvrZ85Wv2BTdPFs0bDQI67D2jhNrU9w1w_iYtZxB_cinFf0xYYORKNu39Ss2odZ0fLbHqO1rNho4yRk1Pl7ch9mVfPQHsN0sG_8IWG2y2lpjbaas0Lm2JOmQ53la80ph8BvBXv-1y9OWsaPETIfMUQZzRPCObDAMEsF2eLFcfSm79VrKLOXOKOMzzDg90f4VzqrE2YJTZdRFsNtbsOxW_bq8xGgEz4dhXK9UhDGNX57Tb2Q4GZBpBA9bgw8zyYgrv8izCPQGFDamujnSLL4GTnDME7XIlH78_3k9g-vj6fGknByeHD2BGymJGSc8TvkubCNg_FOMsM7sXoA1g89XvY5-A8_ZPig
linkToPdf http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwtV1ta9swED5KCmEwxt7rres02D6axLIsyYNR2jWhXbsQRgP95kmyvAWG0yUtoz9gf6q_rnd-6zLGvvWLP1ghCN3p_Jzu9DwAb62ShY6cDHVUiFAgog-NdDZULhJeOuKLqRpkJ_JwJj6dJWcbcN3ehaG2yjYmVoE6Xzg6Ix_EEQJbdCCpB0XTFjE9GO-e_wxJQYoqra2cRu0ix_7qF6Zvqw9HB2jrd5yPR6cfD8NGYSB0gvOLUKQucUrlrvCJlV5jbPbCUeXOixwf1qfOWJ0O80QbQ9RgvpBDazziaicskR5g-N9UlBX1YHN_NJl-uT3hEVSSU324T_3W6OkDiXBFrH0AK52Af4Hbv3s0__jojR_Cgwatsr3avR7Bhi8fQ_9zU49_Ar8rTc15dVeKzRoy2sulz9kprhybl2zUyeyw-sYTqxPeFUO0zGoGoXZkuqQ_Jk9hJNH2Y_WenVCjOj7rQ9X6NRt1DOWM2iCvnsLsTpb9GfTKRem3gBUmLqw11sbKCp1oSwIPSZLwIsoVQuEAdtrVzs5r_o4M8x4yR1aZI4B9skE3SHTb1YvF8lvW7N5MxjxyRhkfY_7piQQwdVYNnU0jqpO6ALZbC2ZNDFhltx4bwJtuGHcvlWRM6ReX9BtZnRNIHsDz2uDdTGJizk-TOAC95gprU10fKeffK4ZwzBq1iJV-8f95vYY-7qHs5Ghy_BLucVI2HkYhj7ahh_7iXyHcurA7jV8z-HrXW-kGp4xDww
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Classifying+Unstructured+Text+in+Electronic+Health+Records+for+Mental+Health+Prediction+Models%3A+Large+Language+Model+Evaluation+Study&rft.jtitle=JMIR+medical+informatics&rft.au=Cardamone%2C+Nicholas+C&rft.au=Olfson%2C+Mark&rft.au=Schmutte%2C+Timothy&rft.au=Ungar%2C+Lyle&rft.date=2025-01-21&rft.pub=JMIR+Publications&rft.eissn=2291-9694&rft.volume=13&rft.spage=e65454&rft_id=info:doi/10.2196%2F65454&rft.externalDBID=HAS_PDF_LINK
thumbnail_l http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/lc.gif&issn=2291-9694&client=summon
thumbnail_m http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/mc.gif&issn=2291-9694&client=summon
thumbnail_s http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/sc.gif&issn=2291-9694&client=summon