Building Safety Guards into LLMs for Trustworthy Automatic Simplification of Medical Documents
Project Number1R01LM014600-01
Contact PI/Project LeaderLI, JUNYI JESSY Other PIs
Awardee OrganizationUNIVERSITY OF TEXAS AT AUSTIN
Description
Abstract Text
Abstract
Texts describing medical advances are of keen interest to the general public. However, reliable
medical evidence is largely disseminated in peer-reviewed journal articles describing new
findings. Because such articles are wrięen in technical language intended for experts, this
“primary literature” is effectively inaccessible to the general population. With very large
language models (LLMs) like ChatGPT now widely available, lay people are increasingly
turning to them for medical information. One potentially promising use is for LLMs to provide
simplified versions of medical papers. However, while LLMs can capably simplify texts
automatically, they can also still generate inaccurate, unsupported, and/or potentially
misleading information, posing a risk.
This proposal seeks to develop novel natural language processing (NLP) technologies to
mitigate risks and improve the reliability of LLM outputs for the task of medical text
simplification. Given the high-stakes of healthcare information, we focus on building
controllable, transparent LLMs that are moderately sized, and design tools that enable
communities to strike a balance between using LLMs safely and perceiving their outputs
critically, while (potentially) improving health literacy by eventually empowering the public
with more reliable access to high-quality, newly published medical findings.
We propose several methodological safeguards. To begin, we will design the first error
detection model for LLM-generated simplifications of medical texts, trained with
expert-annotated data focusing on factual correctness. This tool will then allow us to build safer
knowledge distillation methods, i.e., training much more efficient, smaller models on examples
elicited from massive, closed models like GPT-4 calibrated by estimated confidence of their
correctness. With full access to the parameters of the distilled model, we propose innovative
ways to improve the factuality and readability of the output, and to estimate the model’s
(un)certainty of its own output. We will then integrate these safety guards into a prototype,
such that they can be evaluated by medical experts and lay readers.
Public Health Relevance Statement
Project Narrative
This proposal concerns the development of natural language processing methods and tools that
establish essential guardrails to the safe use of large language models (LLMs) when used to
simplify medical texts. The premise is that LLMs offer the tantalizing opportunity to enable lay
access to newly published evidence, but current models risk introducing inaccurate information
into automatically simplified texts; this proposal concerns designing techniques that mitigate
such risk. These kinds of “safe generation” techniques have the potential to empower users with
the ability to critically read LLM outputs; our immediate, important aim is to enable access to
trustworthy plain language summaries of medical evidence, but the developed methods will
have application in similar contexts where “safe” generation is critical.
NIH Spending Category
No NIH Spending Category available.
Project Terms
AwarenessCalibrationChatGPTCommunicationCommunitiesDataData CollectionDetectionDevelopmentDimensionsEquilibriumEvaluationGeneral PopulationGenerationsGrainHealth PersonnelHealthcareHumanIndividualInformation DisseminationInterventionJudgmentKnowledgeLanguageLiteratureLlamaMedicalMedicineMethodologyMethodsModelingModernizationNatural Language ProcessingOutputPaperPatientsPeer ReviewPersonsProductionPublic HealthPublishingReadabilityReaderReportingRiskSafetySamplingSeveritiesSourceSpecialistSystemTaxonomyTechniquesTechnologyTerminologyTextTrainingUncertaintyVariantVentVisionWorkWritingannotation systemdesignempowermenthealth literacyimprovedinnovationinterestjournal articlelarge language modelmembernovelnovel strategiesopen sourcepredictive modelingprototyperisk mitigationstudent trainingsynergismteachertooltrustworthiness
No Sub Projects information available for 1R01LM014600-01
Publications
Publications are associated with projects, but cannot be identified with any particular year of the project or fiscal year of funding. This is due to the continuous and cumulative nature of knowledge generation across the life of a project and the sometimes long and variable publishing timeline. Similarly, for multi-component projects, publications are associated with the parent core project and not with individual sub-projects.
No Publications available for 1R01LM014600-01
Patents
No Patents information available for 1R01LM014600-01
Outcomes
The Project Outcomes shown here are displayed verbatim as submitted by the Principal Investigator (PI) for this award. Any opinions, findings, and conclusions or recommendations expressed are those of the PI and do not necessarily reflect the views of the National Institutes of Health. NIH has not endorsed the content below.
No Outcomes available for 1R01LM014600-01
Clinical Studies
No Clinical Studies information available for 1R01LM014600-01
News and More
Related News Releases
No news release information available for 1R01LM014600-01
History
No Historical information available for 1R01LM014600-01
Similar Projects
No Similar Projects information available for 1R01LM014600-01