A paper titled, “Introducing Answered with Evidence – A framework for evaluating whether large language model (LLM) responses to biomedical questions are founded in evidence,” was published in medrxiv by authors affiliated with Atropos Health. 

Short summary:

The growing use of LLMs for biomedical question answering raises concerns about the accuracy and evidentiary support of their responses. To address this, we present Answered with Evidence, a framework for evaluating whether LLM-generated answers are grounded in scientific literature. We analyzed thousands of physician-submitted questions using a comparative pipeline that included: (1) Alexandria™, fka the Atropos Evidence™ Library, a retrieval-augmented generation (RAG) system based on novel observational studies, and (2) two PubMed-based retrieval-augmented systems (System and Perplexity). We found that PubMed-based systems provided evidence-supported answers for approximately 44% of questions, while the novel evidence source did so for about 50%. Combined, these sources enabled reliable answers to over 70% of biomedical queries. As LLMs become increasingly capable of summarizing scientific content, maximizing their value will require systems that can accurately retrieve both published and custom-generated evidence—or generate such evidence in real time.

Read the full paper