Published
May 3, 2024
Updated
May 9, 2024

Can AI Really Cite Sources? A New Benchmark Challenges LLMs

REASONS: A benchmark for REtrieval and Automated citationS Of scieNtific Sentences using Public and Proprietary LLMs
By
Deepa Tilwani|Yash Saxena|Ali Mohammadi|Edward Raff|Amit Sheth|Srinivasan Parthasarathy|Manas Gaur

Summary

Citing sources correctly is crucial for any writer, especially in academic and professional settings. But can large language models (LLMs), designed to generate human-like text, reliably handle this essential task? A new research paper and benchmark called REASONS (Retrieval and Automated CitationS Of Scientific Sentences) puts LLMs to the test, revealing some surprising results. The study explored how well different LLMs, including popular ones like GPT-3.5 and GPT-4, could accurately cite scientific papers. They used two main ways of asking the LLMs: "direct queries," where they asked for the authors of a given paper, and "indirect queries," where they gave the LLM a sentence from one paper and asked it to identify the paper it was citing. The researchers built a massive dataset of scientific papers from various fields like computer vision, robotics, and AI, to see how the LLMs performed across different subjects. What they found was a mixed bag. While some LLMs did reasonably well, especially with direct queries, many struggled, particularly with indirect queries. Often, they would either "pass" on the question entirely or, worse, hallucinate citations, making up references that didn't exist. Interestingly, even LLMs specifically designed for citation generation, like Perplexity.ai, sometimes fell short. The research also tested a technique called Retrieval Augmented Generation (RAG), which allows LLMs to access external information to improve their responses. They found that RAG significantly boosted the performance of some LLMs, making them more accurate and less prone to hallucinations. However, even with RAG, some LLMs still struggled with more complex or nuanced queries. One key takeaway is that simply having access to more information isn't enough. The LLMs need to understand the context and meaning of the text to cite sources correctly. This is where many of them currently fall short. The REASONS benchmark provides a valuable tool for evaluating the citation capabilities of LLMs and highlights the need for further research in this area. As LLMs become more integrated into our lives, ensuring they can accurately and reliably cite sources is crucial for maintaining trust and preventing the spread of misinformation.
🍰 Interesting in building your own agents?
PromptLayer provides the tools to manage and monitor prompts with your whole team. Get started for free.

Question & Answers

What is Retrieval Augmented Generation (RAG) and how does it improve LLM citation accuracy?
RAG is a technique that enhances LLM performance by allowing them to access external information sources during response generation. In the context of citations, RAG works by first retrieving relevant documents from a knowledge base, then using this information to generate more accurate citations. The process involves: 1) Document retrieval from a verified database, 2) Context integration with the LLM's existing knowledge, and 3) Response generation based on combined information. For example, when citing a scientific paper, RAG would first fetch the actual paper details from a database, cross-reference them with the query, and then generate an accurate citation, significantly reducing hallucination risks.
Why is accurate source citation important in the age of AI content generation?
Accurate source citation is crucial in AI content generation because it maintains intellectual integrity and ensures information credibility. It helps readers verify facts, track information sources, and distinguish between verified knowledge and AI-generated content. Key benefits include protecting against misinformation, supporting academic integrity, and building trust in AI-generated content. In practical applications, proper citations help businesses create trustworthy content, assist researchers in validating findings, and enable students to properly attribute information sources in their work.
How can AI citation tools benefit content creators and researchers?
AI citation tools can streamline the research and writing process by automating source attribution and reducing manual citation work. These tools help maintain accuracy, save time, and ensure consistency in citation formats across documents. Benefits include reduced human error, faster content production, and improved citation accuracy. For example, content creators can quickly generate properly formatted citations for multiple sources, while researchers can efficiently manage large reference lists in academic papers. This technology is particularly valuable for digital publishers, academic institutions, and professional writers who handle extensive source documentation.

PromptLayer Features

  1. Testing & Evaluation
  2. Aligns with the paper's systematic evaluation of citation accuracy across different LLMs and query types
Implementation Details
Set up batch testing pipelines for citation accuracy using REASONS benchmark dataset, implement A/B testing between RAG and non-RAG approaches, establish scoring metrics for citation accuracy
Key Benefits
• Systematic evaluation of citation accuracy across models • Quantitative comparison between different prompt strategies • Early detection of hallucination issues
Potential Improvements
• Add domain-specific citation accuracy metrics • Implement automated regression testing for citation quality • Develop specialized hallucination detection scores
Business Value
Efficiency Gains
Reduces manual verification time by 70% through automated testing
Cost Savings
Minimizes resources spent on detecting and correcting citation errors
Quality Improvement
Ensures consistent citation accuracy across all deployments
  1. Workflow Management
  2. Relates to the paper's exploration of RAG implementation and multi-step citation processes
Implementation Details
Create reusable RAG templates for citation tasks, implement version tracking for citation accuracy, build multi-step citation verification workflows
Key Benefits
• Standardized citation workflows across teams • Trackable improvements in citation accuracy • Reproducible RAG implementation
Potential Improvements
• Add specialized citation verification steps • Implement citation context awareness • Develop custom RAG optimization workflows
Business Value
Efficiency Gains
Streamlines citation workflow implementation by 50%
Cost Savings
Reduces development time for citation-heavy applications
Quality Improvement
Ensures consistent citation quality across different implementations

The first platform built for prompt engineering