On January 2025, Anthropic announced an innovative new feature named Citations for its AI model Claude, designed to enhance the accuracy of generated responses while minimizing errors commonly referred to as confabulations or hallucinations. This exciting addition marks a significant move towards making AI models more reliable and trustworthy by referencing source documents directly linked to their generated outputs.
The Citations feature allows developers using Claude models to integrate source documents into the model’s context window. This is accomplished by enabling the model to automatically cite specific segments of text that are utilized to provide answers. According to Anthropic’s announcement, when Citations is enabled, the API processes user-provided documents, which can be in PDF or plaintext format, by segmenting them into sentences. These segmented sentences, along with any additional context provided by the user, are then fed into the model, allowing it to generate responses based on solid references.
This new feature has a broad spectrum of potential applications. Developers can use Citations for a variety of purposes, such as summarizing legal case files with key points accurately linked to original documents, answering inquiries related to financial data with well-sourced references, and providing customer support that cites specific product documentation. This could greatly improve the utility of AI systems in professional environments where verified information is crucial.
In internal testing, the introduction of Citations reportedly improved recall accuracy by up to 15% compared to more traditional custom citation implementations developed by users through prompts. Although a 15% boost may initially seem modest, the feature has already attracted attention from AI researchers and developers for its fundamental incorporation of Retrieval Augmented Generation (RAG) techniques. Simon Willison, a prominent figure in the AI community, elaborated on the importance of citation features in his blog, illustrating how they bolster accuracy by providing reliable document excerpts in response to user inquiries.
Willison emphasized that the core principle behind the RAG method is to take a user’s question, extract relevant text fragments from documents, and formulate an accurate answer that incorporates these excerpts, thereby enhancing the reliability of the model’s answers. However, he warned of the ongoing risks that AI models may still hallucinate incorrect details drawn from other, unrelated training data. Citing sources helps alleviate some of this risk, but the complexity involved in establishing an effective citation system should not be understated.
Anthropic’s Alex Albert provided insights into the technology, stating that Claude has been engineered to cite sources effectively. The launch of Citations now exposes this capability to developers, allowing them to activate it through the application programming interface (API) by passing a specific parameter. This parameter signals the model to cite sources whenever relevant documents are provided.
Citations have been integrated into Anthropic’s Claude 3.5 Sonnet and Claude 3.5 Haiku models, both accessible via the Anthropic API and Google’s Vertex AI platform. Evidence of the feature’s practical use is already emerging; for instance, Thomson Reuters, which employs Claude in its CoCounsel legal AI reference platform, is eager to leverage Citations to bolster trust in the reliability of AI-generated content while reducing the risk of hallucinations.
Furthermore, financial technology company Endex reported impressive results, claiming that the usage of Citations effectively eliminated their source confabulation issues from 10% down to zero, while concurrently increasing the number of references provided in responses by 20%. This point highlights the potential for AI tools to become robust pillars of information fidelity in the financial sector.
However, despite the promising advancements related to accuracy and citation, it is crucial to remain cautious about relying wholly on any large language model to present accurate reference information until further extensive studies are conducted. The risk of transmitting inaccurate details cannot be discounted, especially as the technology continues to evolve.
For developers interested in utilizing Citations, Anthropic has confirmed that it will adhere to its standard token-based pricing model; notably, any quoted text included in responses will not contribute to output token costs. For instance, sourcing a 100-page document as a reference, utilizing Claude 3.5 Sonnet will cost approximately $0.30, and Claude 3.5 Haiku will cost around $0.08, thus providing an economical method for leveraging comprehensive research documents.
In summary, Anthropic’s new Citations feature represents a substantial leap forward in fostering greater reliability and trust in AI-generated responses. By linking model outputs to verified source materials, Claude could redefine the parameters of accuracy in artificial intelligence applications, deepening its integration into professional use while paving the way for even more advanced features in the future.