Platform - long context vs RAG + Hallucination

RAG LongContext

In this session, our readings cover:

Required Readings:

Importing Phantoms: Measuring LLM Package Hallucination Vulnerabilities

YaRN: Efficient Context Window Extension of Large Language Models

Long Context vs. RAG for LLMs: An Evaluation and Revisits

More reading:

LazyLLM: Dynamic Token Pruning for Efficient Long Context LLM Inference

Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention

Don’t Do RAG: When Cache-Augmented Generation is All You Need for Knowledge Tasks