Skip to content

Latest commit

 

History

History
24 lines (18 loc) · 1.52 KB

Retrieval Meets Long Context LLMs - 2023.md

File metadata and controls

24 lines (18 loc) · 1.52 KB

Retrieval Meets Long Context Large Language Models


Abstract

  • The paper discusses the popularity of extending the context window of large language models (LLMs).
  • It explores the comparison between retrieval-augmentation and long context window extension for LLMs.
  • The study uses two pretrained LLMs, the proprietary 43B GPT and LLaMA2-70B, for experimentation.
  • Surprisingly, the research finds that a 4K context window LLM with retrieval-augmentation can achieve comparable performance to a finetuned LLM with a 16K context window, while being computationally more efficient.
  • Retrieval is shown to significantly enhance the performance of LLMs, regardless of their extended context window sizes.
  • The best-performing model in the study is a retrieval-augmented LLaMA2-70B with a 32K context window, outperforming other models in various long context tasks, such as question answering and query-based summarization.
  • This retrieval-augmented model also surpasses its non-retrieval baseline while being faster at generation.
  • The study offers valuable insights for practitioners on choosing between retrieval-augmentation and long context extension for LLMs.

Results

ret1

ret3

ret2


Paper: https://arxiv.org/abs/2310.03025