Sign up to save tools and stay up to date with the latest in AI
bg
bg
1

Giraffe – Long Context LLMs

Aug 24, 2023 - blog.abacus.ai
The article announces the release of an arXiv paper titled “Giraffe: Adventures in Expanding Context Lengths in LLMs”, which introduces a new family of models called Giraffe that are finetuned from base LLaMA and LLaMA2. The paper explores the topic of context length extrapolation in Large Language Models (LLMs), a method that uses an LLM trained on a short context length for evaluation on longer context lengths. The authors also release the weights of the Giraffe models on HuggingFace, along with their training code, evaluation datasets, and evaluation scripts.

The paper discusses the challenges of training models on longer contexts due to the quadratic scaling of self-attention in memory and compute as context length increases. It reviews various methods proposed for context length extrapolation, and introduces a new approach called truncation. The authors also critique the commonly used next-token perplexity metric for evaluating LLM performance, and introduce new tasks focused on model recall accuracy. The paper concludes by acknowledging the ongoing challenges in context length extrapolation of LLMs, and expresses interest in further research on the topic.

Key takeaways:

  • The article introduces Giraffe, a new family of models finetuned from base LLaMA and LLaMA2, with a focus on context length extrapolation in Large Language Models (LLMs).
  • Context length extrapolation is crucial for tasks that require the model to attend to a larger corpus of data, such as information retrieval, maintaining long conversations, or coding assistance.
  • The paper explores various methods for context length extrapolation, including a new approach called truncation, and introduces new tasks focused on the accuracy of model recall rather than text coherence.
  • Despite the advancements, the paper acknowledges that none of the methods fully satisfy the need for true extrapolation without performance degradation, indicating a need for further research in this area.
View Full Article

Comments (0)

Be the first to comment!