LM Context Extending

Creator
Creator
Seonglae ChoSeonglae Cho
Created
Created
2023 Jul 9 12:44
Editor
Edited
Edited
2025 Jan 12 15:47
Scaling context-size can flexibly re-organize model representations, possibly unlocking novel capabilities. This multi-faceted examination of LLM capabilities suggests that LLMs can achieve scaling through
Language Model Context
scaling and
LM Context Extending
, demonstrating that this scaling improves connectivity and reconstruction capabilities between concepts in in-context learning and representation.
While long-context LLMs provide simplicity, their ability to accurately recall factual information can diminish. In other words, the attention mechanism struggles to manage complexity when it scales.
Language Model Context Extending Methods
 
 
 
arxiv.org
Paper page - Extending Context Window of Large Language Models via Positional Interpolation
Join the discussion on this paper page
Paper page - Extending Context Window of Large Language Models via Positional   Interpolation
Hugging Face Reads, Feb. 2021 - Long-range Transformers
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
Hugging Face Reads, Feb. 2021 - Long-range Transformers
 
 

Recommendations