LM Context Extending

Creator
Creator
Seonglae Cho
Created
Created
2023 Jul 9 12:44
Editor
Edited
Edited
2025 Jan 12 15:47
Scaling context-size can flexibly re-organize model representations, possibly unlocking novel capabilities. This multi-faceted examination of LLM capabilities suggests that LLMs can achieve scaling through
Language Model Context
scaling and
LM Context Extending
, demonstrating that this scaling improves connectivity and reconstruction capabilities between concepts in in-context learning and representation.
While long-context LLMs provide simplicity, their ability to accurately recall factual information can diminish. In other words, the attention mechanism struggles to manage complexity when it scales.
Language Model Context Extending Methods
 
 
 
 
 

Recommendations