LCIRC

Created
Created
2025 Feb 11 13:0
Creator
Creator
Seonglae Cho
Editor
Edited
Edited
2025 Feb 11 13:9
Refs
Refs

Long-form Context Injection with Recurrent Compression

  1. Recurrent Context Compression
  1. Compressed Context Injection
notion image

Perceiver module

The compressed features are obtained through the Perceiver module as follows
where represents the Perceiver module with query and input features

Recurrent compression process

Segments are sequentially fed into the Perceiver module, with the compressed features from the previous segment serving as the query features for the next segment:

Compressed Context Injection

Injected this compressed information into the pretrained transformer using gated cross-attention layers with residual connections. The embeddings are then contextualized through Gated Cross Attention Block.
Query Dependent Context Modeling
Addition of query dependent modeling through an extra GCA block in the recurrent compression step
Addition of query dependent modeling through an extra GCA block in the recurrent compression step
 
 
 
 
 

Recommendations