Texonom
Texonom
/
Engineering
Engineering
/Data Engineering/Artificial Intelligence/AI Object/NLP/Language Model/Language Model Context/LM Context Extending/
LongNet
Search

LongNet

Created
Created
2023 Jul 13 9:1
Creator
Creator
Seonglae Cho
Editor
Editor
Seonglae Cho
Edited
Edited
2024 Mar 2 6:2
Refs
Refs
1B sequence length
notion image
Dilated Attention
notion image
 
 
 
Microsoft’s LongNet Scales Transformer to One Billion Tokens
Scaling sequence length is of paramount importance for large language models, as it brings about singnificant benefits. These advantages…
Microsoft’s LongNet Scales Transformer to One Billion Tokens
https://medium.com/syncedreview/microsofts-longnet-scales-transformer-to-one-billion-tokens-af02ff657d87
Microsoft’s LongNet Scales Transformer to One Billion Tokens
arxiv.org
https://arxiv.org/pdf/2307.02486.pdf
 
 

Recommendations

Texonom
Texonom
/
Engineering
Engineering
/Data Engineering/Artificial Intelligence/AI Object/NLP/Language Model/Language Model Context/LM Context Extending/
LongNet
Copyright Seonglae Cho