Texonom
Texonom
/
Engineering
Engineering
/Data Engineering/Artificial Intelligence/Machine Learning/Neural Network/Neural Network Structure/Seq2Seq/Attention Mechanism/Attention Mechanism Optimization/Multi-head Attention/
Attention head pruning
Search

Attention head pruning

Creator
Creator
Seonglae Cho
Created
Created
2025 Mar 13 17:18
Editor
Editor
Seonglae Cho
Edited
Edited
2025 Mar 13 17:18
Refs
Refs
 
 
 
 
 
 
Layer-wise Pruning of Transformer Attention Heads for Efficient...
While Transformer-based models have shown impressive language modeling performance, the large computation cost is often prohibitive for practical use. Attention head pruning, which removes...
Layer-wise Pruning of Transformer Attention Heads for Efficient...
https://arxiv.org/abs/2110.03252
Layer-wise Pruning of Transformer Attention Heads for Efficient...
 
 

Recommendations

Texonom
Texonom
/
Engineering
Engineering
/Data Engineering/Artificial Intelligence/Machine Learning/Neural Network/Neural Network Structure/Seq2Seq/Attention Mechanism/Attention Mechanism Optimization/Multi-head Attention/
Attention head pruning
Copyright Seonglae Cho