Skip to content

weigao266/Awesome-Efficient-Arch

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

24 Commits
 
 
 
 

Repository files navigation

Awesome-Efficient-Arch

Awesome arXiv Paper Last Commit Contribution Welcome

🔥 News

✨ Overview

📄 Paper List

Table of Contents


Linear Sequence Modeling

Linear Attention

Linear RNN

State Space Model

Test-Time-Training RNN

Unified Linear Sequence Modeling

Linearization

Hardware-efficient Implementation

Sparse Sequence Modeling

Static Sparse Attention

Dynamic Sparse Attention

Training-free Sparse Attention

Hardware-efficient Implementation

Efficient Full Attention

IO-Aware Attention

Grouped Attention

Mixture of Attention

Quantized Attention

Sparse Mixture-of-Experts

Routing Mechanisms

Expert Architectures

MoE Conversion

Hybrid Architectures

Inter-layer Hybrid

Intra-layer Hybrid

Diffusion Large Language Models

Non-Autoregressive Diffusion LLM

Bridging Diffusion LLM and Autoregressive

Extending Diffusion LLM to Multimodality

Applications to Other Modalities

Vision

Audio

Multimodality

🤝🏻 Contribution

Contributing to the paper list

Join us to improve this repo! ⭐ If you know of any Awesome-Efficient-Arch work we've missed, please contribute via PR or raise an issue. Your contributions are very welcomed!

Contributors

🖤 Citation

If you find this survey useful, please consider citing our paper:

@article{sun2025survey,
  title={Speed Always Wins: A Survey on Efficient Architectures for Large Language Models},
  author={Sun, Weigao and Hu, Jiaxi and Zhou, Yucheng and Du, Jusen and Lan, Disen and Wang, Kexin and Zhu, Tong and Qu, Xiaoye and Zhang, Yu and Mo, Xiaoyu and Liu, Daizong and Liang, Yuxuan and Chen, Wenliang and Li, Guoqi and Cheng, Yu},
  journal={arXiv preprint arXiv:2508.09834},
  year={2025}
}

Releases

No releases published

Packages

No packages published

Contributors 7