xlite-dev/Awesome-LLM-Inference
Inference Engines📚A curated list of Awesome LLM/VLM Inference Papers with Codes: Flash-Attention, Paged-Attention, WINT8/4, Parallelism, etc.🎉
GitHub Metrics
Stars
5.0k
Forks
346
Open Issues
—
Watchers
134
Contributors
33
Weekly Commits
0
Language
Python
License
GPL-3.0
Last Commit
Feb 28, 2026
Created
Aug 27, 2023
Latest Release
v2.6.20
Release Date
Jun 17, 2025
Synced: Mar 3, 2026
Quality Scores
Documentation Qualityw: 20%
0.0
Community Healthw: 20%
0.0
Maintenance Velocityw: 15%
0.0
API Design & DXw: 20%
0.0
Production Readinessw: 15%
0.0
Ecosystem Integrationw: 10%
0.0
Tags
awesome-llmdeepseekdeepseek-r1deepseek-v3flash-attentionflash-attention-3flash-mlallm-inferenceminimax-01mla
Radar
No scores yet