Survey paper for resource-efficient LLMs and multimodal foundation models
Top 99.5% on sourcepulse
This repository serves as a comprehensive survey of resource-efficient techniques for Large Language Models (LLMs) and multimodal foundation models. It targets researchers and engineers working with large-scale AI models, providing a structured overview of algorithmic and system-level innovations aimed at reducing computational, memory, and storage demands. The primary benefit is a consolidated reference for understanding and implementing efficiency strategies in foundation models.
How It Works
The survey categorizes research into several key areas: resource-efficient architectures (e.g., efficient attention mechanisms, dynamic networks, Mixture-of-Experts), resource-efficient algorithms (pre-training, fine-tuning, inference, model compression), and resource-efficient systems (distributed training, federated learning, serving on cloud/edge). It focuses on papers from top-tier CS conferences and arXiv, published primarily after 2020, to capture the latest advancements.
Quick Start & Requirements
This repository is a curated list of papers and does not require installation or execution. It serves as a reference guide.
Highlighted Details
Maintenance & Community
The repository is actively maintained by UbiquitousLearning. Contributions are welcomed via GitHub Issues for paper suggestions.
Licensing & Compatibility
The repository itself contains links to research papers and code repositories, each with their own licenses. The survey content is likely under a permissive license, but users should verify individual paper/code licenses.
Limitations & Caveats
The survey explicitly excludes hardware design innovations. It focuses on algorithmic and system aspects, and the definition of "resource" is limited to physical resources, excluding data or privacy considerations.
10 months ago
1 week