Collection of pre-trained encoder-decoder models and optimization techniques
Top 22.2% on sourcepulse
AliceMind is a comprehensive collection of pre-trained encoder-decoder models and optimization techniques from Alibaba's MinD Lab, targeting researchers and developers in NLP and multimodal AI. It offers a wide array of models for tasks spanning text, image, and video understanding and generation, alongside efficient fine-tuning and compression methods.
How It Works
AliceMind provides a modularized foundation for large multimodal language models (LMMs), enabling modal collaboration. Its models are pre-trained on large-scale datasets using both discriminative and generative objectives. Key innovations include parameter-efficient fine-tuning methods like ChildTuning and PST, and compression techniques like ContrastivePruning, all designed to enhance generalization and reduce resource requirements.
Quick Start & Requirements
Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The README lists numerous models and techniques, but specific installation instructions, hardware requirements, and detailed benchmarks for each are not consolidated in one place, potentially requiring users to consult individual model papers or documentation.
1 year ago
1 week