Discover and explore top open-source AI tools and projects—updated daily.
Curated list for large model trustworthiness
Top 25.4% on SourcePulse
This repository is a curated reading list for large model trustworthiness, focusing on safety, security, and privacy, with a special emphasis on multi-modal LMs. It serves researchers and practitioners in the field by providing a structured collection of papers, toolkits, and surveys.
How It Works
The project maintains a manually collected and categorized list of resources, primarily academic papers, organized by specific sub-topics within LM safety, security, and privacy. It includes classifications for jailbreaking, prompt injection, adversarial examples, data privacy, copyright, and more, facilitating targeted research.
Quick Start & Requirements
Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The repository is explicitly marked as "in progress" and relies on manual collection, which may lead to occasional omissions or categorization nuances. Licenses for linked external resources are not consolidated.
4 days ago
1 day