Awesome-LVLM-Attack  by liudaizong

Curated list of attacks on large vision-language models (LVLMs)

Created 1 year ago
383 stars

Top 74.6% on SourcePulse

GitHubView on GitHub
Project Summary

This repository is a curated list of research papers focusing on attacks against Large Vision-Language Models (LVLMs). It serves as a comprehensive resource for researchers and practitioners interested in the security vulnerabilities and adversarial robustness of multimodal AI systems. The collection aims to track the latest advancements in LVLM attack methodologies, including adversarial attacks, prompt injection, data poisoning, and specialized attacks for LVLM applications.

How It Works

The repository functions as a dynamic, continuously updated bibliography. It categorizes LVLM attacks into distinct types, providing links to relevant research papers, often with associated GitHub repositories for code implementations or further details. The primary goal is to consolidate and organize the rapidly evolving field of LVLM security research, making it accessible to the community.

Quick Start & Requirements

This repository is a curated list of papers and does not have a direct installation or execution process. It requires no specific software or hardware to access.

Highlighted Details

  • Comprehensive categorization of LVLM attacks: Adversarial Attacks, Jailbreak Attacks, Prompt Injection, Data Poisoning, and Special Attacks for LVLM Applications.
  • Links to numerous research papers, many with associated GitHub repositories for code and further details.
  • Includes a survey paper, "A Survey of Attacks on Large Vision-Language Models: Resources, Advances, and Future Trends," for a structured overview.
  • Covers attacks across various domains, including robotics and medical applications.

Maintenance & Community

The repository is maintained by liudaizong and welcomes contributions for missed papers via email (dzliu@stu.pku.edu.cn). The primary citation is provided for the survey paper.

Licensing & Compatibility

The repository itself is a list of links and does not impose a license. Individual linked papers and code repositories will have their own respective licenses.

Limitations & Caveats

This is a curated list and does not provide tools or implementations for performing the attacks. The focus is purely on the research literature. The rapid pace of research means the list may not be exhaustive at any given moment.

Health Check
Last Commit

2 weeks ago

Responsiveness

1 day

Pull Requests (30d)
0
Issues (30d)
0
Star History
22 stars in the last 30 days

Explore Similar Projects

Starred by Elie Bursztein Elie Bursztein(Cybersecurity Lead at Google DeepMind), Chip Huyen Chip Huyen(Author of "AI Engineering", "Designing Machine Learning Systems"), and
6 more.

llm-attacks by llm-attacks

0.2%
4k
Attack framework for aligned LLMs, based on a research paper
Created 2 years ago
Updated 1 year ago
Feedback? Help us improve.