ChartLlama-code  by tingxueronghua

Multimodal LLM for chart interaction

Created 2 years ago
251 stars

Top 99.9% on SourcePulse

GitHubView on GitHub
Project Summary

ChartLlama: A Multimodal LLM for Chart Understanding and Generation

ChartLlama is a multimodal LLM focused on chart understanding and generation. It addresses the need for automated chart interpretation, manipulation, and creation, targeting researchers and developers in data visualization and analysis. The model offers capabilities to redraw charts from visual input, edit them via instructions, and generate new charts from raw data, aiming to streamline complex data visualization workflows.

How It Works

The core approach involves training a multimodal LLM on a custom-generated instruction-tuning dataset. This enables ChartLlama to process visual chart inputs alongside textual commands, facilitating sophisticated chart manipulation. Key functionalities include redrawing charts based on visual examples, editing charts according to specific instructions, and generating novel charts from raw data, leveraging instruction tuning for precise control.

Quick Start & Requirements

Installation is straightforward via pip install -e .. A critical prerequisite is the prior setup of LLaVA-1.5. Inference relies on LLaVA's model_vqa_lora module and requires specific command-line configurations detailed in the repository. Users should anticipate managing LLaVA-1.5 dependencies and potential CUDA requirements for GPU acceleration.

Highlighted Details

  • Multimodal LLM for chart understanding and generation.
  • Chart redrawing from visual input.
  • Instruction-based chart editing.
  • New chart generation from raw data.

Maintenance & Community

Developed by researchers from Tencent and Nanyang Technological University. Specific community channels or active maintenance indicators are not detailed in the provided README.

Licensing & Compatibility

Licensed strictly for "RESEARCH purposes" and limited to "personal/research/non-commercial purposes." This restriction prohibits commercial use and integration into proprietary systems.

Limitations & Caveats

Training scripts and the full dataset are not yet open-sourced, impacting full reproducibility and community-driven development. The project's reliance on LLaVA-1.5 may introduce version dependencies and potential integration complexities.

Health Check
Last Commit

2 years ago

Responsiveness

Inactive

Pull Requests (30d)
0
Issues (30d)
0
Star History
2 stars in the last 30 days

Explore Similar Projects

Starred by Chip Huyen Chip Huyen(Author of "AI Engineering", "Designing Machine Learning Systems"), Elvis Saravia Elvis Saravia(Founder of DAIR.AI), and
1 more.

InternGPT by OpenGVLab

0.0%
3k
Interactive demo platform for showcasing AI models
Created 2 years ago
Updated 1 year ago
Feedback? Help us improve.