Graph instruction tuning research paper for LLMs
Top 46.4% on sourcepulse
GraphGPT is a framework for aligning Large Language Models (LLMs) with graph structural knowledge through a dual-stage graph instruction tuning paradigm. It targets researchers and practitioners working with graph-structured data who want to leverage LLMs for tasks like node classification and link prediction, enabling LLMs to understand and reason about graph properties.
How It Works
GraphGPT employs a text-graph grounding paradigm to encode graph structures into the LLM's natural language space. This involves a graph transformer that is pre-trained to align textual descriptions with graph structures. The core innovation is a dual-stage instruction tuning process: first, self-supervised tuning on general graph instructions, followed by task-specific tuning (e.g., node classification, link prediction) using Chain-of-Thought (CoT) distillation for improved reasoning.
Quick Start & Requirements
pip install -r requirements.txt
.torch_geometric
and related packages, fschat
for Vicuna base model.all_graph_data.pt
).graphgpt_stage1.sh
, graphgpt_stage2.sh
) orchestrate the two-stage tuning process.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
1 year ago
1 week