Awesome
InstructGraph
This repository is implemented for our paper InstructGraph: Boosting Large Language Models via Graph-centric Instruction Tuning and Preference Alignment
.
🆕 [24-05-16] Our paper has been accepted to the Findings of ACL 2024.
What's InstructGraph?
InstructGraph is a framework for empowering large language models (LLMs) on graph-centric tasks via graph instruction tuning and preference alignment. We collect 29 standard graph datasets and decompose them into four groups, including graph structure modeling, graph language modeling, graph generation modeling, and graph thought modeling.
To better bridge the gap between textual LLMs with the graph data, we introduce a structured format verbalizer, which aims to transform the graph data into a code-like format. This interface can enable the LLM to reuse the ability of code understanding and generation. In addition, the LLM can generate a graph by outputting a code-like sequence.
We also explore four hallucination problems in graph reasoning and generation, respectively. We use direct preference optimization (DPO) to perform preference alignment.
More details can be found in our paper.
<!-- ### Released Resource Download -->Quick Start
Download the open-resource llama2-7b to a folder, e.g., "./pre-trained-lm/Llama-2-7b-hf".
We release the instruction corpus at: HuggingFace.
Step1: Perform graph instruction tuning by llama2-7b with lora:
bash examples/instruction_tuning/run_llama2_flashattn.sh
You can obtain a resulting folder in "./output/" with two files, i.e., "adapter_config.json" and "adapter_model.bin".
Step2: Perform graph preference alignment by llama2-7b with lora:
You must first set the argument "--peft_model" as the folder of instruction tuning checkpoint, and then:
bash examples/preference_tuning/run_llama2_flashattn.sh
Step3: Perform inference on graph instruction tasks:
bash examples/inference/run_llama2.sh
Step4: perform inference on preference task:
bash examples/inference/run_llama2_for_preference.sh
Step5: Calculate metrics on graph instruction tasks, e.g., "graph-language-modeling-graph-question-answering-pathquestion":
python3 examples/inference/calculate_metrics.py \
--model_name_or_path meta-llama/Llama-2-7b-hf \
--inference_save_dir output/instruction_tuning/fsdp_peft_flash_1500k/llama2-peft-2epoch/predictions \
--is_graph_instruction \
--inference_task graph-language-modeling-graph-question-answering-pathquestion
Step5: Calculate metrics on graph preference tasks.
python3 examples/inference/calculate_preference_metrics.py \
--inference_save_dir output/preference_tuning/llama2/instructgraph_hallucination_predictions \
--inference_task all
Demo Play
Please see in the jupyter file instruction.ipynb.
<!-- ### Acknowledgement -->Citation
@article{Wang2024InstructGraph,
author = {Jianing Wang and
Junda Wu and
Yupeng Wu and
Yao Liu and
Ming Gao and
Julian McAuley},
title = {InstructGraph: Boosting Large Language Models via Graph-centric Instruction Tuning and Preference Alignment},
eprinttype = {arXiv},
eprint = {2402.08785},
}