You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
From Tencent and Nanyang Technological University.
🔆 Introduction
🤗🤗🤗 We first create an instruction-tuning dataset based on our proposed data generation pipeline. Then, we train ChartLlama on this dataset and achieve the abilities shown in the figure.
Examples about the abilities of ChartLlama.
Redraw the chart according to the given chart, and edit the chart following instructions.
Draw a new chart based on given raw data and instructions
📝 Changelog
[2023.11.27]: 🔥🔥 Update the inference code and model weights.
[2023.11.27]: Create the git repository.
⚙️ Setup
Refer to the LLaVA-1.5.
Since I have uploaded the code, you can just install by
pip install -e .
💫 Inference
You need to first install LLaVA-1.5, then use model_vqa_lora to do inference. The model_path is the path to our Lora checkpoints, the question-file is the json file containing all questions, the image-folder is the folder containing all your images and the answers-file is the output file name.
Create and open source a new chart dataset in Chinese.
Open source the training scripts and the dataset.
Open source the evaluation scripts.
Open source the evaluation dataset.
Open source the inference script.
Open source the model.
Create the git repository.
😉 Citation
@misc{han2023chartllama,
title={ChartLlama: A Multimodal LLM for Chart Understanding and Generation},
author={Yucheng Han and Chi Zhang and Xin Chen and Xu Yang and Zhibin Wang and Gang Yu and Bin Fu and Hanwang Zhang},
year={2023},
eprint={2311.16483},
archivePrefix={arXiv},
primaryClass={cs.CV}
}
📢 Disclaimer
We develop this repository for RESEARCH purposes, so it can only be used for personal/research/non-commercial purposes.