You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
You can use this repository to evaluate the models. To reproduce the models, use SKGInstruct in your preferred finetuning framework. The checkpoitns are being released on Huggingface.
The processed test data is already provided, but the prompts used for training and testing can be found in /prompts
Easy reproduction can be done with the Llama-Factory.
Follow the data preparation steps on their repo to add one of the StructLM datasets from huggingface
use the parameters in the bash script StructLM_finetune.yaml, as a reference replacing the parametres in block quotes [] with your paths. Then start the training like
llamafactory-cli train StructLM_finetuning.yaml, as such
Evaluate StructLM-7B
Install Requirements
Requirements:
Python 3.10
Linux
support for CUDA 11.8
pip install -r requirements.txt
Download files
./download.sh
this will download
The raw data required for executing evaluation
The processed test data splits ready for evaluation
@misc{zhuang2024structlm,
title={StructLM: Towards Building Generalist Models for Structured Knowledge Grounding},
author={Alex Zhuang and Ge Zhang and Tianyu Zheng and Xinrun Du and Junjie Wang and Weiming Ren and Stephen W. Huang and Jie Fu and Xiang Yue and Wenhu Chen},
year={2024},
eprint={2402.16671},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
About
Code and data for "StructLM: Towards Building Generalist Models for Structured Knowledge Grounding" (COLM 2024)