You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository was archived by the owner on Jul 22, 2024. It is now read-only.
Code to create pre-training data for a span selection pre-training task inspired by reading comprehension and an effort to avoid encoding general knowledge in the transformer network itself.
Code to create pre-training data for a span selection pre-training task inspired by reading comprehension and an effort to avoid encoding general knowledge in the transformer network itself.
Pre-trained Models
Available through Hugging Face as:
michaelrglass/bert-base-uncased-sspt
michaelrglass/bert-large-uncased-sspt
Load with: AutoConfig.from_pretrained , AutoTokenizer.from_pretrained , AutoModelForQuestionAnswering.from_pretrained.
See run_qa.py for example code.
Installation
python setup.py
build irsimple.jar (or use pre-built com.ibm.research.ai.irsimple/irsimple.jar)
FIXME: rc_data and span_selection_pretraining require a modified version of pytorch-transformers
The adaptations needed are in the process of being worked into this repo and a pull request for pytorch-transformers.
Hopefully it is relatively clear how it should work.
Code to create pre-training data for a span selection pre-training task inspired by reading comprehension and an effort to avoid encoding general knowledge in the transformer network itself.