Skip to content

Code for paper NarrowBERT: Accelerating Masked Language Model Pretraining and Inference

Notifications You must be signed in to change notification settings

lihaoxin2020/narrowbert

Repository files navigation

NarrowBERT

This reporsitory provides an implementation of the methods described in NarrowBERT: Accelerating Masked Language Model Pretraining and Inference.

Pretrained and MNLI finetuned models could be downloaded here.

Dependencies

This implementation is mainly based on Huggingface Transformers with the optimization package DeepSpeed. Use ./requirements.txt to make sure these packages are installed.

We provide a DeepSpeed example configuration ./ds_zero2_1gpu.json, but feel free to use your own.

Implementation

Configuration and model implementation can be found within ./narrowbert. The code is mainly adapted from BERT provided on Huggingface.

We provide training scripts for all tasks we mentioned in the paper, but you can also take the model and train it with your own scripts.

We provide models for MLM pretraining, sequence classification, and token classification, which cover the experiments mentioned in the paper.

For the tokenizer, we reuse the BERT tokenizer provided on Huggingface. In all of our experiments, we used the pretrained tokenizer from bert-base-uncased.

Pretraining

./run_mlm_narrowbert.py is the script for pretraining, and is adapted from Huggingface exampla run_mlm.py. You can run it with command

python ./run_mlm_narrowbert.py ./run_mlm_config.json

where ./run_mlm_config.json contains the hyperparameters that were used.

GLUE/IMDB/Amazon Tests

We adapt the Huggingface example run_glue.py and provide ./run_glue_narrowbert.py with the corresponding configuration. To run the script:

python ./run_glue_narrowbert.py [config]

where you can replace [config] with ./config_glue.json or ./config_imdb.json. Again, you can modify hyperparameters or choose different tasks of GLUE using these json files.

Note that Amazon2/5 requires some data preprocessing. We provide the script we used amazon.py. To run preprocessing:

python ./amazon.py [cache_path] [amazon2_save_path] [amazon5_save_path]

NER Tests

We use ./run_ner_narrowbert.pt which is adapted from the Huggingface example run_ner.py. To run it:

python ./run_ner_narrowbert.py ./config_ner.json

About

Code for paper NarrowBERT: Accelerating Masked Language Model Pretraining and Inference

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages