These scripts perform vocabulary pruning on the classification model (XLMRobertaForSequenceClassification
) and evaluate the performance.
We use a subset of XNLI English training set as the vocabulary file.
Download the fine-tuned model or train your own model on PAWS-X dataset, and save the files to ../models/xlmr_pawsx
.
Download link: * Google Drive * Hugging Face Models
- Pruning with the textpruner-CLI tool:
bash vocabulary_pruning.sh
- Pruning with the python script:
VOCABULARY_FILE=../datasets/xnli/en.tsv
MODEL_PATH=../models/xlmr_pawsx
python vocabulary_pruning.py $MODEL_PATH $VOCABULARY_FILE
- Evaluate the model:
Set $PRUNED_MODEL_PATH
to the directory where the pruned model is stored.
python measure_performance.py $PRUNED_MODEL_PATH
This script prunes the pre-trained models for MLM with a vocabulary limited to the SST-2 training set.
Set $MODEL_PATH
to the directory where the pre-trained model (BERT, RoBERTa, etc.) is stored.
python MaskedLM_vocabulary_pruning.py $MODEL_PATH