https://github.com/huggingface/transformers
# following has been validated with transformers v4.18
# 24 Layers
# https://huggingface.co/bert-large-uncased-whole-word-masking-finetuned-squad
model=bert-large-uncased-whole-word-masking-finetuned-squad
# 12 Layers
# https://huggingface.co/vuiseng9/bert-base-uncased-squad
model=vuiseng9/bert-base-uncased-squad
# 6 Layers
# https://huggingface.co/distilbert-base-cased-distilled-squad
model=distilbert-base-cased-distilled-squad
batch_size=64
cd transformers/examples/pytorch/question-answering/
python run_qa.py \
--model_name_or_path $model \
--dataset_name squad \
--do_eval \
--per_device_eval_batch_size $batch_size \
--max_seq_length 384 \
--doc_stride 128 \
--output_dir /tmp/eval-$model \
--overwrite_output_dir
plot of events per QPS