c794b711ed | ||
---|---|---|
.. | ||
README.md | ||
cola.sh | ||
config.json | ||
download_data.sh | ||
mnli.sh | ||
mrpc.sh | ||
patch.diff | ||
qnli.sh | ||
qqp.sh | ||
rte.sh | ||
sst2.sh | ||
stsb.sh |
README.md
GLUE fine-tuning task
To run the experiment, you need to
run ./mnli.sh
for fine-tuning mnli base model,
run ./mnli.sh
for fine-tuning mnli large model.
run ./cola.sh
for fine-tuning cola large model.
run ./sst2.sh
for fine-tuning sst2 large model.
run ./stsb.sh
for fine-tuning stsb large model.
run ./rte.sh
for fine-tuning rte large model.
run ./qqp.sh
for fine-tuning qqp large model.
run ./qnli.sh
for fine-tuning qnli large model.
run ./mrpc.sh
for fine-tuning mrpc large model.
Export model to ONNX format and quantization
To export model to onnx format during evaluation, use argument --export_ort_model True
.
To export quantized model, use --fp16 False --export_ort_model True
.
The exported model will be under output folder, and end with
<prefix>__onnx_fp16.bin
if fp16 is True, otherwise the outputs will be <prefix>__onnx_fp32.bin
and <prefix>__onnx_qt.bin
.
Please check ONNX document for more details.