Export model with torch.jit.script()
In this section, we describe how to export a model via
torch.jit.script()
.
When to use it
If we want to use our trained model with torchscript,
we can use torch.jit.script()
.
Hint
See Export model with torch.jit.trace()
if you want to use torch.jit.trace()
.
How to export
We use https://github.com/k2-fsa/icefall/tree/master/egs/librispeech/ASR/pruned_transducer_stateless3 as an example in the following.
cd egs/librispeech/ASR
epoch=14
avg=1
./pruned_transducer_stateless3/export.py \
--exp-dir ./pruned_transducer_stateless3/exp \
--tokens data/lang_bpe_500/tokens.txt \
--epoch $epoch \
--avg $avg \
--jit 1
It will generate a file cpu_jit.pt
in pruned_transducer_stateless3/exp
.
Caution
Don’t be confused by cpu
in cpu_jit.pt
. We move all parameters
to CPU before saving it into a pt
file; that’s why we use cpu
in the filename.
How to use the exported model
Please refer to the following pages for usage:
https://k2-fsa.github.io/sherpa/python/streaming_asr/emformer/index.html
https://k2-fsa.github.io/sherpa/python/streaming_asr/conv_emformer/index.html
https://k2-fsa.github.io/sherpa/python/streaming_asr/conformer/index.html
https://k2-fsa.github.io/sherpa/python/offline_asr/conformer/index.html
https://k2-fsa.github.io/sherpa/cpp/offline_asr/gigaspeech.html
https://k2-fsa.github.io/sherpa/cpp/offline_asr/wenetspeech.html