How did you export the tokenizer to ONNX?
#1
by
Franck-Dernoncourt
- opened
Could you please share how you exported the tokenizer to ONNX? Thanks!
Hi @Franck-Dernoncourt - apologies for the slow reply. https://github.com/onnx/tensorflow-onnx can do it when exporting the model using an extra contrib opset that includes the operations needed for the tokenizer (SentencePiece I think from memory)
E.g. python -m tf2onnx.convert --saved-model /content/use --output /content/use_large_v5.onnx --opset 12 --extra_opset ai.onnx.contrib:1 --tag serve (where /content/use is the model downloaded from TFHub, then uncompressed)