This is an ONNX-optimized version of cardiffnlp/tweet-topic-21-multi for efficient inference.
Model Description
This repository contains:
- model.onnx: Regular ONNX exported model
- model_quantized.onnx: INT8 dynamically quantized model for faster inference with minimal accuracy loss
The model is optimized for zero-shot classification tasks across multiple languages.
Acknowledgments
This ONNX optimization was created for efficient deployment in production environments. Special thanks to the original model authors and the Hugging Face Optimum team.
- Downloads last month
- 475
Model tree for richardr1126/tweet-topic-21-multi-ONNX
Base model
cardiffnlp/tweet-topic-21-multi