This is an ONNX-optimized version of cardiffnlp/tweet-topic-21-multi for efficient inference.

Model Description

This repository contains:

  • model.onnx: Regular ONNX exported model
  • model_quantized.onnx: INT8 dynamically quantized model for faster inference with minimal accuracy loss

The model is optimized for zero-shot classification tasks across multiple languages.

Acknowledgments

This ONNX optimization was created for efficient deployment in production environments. Special thanks to the original model authors and the Hugging Face Optimum team.

Downloads last month
475
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for richardr1126/tweet-topic-21-multi-ONNX

Quantized
(2)
this model