--- tags: - text-classification - toxicity-detection model-index: - name: toxicity-classifier results: [] --- # Toxicity Classifier ## Overview This model is a fine-tuned BERT model designed for detecting toxicity in text. It classifies input text as either "toxic" or "non-toxic" based on learned patterns from a diverse dataset of online comments and discussions. The model achieves high accuracy in identifying harmful language, making it suitable for content moderation tasks. ## Model Architecture The model is based on the BERT (Bidirectional Encoder Representations from Transformers) architecture, specifically `bert-base-uncased`. It consists of 12 transformer layers, each with 12 attention heads and a hidden size of 768. The final layer is a classification head that outputs probabilities for the two classes: non-toxic (0) and toxic (1). ## Intended Use This model is intended for use in applications requiring automated toxicity detection, such as: - Social media platforms for moderating user comments. - Online forums to flag potentially harmful content. - Customer support systems to identify abusive language in queries. It can be integrated into pipelines using the Hugging Face Transformers library. Example usage: ```python from transformers import pipeline classifier = pipeline("text-classification", model="your-username/toxicity-classifier") result = classifier("This is a harmful comment.") print(result)