|
|
--- |
|
|
license: apache-2.0 |
|
|
datasets: |
|
|
- gexai/inquisitiveqg |
|
|
language: |
|
|
- en |
|
|
metrics: |
|
|
- accuracy |
|
|
base_model: |
|
|
- distilbert/distilbert-base-uncased |
|
|
pipeline_tag: text-classification |
|
|
--- |
|
|
|
|
|
## Model Details |
|
|
Text classification model for ambiguity in questions. Classifies questions as ambiguous or clear. |
|
|
Based on distilbert/distilbert-base-uncased. |
|
|
|
|
|
**Example:** |
|
|
|
|
|
"Did he do it?" {'label': 'AMBIG', 'score': 0.9029870629310608} |
|
|
|
|
|
"Did Peter win the game?" {'label': 'CLEAR', 'score': 0.8900136351585388} |
|
|
|
|
|
## Out-of-Scope Use |
|
|
The model was only trained to classify single questions. Other kinds of data are not tested. |
|
|
|
|
|
### Training Data |
|
|
I manually labeled a small part of the inquisitiveqg dataset mixed with a private dataset to train the model to recognize ambiguity in questions. A satisfactory model with 85.5% accuracy was created. |
|
|
|
|
|
#### Metrics |
|
|
"eval_accuracy": 0.8551401869158879, |
|
|
|
|
|
"eval_loss": 0.3658725619316101, |