metadata
license: apache-2.0
datasets:
- tomg-group-umd/wikipedia-en-2k-samples
- BASF-AI/WikipediaEasy10Classification
language:
- en
SLiNeP
Super Low Parameter Wikipedia-based Neural Predictor Super small gpt 2 based model trained from 11.6 Mb of wikipedia data.
Each topic starts with Topic Name : Topic Text and ends with <|endmsg|>.
This is the medium version repo.
Since the model is small and trained on minimal data
it doesnt know many facts, but it learned some english.
Its recommended to finetune it.
I also didn't test it yet but it probably needs low temp to stop saying nonsensical things