docs: mention tokenizer fix
Browse files
README.md
CHANGED
|
@@ -26,7 +26,7 @@ inference: false
|
|
| 26 |
|
| 27 |
## Tokenizer Fix
|
| 28 |
|
| 29 |
-
This repository is a fork of the original [
|
| 30 |
|
| 31 |
Due to not optimal performance on token classification tasks - as reported [here](https://github.com/AnswerDotAI/ModernBERT/issues/149) - the following fixes w.r.t. the tokenizer are applied:
|
| 32 |
|
|
|
|
| 26 |
|
| 27 |
## Tokenizer Fix
|
| 28 |
|
| 29 |
+
This repository is a fork of the original [Ettin 400M](https://huggingface.co/jhu-clsp/ettin-encoder-400m).
|
| 30 |
|
| 31 |
Due to not optimal performance on token classification tasks - as reported [here](https://github.com/AnswerDotAI/ModernBERT/issues/149) - the following fixes w.r.t. the tokenizer are applied:
|
| 32 |
|