Upload complete model
Browse files
README.md
CHANGED
|
@@ -24,7 +24,8 @@ tags:
|
|
| 24 |
|
| 25 |
* Tested on a MacBook Pro connecting to a M3 Ultra 512GB RAM over the internet using [Inferencer app v1.5.4](https://inferencer.com)
|
| 26 |
* Memory usage: ~175 GB
|
| 27 |
-
* Expect
|
|
|
|
| 28 |
* Quantized with a modified version of [MLX](https://github.com/ml-explore/mlx) 0.28
|
| 29 |
* For more details see [demonstration video](https://youtu.be/DCVKP_o2HU0) or visit [MiniMax-M2](https://huggingface.co/MiniMaxAI/MiniMax-M2).
|
| 30 |
|
|
|
|
| 24 |
|
| 25 |
* Tested on a MacBook Pro connecting to a M3 Ultra 512GB RAM over the internet using [Inferencer app v1.5.4](https://inferencer.com)
|
| 26 |
* Memory usage: ~175 GB
|
| 27 |
+
* Expect 42 tokens/s for small contexts (200 tokens) down to 12 token/s for large (6800 tokens)
|
| 28 |
+
** Note: Performance has been improved since original tests by 16.7% see: [github.com/inferencer/issues/46](https://github.com/inferencerlabs/inferencer-feedback/issues/46)
|
| 29 |
* Quantized with a modified version of [MLX](https://github.com/ml-explore/mlx) 0.28
|
| 30 |
* For more details see [demonstration video](https://youtu.be/DCVKP_o2HU0) or visit [MiniMax-M2](https://huggingface.co/MiniMaxAI/MiniMax-M2).
|
| 31 |
|