inferencerlabs commited on
Commit
93201cc
·
verified ·
1 Parent(s): 9da1ce7

Upload complete model

Browse files
Files changed (1) hide show
  1. README.md +2 -1
README.md CHANGED
@@ -24,7 +24,8 @@ tags:
24
 
25
  * Tested on a MacBook Pro connecting to a M3 Ultra 512GB RAM over the internet using [Inferencer app v1.5.4](https://inferencer.com)
26
  * Memory usage: ~175 GB
27
- * Expect 36 tokens/s for small contexts (200 tokens) down to 11 token/s for large (6800 tokens)
 
28
  * Quantized with a modified version of [MLX](https://github.com/ml-explore/mlx) 0.28
29
  * For more details see [demonstration video](https://youtu.be/DCVKP_o2HU0) or visit [MiniMax-M2](https://huggingface.co/MiniMaxAI/MiniMax-M2).
30
 
 
24
 
25
  * Tested on a MacBook Pro connecting to a M3 Ultra 512GB RAM over the internet using [Inferencer app v1.5.4](https://inferencer.com)
26
  * Memory usage: ~175 GB
27
+ * Expect 42 tokens/s for small contexts (200 tokens) down to 12 token/s for large (6800 tokens)
28
+ ** Note: Performance has been improved since original tests by 16.7% see: [github.com/inferencer/issues/46](https://github.com/inferencerlabs/inferencer-feedback/issues/46)
29
  * Quantized with a modified version of [MLX](https://github.com/ml-explore/mlx) 0.28
30
  * For more details see [demonstration video](https://youtu.be/DCVKP_o2HU0) or visit [MiniMax-M2](https://huggingface.co/MiniMaxAI/MiniMax-M2).
31