Also providing MLX quants?

#6
by bibproj - opened

Hi CohereLabs

Congratulations to get 67 languages into such a 'tiny' model!

Thank you for already providing the GGUF quants of your new model. πŸ‘ Would you perhaps consider to also provide MLX quants for people with Apple Macs and iPhones?

I would suggest BF16 for Apple Macs, and 4-bit versions for iPhones.

# Install MLX
pip install -U mlx-lm

# BF16 quants for Apple Macs
mlx_lm.convert --hf-path CohereLabs/tiny-aya-global --mlx-path tiny-aya-global-mlx-bf16 --dtype bfloat16
mlx_lm.convert --hf-path CohereLabs/tiny-aya-water --mlx-path tiny-aya-water-mlx-bf16 --dtype bfloat16
mlx_lm.convert --hf-path CohereLabs/tiny-aya-earth --mlx-path tiny-aya-earth-mlx-bf16 --dtype bfloat16
mlx_lm.convert --hf-path CohereLabs/tiny-aya-fire --mlx-path tiny-aya-fire-mlx-bf16 --dtype bfloat16

# 4-bit quants for Apple iPhones
mlx_lm.convert --hf-path CohereLabs/tiny-aya-global --mlx-path tiny-aya-global-mlx-4bit -q --q-bits 4
mlx_lm.convert --hf-path CohereLabs/tiny-aya-water --mlx-path tiny-aya-water-mlx-4bit -q --q-bits 4
mlx_lm.convert --hf-path CohereLabs/tiny-aya-earth --mlx-path tiny-aya-earth-mlx-4bit -q --q-bits 4
mlx_lm.convert --hf-path CohereLabs/tiny-aya-fire --mlx-path tiny-aya-fire-mlx-4bit -q --q-bits 4

Sign up or log in to comment