Any Plans for an Instruct Model?

#15
by Ashacorporation - opened

This is a very capable reasoning model. I only did some light fine-tuning, but the performance improvement for my use case has been significant. In my experience, it feels comparable to GPT Oss 20B (medium setting). Really impressive work.

By the way, is there an instruct version available? If it performs this “magically” ini in instruct mode, it could potentially become a strong alternative to GPT- 4.1 nano or Gemini 2.5 Flash-Lite.

I wouldn't expect an Instruct version to perform so well, as the extended test computation time (spending a lot of thinking tokens) is very likely the secret sauce behind the model's performance

Nanbeige LLM Lab org

Thanks for the feedback — really glad it’s working well for you!

Yes, Nanbeige4.2 will include an instruct version. We’re also working on making the model smart without excessive thinking tokens.

you got us hooked now , cant wait for the release of the 4.2 version . could you please provide any ETA or approximations about when it MIGHT release ?

Thanks for the feedback — really glad it’s working well for you!

Yes, Nanbeige4.2 will include an instruct version. We’re also working on making the model smart without excessive thinking tokens.

It is a really exciting model! I am honestly surprised by how capable it is for such a small model. It really is comparable to MUCH larger models and if you get the insane amount of overthinking under control, it could be a true challenger for edge applications. Really great job! Congrats to you and your team.

Thanks for the feedback — really glad it’s working well for you!

Yes, Nanbeige4.2 will include an instruct version. We’re also working on making the model smart without excessive thinking tokens.

Have you tried Chain of Draft? https://arxiv.org/abs/2502.18600

This came out a while ago; apparently these models don't actually need full reasoning chains to improve their performance, but for some reason it fell out of favor.

Sign up or log in to comment