Which reasoning parser to use?

#33
by g-a-b-y - opened

The docs indicate we should be running the model with --reasoning-parser minimax_m2_append_think. This works, but causes the <think> tags to be added in the prompt and get displayed in some clients.

A recent issue on vLLM's GitHub says to use the deepseek_r1. Issue here https://github.com/vllm-project/vllm/issues/34625

Can someone from the MiniMax team provide some guidance on this? Maybe there's a bug in the minimax_m2_append_think parser.

From what I saw, the minimax_m2_append_think parser works with models that only send the end token and assume that everything is before this token is reasoning.

Minimax M2.5 send both start token and end token, and nothing is sent through reasoning when using the minimax_m2_append_think parser, that is weird, but using the deepseek_r1 works well as it handles models that send both start and end tokens.

Maybe update the documentation as it is useless to have multiple parsers on vLLM that do the same thing

Sign up or log in to comment