Only end </think> tag but no start <think> tag.

#5
by zhangziji1021 - opened

I used vllm to deploy Qwen/Qwen3-4B-Thinking-2507 model, and used langchain framework to develop an agent with workflow. However, whether I use simple chat flow or the workflow, the returned responses always have no start tag but end tag.

Why is that?

If the response message is too long and exceeds Max_Context_Length, the beginning of the message may be truncated and output.

or maybe this

NOTE: This model supports only thinking mode. Meanwhile, specifying enable_thinking=True is no longer required.
Additionally, to enforce model thinking, the default chat template automatically includes . Therefore, it is normal for the model's output to contain only without an explicit opening tag.

how can i prevent this part from printing? i need only the output

Sign up or log in to comment