6c1d151e57475ef1356b0af5656749635f4baa5322f9cdfc77cc252567d845c3.png


From the huggingface repo:


QwenLong-L1: Towards Long-Context Large Reasoning Models with Reinforcement Learning


We release 🤗 QwenLong-L1-32B, which is the first long-context LRM trained with reinforcement learning for long-context reasoning. Experiments on seven long-context DocQA benchmarks demonstrate that QwenLong-L1-32B outperforms flagship LRMs like OpenAI-o3-mini and Qwen3-235B-A22B, achieving performance on par with Claude-3.7-Sonnet-Thinking, demonstrating leading performance among state-of-the-art LRMs.

We have it on Blablador. I don’t have yet a perfect system prompt for it, so you might see some funny tokens there, like <|im_end|> and so on.

In any case, it would be nice to see if it’s useful for you, and then I would keep it running. Please let me know!


Let’s bark!

Alex