
 From the huggingface repo: QwenLong-L1: Towards Long-Context Large Reasoning Models with Reinforcement Learning We release 🤗 QwenLong-L1-32B, which is the first long-context LRM trained with reinforcement learning for long-context reasoning. Experiments on seven long-context DocQA benchmarks demonstrate that QwenLong-L1-32B outperforms flagship LRMs like OpenAI-o3-mini and Qwen3-235B-A22B, achieving performance on par with Claude-3.7-Sonnet-Thinking, demonstrating leading performance among state-of-the-art LRMs. We have it on Blablador. I don’t have yet a perfect system prompt for it, so you might see some funny tokens there, like <|im_end|> and so on. In any case, it would be nice to see if it’s useful for you, and then I would keep it running. Please let me know! Let’s bark! Alex
participants (1)
-
Strube, Alexandre