An analysis of 100T+ tokens from the past year shows reasoning models now represent over half of all usage, open-weight model use has grown steadily, and more
this is not a model I hear much about. [image] @openrouterai : We collaborated with @a16z to publish the **State of AI** - an empirical report on how LLMs have been used on OpenRouter. After analyzing...
Alibaba releases the Qwen3-VL vision models, the Qwen3Guard “safety moderation” models, and three closed-weight models, including Qwen3-Max with 1T+ parameters
Qwen 50.6k — Safetensors qwen3_vl_moe Julian Nabil / Forbes Middle East : Alibaba Introduces Qwen3-Max AI Model With Over 1T Parameters Markus Kasanmascheff / WinBuzzer : Alibaba Releases Qwen3-VL O...
Alibaba debuts the Qwen3-Coder model for agentic coding, including a 480B-parameter MoE variant, and open sources Qwen Code, a CLI tool adapted from Gemini CLI
Qwen 39.4k — Text Generation Transformers Safetensors qwen3_moe conversational Coco Feng / South China Morning Post : Alibaba upgrades flagship Qwen3 model to outperform OpenAI, DeepSeek in maths, c...
DeepSeek quietly open sources Prover-V2, a math-focused, 671B-parameter AI model using mixture-of-experts, on Hugging Face, one day after Alibaba released Qwen3
The Chinese start-up has released the Prover-V2 model a day after Alibaba released Qwen3, and ahead of an anticipated release of DeepSeek-R2
Alibaba debuts its Qwen3 family of open-weight “hybrid” AI reasoning models, including Qwen3-235B-A22B, with 235B total parameters and 22B activated parameters
Chinese tech company Alibaba on Monday released Qwen3, a family of AI models the company claims matches …