Alibaba debuts Qwen3.5, a 397B-parameter open-weight multimodal AI model that it says is 60% cheaper to use and 8x better at large workloads than Qwen3
Alibaba debuts Qwen3.5, a 397B-parameter open-weight multimodal AI model that it says is 60% cheaper to use and 8x better at large workloads than Qwen3
MiniMax releases M2.5, claiming the model delivers on the “intelligence too cheap to meter” promise, priced at $0.30/1M input tokens and $1.20/1M output tokens
Today we're introducing our latest model, MiniMax-M2.5. — Extensively trained with reinforcement learning …
MiniMax releases M2.5, claiming the model delivers on the “intelligence too cheap to meter” promise, priced at $0.30/1M input tokens and $1.20/1M output tokens
Today we're introducing our latest model, MiniMax-M2.5. — Extensively trained with reinforcement learning …
Source: Chinese AI startup Zhipu anonymously released its new AI model GLM-5 on OpenRouter under the name Pony Alpha; Zhipu plans to debut GLM-5 later this week
Zhipu, one of China's prominent AI developers, has anonymously released its new large language model under a different name on OpenRouter …
Alibaba and Tencent are releasing new models and spending millions on “red envelope” freebies to woo users ahead of the Lunar New Year
HONG KONG — China's biggest AI companies are releasing new models and handing out “red envelope” …
An analysis of 100T+ tokens from the past year shows reasoning models now represent over half of all usage, open-weight model use has grown steadily, and more
this is not a model I hear much about. [image] @openrouterai : We collaborated with @a16z to publish the **State of AI** - an empirical report on how LLMs have been used on OpenRou...
An analysis of 100T+ tokens from the past year shows reasoning models now represent over half of all usage, open-weight model use has grown steadily, and more
this is not a model I hear much about. [image] @openrouterai : We collaborated with @a16z to publish the **State of AI** - an empirical report on how LLMs have been used on OpenRou...
An analysis of 100T+ tokens from the past year shows reasoning models now represent over half of all usage, open-weight model use has grown steadily, and more
this is not a model I hear much about. [image] @openrouterai : We collaborated with @a16z to publish the **State of AI** - an empirical report on how LLMs have been used on OpenRou...
An analysis of 100T+ tokens from the past year shows reasoning models now represent over half of all usage, open-weight model use has grown steadily, and more
this is not a model I hear much about. [image] @openrouterai : We collaborated with @a16z to publish the **State of AI** - an empirical report on how LLMs have been used on OpenRou...
An analysis of 100T+ tokens from the past year shows reasoning models now represent over half of all usage, open-weight model use has grown steadily, and more
this is not a model I hear much about. [image] @openrouterai : We collaborated with @a16z to publish the **State of AI** - an empirical report on how LLMs have been used on OpenRou...
An analysis of 100T+ tokens from the past year shows reasoning models now represent over half of all usage, open-weight model use has grown steadily, and more
An Empirical 100 Trillion Token Study with OpenRouter — Malika Aubakirova*Alex Atallah†Chris Clark†Justin Summerville†Anjney Midha*
An analysis of 100T+ tokens from the past year shows reasoning models now represent over half of all usage, open-weight model use has grown steadily, and more
An Empirical 100 Trillion Token Study with OpenRouter — Malika Aubakirova*Alex Atallah†Chris Clark†Justin Summerville†Anjney Midha*
An analysis of 100T+ tokens from the past year shows reasoning models now represent over half of all usage, open-weight model use has grown steadily, and more
An Empirical 100 Trillion Token Study with OpenRouter — Malika Aubakirova*Alex Atallah†Chris Clark†Justin Summerville†Anjney Midha*
An analysis of 100T+ tokens from the past year shows reasoning models now represent over half of all usage, open-weight model use has grown steadily, and more
An Empirical 100 Trillion Token Study with OpenRouter — Malika Aubakirova*Alex Atallah†Chris Clark†Justin Summerville†Anjney Midha*
An analysis of 100T+ tokens from the past year shows reasoning models now represent over half of all usage, open-weight model use has grown steadily, and more
An Empirical 100 Trillion Token Study with OpenRouter — Malika Aubakirova*Alex Atallah†Chris Clark†Justin Summerville†Anjney Midha*
Alibaba debuts Qwen3-Max-Preview, its largest AI model with over 1T parameters, showcasing strong benchmark performance; the model is not open source
Chinese e-commerce giant Alibaba's “Qwen Team” of AI researchers has done it again. After a busy summer in which the AI lab released …
xAI launches Grok Code Fast 1, a “speedy and economical” agentic coding model, available for free for a limited time on GitHub Copilot, Cursor, Windsurf, more
Elon Musk's artificial intelligence startup, xAI, on Thursday released a new “speedy and economical” agentic coding model …
Z.ai, formerly known as Zhipu and that has raised $1.5B from Tencent and others, releases GLM-4.5, an open-source AI model that it says is cheaper than DeepSeek
chinese models really are taking over huh Simon Willison / @simonwillison.net : Pretty decent pelicans from the new GLM-4.5 and GLM-4.5 Air models. Both models are MIT licensed, r...
Moonshot's Kimi K2 uses a 1T-parameter MoE architecture with 32B active parameters and outperforms models like GPT-4.1 and DeepSeek-V3 on key benchmarks
Moonshot AI, the Chinese artificial intelligence startup behind the popular Kimi chatbot, released an open-source language model on Friday …