A Trump administration official says DeepSeek's new model, expected next week, was trained on Nvidia Blackwell chips, in a potential US export control violation
Anthropic says DeepSeek, MiniMax, and Moonshot violated its ToS by prompting Claude a combined 16M+ times and using distillation to train their own products
The allegations mirror those of OpenAI, which told House lawmakers that DeepSeek used ‘distillation’ to improve models
Prime Intellect debuts INTELLECT-3, an RL-trained 106B parameter open source MOE model it claims outperforms larger models across math, code, science, reasoning
Today, we release INTELLECT-3, a 100B+ parameter Mixture-of-Experts model trained on our RL stack, achieving state …
OpenAI touts GPT-5's scores on math, coding, and health benchmarks: 94.6% on AIME 2025 without tools, 74.9% on SWE-bench Verified, and 46.2% on HealthBench Hard
After literally years of hype and speculation, OpenAI has officially launched a new lineup of large language models (LLMs) …
OpenAI says GPT-5 is a unified system with an efficient model for most questions, a reasoning model for harder problems, and a router that decides which to use
All You Need To Know Lakshay Kumar / Business Today : What is GPT-5? How OpenAI is upgrading your ChatGPT experience Tsveta Ermenkova / PhoneArena : You can now chat with a PhD-lev...
Anthropic and other researchers detail “subliminal learning”, where LLMs learn traits from model-generated data that is semantically unrelated to those traits
We study subliminal learning, a surprising phenomenon where language models learn traits from model-generated data that is semantically unrelated to those traits.
[Thread] An OpenAI researcher says the company's latest experimental reasoning LLM achieved gold medal-level performance on the 2025 International Math Olympiad
1/N I'm excited to share that our latest @OpenAI experimental reasoning LLM has achieved a longstanding grand challenge in AI: gold medal-level performance on the world's most pres...
In a video and a letter signed “Sam & Jony”, Altman and Ive say io, founded in 2024 by Ive, Scott Cannon, Evans Hankey, and Tang Tan will develop new products
Update! It's this: — openai.com/sam-and-jony/ Emma Jacobs / @emmavj : Your dad introducing his new partner — openai.com/sam-and-jony/ Kate Knibbs / @knibbs : but what are they...
OpenAI acquires io, Jony Ive's secretive AI startup, for nearly $6.5B in stock; Ive and LoveFrom will remain independent but take over design for all of OpenAI
OpenAI will acquire the AI device startup co-founded by Apple Inc. veteran Jony Ive in a nearly $6.5 billion all-stock deal …
LMArena says it is updating its leaderboard policies after a Llama 4 Maverick version, which Meta said in fine print is not public, secured the number two spot
With Llama 4, Meta fudged benchmarks to appear as though its new AI model is better than the competition.
DeepSeek releases DeepSeek-V3, an open-source MoE model of 671B total parameters, with 37B activated per token, claiming it outperforms top models like GPT-4o
Chinese AI startup DeepSeek, known for challenging leading AI vendors with its innovative open-source technologies, today released a new ultra-large model: DeepSeek-V3.
DeepSeek releases DeepSeek-V3, an open-source MoE model of 671B total parameters, with 37B activated per token, claiming it outperforms top models like GPT-4o
Chinese AI startup DeepSeek, known for challenging leading AI vendors with its innovative open-source technologies, today released a new ultra-large model: DeepSeek-V3.
DeepSeek releases DeepSeek-V3, an open-source MoE model of 671B total parameters, with 37B activated per token, claiming it outperforms top models like GPT-4o
Chinese AI startup DeepSeek, known for challenging leading AI vendors with its innovative open-source technologies, today released a new ultra-large model: DeepSeek-V3.