Alibaba debuts Qwen3.5, a 397B-parameter open-weight multimodal AI model that it says is 60% cheaper to use and 8x better at large workloads than Qwen3
Alibaba debuts Qwen3.5, a 397B-parameter open-weight multimodal AI model that it says is 60% cheaper to use and 8x better at large workloads than Qwen3
Ai2 launches Open Coding Agents, starting with SERA, an open-source family that includes 32B and 8B parameter models designed to adapt to private codebases
Ai2 launches Open Coding Agents, starting with SERA, an open-source family that includes 32B and 8B parameter models designed to adapt to private codebases
Artificial intelligence is moving swiftly, changing how developers craft, as code flows ever faster into repositories such as GitHub …
Nvidia launches Nemotron 3, a family of AI models using a hybrid mixture-of-experts architecture and the Mamba-Transformer design, in 30B, 100B, and ~500B sizes
Nvidia launched the new version of its frontier models, Nemotron 3, by leaning in on a model architecture that the world's …
SemiAnalysis launches InferenceMAX, an open-source benchmark that automatically tracks LLM inference performance across AI models and frameworks every night
vendor-neutral suite runs nightly and tracks performance changes over time Tae Kim / Barron's Online : Nvidia Touts Software Advantage in Beating Rivals Like AMD Dion Harris / NVID...
SemiAnalysis launches InferenceMAX, an open-source benchmark that automatically tracks LLM inference performance across AI models and frameworks every night
vendor-neutral suite runs nightly and tracks performance changes over time Tae Kim / Barron's Online : Nvidia Touts Software Advantage in Beating Rivals Like AMD Dion Harris / NVID...
Nvidia launches open-source transcription model Parakeet-TDT-0.6B-V2, topping the Hugging Face Open ASR Leaderboard with a word error rate of 6.05%
High accuracy and optimized performance for transcription in 25 languages Asif Razzaq / MarkTechPost : NVIDIA Open Sources Parakeet TDT 0.6B: Achieving a New Standard for Automatic...
Nvidia updates the DGX Station, begins taking reservations for the DGX Spark box, formerly Project Digits, and unveils the RTX PRO workstation and server GPUs
GTC After a Hopper hiatus, Nvidia's DGX Station returns, now armed with an all-new desktop-tuned Grace-Blackwell Ultra Superchip capable …
A look at Nvidia's GTC 2025 announcements, including a focus on addressing pre-training and post-training scaling and inference time scaling working in tandem
The Reasoning Token Explosion — AI model progress has accelerated tremendously, and in the last six months, models have improved more than in the previous six months.
Sources: Tencent, Alibaba, ByteDance, and other Chinese companies are ramping up orders for Nvidia's H20 AI chip due to booming demand for DeepSeek's models
Chinese companies are ramping up orders for Nvidia's (NVDA.O) H20 artificial intelligence chip due to booming demand …
Nvidia unveils Fugatto, an AI model for generating music and audio that can also modify voices, trained on open-source data, and weighs whether to release it
an impressive new AI sound model from Nvidia Mandy Dalugdug / Music Business Worldwide : Nvidia unveils AI audio generator ‘Fugatto’ that can produce ‘sounds never heard before’ Su...
Meta says its Llama models were downloaded almost 350M times, are used by AT&T and others, and usage via cloud providers more than doubled from May to July 2024
we just published a bunch of updates on the adoption we're seeing. And yes, we have a lot more work to do on dev tools and resources which we're bringing online as quickly as we c...
Microsoft announces the general availability of its Phi-3 models, including Phi-3-Silica, a 3.3B parameter model that will be embedded on all Copilot+ PCs
here's what you can use it for Pradeep Viswav / MSPoweruser : Microsoft and Khan Academy announce AI partnership Kevin Okemwa / Windows Central : Microsoft ships Azure AI Studio in...
Google adds Gemma variants CodeGemma, for code completion and generation tasks, and RecurrentGemma, to offer researchers faster inference at higher batch sizes
In February we announced Gemma, our family of lightweight, state-of-the-art open models built from the same research and technology used to create the Gemini models.
Nvidia Research announces Eureka, an AI agent powered by GPT-4 that autonomously writes reward algorithms to teach robots to perform complex skills like a human
Sharon Goldman / VentureBeat :
Nvidia claims TensorRT-LLM will double the H100's performance for running inference on leading LLMs when the open-source library arrives in NeMo in October
Dylan Martin / CRN :
How Nvidia researchers generated pictures of faces that appear to be real but aren't by analyzing photos of celebrities and detecting patterns
To create the final image in this set, the system generated 10 million revisions over 18 days. — The woman in the photo seems familiar. Tweets: @alexsteffen , @pnhoward , @rafael...