Chinese AI startup Z.ai releases GLM-4.7, an open-weight model that Z.ai says delivers significant improvements in coding performance compared to GLM-4.6
like 210 — Z.ai 6.24k — Text Generation Transformers Safetensors English Chinese glm4_moe conversational eWeek : Chinese AI Startup Z.ai Takes On OpenAI Via Cheaper Prices Vincent Chow / South Chi...
Alibaba debuts the Qwen3-Coder model for agentic coding, including a 480B-parameter MoE variant, and open sources Qwen Code, a CLI tool adapted from Gemini CLI
Qwen 39.4k — Text Generation Transformers Safetensors qwen3_moe conversational Coco Feng / South China Morning Post : Alibaba upgrades flagship Qwen3 model to outperform OpenAI, DeepSeek in maths, c...
Alibaba releases open-source reasoning model QwQ-32B on Hugging Face and ModelScope, claiming comparable performance to DeepSeek-R1 but with lower compute needs
Introduction QwQ is the reasoning model of the Qwen series. Paul Barker / InfoWorld : Alibaba says its new AI model rivals DeepSeeks's R-1, OpenAI's o1 Jose Antonio Lanz / Decrypt : Alibaba's Latest A...
Meta announces Llama 3.3 70B, a text-only model that Meta claims can deliver the performance of its largest Llama model at a lower cost
7.0M — 2,040 … The fine-tuning data includes publicly available … Markus Kasanmascheff / WinBuzzer : Meta Unveils New Llama 3.3 70B AI Model with Higher Cost-Efficiency Carl Franzen / VentureBeat : ...
Meta releases Code Llama 70B, a new version of its code generation model, featuring improved code correctness, a variant optimized for Python, and more
available under the same license as previous Code Llama models. Download the models ➡️ https://ai.meta.com/... • CodeLlama-70B • CodeLlama-70B-Python • CodeLlama-70B-Instruct [image] Awni Hannun / @aw...