Nvidia
Criticized for pushing annual GPU release cycles that accelerate depreciation pressures on hyperscalers
How media typically covers Nvidia
Based on 12 scored articles
Nvidia as author
“Published by NVIDIA: "Overcoming Compute and Memory Bottlenecks With FlashAttention-4 on NVIDIA Blackw"”
SurgWorld, a world model trained on surgical videos, can generate synthetic paired video-action data that enables surgical robot policies to significantly outperform models trained only on real demonstrations.
“Published by NVIDIA: "SurgWorld: Learning Surgical Robot Policies from Videos via World Modeling"”
Efficient-DLM converts pretrained autoregressive language models into faster diffusion language models while preserving accuracy, achieving 4.5x higher throughput on 8B compared to Dream 7B with 5.4% better accuracy.
“Published by NVIDIA: "Efficient-DLM: From Autoregressive to Diffusion Language Models, and Beyond in S"”
NVIDIA introduces Nemotron 3, an open family of efficient models using hybrid Mamba-Transformer architecture with up to 1M token context lengths, designed for agentic AI and reasoning tasks with state-of-the-art accuracy.
“Published by NVIDIA: "NVIDIA Nemotron 3: Efficient and Open Intelligence"”
cuTile Python is a new NVIDIA programming language for GPU compute that simplifies kernel development through tile-based abstractions, available via PyPI with support for Blackwell GPUs and CUDA Toolkit 13.1+.
“Author of "cuTile Python" in GitHub”
NVIDIA's ToolOrchestra enables an 8B parameter orchestrator model to outperform GPT-5 on complex reasoning tasks while achieving 2.5x better efficiency through intelligent coordination of smaller models and tools.
“Author of "ToolOrchestra: Elevating Intelligence via Efficient Model and Tool Orchestration"”
NVIDIA introduces Llama-Embed-Nemotron-8B, an open-weights text embedding model achieving state-of-the-art performance on the Multilingual Massive Text Embedding Benchmark with superior multilingual and cross-lingual capabilities.
“Published by NVIDIA: "Llama-Embed-Nemotron-8B: A Universal Text Embedding Model for Multilingual and C"”
ProfBench, a benchmark of 7,000+ expert-evaluated response pairs across Physics, Chemistry, and Finance domains, reveals that state-of-the-art LLMs like GPT-4 achieve only 65.9% performance and shows significant gaps between proprietary and open-weight models.
“Author of "ProfBench: Multi-Domain Rubrics Requiring Professional Knowledge to Answer and Judge"”
NVIDIA shipped DGX Spark, a desktop-sized AI supercomputer delivering 1 petaflop of performance with 128GB unified memory, enabling developers to locally run inference on 200B-parameter models and fine-tune 70B-parameter models.
“Published by NVIDIA: "Nvidia Ships DGX Spark, World's Smallest Supercomputer"”
Front-loading reasoning data during pretraining yields 19% average performance gains and establishes foundational capabilities that cannot be fully replicated by post-training alone, with pretraining benefiting from diverse reasoning patterns while SFT is more sensitive to data quality.
“Published by NVIDIA: "Front-Loading Reasoning: The Synergy between Pretraining and Post-Training Data"”
NVIDIA and Intel announced a multi-generational collaboration to jointly develop custom data center and PC products using NVIDIA NVLink and Intel x86 CPUs, with NVIDIA investing $5 billion in Intel stock.
“Published by NVIDIA: "Nvidia Teams Up With Intel"”