Nvidia
Latest news, analysis, and insights about Nvidia.
Nvidia Releases Open-Weight Model With Learned Memory Compression That Cuts Context Costs 8x
Nvidia just dropped an open-weight 8B model with a technique that compresses key-value cache by 8x. For anyone running inference at scale—or trying to squeeze longer contexts onto consumer GPUs—this matters.
Nvidia's New China Rule: Pay First for H200 AI Chips Amid Export Uncertainty
Nvidia is demanding full upfront payment from Chinese customers for its H200 AI chips—a striking signal of how export controls are reshaping the AI hardware market. The move comes as regulatory approval from both Washington and Beijing remains in limbo.
NVIDIA Unveils Rubin Platform at CES — A Six-Chip Architecture Built for Agentic AI
NVIDIA just revealed its post-Blackwell roadmap at CES. The Rubin platform combines six purpose-built chips — including the new Vera CPU and Rubin GPU — explicitly designed for agentic AI and mixture-of-experts models. This is NVIDIA betting on where AI is headed, not where it's been.
NVIDIA's 72GB Desktop GPU Makes Running Large Language Models Locally Practical
NVIDIA just shipped a desktop GPU with 72GB of VRAM. For AI developers tired of cloud latency and API costs, this Blackwell-based workstation card finally makes running large language models locally realistic.
NVIDIA Buys Slurm Creator SchedMD, Promising Open-Source Continuity for HPC Workloads
NVIDIA has acquired SchedMD, the company behind Slurm—the workload management system running on more than half of the world's top 100 supercomputers. The move extends NVIDIA's reach deeper into the AI training stack while raising questions about infrastructure consolidation.