Learn languages naturally with fresh, real content!

tap to translate recording

Explore By Region

flag Nvidia launches Groq 3 LPU racks for faster, more efficient AI inference, shipping late 2026.

flag Nvidia has launched the Groq 3 Language Processing Unit (LPU) and associated LPX server racks, integrating Groq’s technology into its Vera Rubin platform to boost AI inference speed and efficiency. flag The system, featuring 256 LPUs per rack, delivers up to 1,500 tokens per second with 35 times higher throughput per watt, targeting trillion-parameter models and agentic AI workloads. flag Designed to complement Nvidia’s Rubin GPUs and Vera CPUs, the platform aims to reduce latency and power use while enabling higher revenue per million tokens. flag The Groq 3 LPX racks are expected to ship in late 2026, with Nvidia also introducing the open-source Dynamo 1.0 software platform to streamline large-scale AI inference. flag The move marks a strategic shift toward specialized inference hardware amid growing competition and rising demand from hyperscalers and AI service providers.

55 Articles