Learn languages naturally with fresh, real content!

Popular Topics
Explore By Region
Nvidia launches Groq 3 LPU racks for faster, more efficient AI inference, shipping late 2026.
Nvidia has launched the Groq 3 Language Processing Unit (LPU) and associated LPX server racks, integrating Groq’s technology into its Vera Rubin platform to boost AI inference speed and efficiency.
The system, featuring 256 LPUs per rack, delivers up to 1,500 tokens per second with 35 times higher throughput per watt, targeting trillion-parameter models and agentic AI workloads.
Designed to complement Nvidia’s Rubin GPUs and Vera CPUs, the platform aims to reduce latency and power use while enabling higher revenue per million tokens.
The Groq 3 LPX racks are expected to ship in late 2026, with Nvidia also introducing the open-source Dynamo 1.0 software platform to streamline large-scale AI inference.
The move marks a strategic shift toward specialized inference hardware amid growing competition and rising demand from hyperscalers and AI service providers.
Nvidia lanza racks Groq 3 LPU para inferencias de IA más rápidas y eficientes, que se enviarán a fines de 2026.