Nvidia is working on a chip for AI inferencing with Groq technology
Briefly

Nvidia is working on a chip for AI inferencing with Groq technology
"Groq's Language Processing Units (LPUs) are built on a completely new architecture that performs inferencing with significantly less energy consumption. The company has not yet announced exactly how Nvidia will integrate the technology. The GTC 2026 conference starts on March 16 in San Jose, so we expect more clarity then."
"In December, the company signed a $20 billion licensing deal with Groq, also hiring founder Jonathan Ross and president Sunny Madra. When it comes to inferencing, there are numerous options. AWS and Google Cloud, for example, have chips for this in the public cloud, while many startups are trying to provide a cheaper and more efficient alternative for inferencing."
"It is noteworthy that OpenAI is an early customer, as it had been looking for faster alternatives to Nvidia's GPUs for some time due to dissatisfaction with the inferencing speed for specific tasks, including software development. Last month, it signed a deal with Cerebras for inferencing chips."
Nvidia is expanding beyond GPU training dominance into AI inference processing by acquiring Groq's Language Processing Unit technology through a $20 billion licensing deal. The new chip aims to provide energy-efficient alternatives for running AI services daily, addressing competition from AWS, Google Cloud, and various startups in the inference market. OpenAI has received early access to the processor and signed separate deals with both Nvidia and Cerebras for inference solutions. Nvidia hired Groq's founder Jonathan Ross and president Sunny Madra to lead this initiative. The company plans to unveil details about the chip integration at GTC 2026 in March, as it seeks to establish market leadership in inference processing.
Read at Techzine Global
Unable to calculate read time
[
|
]