Broadcom and OpenAI are building a custom chip for ChatGPT
Briefly

Broadcom and OpenAI are building a custom chip for ChatGPT
"10 gigawatts of power will be used to run ChatGPT on a custom piece of silicon. The chip, the result of a collaboration between Broadcom and OpenAI, is intended to reduce the latter's dependence on Nvidia. It's raining mega deals in the AI landscape. In a somewhat curious sequence of events, Nvidia appears to be investing billions in OpenAI, which in turn is opening its wallet for AMD. The company behind ChatGPT wants to diversify."
"However, inferencing, the name for AI workloads in which LLMs run daily, is considerably easier to run than training. For AI training, OpenAI also appears to be completely dependent on other people's processors in the near future. OpenAI CEO Sam Altman talks about a "broader ecosystem" in which the new AI chip fits. The accelerator is expected to see the light of day at the end of 2026."
"OpenAI is in a unique position. As AI model builders, the company's experts know better than anyone where the limitations of current chip technology most hinder AI. By building a processor that (unlike Nvidia GPUs) excels in only one respect, the end product can be designed precisely to OpenAI's requirements. How big the actual expansion for AI will be remains to be seen."
OpenAI and Broadcom are collaborating on a custom silicon accelerator intended primarily for inferencing workloads to run ChatGPT, with projected power demand up to 10 gigawatts. The chip aims to reduce OpenAI's dependence on Nvidia GPUs and is part of a broader system Broadcom built around AI inferencing. Inferencing is easier to run than training, and OpenAI expects to remain dependent on third-party processors for training. The accelerator targets availability by the end of 2026. OpenAI plans to leverage internal model expertise to design a processor optimized for specific chip limitations, while major infrastructure deals and purchases will influence overall AI expansion.
Read at Techzine Global
Unable to calculate read time
[
|
]