Amazon to buy one million Nvidia chips, focusing on inference and Groq

Nvidia’s newly released Groq 3 LPX servers are already in demand. (Picture: Amazon)
Nvidia Executive Ian Buck confirms to Reuters that the company will sell the chips to Amazon starting this year and closing in 2027.

The main focus on the deal is on inference workloads, the process of completing tasks and answers from an AI query — which is growing at pace with AI’s general expansion.

— Inference is hard. ⁠It’s wickedly hard, Buck told Reuters. — To be the best at inference, it is not a one chip pony. We actually ​use all seven chips.

Amazon is betting on a broad mix of chips, Reuters reports, and says in their press release that they are buying Blackwell and Vera Rubin chips.

From what Reuters understands, they will also be buying a number of the newly released Groq 3 LPX servers — which are optimized for inference and can do 700 million tokens per second.

Read more: Reuters report, Amazon press release.

Nvidia will sell $1 trillion of its AI chips by 2027, launches inference rack

The Groq 3 LPU has only 500 MB of memory, but it’s SRAM flying at 150 TB/s. (Picture: Nvidia)
The Blackwell and Rubin series of chips are selling like hotcakes, the Nvidia CEO says at the Games Developer Conference, as he doubles the previous guidance of $500 billion in sales and justifies a market valuation topping $4 trillion.

Huang’s most interesting offering at the show was the new Groq 3 LPX, a custom rack made for inference loads.

Continue reading “Nvidia will sell $1 trillion of its AI chips by 2027, launches inference rack”