Amazon to buy one million Nvidia chips, focusing on inference and Groq

Nvidia’s newly released Groq 3 LPX servers are already in demand. (Picture: Amazon)
Nvidia Executive Ian Buck confirms to Reuters that the company will sell the chips to Amazon starting this year and closing in 2027.

The main focus on the deal is on inference workloads, the process of completing tasks and answers from an AI query — which is growing at pace with AI’s general expansion.

— Inference is hard. ⁠It’s wickedly hard, Buck told Reuters. — To be the best at inference, it is not a one chip pony. We actually ​use all seven chips.

Amazon is betting on a broad mix of chips, Reuters reports, and says in their press release that they are buying Blackwell and Vera Rubin chips.

From what Reuters understands, they will also be buying a number of the newly released Groq 3 LPX servers — which are optimized for inference and can do 700 million tokens per second.

Read more: Reuters report, Amazon press release.

Nvidia will sell $1 trillion of its AI chips by 2027, launches inference rack

The Groq 3 LPU has only 500 MB of memory, but it’s SRAM flying at 150 TB/s. (Picture: Nvidia)
The Blackwell and Rubin series of chips are selling like hotcakes, the Nvidia CEO says at the Games Developer Conference, as he doubles the previous guidance of $500 billion in sales and justifies a market valuation topping $4 trillion.

Huang’s most interesting offering at the show was the new Groq 3 LPX, a custom rack made for inference loads.

Continue reading “Nvidia will sell $1 trillion of its AI chips by 2027, launches inference rack”

AI lab Thinking Machines gets investment, gigawatt compute from Nvidia

With no product and an experienced team, Murati’s Thinking Machines lab is rounding up funding. (Picture: Nvidia)
Founded by OpenAI’s former Chief Technology Officer after the 2024 leadership spat, Mira Murati’s Thinking Machines has scored a big deal with Nvidia.

The startup has entered into «a multiyear strategic partnership» that will provide them with both money and significant Vera Rubin compute early next year — about the same level the first version of Grok was trained on.

The parties are not disclosing a sum total, but 1 GW of Nvidia compute is estimated to be worth about $50 billion, Reuters notes.

Murati’s AI lab has been largely secretive about its actual products, releasing a configurable API in December 2025 and vowing to make AI models more accessible, capable and, yes, customizable.

They raised $2 billion at a $12 billion valuation from Andreessen Horowitz and Nvidia in July 2025.

Read more: Joint press release, writeups on Reuters, CNBC and TechCrunch.

Jensen Huang says Nvidia’s investment opportunity in AI labs is closing

Huang figures the privately owned AI labs era might be finished. (Picture: Nvidia)
The Nvidia CEO says the opportunity to invest might soon end, Reuters reports.

The reason for this is straightforward, suspecting that Anthropic and OpenAI going public «later this year» will shutter the window to private equity deals.

The latest deal to fund OpenAI with $30 billion «might be the last time» to «invest in a consequential company like this,» Huang admits.

Nvidia has invested some $130 billion in OpenAI in two rounds, the recent straight up investment, and one circular deal where they paid $100 billion in return for OpenAI buying $100 billion in chips from them.

Likewise, Nvidia was an investor in a November funding round for Anthropic, buying $15 billion in shares from the company.

Read more: Reuters, CNBC and TechCrunch.

OpenAI closes record funding round: $110 billion invested at $840B valuation

At more than double the cost of last years record deal, the funding highly values OpenAI by investors, according to Reuters.

Amazon invested $50 billion, Nvidia put up $30 billion and SoftBank shelled out $30 billion, agreeing to a $840 billion valuation, the largest of any frontier AI lab by far.

«Strategic partnership» with Amazon
Amazon’s deal structure is slightly different, as it comes in the terms of a strategic partnership where OpenAI will receive $15 billion up front, and then qualify for the rest $35 billion «over the coming months.»

OpenAI has committed to using 2 gigawatts of capacity on AWS’ Trainium platform and will make their models available on Amazon’s services.

None of this is said to change OpenAI’s relationship with Microsoft, OpenAI says in a release.

At the same time, OpenAI’s Nick Turley says they have surpassed 900 million weekly users and has 50 million paying subscribers, up from roughly 800 million before.

Read more: Sama’s thank you thread, writeups on Reuters, CNBC

Nvidia’s Data Center unit up 75% YoY, Q4 profits beat estimates yet again

Markets have become accustomed to roaring earnings beats from Nvidia. (Picture: Nvidia)
Markets were lackluster on the last quarterly report of $68.13 billion in revenue for the AI chipmaker, as revenue growth seems to be slipping, Reuters reports.

The full year revenue hit $215.9 billion, up 65% year-on-year, with Data Center revenue hitting a record of $62.3 billion — which is responsible for their AI chips.

Nvidia also raised its guidance for Q1 2026, and is certainly not seeing any slowdown:

Continue reading “Nvidia’s Data Center unit up 75% YoY, Q4 profits beat estimates yet again”

Nvidia strikes «multi-year strategic partnership» with Meta for AI chips

Likely costing a significant measure of Meta’s capital expenditures, the deal is expected to be in tens of billions dollars or more.
Both Meta and Nvidia are announcing a long-term, multi-generational strategic partnership today — without mentioning the price.

Meta, already a top customer for Nvidia, will use their chips in a «large-scale deployment» to build out data centers «optimized for AI training and inference,» they say.

The cost of the deal will likely run into the tens of billions of dollars or more, CNBC reckons, and includes access to future chips as well as the current Blackwell and Vera Rubin generations.

— We do expect a good portion of Meta’s capex to go toward this Nvidia build-out, chip analyst Ben Bajarin of Creative Strategies tells CNBC.

Reuters notes that Meta is likely one of the top three customers accounting for more than half of Nvidia’s sales.

Read more: Meta announcement, Nvidia announcement. Writeups on CNBC, Reuters and The Verge.

OpenAI closing in on $100 billion funding round at $830B valuation

The big guns are all out for OpenAI’s latest funding round. (Picture: generated)
In what looks like one of the strongest funding rounds in history, OpenAI is getting investments from SoftBank and half of the Magnificent Seven.

SoftBank and Nvidia will be the largest investors, clocking in at $30 billion each, while Amazon will pitch in «potentially» $20 billion and Microsoft will contribute «less than» $10 billion, according to Reuters and The Information.

Apparently, Amazon’s investment could come with a caveat that OpenAI expands its cloud server rental with the company, which will likely not be a large hitch.

This will also be SoftBank’s second investment in OpenAI, after recently completing a $41 billion investment, and selling out Nvidia.

That would bring their holdings to $71 billion, which is still short of Microsoft’s reported stake of $135 billion.

Read more: Reuters, and The Information, summarized by Reuters.

Reuters sources: Nvidia H200 not allowed in China

The H200 was getting popular in China, being miles ahead on performance. (Picture: generated)
Several sources are telling Reuters that the H200 chips are not permitted to enter, and authorities have told technology execs explicitly to not purchase the chips.

The H200 was cleared by Commerce for export to China in December and got finally approved this week.

They are much more powerful than anything on the Chinese market, and Nvidia received so many orders, they ran out.

Though the sources from Reuters say officials have not given any reason for this ban, it is likely in order to protect their domestic chip industry.

They do say that chips can be bought «when necessary» or for research and development programs with Chinese universities.

Read the full scoop at Reuters.

Nvidia releases the Vera Rubin platform: three and a half times faster

Nvidia’s new system will drastically reduce training time. (Picture: Nvidia)
The Rubin platform is that much faster on training, and also five times quicker on inference tasks. It wasn’t expected until later this year, writes The Verge.

The system actually consists of six chips, including a CPU, a GPU, an NVlink chip, a NIC, and a DPU and an optics chip.

It can train a «mixture of experts» model with 10x less inference token cost, and a 4x reduction in GPUs compared to the Blackwell platform, Nvidia says.

The platform is now rolling out to nearly every cloud provider, including to Nvidia partners Anthropic, OpenAI and Amazon, according to TechCrunch.

Jensen Huang estimates that AI companies will be spending between $3 and $4 trillion on infrastructure over the next five years.

Read more: Nvidia’s press release, Nvidia’s CES keynote. Writeups on The Verge, TechCrunch.

Nvidia struggles with huge Chinese orders for the H200, asks TSMC for help

The H200 chip is a huge leap forward for Chinese infrastructure firms, compared to local capacity. (Picture: Generated)
Citing five sources familiar with the matter, Reuters reports a boom in Chinese chip orders from Nvidia, which far outstrips supply.

Nvidia currently sits on some 600K high performance H200 chips, recently cleared for China, but Chinese companies have placed orders for a whopping 2 million of them for 2026.

This has led Nvidia to re-approach TSMC for another production run, Reuters reports.

This is notwithstanding regulatory pressure from the Chinese government, who have not said if they will allow the chips in the country.

They are instead considering bundling H200 purchases with domestically produced chips, Reuters says, in order to boost their internal industry.

Read the scoop at Reuters.

Nvidia gets all clear from Trump Commerce to sell H200 chip in China

Nvidia gets approval to ship powerful H200 chips to China.
Nvidia’s more powerful H200 chips can now be sold in China. (Picture: Nvidia)
The powerful chip can be sold to vetted partners with a tax of 25% to the U.S. government, Trump said in a statement.

Nvidia’s H200 is a much more advanced chip than the custom H20 that the company was allowed to export to China earlier, with the newer Blackwell chips being only about 1.5 times faster, Reuters notes.

The H20 chips were recently banned in China, where the authorities instead opted for Huawei chips for data center supplies.

Continue reading “Nvidia gets all clear from Trump Commerce to sell H200 chip in China”

SoftBank sold $5.83 billion of its Nvidia stock in October: Here’s why

People were thinking SoftBank had seen a bubble forming. Far from it, according to its CEO.
The massive stock sale fueled a lot of rumors and worries. (Picture: generated)
Known for its massive stakes in emerging technologies, SoftBank’s disclosure of the sale caused a 2% dip in Nvidia’s share price and inflated talk of an AI bubble.

Now CEO Masayoshi Son has revealed that it certainly wasn’t because they were taking profit before any downturn. On the contrary:

— I was crying to sell Nvidia shares, he told the FII Priority Asia forum in Tokyo, and says he «didn’t want to sell a single share,» adding that «I respect Jensen, I respect Nvidia so much,» according to Business Insider.

Continue reading “SoftBank sold $5.83 billion of its Nvidia stock in October: Here’s why”

Nvidia responds to report that Meta might use Google’s TPU chips

A deal between Meta and Google might see custom TPU chips installed on Meta's data centers, as the industry looks for Nvidia alternatives.
Talks between Meta and Google on custom chips has Nvidia slightly worried, but not much. (Picture: generated)
News on the talks was posted by The Information on tuesday, and caused Nvidia to drop 3% and Google to tease a $4 billion valuation in the markets.

Under the deal being discussed, Meta would start renting compute on Google’s Tensor Processing Units as early as next year, Reuters reports.

Continue reading “Nvidia responds to report that Meta might use Google’s TPU chips”

Anthropic partners with Nvidia, Microsoft in $30 billion Azure deal

Anthropic nets a Microsoft investment, and promises to use it all on Microsoft's Azure.
Nvidia and Anthropic will optimize for each other, and use Microsoft’s cloud. (Picture: generated)
The companies will invest some $15 billion in Anthropic, while the AI lab commits to $30B in spending on Azure, Microsoft’s cloud offering.

— We’re increasingly going to be customers of each other. We will use Anthropic models, they will use our infrastructure and we’ll go to market together, Microsoft CEO Satya Nadella said, according to Reuters.

The agreement means Anthropic will optimize their software stack to better run on Nvidia’s hardware, while Nvidia will «optimize for Anthropic workloads.»

Continue reading “Anthropic partners with Nvidia, Microsoft in $30 billion Azure deal”