How DeepSeek's new AI model upended industry assumptions about the price of building leading-edge AI models, the U.K. will consider remedies to address cloud competition involving AWS and Microsoft, and the latest funding rounds in enterprise tech.
Today on Product Saturday: Jetbrains has a take on the inevitable collision of AI and IDEs, Qdrant's vector search performance gets a GPU upgrade, and the quote of the week.
Today: OpenAI unveils its take on AI agents that promise to take all the drudgery out of using a computer, more on the massive Project Stargate circus, and this week's enterprise moves.
Today: why tech came to a standstill Monday afternoon to watch Nvidia unveil its latest AI chip, Microsoft puts another outsider at the helm of its AI strategy, and the latest funding rounds in enterprise tech.
Welcome to Runtime! Today: why tech came to a standstill Monday afternoon to watch Nvidia unveil its latest AI chip, Microsoft puts another outsider at the helm of its AI strategy, and the latest funding rounds in enterprise tech.
(Was this email forwarded to you?Sign up here to get Runtime each week.)
Let it B
It's been a year to forget for the regular occupants of San Jose's SAP Center, with the Sharks all but guaranteed to finish the season as the worst team in professional hockey. Nvidia CEO Jensen Huang's performance in that arena Monday showcased a company on the opposite end of that spectrum.
Nvidia unveiled its Blackwell GPU architecture Mondaybefore 11,000 people gathered to watch Huang show off the company's latest creation. Blackwell is the next generation of Nvidia's Hopper architecture, which produced the H100 and H200 GPUs that powered the generative AI boom.
The most powerful iteration of the Blackwell line, the B200, will be available from cloud providers and server vendors later this year.
According to Tom's Hardware, it will offer 20 petaflops of computing performance, compared to 4 petaflops provided by the H100.
And Nvidia continued its push into the CPU side of the data center with Grace Blackwell, a "superchip" consisting of two B200s and its Grace CPU.
The Hopper architecture had an extraordinary run, with demand far outpacing supply for most of 2023. While the frenzied part of the rise of generative AI might be settling down as business realities creep in, there will still be enormous demand for Blackwell chips in the short term.
Nvidia did not disclose pricing, but Huang told CNBC that Blackwell GPUs will cost between $30,000 and $40,000 per chip, roughly in line with the Hopper generation.
The Blackwell architecture actually consists of two chips squashed together in a single package, however, so it's not clear exactly what he meant by those prices.
The B100 version of the new Blackwell design is compatible with Nvidia's existing Hopper systems but B200 customers will have to upgrade to new server designs.
That's because the B200 demands more power than the H200, although Nvidia promised that overall, servers built around Blackwell chips will be far cheaper to run and use less power than servers built around the Hopper generation.
It could take up to a year before the Blackwell GPUs start to make an impact on AI computing, but they could represent a milestone in the advancement of generative AI.
In between grand, sweeping promises about its potential to change the world, generative AI vendors readily acknowledge that there are still a lot of issues to be ironed out, most notably (but not limited to) the hallucination problem.
A big part of the overall pitch centers around the promise that these models will continue to get better over time as companies like Nvidia ramp up their silicon capabilities and model builders learn how to take advantage of that increased horsepower.
When Blackwell arrives, those companies will be under pressure to show significant improvements in the performance and reliability of their models given how many billions of dollars have already been invested in these efforts.
And as they continue to rely heavily on Nvidia (and TSMC's) ability to deliver sufficient supplies of those chips, the small market for Nvidia alternatives will continue to simmer.
Like its unusual arrangement with Sam Altman and OpenAI, the creation of Microsoft's new business unit appears to have been designed to avoid the antitrust scrutiny that would have come along with an outright acquisition of Inflection. The facilitator was likely Reid Hoffman, a member of Microsoft's board of directors and a co-founder of Inflection who will continue to serve on its board, the company said.
So Microsoft has now placed enormous trust in two men who were each fired from previous roles over their difficult management styles to shape the future of its AI development. Startups seeking investment in the future might be skeptical about taking money from Microsoft, given that it just hollowed out one of its portfolio companies less than six months after trying to do the same thing with OpenAI.
Enterprise funding
Together AI raised $106 million in Series C funding from Salesforce Ventures and existing investors to build enterprise-class features for its AI cloud services platform.
Unstructured landed $40 million in Series B funding to help companies make sense of corporate data that is, well, unstructured.
WarpStream raised $20 million in new funding to build out a streaming data platform the company claims is easier to use than Apache Kafka.
Euno launched with $6.25 million in seed funding to help customers manage business logic around their data pipelines.
Broadcom floated plans to let smaller VMware cloud services providers continue to manage SaaS versions of VMware products after cutting them off earlier this year, according to The Register.
Databricks acquired Lilac, a startup working on a tool for data scientists dealing with unstructured data, for an undisclosed amount.
Tom Krazit has covered the technology industry for over 20 years, focused on enterprise technology during the rise of cloud computing over the last ten years at Gigaom, Structure and Protocol.
How DeepSeek's new AI model upended industry assumptions about the price of building leading-edge AI models, the U.K. will consider remedies to address cloud competition involving AWS and Microsoft, and the latest funding rounds in enterprise tech.
Today on Product Saturday: Jetbrains has a take on the inevitable collision of AI and IDEs, Qdrant's vector search performance gets a GPU upgrade, and the quote of the week.
Today: OpenAI unveils its take on AI agents that promise to take all the drudgery out of using a computer, more on the massive Project Stargate circus, and this week's enterprise moves.
Today: OpenAI reveals its plan to build $500 billion worth of data centers in the U.S., why Oracle's decision to keep TikTok running could backfire down the road, and the latest funding rounds in enterprise tech.