Get all your news in one place.
100’s of premium titles.
One app.
Start reading
Fortune
Fortune
Sharon Goldman

Amazon is making its own AI power play with new Nova models

Amazon's Rohit Prasad sitting in a chair on stage at a tech conference. (Credit: Rita Franca—NurPhoto via Getty Images)

After nearly two years of speculation about when—or even if—Amazon would launch its own competitive family of AI models to challenge OpenAI, Google, and Anthropic, the company finally delivered a mic drop moment at its AWS re:Invent conference. Amazon announced a new family of AI models for text, images and video, called Amazon Nova, that it claimed have “state-of-the-art intelligence” across many tasks, are 75% cheaper than top-performing industry competitors.

The question on many Amazon-watchers minds, however, was: Why? After all, Amazon has invested $8 billion into Anthropic, including an additional $4 billion announced just last week. As part of the deal, Anthropic has also committed to using Amazon’s AWS cloud and Amazon’s custom AI computer chip, called Trainium, to train and run its models. 

Amazon wants control over its own AI destiny

But Amazon clearly has no intention of relying solely on external partners when it comes to its AI strategy. It does not want to be beholden to Anthropic or any other third party. It wants to drive the cost of its AI offerings to cloud customers way down—a low-cost strategy that Amazon's AWS cloud computing service has long-pursued. Doing so would be harder if it was only using Anthropic's models. Finally, the company says its customers wanted capabilities, such as video, that Anthropic does not currently offer.

The Nova releases are part of a larger, perhaps master plan to blaze its own path towards AI dominance—which also includes building what it says will be the world's largest AI supercomputer, called Project Rainier, which will include hundreds of thousands of Amazon’s Trainium chips working in unison. 

In a keynote speech yesterday at re:Invent, Amazon CEO Andy Jassy’s cited three lessons about how Amazon is developing its AI strategy. One, he said, is that as organizations scale their generative AI applications, “the cost of compute really matters”—so a cheaper, commoditized model becomes increasingly important. 

Next, Jassy said AI success won’t simply come from a capable model, but one that tackles issues like latency—the time it takes to get the information or results you need from a model—as well as user experience and infrastructure issues. In addition, he insisted that Amazon wants to give customers, both internal and external, a diversity of models to choose from. 

“We have a lot of our internal builders using [Anthropic’s] Claude,” he said. “But, they’re also using Llama models. They’re also using Mistral models, and they’re also using some of our own models, and they’re also using home-grown models themselves.” 

This was surprising at first, he said, but added that “we keep learning the same lesson over and over and over again, which is that there is never going to be one tool to rule the world.” 

This appears to be the crux of Amazon’s master AI plan: Ben Thompson, a business and tech analyst and author of Stratechery, wrote yesterday that Jassy’s comments about AWS’s AI strategy “looks a lot like the AWS strategy generally.” Just as AWS offers plenty of choice in processing or databases, AWS will offer AI model choice on its Bedrock service. And that will include Amazon’s own Nova models, which just so happen to be the likely-cheapest option for third-party developers. Amazon is betting that AI will become a commodity, he said: “AWS’s bet is that AI will be important enough that it won’t, in the end, be special at all, which is very much Amazon’s sweet spot.” 

Amazon needs to strike a balance between cost and performance

However, some argue that lower-cost models aren't the key to building reliable AI applications. Instead, they maintain that performance takes precedence over cost when it comes to creating effective, high-quality solutions. There's a question whether Amazon's Nova models, which it claims are "as good or better" than rival AI software across many, but not all, benchmark tests, will be seen as good enough by developers to persuade them to make the jump.

This might be the balance Amazon is aiming for, but it's unclear whether it has struck the right one. I spoke yesterday to Rohit Prasad, Amazon’s SVP and head scientist for AGI (artificial general intelligence), who told me that the name Nova was purposeful—signalling a new and very different generation of AI models of “exceptional quality.” 

When I asked why Amazon did not turn to Anthropic to build new models for it, Prasad pointed to Amazon’s own “urgent” internal customer needs such as generating videos—something he said Anthropic does not offer, to the best of his knowledge. “We have our Prime Video team that wants to recap seasons, and they can do it with the video understanding capability of the model,” he said. “Amazon ads needed models that can generate videos.” 

Prasad would not comment on Amazon’s long-term roadmap when it comes to its AI models, but he did say that there will be “more paradigm changes to come,” including more capable models. In the meantime, he said that the Nova models are available to all internal Amazon teams, including the one the working on a new generative Amazon's Alexa digital assistant (which, as I reported back in June, has been a long, less-than-successful slog). 

Amazon wants to give its customers choice among models from a variety of different vendors, he emphasized. The question is whether the same strategy that worked so well for Amazon’s AWS in the past—offering low cost, product choice, and flexibility—will pay off again in this new era of AI?

Sign up to read this article
Read news from 100’s of titles, curated specifically for you.
Already a member? Sign in here
Related Stories
Top stories on inkl right now
One subscription that gives you access to news from hundreds of sites
Already a member? Sign in here
Our Picks
Fourteen days free
Download the app
One app. One membership.
100+ trusted global sources.