Why the ‘one AI model to rule them all’ myth needs to die
[Photo: NicoElNino/Adobe Stock]

Why the ‘one AI model to rule them all’ myth needs to die

Welcome to AI Decoded, Fast Company’s weekly newsletter that breaks down the most important news in the world of AI. I’m Mark Sullivan, a senior writer at Fast Company, covering emerging tech, AI, and tech policy.

This week, I’m focusing on the growing consensus that huge foundation models alone won’t deliver AGI. I also look at Amazon’s new family of models, and offer a glimpse at our second annual AI 20 series.

Sign up to receive this newsletter every week via email here. And if you have comments on this issue and/or ideas for future ones, drop me a line at sullivan@fastcompany.com, and follow me on X (formerly Twitter) @thesullivan


The myth of scaling to reach AGI is wearing thin

The first two years of the AI boom were all about monolithic large language models (LLMs) such as those OpenAI used to power ChatGPT. But the trick that produced the big intelligence gains in those models—scaling up their size and giving them more compute power—is now yielding diminishing returns. It turns out that models can learn only so much from pretraining on huge swaths of internet content. Yes, models continue getting smarter, but experts say that the intelligence gains are now coming mainly from human intelligence, such as human-labeled training data or human feedback on outputs.

It appears, then, that single-frontier models are not the path to artificial general intelligence (AGI), in which AI is smarter than humans across a wide variety of tasks. Rather, as Google DeepMind CEO Demis Hassabis suggested to me last year, AGI might be achieved using a system of AI models, in which LLMs would work in concert with other kinds of AI models—some of them generative, some not.

Large language models are built on top of neural networks, large math equations that take their inspiration from the workings of the neurons in the human brain. Many thinkers in the industry believe that AI systems must borrow a lot more from the human brain in order to reach higher levels of synthetic intelligence. (See: Numenta’s Thousand Brains Project.) 

Andrew Filev, founder and CEO of Zencoder, points out that the human brain is a far more diverse system than the LLMs we have today. “It’s not a unified system. . . . There is some specialization in the brain,” Filev says. “What the hippocampus does is very different than what the prefrontal cortex does. It’s a complex system with multiple feedback loops, multiple different levels with which it operates, multiple independent agents working in the background . . . so it’s only logical that AI is going to end up being similarly complex.” 

Click here to read more about the push to achieve AGI through larger LLM systems.


Amazon debuts a new family of models, emphasizing “choice”

Amazon was not initially one of the star players in the generative AI boom. On Tuesday, it showed signs of catching up. The company announced a new family of AI models called “Nova” that perform at near-state-of-the-art levels compared to models from OpenAI, Anthropic, and Google. Amazon AGI VP Vishal Sharma tells Fast Company that his company has had its foot on the gas over the past year to push the performance of Amazon’s AI models, and has in fact been using the Nova models to run aspects of Amazon’s business. 

Sharma wouldn’t divulge the size of the Nova models in parameter count, but the Nova Premier model, the largest, is likely in the hundreds of billions of parameters. That model can be used for more advanced AI tasks that require reasoning and agentic capabilities. The Nova line also includes smaller models (Lite and Micro) that are faster and less expensive to operate, as well as a new image generator (Canvas) and video generator (Reel).

Amazon operates differently than pure AI labs like OpenAI and Anthropic (though it is a major investor in Anthropic). Its first business is selling Amazon Web Services cloud data storage and app hosting. But AI is increasingly powering apps used by AWS customers, so AWS needs to provide a broad selection of models, including bleeding-edge frontier AI models. AWS has said that selling AI services is already a “multibillion-dollar” business. 

Click here to read more about Amazon’s new “Nova” AI models.


A look at our AI 20 series

We’re proud to announce the launch of our second annual AI 20 series, which spotlights the most interesting technologists, entrepreneurs, corporate leaders, and creative thinkers shaping the world of artificial intelligence. Among those profiled as part of the series is AI pioneer Noam Shazeer, who returned to Google this year to help guide the company’s Gemini models toward artificial general intelligence. 

Article content
Noam Shazeer

When I spoke to Shazeer for the profile, he expressed confidence that Google is in a strong position to emerge as the leader in large frontier models. After all, he pointed out, the company developed and owns much of the technology behind the LLM revolution.

Click here to read more about Fast Company’s AI 20 series.


More AI coverage from Fast Company:


Want exclusive reporting and trend analysis on technology, business innovation, future of work, and design? Sign up for Fast Company Premium.

sohrab ghafari

Electrical Inspector | MEng in power electricity

8mo

👌 👌 👌 👌

Like
Reply
Astrid van Dorst

I help leaders grow their business with #CRM #Automation & #AI | CEO CloudAnalysts | Customer Rating: Excellent

8mo

Insightful breakdown: the shift away fr/ monolithic AI models toward diverse, specialised systems mirrors the complexity of the human brain.

Like
Reply

To view or add a comment, sign in

Others also viewed

Explore topics