What’s next for AI?

While large language models continue to advance, new models and agents are proving to be more effective at discrete tasks. AI needs different horses for different courses.

Kelly Raskovich

United States

Bill Briggs

United States

Mike Bechtel

United States

Abhijith Ravinutala

United States

Jim Rowan

United States

Nitin Mittal

United States

Lou DiLorenzo Jr.

United States

Blink and you’ll miss it: The speed of artificial intelligence’s advancement is outpacing expectations. Last year, as organizations scrambled to understand how to adopt generative AI, we cautioned Tech Trends 2024 readers to lead with need as they differentiate themselves from competitors and adopt a strategic approach to scaling their use of large language models (LLMs). Today, LLMs have taken root, with up to 70% of organizations, by some estimates, actively exploring or implementing LLM use cases.1

But leading organizations are already considering AI’s next chapter. Instead of relying on foundation models built by large players in AI, which may be more powerful and built on more data than needed, enterprises are now thinking about implementing multiple, smaller models that can be more efficient for business requirements.2 LLMs will continue to advance and be the best option for certain use cases, like general-purpose chatbots or simulations for scientific research, but the chatbot that peruses your financial data to think through missed revenue opportunities doesn’t need to be the same model that replies to customer inquiries. Put simply, we’re likely to see a proliferation of different horses for different courses.

A series of smaller models working in concert may end up serving different use cases than current LLM approaches. New open-source options and multimodal outputs (as opposed to just text) are enabling organizations to unlock entirely new offerings.3

In the years to come, the progress toward a growing number of smaller, more specialized models could once again move the goalposts of AI in the enterprise. Organizations may witness a fundamental shift in AI from augmenting knowledge to augmenting execution. Investments being made today in agentic AI, as this next era is termed, could upend the way we work and live by arming consumers and businesses with armies of silicon-based assistants. Imagine AI agents that can carry out discrete tasks, like delivering a financial report in a board meeting or applying for a grant. “There’s an app for that” could well become “There’s an agent for that.”

Now: Getting the fundamentals right

LLMs are undoubtedly exciting but require a great deal of groundwork. Instead of building models themselves, many enterprises are partnering with companies like Anthropic or OpenAI or accessing AI models through hyperscalers.4 According to Gartner®, AI servers will account for close to 60% of hyperscalers’ total server spending.5 Some enterprises have found immediate business value in using LLMs, while others have remained wary about the accuracy and applicability of LLMs trained on external data.6 On an enterprise time scale, AI advancements are still in a nascent phase (crawling or walking, as we noted last year). According to recent surveys by Deloitte and Fivetran and Vanson Bourne, in most organizations, fewer than a third of generative AI experiments have moved into production, often because organizations struggle to access or cleanse all the data needed to run AI programs.7 To achieve scale, organizations will likely need to further think through data and technology, as well as strategy, process, and talent, as outlined in a recent Deloitte AI Institute report.

According to Deloitte’s Q3 2024 State of generative AI in the enterprise report, 75% of surveyed organizations have increased their investments in data-life-cycle management due to generative AI.8 Data is foundational to LLMs, because bad inputs lead to worse outputs (in other words, garbage in, garbage squared). That’s why data-labeling costs can be a big driver of AI investment.9 While some AI companies scrape the internet to build the largest models possible, savvy enterprises create the smartest models possible, which requires better domain-specific “education” for their LLMs. For instance, LIFT Impact Partners, a Vancouver-based organization that provides resources to nonprofits, is fine-tuning its AI-enabled virtual assistants on appropriate data to help new Canadian immigrants process paperwork. “When you train it on your organization’s unique persona, data, and culture, it becomes significantly more relevant and effective,” says Bruce Dewar, president and CEO of LIFT Impact Partners. “It brings authenticity and becomes a true extension of your organization.”10

Data enablement issues are dynamic. Organizations surveyed by Deloitte said new issues could be exposed by the scale-up of AI pilots, unclear regulations around sensitive data, and questions around usage of external data (for example, licensed third-party data). That’s why 55% of organizations surveyed avoided certain AI use cases due to data-related issues, and an equal proportion are working to enhance their data security.11 Organizations could work around these issues by using out-of-the-box models offered by vendors, but differentiated AI impact will likely require differentiated enterprise data.

Thankfully, once the groundwork is laid, the benefits are clear: Two-thirds of organizations surveyed say they’re increasing investments in generative AI because they’ve seen strong value to date.12 Initial examples of real-world value are also appearing across industries, from insurance claims review to telecom troubleshooting and consumer segmentation tools.13 LLMs are also making waves in more specialized use cases, such as space repairs, nuclear modeling, and material design.14

As underlying data inputs improve and become more sustainable, LLMs and other advanced models (like simulations) may become easier to spin up and scale. But size isn’t everything. Over time, as methods for AI training and implementation proliferate, organizations are likely to pilot smaller models. Many may have data that can be more valuable than previously imagined, and putting it into action through smaller, task-oriented models can reduce time, effort, and hassle. We’re poised to move from large-scale AI projects to AI everywhere, as discussed in this year’s introduction.

New: Different horses for different courses

While LLMs have a vast array of use cases, the library is not infinite (yet). LLMs require massive resources, deal primarily with text, and are meant to augment human intelligence rather than take on and execute discrete tasks. As a result, says Vivek Mohindra, senior vice president of corporate strategy at Dell Technologies, “there is no one-size-fits-all approach to AI. There are going to be models of all sizes and purpose-built options—that’s one of our key beliefs in AI strategy.”15

Over the next 18 to 24 months, key AI vendors and enterprise users are likely to have a toolkit of models comprising increasingly sophisticated, robust LLMs along with other models more applicable to day-to-day use cases. Indeed, where LLMs are not the optimal choice, three pillars of AI are opening new avenues of value: small language models, multimodal models, and agentic AI (figure 1).

Small language models

LLM providers are racing to make AI models as efficient as possible. Instead of enabling new use cases, these efforts aim to rightsize or optimize models for existing use cases. For instance, massive models are not necessary for mundane tasks like summarizing an inspection report—a smaller model trained on similar documents would suffice and be more cost-efficient.

Small language models (SLMs) can be trained by enterprises on smaller, highly curated data sets to solve more specific problems, rather than general queries. For example, a company could train an SLM on its inventory information, enabling employees to quickly retrieve insights instead of manually parsing large data sets, a process that can sometimes take weeks. Insights from such an SLM could then be coupled with a user interface application for easy access.

Naveen Rao, vice president of AI at Databricks, believes more organizations will take this systems approach with AI: “A magic computer that understands everything is a sci-fi fantasy. Rather, in the same way we organize humans in the workplace, we should break apart our problems. Domain-specific and customized models can then address specific tasks, tools can run deterministic calculations, and databases can pull in relevant data. These AI systems deliver the solution better than any one component could do alone.”16

An added benefit of smaller models is that they can be run on-device and trained by enterprises on smaller, highly curated datasets to solve more specific problems, rather than general queries, as discussed in “Hardware is eating the world.” Companies like Microsoft and Mistral are currently working to distill such SLMs, built on fewer parameters, from their larger AI offerings, and Meta offers multiple options across smaller models and frontier models.17

Finally, much of the progress happening in SLMs is through open-source models offered by companies like Hugging Face or Arcee.AI.18 Such models are ripe for enterprise use since they can be customized for any number of needs, as long as IT teams have the internal AI talent to fine-tune them. In fact, a recent Databricks report indicates that over 75% of organizations are choosing smaller open-source models and customizing them for specific use cases.19 Since open-source models are constantly improving thanks to the contributions of a diverse programming community, the size and efficiency of these models are likely to improve at a rapid clip.

Multimodal models

Humans interact through a variety of mediums: text, body language, voice, videos, among others. Machines are now hoping to catch up.20 Given that business needs are not contained to text, it’s no surprise that companies are looking forward to AI that can take in and produce multiple mediums. In some ways, we’re already accustomed to multimodal AI, such as when we speak to digital assistants and receive text or images in return, or when we ride in cars that use a mix of computer vision and audio cues to provide driver assistance.21

Multimodal generative AI, on the other hand, is in its early stages. The first major models, Google's Project Astra and OpenAI’s GPT-4 Omni, were showcased in May 2024, and Amazon Web Services’ Titan offering has similar capabilities.22 Progress in multimodal generative AI may be slow because it requires significantly higher amounts of data, resources, and hardware.23 In addition, the existing issues of hallucination and bias that plague text-based models may be exacerbated by multimodal generation.

Still, the enterprise use cases are promising. The notion of “train once, run anywhere (or any way)” promises a model that could be trained on text, but deliver answers in pictures, video, or sound, depending on the use case and the user’s preference, which improves digital inclusion. Companies like AMD aim to use the fledgling technology to quickly translate marketing materials from English to other languages or to generate content.24 For supply chain optimization, multimodal generative AI can be trained on sensor data, maintenance logs, and warehouse images to recommend ideal stock quantities.25 This also leads to new opportunities with spatial computing, which we write about in “Spatial computing takes center stage.” As the technology progresses and model architecture becomes more efficient, we can expect to see even more use cases in the next 18 to 24 months.

Agentic AI

The third new pillar of AI may pave the way for changes to our ways of working over the next decade. Large (or small) action models go beyond the question-and-answer capabilities of LLMs and complete discrete tasks in the real world. Examples range from booking a flight based on your travel preferences to providing automated customer support that can access databases and execute needed tasks—likely without the need for highly specialized prompts.26 The proliferation of such action models, working as autonomous digital agents, heralds the beginnings of agentic AI, and enterprise software vendors like Salesforce and ServiceNow are already touting these possibilities.27

Chris Bedi, chief customer officer at ServiceNow, believes that domain- or industry-specific agentic AI can change the game for humans and machine interaction in enterprises.28 For instance, in the company’s Xanadu platform, one AI agent can scan incoming customer issues against a history of incidents to come up with a recommendation for next steps. It then communicates to another autonomous agent that’s able to execute on those recommendations, and a human in the loop reviews those agent-to-agent communications to approve the hypotheses. In the same vein, one agent might be adept at managing workloads in the cloud, while another provisions orders for customers. As Bedi says, “Agentic AI cannot completely take the place of a human, but what it can do is work alongside your teams, handling repetitive tasks, seeking out information and resources, doing work in the background 24/7, 365 days a year.”29

Finally, aside from the different categories of AI models noted above, advancements in AI design and execution can also impact enterprise adoption—namely, the advent of liquid neural networks. “Liquid” refers to the flexibility in this new form of training AI through a neural network, a machine learning algorithm that mimics the human brain’s structure. Similar to how quantum computers are freed from the binary nature of classical computing, liquid neural networks can do more with less: A couple dozen nodes in the network might suffice, versus 100,000 nodes in a more traditional network. The cutting-edge technology aims to run on less computing power, with more transparency, opening up possibilities for embedding AI into edge devices, robotics, and safety-critical systems.30 In other words, it’s not just the applications of AI but also its underlying mechanisms that are ripe for improvement and disruption in the coming years.

Next: There’s an agent for that

In the next decade, AI could be wholly focused on execution instead of human augmentation. A future employee could make a plain-language request to an AI agent, for example, “close the books for Q2 and generate a report on EBITDA.” Like in an enterprise hierarchy, the primary agent would then delegate the needed tasks to agents with discrete roles that cascade across different productivity suites to take action. As with humans, teamwork could be the missing ingredient that enables the machines to improve their capabilities.31 This leads to a few key considerations for the years to come (figure 2):

  • AI-to-AI communication. Agents will likely have a more efficient way of communicating with each other than human language, as we don’t need human-imitating chatbots talking to each other.32 Better AI-to-AI communication can enhance outcomes, as fewer people will need to become experts to benefit from AI. Rather, AI can adapt to each person’s communication style.33
  • Job displacement and creation. Some claim that roles such as prompt engineer could become obsolete.34 However, the AI expertise of those employees will remain pertinent as they focus on managing, training, and collaborating with AI agents as they do with LLMs today. For example, a lean IT team with AI experts might build the agents it needs in a sort of “AI factory” for the enterprise. The significant shift in the remaining workforce’s skills and education may ultimately reward more human skills like creativity and design, as mentioned in previous Tech Trends.
  • Privacy and security. The proliferation of agents with system access is likely to raise broad concerns about cybersecurity, which will only become more important as time progresses and more of our data is accessed by AI systems. New paradigms for risk and trust will be required to make the most out of applying AI agents.
  • Energy and resources. AI’s energy consumption is a growing concern.35 To mitigate environmental impacts, future AI development will need to balance performance with sustainability. It will need to take advantage of improvements in liquid neural networks or other efficient forms of training AI, not to mention the hardware needed to make all of this work, as we discuss in “Hardware is eating the world.” 
  • Leadership for the future. AI has transformative potential, as everyone has heard plenty over the last year, but only insofar as leadership allows. Applying AI as a faster way of doing things the way they’ve always been done will result in, at best, missed potential, and, at worst, amplified biases.36 Imaginative, courageous leaders should dare to take AI from calcified best practices to the creation of “next practices,” where we find new ways of organizing ourselves and our data toward an AI-enabled world.

When it comes to AI, enterprises will likely have the same considerations in the future that they do today: data, data, and data. Until AI systems can reach artificial general intelligence or learn as efficiently as the human brain,37 they will be hungry for more data and inputs to help them be more powerful and accurate. Steps taken today to organize, streamline, and protect enterprise data could pay dividends for years to come, as data debt could one day become the biggest portion of technical debt. Such groundwork should also help enterprises prepare for the litany of regulatory challenges and ethical uncertainties (such as data collection and use limitations, fairness concerns, lack of transparency) that come with shepherding this new, powerful technology into the future.38 The stakes of garbage in, garbage out are only going to grow: It would be much better to opt for genius in, genius squared.39

BY

Kelly Raskovich

United States

Bill Briggs

United States

Mike Bechtel

United States

Abhijith Ravinutala

United States

Endnotes

  1. Carl Franzen, “More than 70% of companies are experimenting with generative AI, but few are willing to commit more spending,” VentureBeat, July 25, 2023.

    View in Article
  2. Tom Dotan and Deepa Seetharaman, “For AI giants, smaller is sometimes better,” The Wall Street Journal, July 6, 2024.

    View in Article
  3. Google Cloud, “Multimodal AI,” accessed October 2024.

    View in Article
  4. Silvia Pellegrino, “Which companies have partnered with OpenAI?,” Tech Monitor, May 15, 2023; Maxwell Zeff, “Anthropic launches Claude Enterprise plan to compete with OpenAI,” TechCrunch, September 4, 2024; Jean Atelsek and William Fellows, “Hyperscalers stress AI credentials, optimization and developer empowerment,” S&P Global Market Intelligence, accessed October 2024.

    View in Article
  5. Gartner, “Gartner forecasts worldwide IT spending to grow 8% in 2024,” press release, April 17, 2024. GARTNER is a registered trademark and service mark of Gartner, Inc. and/or its affiliates in the U.S. and internationally and is used herein with permission. All rights reserved.

    View in Article
  6. Patricia Licatta, “Between sustainability and risk: Why CIOs are considering small language models,” CIO, August 1, 2024.

    View in Article
  7. Jim Rowan et al., “Now decides next: Moving from potential to performance,” Deloitte’s State of Generative AI in the Enterprise Q3 report, August 2024; Mark Van de Wiel, “New AI survey: Poor data quality leads to $406 million in losses,” Fivetran, March 20, 2024.

    View in Article
  8. Rowan et al., “Now decides next: Moving from potential to performance.”

    View in Article
  9. Sharon Goldman, “The hidden reason AI costs are soaring—and it’s not because Nvidia chips are more expensive,” Fortune, August 23, 2024.

    View in Article
  10. Deloitte Insights, "Lifting up the nonprofit sector through generative AI," September 23, 2024.

    View in Article
  11. Rowan et al., “Now decides next: Moving from potential to performance.”

    View in Article
  12. Ibid.

    View in Article
  13. Ibid.

    View in Article
  14. Sandra Erwin, “Booz Allen deploys advanced language model in space,” SpaceNews, August 1, 2024; Argonne National Laboratory, “Smart diagnostics: How Argonne could use Generative AI to empower nuclear plant operators,” press release, July 26, 2024; Kevin Maik Jablonka et al., “14 examples of how LLMs can transform materials science and chemistry: A reflection on a large language model hackathon,” Digital Discovery 5 (2023).

    View in Article
  15. Phone interview with Vivek Mohindra, senior vice president of corporate strategy, Dell Technologies, October 11, 2024.

    View in Article
  16. Phone interview with Naveen Rao, vice president of AI at Databricks, October 2, 2024.

    View in Article
  17. YouTube, “Introducing the next evolution of generative AI: Small language models,” Microsoft Dynamics 365, video, May 9, 2024; Llama team, “The Llama 3 herd of models,” Meta, July 23, 2024.

    View in Article
  18. Rachel Metz, “In AI, smaller, cheaper models are getting big attention,” Bloomberg, August 8, 2024.

    View in Article
  19. Databricks, “AI is in production,” accessed October 2024.

    View in Article
  20. MIT Technology Review Insights, “Multimodal: AI’s new frontier,” May 8, 2024.

    View in Article
  21. Akesh Takyar, “Multimodal models: Architecture, workflow, use cases and development,” LeewayHertz, accessed October 2024.

    View in Article
  22. NeuronsLab, “Multimodal AI use cases: The next opportunity in enterprise AI,” May 30, 2024.

    View in Article
  23. Ellen Glover, “Multimodal AI: What it is and how it works,” Built In, July 1, 2024.

    View in Article
  24. Mary E. Morrison, “At AMD, opportunities, challenges of using AI in marketing,” Deloitte’s CIO Journal for The Wall Street Journal, July 2, 2024.

    View in Article
  25. NeuronsLab, “Multimodal AI use cases: The next opportunity in enterprise AI.”

    View in Article
  26. Oguz A. Acar, “AI prompt engineering isn’t the future,” Harvard Business Review, June 6, 2023.

    View in Article
  27. Salesforce, “Agentforce,” accessed October 2024; ServiceNow, “Our biggest AI release is here,” accessed October 2024.

    View in Article
  28. Phone interview with Chris Bedi, chief customer officer at ServiceNow, September 30, 2024.

    View in Article
  29. Ibid.

    View in Article
  30. Brian Heater, “What is a liquid neural network, really?,” TechCrunch, August 17, 2023.

    View in Article
  31. Edd Gent, “How teams of AI agents working together could unlock the tech’s true power,” Singularity Hub, June 28, 2024.

    View in Article
  32. Will Knight, “The chatbots are now talking to each other,” WIRED, October 12, 2023.

    View in Article
  33. David Ellis, “The power of AI in modeling healthy communications,” Forbes, August 17, 2023.

    View in Article
  34. Acar, “AI prompt engineering isn’t the future.”

    View in Article
  35. James Vincent, “How much electricity does AI consume?,” The Verge, February 16, 2024.

    View in Article
  36. IBM, “Shedding light on AI bias with real world examples,” October 16, 2023.

    View in Article
  37. University of Oxford, “Study shows that the way the brain learns is different from the way that artificial intelligence systems learn,” January 3, 2024.

    View in Article
  38. Nestor Maslej et al., The AI Index 2024 annual report, AI Index Steering Committee, Institute for Human-Centered AI, Stanford University, Stanford, CA, April 2024.

    View in Article
  39. Deloitte, Work Re-Architected video series, accessed October 2024.

    View in Article

Acknowledgments

The authors would like to thank the Office of the CTO Market-Making team, without whom this report would not be possible: Caroline Brown, Ed Burns, MacKenzie Hackathorn, Stefanie Heng, Bri Henley, Dana Kublin, Haley Gove Lamb, Kiran Makhijani, Sangeet Mohanty, Heidi Morrow, Sarah Mortier, Abria Perry, Abhijith Ravinutala, and Bella Stash.

Much gratitude goes to the many subject matter leaders across Deloitte that contributed to our research for the Information chapter: Lou DiLorenzo, Lena La, Nitin Mittal, Sanghamitra Pati, Jim Rowan, and Baris Sarer.

Additionally, the authors would like to acknowledge and thank Deanna Gorecki, Ben Hebbe, Tracey Parry, Mikaeli Robinson, and Madelyn Scott, as well as the Deloitte Insights team, the Marketing Excellence team, the NExT team, and the Knowledge Services team.

Cover image by: Sylvia Yoon Chang, Manya Kuzemchenko, and Heidi Morrow; Getty Images, Adobe Stock