The great Yogi Berra once said, “It’s difficult to make predictions, especially about the future.” Truer words have never been spoken, especially in the rapidly evolving world of AI.
With the new Gartner 2024 Hype Cycle for Emerging Technologies providing a roadmap for the near future, it’s time to look ahead to 2025. Generative AI has crested the Peak of Inflated Expectations and moves ever closer to the Trough of Disillusionment. 2024 was a year of generative AI experimentation, and IT leaders are learning what should and shouldn’t go into production at scale based on the tangible reality of improved business outcomes.
Meanwhile, AI agents gain awareness as they move up the Innovation Trigger slope. For IT leaders who are unclear about the differences between generative AI and agentic AI, it can be broken down as generative AI creates, and agentic AI acts.
Humans partner closely with generative AI to generate and use outreach emails, source code, business plans, and marketing strategies, but the human is the one who activates the generative AI – and it’s also the human who acts upon the generative AI’s output.
In the case of agentic AI, agents can activate generative AI, as well as external workflows and systems of record. Agents can also figure out and execute the next steps, which may or may not require human intervention.
While the business value of generative AI may be difficult to assess, the benefit of agentic AI is more straightforward as there’s a human cost of “jobs to be done,” which can be compared with the cost of the job being done by an agent. Furthermore, agentic AI is elastic, which allows for surges of AI agents to cost-effectively scale up and down based on predictable seasonality, e.g. Black Friday sales, Tax Day support or unexpected events like major service outages or natural disasters.
5 AI predictions for 2025The landscape of artificial intelligence is shifting rapidly, with new advancements and trends emerging constantly. As we approach another year, it’s crucial for technology leaders to anticipate and prepare for these changes.
Here are five AI predictions that will equally excite and challenge technology leaders in 2025:
For the past year, industry leaders have been repeating the mantra, “Your AI strategy is only as good as your data strategy.” This is true because if your data is trapped and inaccessible by your AI, your generative AI results will be generic and disappointing at best and harmfully hallucinated at worst. As our AI journey expands to the world of agents, we’ll quickly learn the corollary: “Your agent strategy is only as good as your integration strategy.”
Why is this the case? Just as a well-oiled machine relies on the seamless interplay of its parts, the success of AI agents will depend heavily on their integration within the broader technological ecosystem.
AI agents, with their ability to automate tasks and make decisions, are poised to revolutionize various aspects of business operations. However, their true potential can only be unlocked when seamlessly integrated with existing systems and data sources across the enterprise.
Imagine an AI agent tasked with customer service. To be truly effective, it needs access to customer data, order history, product information, shipping information, and even real-time inventory levels. Without this integration, the agent’s usefulness would be severely limited, resulting in fewer case deflections and higher support costs.
What IT leaders should do in 2025By prioritizing integration, IT leaders can entrust AI agents to perform at their best and maximize business value for their organizations.
2. LAMs will be lionsMove over LLMs – there’s a new acronym in town. While large language models (LLMs) excel at understanding and generating human language, large action models (LAMs) take it further by translating that understanding into action.
Think of it this way: LLMs are like the brains of the operation, capable of processing information and generating insights. LAMs are the hands, putting those insights into action.
Why LAMs are critical to agentic AI successLAMs are poised to be the driving force behind the success of agentic AI, which focuses on creating AI agents that can autonomously perform tasks and achieve goals. By integrating with APIs across the enterprise, LAMs can interact with various systems and applications, enabling them to carry out complex actions.
For example, an LAM-powered AI agent could automate the onboarding process for new employees. It could gather the necessary information, create accounts, grant access to relevant systems, and even schedule introductory meetings.
Integrating LAMs with APIsThe key to unlocking the full potential of LAMs lies in their integration with APIs. APIs (application programming interfaces) act as the connective tissue between different software applications, allowing them to communicate and exchange data.
By integrating LAMs with APIs, organizations can enable AI agents to interact with a wide range of cloud-based and on-premise systems and services, from CRM and ERP systems to marketing automation platforms and ecommerce platforms. This integration opens up a world of possibilities for automating tasks, streamlining processes, and improving efficiency.
What IT leaders should do in 2025By embracing LAMs and prioritizing their integration with APIs, IT leaders can position their organizations to reap the full benefits of agentic AI, resulting in new levels of automation and efficiency.
3. Small language models: Does size really matter?While large language models have dominated the AI landscape with their impressive capabilities, a new trend is emerging: small language models (SLMs). These compact and efficient models are gaining traction as organizations recognize their svelte benefits.
SLMs are smaller versions of their LLM counterparts. They have significantly fewer parameters, which translates to several advantages:
Researchers employ various techniques to create SLMs that maintain performance while reducing size:
SLMs hold particular promise for the public sector, especially for applications that require on-device operation. This is crucial for forward-deployed scenarios or classified applications running on systems without internet access to cloud-based frontier large language models.
What IT leaders should do in 2025Small is big. By embracing the “thin is in” approach and leveraging the advantages of SLMs, IT leaders can optimize their AI initiatives for cost-effectiveness, speed, privacy, and accessibility.
4. Energy efficiency will be as important as performanceAs generative AI enters the mainstream, its environmental impact is coming under increased scrutiny. Organizations realize that the widespread use of large language models (LLMs) can significantly affect their environmental sustainability metrics and goals.
The energy consumption of these powerful models is a growing problem. Training and running LLMs requires vast computational power, translating to a substantial carbon footprint. This has prompted a call for greater transparency and accountability regarding the environmental costs of AI.
Efforts toward transparencyFortunately, efforts are underway to address this issue. Researchers and industry leaders are working to develop an “Energy Star” equivalent for AI models, providing clear metrics to assess their environmental impact. This will enable organizations to make informed decisions about their chosen models, balancing performance with sustainability considerations.
Optimizing LLM selectionInstead of defaulting to the most advanced and resource-intensive frontier models for every application, organizations will increasingly prioritize the optimal LLM for their specific needs. Just as they consider factors like price and performance, they’ll also factor in the model’s environmental impact.
This shift toward energy efficiency will help to encourage innovation and the development of more sustainable AI models. Researchers will focus on creating models that deliver high performance with a lower carbon footprint, using techniques like model compression, quantization, and efficient training methods.
What IT leaders should do in 2025By prioritizing energy efficiency alongside performance, IT leaders can ensure that their AI initiatives align with their organization’s sustainability goals and contribute to a greener future.
5. Multiple AI model rings will rule them allAs organizations increasingly rely on large language models (LLMs) to power their AI applications, efficient and cost-effective LLM management becomes paramount. This is where LLM routers act as intelligent traffic directors for LLM requests.
LLM routers optimize LLM usage by analyzing incoming prompts and routing them to the most suitable model based on complexity, performance requirements, and cost. This ensures that less expensive models handle simpler tasks while more complex tasks are directed to the most powerful models.
RouteLLM: An example of LLM routing in actionOne example of an LLM router is RouteLLM, an open-source framework designed to optimize LLM serving. RouteLLM analyzes incoming queries and dynamically selects the best model to handle each request, considering factors like cost and performance. This approach not only reduces costs by up to 85% while maintaining 95% of the performance of frontier models but can also improve overall efficiency and response times.
Expanding LLM routers’ roleWhile current LLM routers primarily focus on optimizing for price and performance, their potential extends far beyond these dimensions. As awareness of the environmental impact of LLMs grows, LLM routers can play a crucial role in promoting sustainability.
Imagine an LLM router that considers not only price and performance but also the environmental impact of each model. This would enable organizations to optimize their LLM usage for sustainability, selecting the most energy-efficient model for each prompt.
What IT leaders should do in 2025By embracing LLM routers and expanding their capabilities to include sustainability considerations, IT leaders can ensure that their AI initiatives are both efficient and environmentally responsible.
The AI future ain’t what it used to beThe future of AI is filled with exciting possibilities, and these five predictions offer a glimpse into the trends and challenges that await technology leaders in 2025. By staying informed and prepared, organizations can navigate this evolving landscape and harness the power of AI to drive innovation, achieve their business goals, and be the disruptor instead of the disrupted.
Want to explore more? We’ve got you covered.
Together, let’s navigate the future of AI and unlock its full potential for a brighter tomorrow. To keep on learning, download our 2025 digital trends report to see what seven trends are shaping digital transformation in the upcoming year.
RetroSearch is an open source project built by @garambo | Open a GitHub Issue
Search and Browse the WWW like it's 1997 | Search results from DuckDuckGo
HTML:
3.2
| Encoding:
UTF-8
| Version:
0.7.4