Hallucinations to agents: Here’s how AI language models evolved in 2024

I research the intersection of artificial intelligence, natural language processing, and human reasoning as the director of the Advancing Human and Machine Reasoning lab at the University of South Florida. I am also commercializing this research in an AI startup that provides a vulnerability scanner for language models.

From my vantage point, I observed significant developments in the field of AI language models in 2024, both in research and the industry.

Perhaps the most exciting of these are the capabilities of smaller language models, support for addressing AI hallucination, and frameworks for developing AI agents.

Small AIs make a splash

At the heart of commercially available generative AI products like ChatGPT are large language models, or LLMs, which are trained on vast amounts of text and produce convincing humanlike language. Their size is generally measured in parameters, which are the numerical values a model derives from its training data. The larger models like those from the major AI companies have hundreds of billions of parameters.

There is an iterative interaction between ">large language models and smaller language models, which seems to have accelerated in 2024.

First, organizations with the most computational resources experiment with and train increasingly larger and more powerful language models. Those yield new large language model capabilities, benchmarks, training sets, and training or prompting tricks. In turn, those are used to make smaller language models—in the range of three billion parameters or less—which can be run on more affordable computer setups, require less energy and memory to train, and can be fine-tuned with less data.

It’s no surprise, then, that developers have released a host of powerful smaller language models—although the definition of small keeps changing: Phi-3 and Phi-4 from Microsoft, Llama-3.2 1B and 3B, and Qwen2-VL-2B are just a few examples.

These smaller language models can be specialized for more specific tasks, such as rapidly summarizing a set of comments or fact-checking text against a specific reference. They can work with their larger cousins to produce increasingly powerful hybrid systems. https://www.youtube.com/embed/zDj24etsRZ4?wmode=transparent&start=0 What are small language model AIs – and why would you want one?

Wider access

Increased access to highly capable language models large and small can be a mixed blessing. As there were many consequential elections around the world in 2024, the temptation for the misuse of language models was high.

Language models can give malicious users the ability to generate social media posts and deceptively influence public opinion. There was a great deal of concern about this threat in 2024, given that it was an election year in many countries.

And indeed, a robocall faking President Joe Biden’s voice asked New Hampshire Democratic primary voters to stay home. OpenAI had to intervene to disrupt over 20 operations and deceptive networks that tried to use its models for deceptive campaigns. Fake videos and memes were created and shared with the help of AI tools.

Despite the anxiety surrounding AI disinformation, it is not yet clear what effect these efforts actually had on public opinion and the U.S. election. Nevertheless, U.S. states passed a large amount of legislation in 2024 governing the use of AI in elections and campaigns.

Misbehaving bots

Google started including AI overviews in its search results, yielding some results that were hilariously and obviously wrong—unless you enjoy glue in your pizza. However, other results may have been dangerously wrong, such as when it suggested mixing bleach and vinegar to clean your clothes.

Large language models, as they are most commonly implemented, are prone to hallucinations. This means that they can state things that are false or misleading, often with confident language. Even though I and ">others continually beat the drum about this, 2024 still saw many organizations learning about the dangers of AI hallucination the hard way.

Despite significant testing, a chatbot playing the role of a Catholic priest advocated for baptism via Gatorade. A chatbot advising on New York City laws and regulations incorrectly said it was “legal for an employer to fire a worker who complains about sexual harassment, doesn’t disclose a pregnancy or refuses to cut their dreadlocks.” And OpenAI’s speech-capable model forgot whose turn it was to speak and responded to a human in her own voice.

Fortunately, 2024 also saw new ways to mitigate and live with AI hallucinations. Companies and researchers are developing tools for making sure AI systems follow given rules pre-deployment, as well as environments to evaluate them. So-called guardrail frameworks inspect large language model inputs and outputs in real time, albeit often by using another layer of large language models.

And the conversation on AI regulation accelerated, causing the big players in the large language model space to update their policies on responsibly scaling and harnessing AI.

But although researchers are continually finding ways to reduce hallucinations, in 2024, research convincingly showed that AI hallucinations are always going to exist in some form. It may be a fundamental feature of what happens when an entity has finite computational and information resources. After all, even human beings are known to confidently misremember and state falsehoods from time to time.

The rise of agents

Large language models, particularly those powered by variants of the transformer architecture, are still driving the most significant advances in AI. For example, developers are using large language models to not only create chatbots, but to serve as the basis of AI agents. The term “agentic AI” shot to prominence in 2024, with some pundits even calling it the third wave of AI.

To understand what an AI agent is, think of a chatbot expanded in two ways: First, give it access to tools that provide the ability to take actions. This might be the ability to query an external search engine, book a flight, or use a calculator. Second, give it increased autonomy, or the ability to make more decisions on its own.

For example, a travel AI chatbot might be able to perform a search of flights based on what information you give it, but a tool-equipped travel agent might plan out an entire trip itinerary, including finding events, booking reservations, and adding them to your calendar.

In 2024, new frameworks for developing AI agents emerged. Just to name a few, LangGraph, CrewAI, PhiData, and AutoGen/Magentic-One were released or improved in 2024.

Companies are just beginning to adopt AI agents. Frameworks for developing AI agents are new and rapidly evolving. Furthermore, security, privacy, and hallucination risks are still a concern.

But global market analysts forecast this to change: 82% of organizations surveyed plan to use agents within 1-3 years, and 25% of all companies currently using generative AI are likely to adopt AI agents in 2025.

John Licato is an associate professor of computer science at the Director of AMHR Lab at University of South Florida.

This article is republished from The Conversation under a Creative Commons license. Read the original article.

https://www.fastcompany.com/91251234/ai-language-models-2024-evolution-hallucinations-agents?partner=rss&utm_source=rss&utm_medium=feed&utm_campaign=rss+fastcompany&utm_content=rss

Created 1d | Dec 23, 2024, 11:20:05 AM


Login to add comment

Other posts in this group

45 years ago, the Walkman changed how we listen to music

Back in 1979, Sony cofounder Masaru Ibuka was looking for a way to listen to classical music on long-haul flights. In response, his company’s engineers dreamed up the Walkman, ordering 30,000 unit

Dec 24, 2024, 3:10:04 PM | Fast company - tech
The greatest keyboard never sold

Even as the latest phones and wearables tout speech recognition with unprecedented accuracy and spatial computing products flirt with replacing tablets and laptops, physical keyboards remain belov

Dec 24, 2024, 12:50:02 PM | Fast company - tech
The 25 best new apps of 2024

One of the most pleasant surprises about this year’s best new apps have nothing to do with AI.

While AI tools are a frothy area for big tech companies and venture capitalists, ther

Dec 24, 2024, 12:50:02 PM | Fast company - tech
The future belongs to systems of action

The world of enterprise tech is built on sturdy foundations. For decades, systems of record—the databases, customer relationship management (CRM), and enterprise resource planning (ERP) platforms

Dec 23, 2024, 10:50:06 PM | Fast company - tech
Banning Chinese-made drones could hurt some Americans

Russell Hedrick, a North Carolina farmer, flies drones to spray fertilizers on his corn, soybean and wheat fields at a fraction of what it

Dec 23, 2024, 8:40:03 PM | Fast company - tech
Fake AI-generated reviews are all over the internet, research shows

The emergence of generative artificial intelligence tools that allow people to efficiently produce novel and detailed online reviews wi

Dec 23, 2024, 8:40:02 PM | Fast company - tech