The most innovative companies in artificial intelligence for 2025

The AI industry hit a significant bend in the road toward artificial general intelligence in 2024. Previously, the stunning intelligence gains that led to chatbots such ChatGPT and Claude had come from supersizing models and the data and computing power used to train them. When the progress from massive scaling leveled off, researchers knew they would need a new strategy, beyond training, to keep moving toward AGI models that are broadly smarter than humans. 

Starting with OpenAI’s pivotal o1 model, researchers began to apply more computing power to the real-time reasoning a model does just after a user prompts it with a problem or question. o1 required more time to produce answers than other models, but its answers were clearly better than those of non-reasoning models. The o1 model rose quickly to the top of the rankings in common benchmark tests, and soon Google DeepMind, Anthropic, DeepSeek and others were training their models for real-time reasoning. 

The big AI labs would now need even more of the Nvidia GPUs they’d been using for training to support all the real-time reasoning their models would be doing. That’s part of the reason Nvidia saw its stock price rise 171% in 2024. The chip company had been reading the tea leaves and built new features into its GPUs that made them even better suited to real-time inference than before. The result was a new architecture called Blackwell, and new chips called the B100 and B200. Nvidia unveiled the new GPUs in March 2024 and quickly sold its entire 2024 supply to the largest data center operators. 

Even before the appearance of new reasoning models, some of AI’s hottest companies produced state-of-the-art new AI systems. Google DeepMind broke through with a family of natively multi-modal models called Gemini that understand imagery and audio as well as they do language. Anthropic continued to put intense pressure on OpenAI, and was first to release an AI model that could perform task on a user’s computer. Mistral released impressive new small language models that can run on laptops and even phones with its Ministral 3B and Ministral 8B, as did Microsoft with its Phi-3 and Phi-4 models. And Runway again made its case for the state-of-the-art with its new Gen-3 Alpha video generation models.    

And all the while a small AI lab in China was quietly developing new AI models, including reasoning models, that would begin sending shockwaves through the AI industry by the end of 2024. DeepSeek—#1 on our list of the most innovative companies in Asia-Pacific—trained state-of-the-art models at far lower cost and with far less GPU power than anyone thought possible. It even showed its work through research papers and by open-sourcing its models. DeepSeek’s breakthroughs caused some angst, but its fresh thinking and openness will likely spur bigger and faster innovations from the world’s top AI companies by this time next year.

1. Nvidia

For arming employees with the tools to get their jobs done

Large-language model makers continue to pour money into realizing the lofty ambitions of their AI systems. But one company is already reaping the rewards. Nvidia kicked off the AI race by providing the computing power with its market-dominating graphics processing units (GPUs). Now, as its customers chase human-level intelligence, its groundbreaking Blackwell processor and platform is ready. 

For general model training tasks, the Blackwell processor is up to 2.5 times more powerful than its predecessor, the H100, and requires significantly less energy to operate. The largest data center operators and AI labs, including Google, Meta, Microsoft, OpenAI, Tesla and xAI, are buying or planning to buy Blackwell GPUs by the hundreds of thousands.

While recent models from China’s DeepSeek and Alibaba have made big strides in wringing cutting-edge AI out of older, less powerful Nvidia GPUs, the company isn’t just cranking out processors and waiting to see what the world does with them. Instead, it’s actively building platforms for everything from drug discovery (Clara for Biopharma) to autonomous vehicles (Drive AGX) to video production (Holoscan) to digital twins (Omniverse). By driving AI progress in an ever-growing array of real-world scenarios, it’s positioning itself for continued growth, even if future models don’t need all the computational muscle they can get.

Read more about Nvidia, honored as No. 2 on Fast Company’s list of the World’s 50 Most Innovative Companies of 2025.

2. OpenAI

For improving AI by giving it more time to think

Since 2019, OpenAI has continually improved its models by giving them more training data and computing power. The rest of the industry has followed its lead. Now that researchers are seeing diminishing returns from this scaling strategy, OpenAI needed an alternative route toward its goal of creating AI models that are smarter than humans in most tasks–in other words, models that can reach artificial general intelligence (AGI).

The company found it, and the resulting model is called o1. Instead of scaling up training data and compute time while the model is pre-trained, OpenAI researchers built o1 to spend more time and computing power at inference, when the model is deployed and answering user prompts. While doing this, the model collects and remembers contextual data, both from the user and from relevant data sources. It takes a trial-and-error approach to figuring out the best route to an answer. The model generates PhD-level answers to complex questions, landing it on top of performance benchmark rankings.

OpenAI o1 “experimental” and “mini” versions are available to ChatGPT Plus users. The company offers a new service called ChatGPT Pro that provides unlimited access to o1 proper for $200 per month. In December of 2024 OpenAI announced o1’s successor, o3, and in February 2025 it gave paid users access to o3-mini, a smaller, faster version that excels in science, math, and coding. The biggest impact of OpenAI’s new reasoning models is proving to the industry that scaling up computing at inference time may be a viable path to making new intelligence breakthroughs on the way to AGI.

3. Google DeepMind

For delivering a truly multi-modal AI model

The foundational research in AI model architecture and model training that led to the chatbots of today happened at Google in the late 2010s. The search giant developed a large language model-powered chatbot long before ChatGPT, but, as the story goes, it hesitated to expose the technology to the public because of concerns over safety, privacy, and legal risks. That hesitation put Google behind in the AI race that ensued with the launch of ChatGPT.

The release of Google DeepMind’s Gemini 2.0 in 2024 marked the moment when Google officially caught up. Gemini 2.0 is the first mass market AI model that is natively “multimodal,” meaning that it can process and generate images, video, audio, and computer code the same way it does text. This makes it possible for the model to watch a video clip, or live video from a phone camera, and quickly analyze and reason about it.

The Gemini model is also notable for being able to control other Google services such as Maps and Search. This is Google beginning to flex–combining its AI research with its legacy information and productivity tools. Gemini is one of the first AI models capable of working autonomously and reasoning its way through complex problems on the behalf of the user. The Gemini 2.0 Flash Thinking Experimental model even shows the user the thought process it used to reach an answer. And in December, Google showed off Project Mariner, a Gemini-based agentic AI feature designed to perform tasks such as online grocery shopping on behalf of the user.

4. Anthropic

For imbuing Claude with the ability to use a computer

Generative AI has so far been used mainly for a few tasks: writing and summarizing text as well as generating images. The next step is making large language models reason and use tools. Anthropic’s “Computer Use” model gave us our first look into that future.

Starting with 2024’s Claude 3.5 Sonnet, Anthropic’s model can perceive what’s happening on the user’s screen, including internet content. It can operate a cursor, click buttons, and type in text. In a video, a Claude researcher demonstrated how Claude can fill out a form based on information available at the websites open in the browser tabs. It can complete tasks such as building a personal website or sorting the logistics of a day trip. It’s wild to watch the AI work, clicking open new tabs, running searches, and filling in data fields.

Right now the model runs slowly and doesn’t always get the right answer, but that’s likely to change quickly as Anthropic finds and fixes the model’s pitfalls. Google’s aforementioned Project Mariner followed Anthropic’s lead in December, and OpenAI introduced its own computer use model, Operator, in January 2025. In February 2025, Anthropic moved on to its next major iteration of Claude, Claude 3.7 Sonnet, a larger model that can automatically switch into reasoning mode for hard questions.

5. Microsoft

For pushing the state-of-the-art in small language models

The development of Microsoft’s Phi models started out with a question posed by the company’s researchers in 2023: “How small can we make a model that shows signs of emerging intelligence?” This turns out to be an important moment in the evolution of “small language models,” or models that perform well in scenarios where memory, processing power, or connectivity are limited, and response speed is important.

During 2024, Microsoft released two generations of small models that exhibited reasoning and logic capabilities that hadn’t been explicitly trained into them. In April the company released a series of Phi-3 models that excel in language, reasoning, coding, and math benchmarks–probably owing to the fact that Microsoft trained them using synthetic data generated by much larger and more capable LLMs. People downloaded variants of the open-source Phi-3 more than 4.5 million times on HuggingFace during 2024.

In late 2024, Microsoft released its Phi-4 small language models, which improved on the Phi-3 models in reasoning-focused tasks and even outperformed OpenAI’s GPT-4o on the GPQA (scientific questions) and MATH benchmarks. Microsoft released the model under an open-source and open-weights license, so developers can use it to develop edge models or apps that run on phones or laptops. In less than a month, Phi-4 was downloaded 375,000 times on HuggingFace.

6. Amazon

For building a cost-effective rival to Nvidia’s GPU

Amazon AWS recently released a new version of its Tranium processor for AI, called Trainium2, which—in some settings—could challenge the dominance of Nvidia GPUs. The Trainium2 chip is built to supply the massive computing power used in training the largest generative AI models, and for inference time work after the model has been deployed. AWS says Trainium is 30% to 40% more cost-effective in jobs normally done by GPUs.

Trainium2 cures the power and software integration deficiencies seen in the first Trainium chip, says Dylan Patel of SemiAnalysis, and puts Amazon “on a path” toward catching up to Nvidia. (Note that AWS itself remains very reliant on NVIDIA for GPUs.) Unseating Nvidia is difficult because customers get locked in to building on top of Nvidia’s CUDA software layer, which lets researchers control how their models use the chip’s resources. Amazon offers its own kernel control software layer called Neuron Kernel Interface (NKI), which, like CUDA, gives researchers granular control over how the chip kernels work together.

Patel points out that Trainium2 is still untested at scale. AWS is now building a server cluster with 400,000 Trainium2 chips for Anthropic, which could teach it a lot about making its AI chips work well together in big numbers.

7. Arm

For making AI data centers more cost effective and efficient

The British semiconductor designer Arm has long provided the tech industry with the architecture used in chips that power small devices such as phones, sensors, and IoT hardware. This takes on new importance as a new era arrives in which edge device chips will run AI models. But data centers will play a huge role in this evolution as well, often handling some or all of the heaviest AI processing and delivering the results down to the edge device.

As data centers expand around the world, the amount of electrical power they use will become a pressing consideration. That’s part of the reason Arm’s latest Neoverse CPU architecture focuses on efficiency. It offers a 50% performance improvement over earlier generations, and 20% better performance per watt than processors using competing x86 architectures, the company says.

Arm says that Amazon, Microsoft, Google, and Oracle have all now adopted Arm Neoverse for both general-purpose computing and CPU-based AI inference and training. For example, in 2024 Micerosoft announced its first custom silicon built for the cloud, the Cobalt 100 processor, was built on Arm Neoverse. Some of the largest AI data centers will rely on NVIDIA’s Grace Hopper Superchip, which contains a Hopper GPU and a Grace CPU that uses Neoverse. Arm is reportedly planning to launch its own CPU this year, with Meta as one of its first customers.

8. Gretel

For feeding the generative AI boom with synthetic data

Over the past year, AI companies have seen diminishing returns from training their models with ever larger amounts of data scraped from the web. They’ve begun focusing less on the quantity of their training data, and more on its quality. That’s why the they’re are spending more and more on non-public and specialty content that the license from publisher partners. AI researchers also must fill gaps or blind spots within their human-generated or human-annotated training data. For this they’ve increasingly turned to synthetic training data generated by special AI models.

Gretel gained a higher profile during 2024 by specializing in creating and curating synthetic training data. The company announced the general availability of its flagship product, Gretel Navigator, which lets developers use natural language or SQL prompts to create, augment, edit, and curate synthetic training datasets, or their fine-tuning and testing datasets. The platform has already attracted a community of more than 150,000 developers who have synthesized more than 350 billion pieces of training data.

Other industry players have taken notice. Gretel partnered with Google to make its synthetic training data easily available to Google Cloud customers. The company announced a similar partnership with Databricks in June, which allows that company’s enterprise customers to access synthetic training data for their models running within the Databricks cloud.

9. Mistral AI

For creating small AI models that run in laptops and phones

Mistral AI, France’s entry into the generative AI race, has constantly put pressure on OpenAI, Anthropic, and Google at the leading edge of frontier AI model development. Mistral AI released a string of new models containing significant technological advances in 2024, and demonstrated rapid growth in its business, both through direct marketing of its APIs and through strategic partnerships.

Early in the year the company produced a pair of open-source models called Mixtral, notable for its innovative use of the “mixture of experts” architecture, in which only a specialized subset of the model’s parameters are put in play to handle a query, improving efficiency. In July 2024 Mistral announced Mistral Large 2, which, at 123 billion parameters, showed significant improvements in code generation, maths, reasoning, and function calling. The French company also released Ministral 3B and Ministral 8B–both of which are smaller models that can run in laptops or phones and store about 50 text pages of context information provided by the user.

Mistral has seen success in Europe by marketing itself as a low-cost and flexible alternative to U.S. AI companies like OpenAI. It also continued its expansion into the U.S. enterprise market during 2024. In June, the company raised a $640 million funding round, led by the venture capital firm General Catalyst. The round increased Mistral’s valuation to roughly $6.2 billion.

10. Fireworks AI

For bringing a plug-and-play model to AI development and deployment

Fireworks provides a custom runtime environment that removes some of the considerable engineering work typically associated with building infrastructure for AI deployments. Using the Fireworks platform, enterprises can plug in any of more than 100 AI models, then customize and fine-tune them for their specific use cases.

The company introduced new products during 2024 that will equip it to ride some important waves in the AI industry. For one, developers have become more concerned about how quickly AI-powered models and apps re

Creată 9d | 18 mar. 2025, 11:50:08


Autentifică-te pentru a adăuga comentarii

Alte posturi din acest grup

AI is painful for journalists but healthy for journalism

As AI takes on a greater role in our media ecosystem, many journalists look at it like a farmer sees an invasive species:

27 mar. 2025, 11:30:06 | Fast company - tech
The Apple Watch faces a wearables wake-up call

Apple Watch sales are enduring a years-long backslide.

While Apple first launched its watch in 2015, sales didn’t spike until the pandemic, when consumers were highly focused on their he

27 mar. 2025, 11:30:06 | Fast company - tech
For Gen Z, viral videos are the new primetime

Gen Z isn’t just watching creators—they’re choosing them over traditional TV and movies.

That’s the big takeaway from Deloitte’s 19th annual

27 mar. 2025, 11:30:05 | Fast company - tech
Simon & Schuster wants to be your favorite content creator

If you were given $100 and five minutes in a bookstore, which titles would you pick? That’s the premise of Simon & Schuster’s upcoming web series, Bookstore Blitz—the publisher’s latest i

27 mar. 2025, 04:30:07 | Fast company - tech
Supreme Court upholds Biden’s ghost gun rule requiring serial numbers and background checks

The Supreme Court on Wednesday upheld a Biden administration regulation on the nearly impossible-to-trace weapons called 

26 mar. 2025, 21:30:06 | Fast company - tech
‘No one I know would’ve been this reckless’: Ex-officials slam Trump team’s use of Signal for war planning

Current and former government technologists reacted with shock and disbelief to reports that top Trump Administration officials used the consumer messaging app Signal to

26 mar. 2025, 16:50:06 | Fast company - tech
The ‘living globe’ that can help drones fly without GPS

Before the advent of GPS, especially at sea, navigation meant finding your position by looking up at the stars. Today, when the Global Positioning System isn’t working—or gets jammed by elec

26 mar. 2025, 14:40:05 | Fast company - tech