Voltron Data harnesses AI-friendly chips for big data crunching

The AI boom has brought renewed attention to graphics processing units, the high-powered computing chips used to power modern generative AI and other machine learning applications.

And a startup called Voltron Data is harnessing some of those same GPUs to efficiently handle massive data analytics tasks, like combing through huge server logs for cybersecurity purposes, analyzing enormous sets of financial data, or processing telemetry data from complex systems like autonomous cars. At large enough scale, starting around 30 terabytes of data, database queries on traditional computing chips can start to bottleneck, with processing time no longer going down in a linear fashion as you add more computing power, says Voltron Data cofounder and field CTO Rodrigo Aramburu.

“When things get really big, they get really weird,” he says. “The unit economics basically break down.”

But, he says, modern GPUs from Nvidia and the servers that house them are designed to quickly shift large volumes of data onto those processing chips, faster than traditional CPU-based systems. GPUs can also quickly process the mathematical operations needed to search and sort through data or combine data from multiple large tables, making database operations efficient even when data sets are mindbogglingly large.

“It’s all of these little things that just add up,” says Aramburu. “From a database operation perspective, if you know how to take advantage of them, and you build something from the ground up to take advantage of all of this hardware acceleration, you’re going to get really, really good performance from the GPU.”

Companies using Voltron Data’s software, called Theseus, have been able to replace fleets of CPU-powered systems with vastly smaller numbers of GPU-centric servers, he says, sometimes substituting as few as one GPU-powered server for every 100 previously in use. That can substantially cut the amount of energy and real estate they require for handling big data—even as data processing tasks run more quickly.

One large retail client had a nightly process designed to predict sales and optimize volumes of perishable goods sent to individual stores. On a CPU-based system, the process typically took almost eight hours to run. With Voltron Data’s GPU-powered architecture, it could run in just 25 minutes. The speedup was a relief to the company, which previously had little margin if anything stalled the process in the eight or so hours it had to run each night, and it also gave developers the ability to test different versions of the code, enabling them to make its predictions more accurate and get more goods to places where they’d be sold.

“They were able to iterate on it enough times that they reasonably impacted the model accuracy by a few percentage points,” he says.

Voltron Data’s own engineers need to know the details of how to optimize data processing tasks for GPUs, but developers and database engineers at the company’s customers can continue to write queries using SQL, the standard database language, or data frame processing libraries in various programming languages. They can also continue to store data in standard data lake environments and common formats, with Voltron Data software processing it essentially as is when the time comes to run queries.

Voltron Data doesn’t host its own cloud systems—instead, customers install the software on computers they already control, either in their own data centers or an existing cloud network. In many cases, they’re able to take advantage of largely idle GPU servers they’ve already acquired for other purposes, like development with previous generations of AI tools, Aramburu says.

The company’s systems aren’t necessary for every business, he freely admits, or for every task. For smaller data sets, traditional database tools still make sense, especially when they’re used efficiently by programmers who know how to optimize performance. But for organizations dealing with data at the right scale, the advantages start to be apparent.

“All those algorithmic tricks that we can engage in start losing their utility, and being able to just brute force process all the data as quickly as possible becomes necessary,” he says. “That’s where an engine like ours, which is able to leverage all this advanced hardware, can just churn through it much faster.”

https://www.fastcompany.com/91240501/voltron-data-harnesses-ai-friendly-chips-for-big-data-crunching?partner=rss&utm_source=rss&utm_medium=feed&utm_campaign=rss+fastcompany&utm_content=rss

Utworzony 3mo | 9 gru 2024, 11:20:06


Zaloguj się, aby dodać komentarz

Inne posty w tej grupie

How ‘lore’ became the internet’s favorite way to overshare

Lore isn’t just for games like The Elder Scrolls or films like The Lord of the Rings—online, it has evolved into something entirely new.

The Old English word made the s

24 lut 2025, 13:20:04 | Fast company - tech
These LinkedIn comedians are leaning into the cringe for clout

Ben Sweeny, the salesman-turned-comedian behind that online persona Corporate Sween, says that bosses should waterboard their employees. 

“Some companies drown their employees with

24 lut 2025, 10:50:08 | Fast company - tech
The best apps to find new books

This article is republished with permission from Wonder Tools, a newsletter that helps you discover the most useful sites and apps. 

24 lut 2025, 06:20:05 | Fast company - tech
5 tips for mastering virtual communication

Andrew Brodsky is a management professor at McCombs School of Business at the University of Texas at Austin. He is also CEO of Ping Group and has received nume

23 lut 2025, 11:50:03 | Fast company - tech
Apple’s hidden white noise feature may be just the productivity boost you need

As I write this, the most pleasing sound is washing over me—gentle waves ebbing and flowing onto the shore. Sadly, I’m not actually on some magnificent tropical beach. Instead, the sounds of the s

22 lut 2025, 12:40:06 | Fast company - tech
The next wave of AI is here: Autonomous AI agents are amazing—and scary

The relentless hype around AI makes it difficult to separate the signal from the

22 lut 2025, 12:40:05 | Fast company - tech
This slick new service puts ChatGPT, Perplexity, and Wikipedia on the map

I don’t know about you, but I tend to think about my favorite tech tools as being split into two separate saucepans: the “classic” apps we’ve known and relied on for ages and then the newer “AI” a

22 lut 2025, 12:40:03 | Fast company - tech