Multiverse Computing is pushing its compact AI models into the mainstream


With company defaults running at Higher than 9.2% – Highest rate in years – VC firm Lux Capital recently advised AI-driven companies to get their computing power commitments Confirmed in writing. As financial instability spreads across the AI ​​supply chain, Lux warns that a handshake agreement is not enough.

But there is another option entirely, which is to stop relying on external computing infrastructure altogether. Smaller AI models that run directly on a user’s own machine — without a data center, cloud provider, or counterparty risk — are now good enough to be worth considering. and Multiverse computing He raises his hand.

The Spanish startup has so far maintained a lower profile than some of its peers, but as demand for AI efficiency grows, this is changing. After compressing models from major AI labs including OpenAI, Meta, DeepSeek, and Mistral AI, it has launched an app showcasing the capabilities of its compressed models and an API gateway — a gateway that allows developers to access and build on these models — making them more widely available.

the Compactive Eye appwhich shares its name with Multiverse’s quantum-inspired compression technology, is an AI-powered chat tool similar to Mistral’s ChatGPT or Le Chat. Ask a question, and the model answers. The difference is that Multiverse is built into Gilda, a model so small that it can run locally and offline, according to the company.

For end users, this is an AI experience at the edge, with data that never leaves their device and requires no connectivity. But there’s a caveat: Their mobile devices must have enough RAM and storage space. If it doesn’t — and many older iPhones won’t — the app will switch back to cloud-based models via an application programming interface (API). The routing between on-premises and cloud processing is handled automatically by a system Multiverse has dubbed Ash Nazg, whose name will ring a bell to Tolkien fans because it refers to the One Ring cameo in “The Lord of the Rings.” But when an app heads to the cloud, it loses a key privacy advantage in the process.

These limitations mean that CompactifAI isn’t quite ready for widespread adoption by customers yet, although that was probably never the goal. According to data from Sensor Tower, the app was Less than 5000 downloads Last month.

The real target is companies. Today, Multiverse launches a Self-service API gateway It gives developers and organizations direct access to their compressed models – without the need for the AWS Marketplace.

TechCrunch event

San Francisco, California
|
October 13-15, 2026

“The CompactifAI API portal (now) gives developers direct access to compact models with the transparency and control needed to run them in production,” CEO Enrique Lizaso said in a statement.

Real-time usage monitoring is one of the main features of the API, and this is no coincidence. Beyond the potential advantages of deploying at the edge, lower compute costs are one of the main reasons organizations are considering smaller models as an alternative to large language models (LLMs).

It also helps that small models are less limited than they used to be. Earlier this week, Mistral updated its mini model family with Mistral Small 4 launchwhich it says is simultaneously optimized for general chat, programming, proxy tasks, and reasoning. French company too Release hima system that allows organizations to create custom models, including small models from which they can choose which trade-offs their use cases can best tolerate.

Recent findings from Multiverse also suggest that the gap with LLMs is narrowing. Its latest compact models, Hypernova 60B 2602is built on gpt-oss-120b – an OpenAI model whose base code is publicly available. The company claims it is now offering Faster responses At a lower cost than the original version from which it was derived, a particularly important feature for agent coding workflows, where AI completes complex, multi-step programming tasks autonomously.

Making models small enough to work on mobile devices while maintaining their usefulness is a huge challenge. Apple intelligence Avoid this problem by combining the on-device model with the cloud model. Multiverse’s CompactifAI app can also route requests to gpt-oss-120b via an API, but its main goal is to show that homegrown models like Gilda and its future replacements have benefits beyond cost savings.

For workers in critical fields, a model that can run on-premises without connecting to the cloud provides greater privacy and flexibility. But the greatest value lies in the commercial use cases this could open up – for example, integrating AI into drones, satellites and other settings where connectivity cannot be taken for granted.

The company already serves more than 100 global clients including Bank of Canada, Bosch and Iberdrola, but expanding its client base could help it unlock more financing. After raising a $215 million Series B Last year, it is now It is rumored to be raising a new funding round worth €500 million With a value exceeding 1.5 billion euros.

Leave a Reply

Your email address will not be published. Required fields are marked *