Multiverse Computing Claims Compressed Model Beats OpenAI Original on Speed and Cost
Spanish quantum-AI firm says its 60-billion-parameter model outperforms the 120B OpenAI base it was derived from, as enterprises weigh smaller alternatives to frontier LLMs.

Multiverse Computing, a Spanish firm specializing in quantum-inspired AI compression, claims its latest model delivers faster responses at lower cost than the OpenAI foundation model it was built upon—a technical milestone that could accelerate enterprise adoption of compressed alternatives to frontier large language models.
The company's HyperNova 60B 2602 is derived from gpt-oss-120b, an OpenAI model whose underlying code is publicly available. Multiverse says the compressed version now outperforms the original on speed and operational expense, advantages the firm positions as particularly relevant for agentic coding workflows where AI autonomously executes complex, multi-step programming tasks.
The announcement arrives as the small-model ecosystem gains momentum. Earlier this week, France's Mistral AI updated its compact model family with Mistral Small 4, optimized simultaneously for general chat, coding, agentic tasks, and reasoning. Mistral also released Forge, a system enabling enterprises to build custom models and select performance tradeoffs tailored to specific use cases.
Multiverse has embedded real-time usage monitoring into its API, reflecting enterprise demand for visibility into compute costs—a central driver behind interest in smaller models as alternatives to resource-intensive LLMs. The firm is positioning edge deployment and reduced infrastructure overhead as strategic advantages over centralized, large-scale architectures.
(Multiverse Computing has not disclosed independent benchmarking or third-party validation of its performance claims. The company operates at the intersection of quantum computing research and classical AI optimization, a niche that remains commercially unproven at scale.)
The competitive landscape for compressed and small models is intensifying as enterprises reassess the cost-performance calculus of frontier LLMs. Mistral's dual release of an updated small model and a customization platform signals a broader industry shift toward modular, use-case-specific AI architectures rather than one-size-fits-all large models. Multiverse's claim to outperform an OpenAI base model—if validated—would represent a technical inflection point, suggesting that compression techniques may now rival or exceed the efficiency of the original architectures they derive from.
Keywords
Sources
https://techcrunch.com/2026/03/19/multiverse-computing-pushes-its-compressed-ai-models-into-the-mainstream/
Focuses on Multiverse's performance claims and real-time monitoring features for enterprise cost control
https://www.adweek.com/media/quicken-is-producing-100-pieces-of-content-every-few-weeks-using-ai/
Examines enterprise shift to LLMs for content production and Generative Engine Optimization challenges
https://www.law.com/corpcounsel/2026/03/18/study-finds-some-elite-professions-vulnerable-to-ai-layoffs/
Reports Anthropic research linking AI deployment to hiring slowdowns in exposed professions
https://medium.com/enrique-dans/the-era-of-free-ai-is-ending-heres-how-you-ll-pay-for-it-2ae819d5e947
Explores broader AI monetization trends and strategic shifts in enterprise AI adoption
