
The artificial intelligence landscape is poised for a significant transformation, with Spanish startup Multiverse Computing spearheading the adoption of its compressed AI model technology. This initiative offers businesses and developers a robust solution to break free from cloud dependency. The strategic expansion arrives at a critical juncture, following warnings from venture capital firm Lux Capital about the fragility of AI computing supply chains and recommendations for enterprises to secure compute commitments in writing amidst rising default rates among private companies. Multiverse's launch of a self-service API portal and its CompactifAI application marks a pivotal step towards efficient, on-device AI processing, promising enhanced data privacy and reduced operational costs.
Multiverse Computing: A Strategic Focus on AI Efficiency
The current AI infrastructure is grappling with severe financial challenges. With private company default rates climbing above 9.2%, reaching multi-year highs, over-reliance on external cloud service providers presents substantial counterparty risk. Consequently, Lux Capital recently issued guidance urging AI-dependent enterprises to formalize agreements with their compute providers. Multiverse Computing offers a tangible solution to this instability by promoting the use of miniaturized, compressed models that can run directly on end-user devices. This paradigm eliminates the need for data centers and cloud providers, fundamentally altering the risk profile for AI integration.
While perhaps not as prominent as industry giants previously, Multiverse is capitalizing on the growing demand for AI efficiency. The company has successfully compressed foundational models from leading AI labs including OpenAI, Meta, DeepSeek, and Mistral AI. Its recent introduction of the consumer-facing CompactifAI application and a dedicated API portal for developers signifies a solid move towards broader market penetration. These tools demonstrate that high-performance AI does not necessitate massive, remote computing clusters.
The Technical Principles of Model Compression
Multiverse's core competency lies in its quantum-inspired compression technology, also named CompactifAI. This technique significantly reduces the footprint of large language models (LLMs) without a substantial sacrifice in performance. This allows complex models to operate on resource-constrained hardware, such as smartphones and edge devices. The company's latest compressed model, HyperNova 60B 2602, built upon OpenAI's open model gpt-oss-120b, is a prime example. Multiverse claims its compressed version offers faster response times and lower costs, crucial advantages for Agentic coding workflows that require AI to autonomously execute multi-step programming tasks.
CompactifAI Application: Showcasing Local AI Prowess
The CompactifAI application serves as Multiverse's public-facing showcase for its technology. It functions similarly to ChatGPT or Mistral's Le Chat, allowing users to interact with AI through a familiar chat interface. Its key innovation lies in the integration of the 'Gilda' model, a compact...

