Machine Learning

ElastixAI Exits Stealth with FPGA AI Supercomputers

Company leverages software-ML-hardware co-design to deliver up to 50x lower TCO and 80% reduced power consumption for LLM inference

ElastixAI Inc. today emerged from stealth to tackle the systemic inefficiencies and high costs of generative AI (GenAI) inference. Founded by former Apple and Meta machine learning (ML) researchers, and backed by $18M in seed funding, the company is launching a novel software platform that converts off-the-shelf FPGA-based servers into high-efficiency AI supercomputers.

The AI inference market is expected to reach $255 billion by 2030, yet existing infrastructure is fundamentally mismatched for GenAI. While LLM inference is memory-bound, standard GPUs are designed for compute-bound tasks like training. This results in very low compute utilization during inference, leading to massive wasted capital and energy. Further, custom silicon is architected years before it reaches production, leaving it years behind current ML innovations. For example, 4-bit quantization—which theoretically doubles performance—often yields only a 10% gain on hardware lacking native support.

“The industry is currently leaving an order-of-magnitude of performance on the table because hardware can’t keep up with the advances in ML,” said Mohammad Rastegari, PhD, co-founder of ElastixAI. “We’re moving away from ‘one-size-fits-all’ hardware. By applying proprietary post-training optimizations to FPGAs, we let hardware adapt to the model rather than forcing the model to struggle on the hardware.”

ElastixAI provides a drop-in replacement for legacy GPU workflows, maintaining workflow compatibility while delivering massive efficiency gains. Key benefits include:

  • Up to 50x TCO Advantage: ElastixAI’s hardware-software co-optimization executes LLM operations with far greater density than general-purpose GPU kernels.
  • 80% Lower Power Consumption: The ElastixAI solution eliminates “dark silicon” by activating only the circuits required for inference.
  • Staying Cutting-Edge: Custom silicon development cycles can exceed 3 years, but AI innovations occur daily. ElastixAI bridges this gap, enabling cutting-edge AI implementations on current hardware.

ElastixAI is available to select enterprise partners, data center operators, and AI model providers. Visit www.elastix.ai for technical specifications or to request a demo.

Business Wire

Business Wire is a trusted source for news organizations, journalists, investment professionals and regulatory authorities, delivering news directly into editorial systems and leading online news sources via its multi-patented NX Network. Business Wire has 18 newsrooms worldwide to meet the needs of communications professionals and news media.

Related posts

Rambus Advances AI/ML Performance

PR Newswire

nOps Secures $30M Series A Funding

Business Wire

Dazl Emerges from Stealth with $10M to Build AI-Driven Apps

PR Newswire