Skip to main content

Jul 02 2025

Cerebras June Highlights

🔥 June at : Speed, Scale & Real-World AI 🔥

  • No more waitlist – the world’s fastest inference API is now open to all
  • SUPERNOVA goes global – next stop: Paris, for the RAISE Summit
  • Live at ICML – meet us in Vancouver for talks, demos, and real-time ML
  • ICYMI – Mistral’s Magistral, IBM enterprise AI, and NinjaTech’s 16-agent super-assistant are all running on Cerebras

Ready to see what next-gen AI really looks like? Let’s dive in 👇

Supernova Takes Paris | July 8–9, 2025


This is your exclusive invitation to Supernova Paris: THE VIP Experience at the RAISE Summit in Paris.

We’re bringing the world’s fastest inference engine and most powerful AI training system to the stage, with speakers including Jessica Liu (Cerebras), Dr. Shant Ayanian (Mayo Clinic), Andrei Papancea (NLX), Robin Rombach (Black Forest Labs), Angela Yeung (Cerebras), and the team from Meta.

Expect two days of provocative talks, live demos, hands-on workshops, and a high-stakes startup competition. From instant UI agents and genomics breakthroughs to Perplexity-style research assistants, this is where the future of AI comes to life. RSVP now

See you in Vancouver!

Come see how we're accelerating model development with wafer-scale compute, making it possible to train smarter and infer faster. Here’s what’s coming:

  • We’re thrilled to have two papers accepted at ICML 2025’s Efficient Systems for Foundation Models Workshop (ES-FoMo III)!
  • Hagay Lupesko’s talk on real-time inference and AI infrastructure
    Café Compute: ICML Edition – swing by for casual conversations and deep ML dives
  • Live demos at our booth – grab your API key and see what 10× inference speed really looks like

Meet Magistral: The Smarter, Faster MoE from Mistral


Real-time reasoning is now on Mistral AI — powered by Cerebras.

Magistral is Mistral AI’s first reasoning model, built for multi-step logic, multilingual tasks, and complex workflows. It’s fast, efficient, and optimized for intelligence at inference.

⚡ Magistral runs 10× faster in Le Chat thanks to Cerebras’ wafer-scale inference — delivering instant, high-quality responses for real-world applications.

🤝 Cerebras x IBM: AI Without Compromise

We’re partnering with IBM to help enterprises accelerate AI adoption—without choosing between bleeding-edge performance and enterprise-grade reliability.

Cerebras brings ultra-fast inference to IBM's watsonx.ai Model Gateway, enabling enterprises to maximize the impact of generative AI on their business.

We join a rich ecosystem of providers—including OpenAI, Anthropic, and more—making it easier than ever to choose the right model for any application, without vendor lock-in.

Get your API Key and get started.

NinjaTech AI, Powered by Cerebras

Going beyond assistance.

Meet Super Ninja — the general AI agent that doesn't just handle one piece of your project; it completes the entire workflow from start to finish.

While other agents get stuck with token limits or require constant hand-holding, Super Ninja runs extensive data analysis, codes and validates full applications, performs comprehensive research, builds websites, and delivers high-quality results in your preferred format.

Powered by Cerebras’ wafer-scale architecture, it operates like having a team of experts working 24/7.

🎥 Watch the demo>

Follow us to get on the list for 2025 events