Modular acquires BentoML to deliver production AI in the cloud!  - Read more

August 21, 2025

Modverse #50: Modular Platform 25.5, Community Meetups, and Mojo's Debut in the Stack Overflow Developer Survey

Caroline Frasca

Community

This past month brought a wave of community projects and milestones across the Modular ecosystem!

Modular Platform 25.5 landed with Large Scale Batch Inference, leaner packages, and new integrations that make scaling AI easier than ever. It’s already powering production deployments like SF Compute’s Large Scale Inference Batch API, cutting costs by up to 80% while supporting more than 15 leading models.

Around the world, the Modular community has been busy: experimenting with Gaussian splatting, building probabilistic data structures in Mojo, digging into GPU puzzles, and hosting meetups. Mojo even made its debut in the 2025 Stack Overflow Developer Survey, just two years after launch, another milestone in its rapid adoption.

Let’s take a look at everything the Modular universe made possible.

Blogs, Tutorials, and Videos

  • At our July Community Meeting, Maxim presented his newly merged work on Hasher-based hashing, and we heard from all three Modular Hack Weekend winners: Martin Vuyk, who implemented Fast Fourier Transform in Mojo, Seth Stadick, who built Mojo-Lapper, a GPU-accelerated interval overlap detection library, and Thomas Trenty, who created QLabs, a GPU-powered quantum circuit simulator.
  • Mojo made its debut in the 2025 Stack Overflow Developer Survey, just two years after launch.
  • We’ve partnered with SF Compute to launch the Large Scale Inference Batch API, offering up to 80% cost savings, support for 15+ leading models, and real-time GPU spot pricing. Watch the launch video, read the blog post, and get in touch if you’d like to try it.
  • Modular Platform 25.5 is here, built for developers who need scale. Highlights include:
    • Large Scale Batch Inference, a high-throughput OpenAI-compatible API powered by Mammoth, already live in production with SF Compute.
    • Standalone Mojo Conda packages, leaner (<700 MB) MAX Serving packages, a fully open-source MAX Graph API, and seamless MAX ↔ PyTorch integration. Full details: MAX changelog | Mojo changelog.
  • The August Community Meeting featured mojo-regex optimizations from Manuel, Apple GPU updates from Amir, and live Q&A with the team.
  • Big things are happening on August 28th at our Los Altos HQ: talks and networking with Modular and Inworld AI! Chris Lattner on the open future of compute and Mojo, Feifan Fan on voice AI in production, and Chris Hoge on matmul optimization. Grab your seat.
__wf_reserved_inherit

Awesome MAX + Mojo

__wf_reserved_inherit
__wf_reserved_inherit

Open-Source Contributions

If you’ve recently had your first PR merged, message the Modular team in the forum to claim your epic Modular swag!

Check out the recently merged contributions from our amazing community members:

Read more from Modular

View all blogs

Build the future of AI with Modular

View Editions
  • Person with blonde hair using a laptop with an Apple logo.

    Get started guide

    Install MAX with a few commands and deploy a GenAI model locally.

    Read Guide
  • Magnifying glass emoji with black handle and round clear lens.

    Browse open models

    500+ models, many optimized for lightning-fast performance

    Browse models
No items found.