Modular partners with AMD, and much more!

Watch the June 10th,2025 announcement below

Unlock speed-of-light inference on GPUs with an open, unified AI stack

Scale your AI workloads from 1 GPU to megaclusters, all without CUDA - all on a single platform that grows with you from startup to enterprise

NEW: Modular Comic Book

GPU pain, perfectly roasted in comic form

What can you do with Modular?

500+ GenAI models

Customizable

Open source implementation

Get going fast

Tiny containers

Multi-hardware support

Multi-cloud deployment

Full hardware control

Great documentation

State-of-the-art performance

Hardware agnostic

Multi hardware

Write once, deploy anywhere

Backwards compatible

"At AWS we are focused on powering the future of AI by providing the largest enterprises and fastest-growing startups with services that lower their costs and enable them to move faster. The MAX Platform supercharges this mission for our millions of AWS customers, helping them bring the newest GenAI innovations and traditional AI use cases to market faster."

Bratin Saha

VP of Machine Learning & AI services

Read More

"We're truly in a golden age of AI, and at AMD we're proud to deliver world-class compute for the next generation of large-scale inference and training workloads… We also know that great hardware alone is not enough. We've invested deeply in open software with ROCm, empowering developers and researchers with the tools they need to build, optimize, and scale AI systems on AMD. This is why we are excited to partner with Modular… and we’re thrilled that we can empower developers and researchers to build the future of AI."

Vamsi Boppana

SVP of AI, AMD

Read More

"Our collaboration with Modular is a glimpse into the future of accessible AI infrastructure. Our API now returns the first 2 seconds of synthesized audio on average ~70% faster compared to vanilla vLLM based implementation, at just 200ms for 2 second chunks. This allowed us to serve more QPS with lower latency and eventually offer the API at a ~60% lower price than would have been possible without using Modular’s stack."

Igor Poletaev

Chief Science Officer - Inworld

Read More

"Modular’s team is world class. Their stack slashed our inference costs by 80%, letting our customer dramatically scale up. They’re fast, reliable, and real engineers who take things seriously. We’re excited to partner with them to bring down prices for everyone, to let AI bring about wide prosperity."

Evan Conrad

CEO - San Francisco Compute

Read More

"Modular allows Qwerky to write our optimized code and confidently deploy our solution across NVIDIA and AMD solutions without the massive overhead of re-writing native code for each system."

Evan Owen

CTO, Qwerky AI

Read More

Unlock speed of light performance at every level

Industry leading performance out-of-the-box

Achieve state-of-the-art speed and efficiency immediately—no complex tuning, just install and run.

Scale from 1 GPU, to unlimited

Kubernetes-native control plane, router, and substrate specially-designed for large-scale distributed AI serving. It supports multi-model management, prefill-aware routing, disaggregated compute and cache.

Biggest open source library of GPU code

Boost performance across the entire AI stack with innovations like compiler, kernels, runtime, and more

Drive Total Cost Savings

MAX's speed is proven to bring your overall AI budget down. Connect with us to find out how much you can save

Benchmark our performance

Run real benchmarks, validate SOTA performance, and confidently scale AI workloads

Native portability, from CPU to GPU and beyond

Write once, deploy everywhere

Forward & backward compatible

Future proof your AI stack

Spend less on compute

Data center & consumer GPUs

Take control, customize at every layer—from model to silicon

Customize down to the silicon

Fine-tune custom ops and runtimes for performance and precision at the deepest levels.

Bring cutting-edge research to production

Boost performance across the entire AI stack with innovations like compiler, kernels, runtime, and more.

Open source

Access, modify, and extend every layer of the stack without restrictions—true freedom to innovate.

Vertically integrated

One cohesive system for better performance, easier debugging, and more consistent behavior across development and deployment.

Start simple and grow

Our engineers have created a platform that works out of the box. Get comfortable with our tools, then optimize further down to the metal.

"after wrestling with CUDA drivers for years, it felt surprisingly… smooth. No, really: for once I wasn’t battling obscure libstdc++ errors at midnight or re-compiling kernels to coax out speed. Instead, I got a peek at writing almost-Pythonic code that compiles down to something that actually flies on the GPU."

Sanika

1/1

"This is about unlocking freedom for devs like me, no more vendor traps or rewrites, just pure iteration power. As someone working on challenging ML problems, this is a big thing."

Jayesh

1/1

“The more I benchmark, the more impressed I am with the MAX Engine.”

justin_76273

1/1

"The performance gains and the whole Mojo direction felt genuinely exciting, especially from the perspective of someone who’s always looking to squeeze more out of Python-based workflows."

Vishnu

1/1

“It’s fast which is awesome. And it’s easy. It’s not CUDA programming...easy to optimize.”

dorjeduck

1/1

"Having a single framework in Mojo that compiles both lower-level code and a higher level metaprogramming layer using dependent types feels like black magic, so exciting!"

Theodore Papamarkou

1/1

Less waiting, more building

Minimize dependencies

Fewer external libraries for faster installs, smaller packages, and smoother deployments.

GPU Acceleration, without CUDA

Get your AI stack running without complex GPU drivers like CUDA.  Say goodbye to setup issues and dependency conflicts.

Lightning-fast compile times

Experience effortless prototyping—no lag, no friction, just fast iteration. Pythonic code that flies on the GPU.

Build with Python, optimize with Mojo🔥

Develop easily with familiar Python, then switch to Mojo for higher speed and precision when needed, without changing platforms.

Best developer tools

With native VS Code extensions (100k installs!), and fully featured LLDB debugger, develop faster than ever.

Developer Approved 👍

“Max installation on Mac M2 and running llama3 in (q6_k and q4_k) was a breeze! Thank you Modular team!”

NL

"Mojo is Python++. It will be, when complete, a strict superset of the Python language. But it also has additional functionality so we can write high performance code that takes advantage of modern accelerators."

jeremyphoward

“Tired of the two language problem. I have one foot in the ML world and one foot in the geospatial world, and both struggle with the 'two-language' problem. Having Mojo - as one language all the way through would be awesome.”

fnands

“Mojo can replace the C programs too. It works across the stack. It’s not glue code. It’s the whole ecosystem.”

scrumtuous

“Max installation on Mac M2 and running llama3 in (q6_k and q4_k) was a breeze! Thank you Modular team!”

NL

"Mojo is Python++. It will be, when complete, a strict superset of the Python language. But it also has additional functionality so we can write high performance code that takes advantage of modern accelerators."

jeremyphoward

“Tired of the two language problem. I have one foot in the ML world and one foot in the geospatial world, and both struggle with the 'two-language' problem. Having Mojo - as one language all the way through would be awesome.”

fnands

“Mojo can replace the C programs too. It works across the stack. It’s not glue code. It’s the whole ecosystem.”

scrumtuous

“Max installation on Mac M2 and running llama3 in (q6_k and q4_k) was a breeze! Thank you Modular team!”

NL

"Mojo is Python++. It will be, when complete, a strict superset of the Python language. But it also has additional functionality so we can write high performance code that takes advantage of modern accelerators."

jeremyphoward

“Tired of the two language problem. I have one foot in the ML world and one foot in the geospatial world, and both struggle with the 'two-language' problem. Having Mojo - as one language all the way through would be awesome.”

fnands

“Mojo can replace the C programs too. It works across the stack. It’s not glue code. It’s the whole ecosystem.”

scrumtuous

“Max installation on Mac M2 and running llama3 in (q6_k and q4_k) was a breeze! Thank you Modular team!”

NL

"Mojo is Python++. It will be, when complete, a strict superset of the Python language. But it also has additional functionality so we can write high performance code that takes advantage of modern accelerators."

jeremyphoward

“Tired of the two language problem. I have one foot in the ML world and one foot in the geospatial world, and both struggle with the 'two-language' problem. Having Mojo - as one language all the way through would be awesome.”

fnands

“Mojo can replace the C programs too. It works across the stack. It’s not glue code. It’s the whole ecosystem.”

scrumtuous

“What @modular is doing with Mojo and the MaxPlatform is a completely different ballgame.”

scrumtuous

“I am focusing my time to help advance @Modular. I may be starting from scratch but I feel it’s what I need to do to contribute to #AI for the next generation.”

mytechnotalent

“Mojo and the MAX Graph API are the surest bet for longterm multi-arch future-substrate NN compilation”

pagilgukey

“A few weeks ago, I started learning Mojo 🔥 and MAX. Mojo has the potential to take over AI development. It's Python++. Simple to learn, and extremely fast.”

svpino

“Mojo destroys Python in speed. 12x faster without even trying. The future is bright!”

svpino

“It’s fast which is awesome. And it’s easy. It’s not CUDA programming...easy to optimize.”

dorjeduck

“I tried MAX builds last night, impressive indeed. I couldn't believe what I was seeing... performance is insane.”

drdude81

“What @modular is doing with Mojo and the MaxPlatform is a completely different ballgame.”

scrumtuous

“I am focusing my time to help advance @Modular. I may be starting from scratch but I feel it’s what I need to do to contribute to #AI for the next generation.”

mytechnotalent

“Mojo and the MAX Graph API are the surest bet for longterm multi-arch future-substrate NN compilation”

pagilgukey

“A few weeks ago, I started learning Mojo 🔥 and MAX. Mojo has the potential to take over AI development. It's Python++. Simple to learn, and extremely fast.”

svpino

“Mojo destroys Python in speed. 12x faster without even trying. The future is bright!”

svpino

“It’s fast which is awesome. And it’s easy. It’s not CUDA programming...easy to optimize.”

dorjeduck

“I tried MAX builds last night, impressive indeed. I couldn't believe what I was seeing... performance is insane.”

drdude81

“What @modular is doing with Mojo and the MaxPlatform is a completely different ballgame.”

scrumtuous

“I am focusing my time to help advance @Modular. I may be starting from scratch but I feel it’s what I need to do to contribute to #AI for the next generation.”

mytechnotalent

“Mojo and the MAX Graph API are the surest bet for longterm multi-arch future-substrate NN compilation”

pagilgukey

“A few weeks ago, I started learning Mojo 🔥 and MAX. Mojo has the potential to take over AI development. It's Python++. Simple to learn, and extremely fast.”

svpino

“Mojo destroys Python in speed. 12x faster without even trying. The future is bright!”

svpino

“It’s fast which is awesome. And it’s easy. It’s not CUDA programming...easy to optimize.”

dorjeduck

“I tried MAX builds last night, impressive indeed. I couldn't believe what I was seeing... performance is insane.”

drdude81

“What @modular is doing with Mojo and the MaxPlatform is a completely different ballgame.”

scrumtuous

“I am focusing my time to help advance @Modular. I may be starting from scratch but I feel it’s what I need to do to contribute to #AI for the next generation.”

mytechnotalent

“Mojo and the MAX Graph API are the surest bet for longterm multi-arch future-substrate NN compilation”

pagilgukey

“A few weeks ago, I started learning Mojo 🔥 and MAX. Mojo has the potential to take over AI development. It's Python++. Simple to learn, and extremely fast.”

svpino

“Mojo destroys Python in speed. 12x faster without even trying. The future is bright!”

svpino

“It’s fast which is awesome. And it’s easy. It’s not CUDA programming...easy to optimize.”

dorjeduck

“I tried MAX builds last night, impressive indeed. I couldn't believe what I was seeing... performance is insane.”

drdude81

"Mojo gives me the feeling of superpowers. I did not expect it to outperform a well-known solution like llama.cpp."

Aydyn

"C is known for being as fast as assembly, but when we implemented the same logic on Mojo and used some of the out-of-the-box features, it showed a huge increase in performance... It was amazing."

Aydyn

“I'm excited, you're excited, everyone is excited to see what's new in Mojo and MAX and the amazing achievements of the team at Modular.”

Eprahim

“I'm very excited to see this coming together and what it represents, not just for MAX, but my hope for what it could also mean for the broader ecosystem that mojo could interact with.”

strangemonad

"It worked like a charm, with impressive speed. Now my version is about twice as fast as Julia's (7 ms vs. 12 ms for a 10 million vector; 7 ms on the playground. I guess on my computer, it might be even faster). Amazing."

Adalseno

“The more I benchmark, the more impressed I am with the MAX Engine.”

justin_76273

"Mojo gives me the feeling of superpowers. I did not expect it to outperform a well-known solution like llama.cpp."

Aydyn

"C is known for being as fast as assembly, but when we implemented the same logic on Mojo and used some of the out-of-the-box features, it showed a huge increase in performance... It was amazing."

Aydyn

“I'm excited, you're excited, everyone is excited to see what's new in Mojo and MAX and the amazing achievements of the team at Modular.”

Eprahim

“I'm very excited to see this coming together and what it represents, not just for MAX, but my hope for what it could also mean for the broader ecosystem that mojo could interact with.”

strangemonad

"It worked like a charm, with impressive speed. Now my version is about twice as fast as Julia's (7 ms vs. 12 ms for a 10 million vector; 7 ms on the playground. I guess on my computer, it might be even faster). Amazing."

Adalseno

“The more I benchmark, the more impressed I am with the MAX Engine.”

justin_76273

"Mojo gives me the feeling of superpowers. I did not expect it to outperform a well-known solution like llama.cpp."

Aydyn

"C is known for being as fast as assembly, but when we implemented the same logic on Mojo and used some of the out-of-the-box features, it showed a huge increase in performance... It was amazing."

Aydyn

“I'm excited, you're excited, everyone is excited to see what's new in Mojo and MAX and the amazing achievements of the team at Modular.”

Eprahim

“I'm very excited to see this coming together and what it represents, not just for MAX, but my hope for what it could also mean for the broader ecosystem that mojo could interact with.”

strangemonad

"It worked like a charm, with impressive speed. Now my version is about twice as fast as Julia's (7 ms vs. 12 ms for a 10 million vector; 7 ms on the playground. I guess on my computer, it might be even faster). Amazing."

Adalseno

“The more I benchmark, the more impressed I am with the MAX Engine.”

justin_76273

"Mojo gives me the feeling of superpowers. I did not expect it to outperform a well-known solution like llama.cpp."

Aydyn

"C is known for being as fast as assembly, but when we implemented the same logic on Mojo and used some of the out-of-the-box features, it showed a huge increase in performance... It was amazing."

Aydyn

“I'm excited, you're excited, everyone is excited to see what's new in Mojo and MAX and the amazing achievements of the team at Modular.”

Eprahim

“I'm very excited to see this coming together and what it represents, not just for MAX, but my hope for what it could also mean for the broader ecosystem that mojo could interact with.”

strangemonad

"It worked like a charm, with impressive speed. Now my version is about twice as fast as Julia's (7 ms vs. 12 ms for a 10 million vector; 7 ms on the playground. I guess on my computer, it might be even faster). Amazing."

Adalseno

“The more I benchmark, the more impressed I am with the MAX Engine.”

justin_76273

Latest Blog Posts

MAX on GPU waiting list

Be the first to get lightning fast inference speed on your GPUs. Be the envy of all your competitors and lower your compute spend.