Blog

Democratizing AI Compute Series
Go behind the scenes of the AI industry with Chris Lattner

Modverse #47: MAX 25.2 and an evening of GPU programming at Modular HQ
MAX 25.2 is turning heads — and for good reason. This powerful update delivers industry-leading performance for large language models on NVIDIA GPUs, all without CUDA. MAX 25.2 builds on the momentum of 25.1 and introduces major upgrades to help you build GenAI systems that are faster, leaner, and easier to scale.

MAX is here! What does that mean for Mojo🔥?
When we started Modular, building a programming language wasn't our goal, it ended up being a solution to a set of problems. Specifically, as we were building our platform to unify the world’s ML/AI infrastructure, we realized that programming across the entire stack was too complicated.

Mojo 🔥 Advent of Code 2023
Advent of Code is an annual online coding event that takes place during the holiday season, starting on December 1st and continuing until December 25th. It consists of a series of small programming puzzles that are released daily, each becoming available at midnight EST (UTC-5). Participants from around the world compete for fun, honing their coding abilities and often learning new programming concepts in the process.

Community Spotlight: How I built llama2.🔥 by Aydyn Tairov
Mojo SDK was released in September 2023. As someone who relies on the simplicity of Python and also cares about high performance delivered by languages like C, I was excited to try out Mojo. I felt the same joy and thrill I had experienced when I first discovered programming and ran “Hello World” in QBasic and Turbo Pascal.
Democratizing Compute Series
Go behind the scenes of the AI industry in this blog series by Chris Lattner. Trace the evolution of AI compute, dissect its current challenges, and discover how Modular is raising the bar with the world’s most open inference stack.
Matrix Multiplication on Blackwell
Learn how to write a high-performance GPU kernel on Blackwell that offers performance competitive to that of NVIDIA's cuBLAS implementation while leveraging Mojo's special features to make the kernel as simple as possible.
No items found within this category
We couldn’t find anything. Try changing or resetting your filters.

Get started guide
Install MAX with a few commands and deploy a GenAI model locally.
Read Guide
Browse open models
500+ models, many optimized for lightning-fast performance
Browse models





.jpeg)



