ModCon '23 Keynote Live

The futureof AI development starts here

Mojo 🔥 - the programming language for all AI developers

Available on Mac 🍎, Linux & Windows (WSL)

Unpacking modular (0.1.0) ... Setting up modular (0.1.0) ...
  __  __           _       _
 |  \/  |         | |     | |
 | \  / | ___   __| |_   _| | __ _ _ __
 | |\/| |/ _ \ / _` | | | | |/ _` | '__|
 | |  | | (_) | (_| | |_| | | (_| | |
 |_|  |_|\___/ \__,_|\__,_|_|\__,_|_|

Thanks for installing the Modular CLI!
To get started, type "modular --help".
To install Mojo, type "modular install mojo".

Visit for our getting started guide.
Join our Discord at,
or file issues on Github at
modular install mojo # Found release for @ 0.3.0--release, installing to /root/.modular/pkg/packages.modular.com_mojo
# Downloads complete, setting configs...
# Configs complete, running post-install hooks...
Testing `MODULAR_HOME=/root/.modular` * `/root/.modular/pkg/packages.modular.com_mojo/bin/mojo`...
TEST: `mojo --help`... OK
TEST: `mojo run --help`... OK
TEST: `mojo build test_mandelbrot.mojo`... OK
TEST: `mojo build test_python.mojo`... OK
TEST: `mojo demangle`... OK
reformatted /tmp/tmpmclc50j5/test_format.mojo

All done! ✨ 🍰 ✨
# Installation complete. Mojo is installed under $HOME/.modular. Set $MODULAR_HOME in your current shell to this location, and add $HOME/.modular/pkg/packages.modular.com_mojo/bin to your PATH.
# Downloads complete, setting configs...
# Configs complete, running post-install hooks...
🔥 Mojo installed! 🔥
python3 Throughput of a 128x128 matrix multiplication in Python:
0.00215 GFLOP/s
mojo matmul.🔥 Throughput of a 512x512 vectorized + parallelized matrix multiplication in Mojo:
79.636 GFLOP/s (36870x speedup over Python)

Our unified, extensible platform superpowers your AI

Modular is an integrated, composable suite of tools that simplifies your AI infrastructure so your team can develop, deploy, and innovate faster.

Ai Frameworks

The world’s fastest unified inference engine

Modular’s inference engine unifies AI industry frameworks and hardware, enabling you to deploy to any cloud or on-prem environment with minimal code changes – unlocking unmatched usability, performance, and portability.

Run your models
anywhere, reduce costs

Seamlessly move your workloads to the best hardware for the job without rewriting or recompiling your models. Avoid lock-in and take advantage of cloud price efficiencies and performance improvements without migration costs.

Mojo 🔥 — a new programming language for all AI developers

Mojo is a programming language that combines the usability of Python with the performance of C, unlocking unparalleled programmability of AI hardware and extensibility of AI models.

def softmax(lst):
  norm = np.exp(lst - np.max(lst))
  return norm / norm.sum()

struct NDArray:
  def max(self) -> NDArray:
    return self.pmap(SIMD.max)

struct SIMD[type: DType, width: Int]:
  def max(self, rhs: Self) -> Self:
    return (self >= rhs).select(self, rhs)

Deploy the largest models in the world on our stack

The Modular Compute Platform dynamically partitions models with billions of parameters and distributes their execution across multiple machines, enabling unparalleled efficiency, scale, and reliability for the largest workloads.


Get help from the people who know Modular best

As a community member, you can chat with the Modular team directly on Discord, and as an enterprise customer, you get direct support from industry experts to keep you running and enable you to scale to your next challenges.

Today at 4:17PM
How can we help you?

Deploy on the fastest unified infrastructure on the planet

Modular unlocks state-of-the-art latency, efficiency, and throughput, helping you productionize larger models and realize massive cost savings on your cloud bill.

125 qps
Modular Engine
* Model
AWS c6g.4xlarge
Batch Size

Modular's cloud
compute platform


Notebooks for training on the largest compute clusters using Python & Mojo 🔥 for highly optimized workloads.


Utilize our managed environment, or Bring your own cloud (BYOC), for seamless workload management.


Detailed machine performance and metrics data to provide end-to-end insight into your AI workloads.


Leverage our easy-to-use web UI or CLI tooling to seamlessly manage your training and deployment workflows.


Enterprise security & encryption for your data to be secured at rest and in transit on your data stores.

Why Modular?

Built by the world’s AI experts,

Our team has built most of the world’s existing AI infrastructure, including TensorFlow, PyTorch, TPUs, and MLIR, and launched software like Swift and LLVM. Now we’re focused on rebuilding AI infrastructure for the world.

Reinvented from the ground up

To unlock the next wave of AI innovation, we need a “first principles” approach to the lowest layers of the AI stack. We can’t pile on more and more layers of complexity on top of already over-complicated existing solutions.

Built with generality in mind

Natively multi-model, multi-framework, multi-hardware, and multi-cloud — our infrastructure scales from the largest clusters down to the smallest edge devices and in-between.

Infrastructure that just works

We build technology that meets you where you are at. You shouldn’t have to rewrite your models or application code, grapple with confusing converters, or be a hardware expert to take advantage of state-of-the-art technology.

Built for you

Move beyond Big Tech’s trickle-down infrastructure. Get direct access to industry experts that will help solve any issue you have with our infrastructure and make sure we’re meeting your SLA/SLOs.

Ready to get started?

Sign up to gain access to Modular’s infrastructure.

Read the docs