Unlock your hardware design with software that provides generality and usability for AI developers.

Modular's software stack will take care of all upstream integrations with AI frameworks, graph optimizations, and more, so you can focus on the differentiating features of your hardware. Own your code generation, performance and feature set. It's still early, but we're excited about the future.

Integrating with the Modular stack gets you:


Effortlessly extend your customer reach to all popular AI frameworks.


Autofusion and graph optimizations boost your hardware performance.


We support the ever-changing AI ecosystem, including the long-tail of operators and models.


You own your performance and can utilize the full capabilities of your hardware.


Your hardware “just works” often with only a few weeks of development.

Integration is simple

Modular handles integration & packaging
end-user tools

Modular handles offload, kernel fusion, compilation,, caching, & developer tools

You provide LLVM or MLIR code generation, and extend the stack with Mojo 🔥 kernels

You design the hardware
and have in-house expertise

Modular handles framework support and end-user tooling

Modular handles the ever-changing world of AI frameworks, models, and operators, providing you with a single cross-framework integration point for your hardware stack.

Full framework, model,
and op support

  1. Integration with TensorFlow, PyTorch, plus variants like ONNX and TorchScript
  2. Full generality of models, including dynamic shapes, sparsity, custom ops, etc.
  3. Thousands of long-tail operators needed for compatibility

Compiler transformations
and developer tooling

  1. Kernel fusion and other performance optimizations
  2. Automatic graph partitioning for distributed inference
  3. Standardized and hackable tools

Plug your hardware in at the graph or code generation level

Bring your own code gen backend for CPU, GPU, or DSPs, and everything just works. For CGRA, FPGA, or other exotic hardware, start with our standardized operator set and extend the system to your needs.

Focus your effort on ML operators you care about

  1. Modular provides a library of customizable kernels and microkernels written in Mojo 🔥
  2. Focus on the subset of the problem for your hardware
  3. Everything else just works — we provide fallback legacy long-tail kernels for compatibility

Extend the system
as needed

  1. Add new Mojo 🔥 kernels if you don’t find what you’re looking for, and enable your customers to the same
  2. Use Mojo 🔥 to directly inject MLIR, C/C++ or assembly code as needed
  3. Leverage standardized operators and tools to write high-level graph transformations

Interested in partnering with us?

Contact us to discuss how we can work together.

Learn more about our platform