All Articles  (X)

Clear
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.

🚨

NEW

Product

Introducing MAX 24.6: A GPU Native Generative AI Platform

MAX 24.6 release bog featuring MAX GPU

December 17, 2024

/

Modular Team

Read

🚨

NEW

Developer

Build a Continuous Chat Interface with Llama 3 and MAX Serve

Build a Chat Application with Llama 3 and MAX Serve

December 17, 2024

/

Ehsan M. Kermani

Read

🚨

NEW

🔥

Popular

Developer

Why Magic?

November 5, 2024

/

Bill Welense

Read

🚨

NEW

Developer

Understanding SIMD: Infinite Complexity of Trivial Problems

A deep dive into the complexities of optimizing code for SIMD instruction sets across multiple platforms.

October 25, 2024

/

Ash Vardanian

Read

🚨

NEW

Community

Community Spotlight: Writing Mojo with Cursor

October 10, 2024

/

Julian Acero

Caroline Frasca

Read

🚨

NEW

Developer

Hands-on with Mojo 24.5

Hands-on with Mojo 24.5 and learn how to apply new language features in your code

October 1, 2024

/

Ehsan M. Kermani

Read

🚨

NEW

Product

MAX 24.5 - With SOTA CPU Performance for Llama 3.1

We’re excited to announce the release of MAX 24.5, which ships with significant improvements to Llama 3.1 CPU performance, new Python graph API bindings, our biggest update to Mojo ever, industry-standard packaging, and a clarified license.

September 13, 2024

/

Modular Team

Read

🚨

NEW

Engineering

Announcing stack-pr: an open source tool for managing stacked PRs on GitHub

We are pleased to announce the release of a new tool aimed at simplifying the management of stacked pull requests (PRs) on GitHub - stack-pr. This tool is still in its early development days, but we are excited to share it with the community and welcome your contributions.

July 23, 2024

/

Mikhail Zolotukhin

Read

🚨

NEW

Engineering

Debugging in Mojo🔥

Developer tooling is a big priority for Mojo and MAX, we want to vastly improve the debugging experience compared to the traditional Python, C++, and CUDA stack. Machine learning often requires inspecting the state of a program after a long running process, requiring more control than what "print debugging" gives you. Over time this tooling will extend to GPUs, allowing you to step through CPU code into GPU calls with the same developer experience.

July 16, 2024

/

Jack Clayton

Walter Erquinigo

Read

🚨

NEW

Product

Develop locally, deploy globally

The recent surge in AI application development can be attributed to several factors: (1) advancements in machine learning algorithms that unlock previously intractable use cases, (2) the exponential growth in computational power enabling the training of ever-more complex models, and (3) the ubiquitous availability of vast datasets required to fuel these algorithms. However, as AI projects become increasingly pervasive, effective development paradigms, like those commonly found in traditional software development, remain elusive.

July 9, 2024

/

Modular Team

Read

🤔

No results for this query