Max Platform accelerates AI.
It's Programmable
We rebuilt the modern AI software stack, from the ground up, to boost any AI pipeline, on any hardware.
Download & install in your terminal now
Available on Linux & Windows (WSL) now, Mac 🍎 soon!
By downloading, you accept our Terms.
Programmable, performant & portable
Full programmability
MAX is built on top of Mojo from the ground up to empower AI engineers to unlock the full potential of AI hardware by combining the usability of Python, the safety of Rust, and the performance of C.
Unparalleled performance
MAX unlocks state-of-the-art performance for your AI models. Extend and optimize your AI pipelines without having to rewrite them, with unparalleled performance using a next generation compiler.
Seamless portability
Seamlessly move your models and AI pipelines to any hardware target, maximizing your performance to cost ratio and avoiding vendor lock-in.
Unparalleled latency & cost savings
MAX unlocks state-of-the-art latency and throughput for your AI pipeline, including generative models, helping you quickly productionize AI pipelines and realize massive cost savings on your cloud bill.
vs
Modular is 1.7x faster than TensorFlow when running [Stable Diffusion-UNet] on [CPU]
vs
Modular is 1.7x faster than PyTorch when running [Stable Diffusion-UNet] on [CPU]
An integrated AI
developer experience
The Modular Accelerated Xecution (MAX) platform is a unified set of tools and libraries that provides everything you need to deploy low-latency, high-throughput, real-time AI inference pipelines into production.
Max components
Mojo
Learn about MojoA programming language that combines the usability of Python with the performance of C, unlocking unparalleled programmability of AI hardware and extensibility of AI models for all AI engineers.
MAX Engine
Learn about MAX EngineA model inference runtime and API library that executes all your AI pipelines on any hardware with unparalleled performance and cost savings.
MAX Serving
Learn about MAX ServingA model serving library for the MAX Engine that provides full interoperability with existing serving systems (e.g. Triton) and that seamlessly deploys within existing container infrastructure (e.g., Kubernetes).
Latest about Modular
Why Modular?
01
BUILT BY THE WORLD’S AI EXPERTS
Our team has built most of the world’s existing AI infrastructure, including TensorFlow, PyTorch, ONNX, and XLA, and we’ve built and scaled dev tools like Swift, LLVM, and MLIR. Now we’re focused on rebuilding AI infrastructure for the world.
02
Reinvented from the ground up
To unlock the next wave of AI innovation, we started with a “first principles” approach to building the lowest layers of the AI stack. We can’t pile on more and more layers of complexity on top of already over-complicated existing solutions.
03
Infrastructure that just works
We build technology that meets you where you are. We don’t require you to rewrite your models, workflows, or application code, grapple with confusing converters, or be a hardware expert to take advantage of bleeding-edge technology.
Try Max right now
Up and running, for free, in 5 minutes.