AI just got faster with MAX 25.1!
Prefix caching and paged attention boost LLM performance, offline batch inference improves latency & load time, and MAX Builds is your go-to hub for GenAI models, recipes, and packages.
Start building smarter today and check out the full release blog post: Modular: MAX 25.1 - Introducing MAX Builds
Donβt miss our upcoming livestream on LinkedIn to learn more about what makes 25.1 awesome: Introducing MAX 25.1 | LinkedIn