During our upcoming community meeting on March 10th, @clattner will be answering questions from the community on our recent Democratizing AI Compute series!
One of the strategies that was outlined for CUDA’s success was that it ran and worked on developer grade hardware so developers could learn how to use it.
Currently only very expensive GPUs are supported for MAX, is Modular looking to follow a similar path to CUDA and support GPUs that people have access to sooner rather than later to speed adoption? Or will premium GPUs be the main focus? Why not start with developer grade hardware and then work up?
What lessons from the Objective C->Swift transition can be applied to CUDA->Mojo/MAX?
Not CUDA specific, but have there been any emergent features in Mojo that you particularly like?
Julia’s ecosystem has offered a mature GPU programming model for NVIDIA and AMD for a few years. Other than “Julia is not Python”, what’s your take on the technical reasons behind Julia not reaching the level of adoption in the AI community to be considered a viable alternative to CUDA?
Many of the AI ASIC vendors like Cerebras, Samba Nova Systems, AWS Trainium, Groq, etc… have built out their own AI graph compilers and low-level kernel programming interfaces. Essentially, they have taken a page out of NVIDIA’s playbook and are building their own “CUDA” specifically for their hardware. What’s your take on how Modular’s vision fits within this growing fragmentation of the accelerated computing software stack?
Question: Can you characterize in a few dot points how exactly MAX (as a project) differs from all of the previous projects that aimed to offer a hardware-independent compute platform? I’ve heard several detailed answers at this point (that includes your blog series), but not a succinct answer that I can easily share with others.
In fact, it would be cool to have those dot points listed on the MAX page of the Modular website. I expect programmers would find such a comparison useful.