As we just announced, Modular’s platform is now available via pip install modular
. This new deployment mechanism unlocks an exciting capability: using MAX in Google Colab.
The stable 25.3 release of modular
will work with the L4 and A100 GPUs in Colab Pro. Thanks to the fantastic work of @ssslakter , introductory support for the T4 GPUs in the free tier of Colab is present in the latest nightlies. Full models don’t yet run on T4 GPUs, but you can run MAX graphs in the free tier of Colab.
This is a sample Jupyter Notebook that you can load into Colab and try out for yourself. This uses the Python interfaces in MAX to construct and run a simple graph, as well as perform inference using a supported large language model. This uses the latest nightly modular
package, so the first section of the notebook will run on T4 GPUs in the free tier of Colab. If you have Colab Pro, you can run the LLM portion at the end of the notebook on an L4 or A100 GPU.
Google Colab support is still experimental and very new, but if you have any questions about the process, feel free to ask them below.