Deploy Gemma on mobile devices

You can run Gemma models on mobile devices with the MediaPipe LLM Inference API. The LLM Inference API acts as a wrapper for large language models, enabling you run Gemma models on-device for common text-to-text generation tasks like information retrieval, email drafting, and document summarization.

The LLM Inference API is available on the following mobile platforms:

To learn more, refer to the MediaPipe LLM Inference documentation.