You can run Gemma models completely on-device with the MediaPipe LLM Inference
API. The
LLM Inference API acts as a wrapper for large language models, enabling you run
Gemma models on-device for common text-to-text generation tasks like information
retrieval, email drafting, and document summarization.
Try the LLM Inference API with MediaPipe
Studio, a
web-based application for evaluating and customizing on-device models.
For more information on deploying Gemma to web browsers with
the LLM Inference API, see the
LLM Inference guide for Web.
To learn more about the MediaPipe LLM Inference capabilities, see the
LLM inference guide.
[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Missing the information I need","missingTheInformationINeed","thumb-down"],["Too complicated / too many steps","tooComplicatedTooManySteps","thumb-down"],["Out of date","outOfDate","thumb-down"],["Samples / code issue","samplesCodeIssue","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2024-12-04 UTC."],[],[],null,["# Deploy Gemma in web browsers\n\nYou can run Gemma models completely on-device with the [MediaPipe LLM Inference\nAPI](https://ai.google.dev/edge/mediapipe/solutions/genai/llm_inference). The\nLLM Inference API acts as a wrapper for large language models, enabling you run\nGemma models on-device for common text-to-text generation tasks like information\nretrieval, email drafting, and document summarization.\n\nTry the LLM Inference API with [MediaPipe\nStudio](https://mediapipe-studio.webapps.google.com/demo/llm_inference), a\nweb-based application for evaluating and customizing on-device models.\n\nFor more information on deploying Gemma to web browsers with\nthe LLM Inference API, see the\n[LLM Inference guide for Web](/edge/mediapipe/solutions/genai/llm_inference/web_js).\nTo learn more about the MediaPipe LLM Inference capabilities, see the\n[LLM inference guide](https://ai.google.dev/edge/mediapipe/solutions/genai/llm_inference)."]]