Efficient conversion, runtime, and optimization for on-device machine learning.
LiteRT isn't just new; it's the next generation of the world's most widely deployed machine learning runtime. It powers the apps you use every day, delivering low latency and high privacy on billions of devices.

Trusted by the most critical Google apps

100K+ applications, billions of global users

LiteRT highlights

Deploy via LiteRT

Use .tflite pre-trained models or convert PyTorch, JAX or TensorFlow models to .tflite
Optionally Quantize the model
Pick desired accelerator and run on LiteRT

Samples, models, and demo

Tutorials show you how to use LiteRT with complete, end-to-end examples.
Pre-trained, out-of-the-box Gen AI models.
A gallery that showcases on-device ML/GenAI use cases