ShieldGemma

ShieldGemma is a set of instruction tuned models for evaluating the safety of text and images against a set of defined safety policies. You can use this model as part of a larger implementation of a generative AI application to help evaluate and prevent generative AI applications from violating safety policies. The ShieldGemma family of models is provided with open weights to allow you to fine-tune it for your specific use case.

ShieldGemma 2 is a 4B parameter model built to label images for safety.

ShieldGemma 1 is built on Gemma 2 in 2B, 9B, and 27B parameter sizes.

  • Evaluate the safety of prompt input and output responses against a set of defined safety policies.
  • ShieldGemma models are provided with open weights and can be fine-tuned for your specific use case.

ShieldGemma 2

ShieldGemma's 2 model card contains detailed information about the model implementation, evaluations, model usage and limitations, and more.
View more code, Colab notebooks, information, and discussions about ShieldGemma on Kaggle.
Run a working example for using ShieldGemma to evaluate images.

ShieldGemma 1

ShieldGemma's model card contains detailed information about the model implementation, evaluations, model usage and limitations, and more.
View more code, Colab notebooks, information, and discussions about ShieldGemma on Kaggle.
Run a working example for using ShieldGemma to evaluate text prompt input and output.