Evolving the Responsible Generative AI Toolkit with new tools for every LLM

Share

By Ryan Mullins, Research Engineer and RAI Toolkit Tech Lead OCT 23, 2024

Building AI responsibly is crucial. That’s why we created the Responsible GenAI Toolkit, providing resources to design, build, and evaluate open AI models. And we’re not stopping there! We’re now expanding the toolkit with new features designed to work with any LLMs, whether it’s Gemma, Gemini, or any other model. This set of tools and features empower everyone to build AI responsibly, regardless of the model they choose.

Here’s what’s new:

SynthID Text: watermarking and detecting AI-generated content

Is it difficult to tell if a text was written by a human or generated by AI? SynthID Text has you covered. This technology allows you to watermark and detect text generated by your GenAI product.

How it works: SynthID watermarks and identifies AI-generated content by embedding digital watermarks directly into AI-generated text.

Open source for developers: SynthID for text is accessible to all developers through Hugging Face and the Responsible GenAI Toolkit.


Learn more:

  • Explore the SynthID Technology Page for a comprehensive overview of SynthID’s capabilities across all modalities.


Use it today:

We invite the open source community to help us expand the reach of SynthID Text across frameworks, based on the implementations above. Reach out on GitHub or Discord with questions.


Model Alignment: refine your prompts with LLM assistance

Crafting prompts that effectively enforce your business policies is crucial for generating high-quality outputs.

The Model Alignment library helps you refine your prompts with support from LLMs.

Provide feedback about how you want your model’s outputs to change as a holistic critique or a set of guidelines.

Use Gemini or your preferred LLM to transform your feedback into a prompt that aligns your model’s behavior with your application’s needs and content policies.


Use it today:

  • Experiment with the interactive demo in Colab and see how Gemini can help align and improve prompts for Gemma.
  • Access the library on PyPI.

Prompt Debugging: streamline LIT deployment on Google Cloud

Debugging prompts is essential for responsible AI development. We’re making it easier and faster with an improved deployment experience for the Learning Interpretability Tool (LIT) on Google Cloud.

  • Efficient, versatile model serving: Leverage LIT’s new model server container to deploy any Hugging Face or Keras LLM with support for generation, tokenization, and salience scoring on Cloud Run GPUs.
  • Expanded connectivity from the LIT App: Seamlessly connect to self-hosted models or Gemini via the Vertex API (generation only).
Learning Interpretability Tool (LIT) on Google Cloud.

Info from Google for Developers

Ryan Mullins

I am a Software Engineer on the People + AI Research team, part of Google Research, where I study human-AI collaboration and AI/ML interpretability and explainability. Previously I was a Senior Researcher at Aptima, where I studied human-AI collaboration for software security, planning, and information analysis.

I research and develop visualizations, experiences, and analytics for Human-AI systems working across a variety of domains, including cybersecurity, planning, and problem-solving.

0 0 votes
Article Rating
Subscribe
Notify of
guest

0 Comments
Inline Feedbacks
View all comments
0
Would love your thoughts, please comment.x
()
x