The Future of AI Reliability Is Open and Collaborative: Introducing Guardrails Hub
February 15, 2024
Categories:
The Future of AI Reliability Is Open and Collaborative: Introducing Guardrails Hub
Guardrails Hub: Where Reliability Meets Community
Community Initiative towards AI reliability
Getting Started with Guardrails Hub
I'm incredibly excited to share a major milestone in Guardrails AI's journey: the launch of Guardrails Hub! This open-source platform is the culmination of our deep belief that for AI to reach its full potential, it needs to be inherently reliable, responsible and open.
The AI Reliability Conundrum
Generative AI and large language models (LLMs) are game-changers. They hold the keys to unlocking incredible human creativity and streamlining countless processes. But as a seasoned AI practitioner, how unpredictable these models can be. Even the most sophisticated LLMs can generate unexpected results, sometimes with factual inaccuracies or biases.
These "surprises" create a trust barrier that's slowing down the wider adoption of AI in critical applications. Guardrails AI was founded to change this.
Guardrails Hub: Where Reliability Meets Community
Think of Guardrails Hub as a one-stop-shop for developers to find, build, and share advanced validation techniques called "validators." These validators act as checks and balances for AI applications, enforcing reliability, correctness, and alignment with your organization's specific standards.
What makes Guardrails Hub truly special is its open-source nature:
- Ready-to-use Validators: Browse a growing collection of pre-built validators, spanning tasks like bias detection to ensure the factuality of AI-generated content. Find what you need and implement it immediately for added assurance.
- Harness Collective Wisdom: We envision a vibrant community of AI developers working together. Contribute your own validators to the hub, expanding the knowledge base and making AI reliability solutions accessible to everyone.
- Build It Your Way: Combine validators like building blocks into "guards." This powerful capability lets you design custom reliability layers tailored to the unique risks and complexities of your AI applications.
Use Cases Across Industries
Guardrails Hub's impact reaches far and wide. Here are just a few ways it can transform how we utilize AI:
- Healthcare: Validate AI-assisted diagnoses to mitigate patient risk
- Finance: Ensure compliance with regulations for critical financial decision-making
- Customer Service: Keep AI responses on-brand and aligned with company values
Community Initiative towards AI reliability
At Guardrails AI, we're firm believers that the path to trustworthy AI is a collaborative one. Guardrails Hub empowers developers everywhere to work together in solving the AI reliability puzzle. This launch is just the beginning. As the hub grows, we hope it becomes an invaluable resource, propelling us toward an AI future characterized by reliable technologies. We invite you to join us on this journey!
- Explore Guardrails Hub: https://hub.guardrailsai.com
- Contribute your validators and shape the future of AI reliability
- Share your stories of how Guardrails Hub is powering your reliable AI initiatives
Getting Started with Guardrails Hub
Here's how to get up and running quickly:
1. Install the Hub CLI
The command-line interface (CLI) lets you download guardrails and manage your Hub configuration. Install it with a simple command:
pip install guardrails-ai
2. Configure Your Settings
Get your API key from the Guardrails Hub website and use the following command to set it up:
guardrails configure
3. Download Your First Guardrail
Guardrails are the core rules that protect your AI applications. Let's install one that validates text formats:
guardrails hub install hub://guardrails/regex_match
4. Put Your Guardrail to Work
Here's a Python example showing how to use the regex_match guardrail to ensure a name starts with a capital letter:
from guardrails.hub import RegexMatch from guardrails import Guard val = Guard().use( RegexMatch(regex="^[A-Z][a-z]*$") ) val.parse("Sarah") # Passes! val.parse("john") # Fails!
5. Supercharge Your Safety with Multiple Guardrails
Combine guardrails to address diverse issues like toxicity and competitor mentions:
from guardrails import Guard from guardrails.hub import CompetitorCheck, ToxicLanguage # ... install guardrails from the Hub ... guard = Guard().use( CompetitorCheck( competitors=["Apple", "Samsung"] ) ).use( ToxicLanguage() ) # Protects against competitor mentions AND toxic content
That's it! You're now using Guardrails Hub to make your AI applications safer and more reliable.
I'm truly grateful for the passionate AI community driving this transformation. Together, we'll make AI applications dependable and ensure this powerful technology remains a force for good in the world.
Sincerely,
Shreya Rajpal
CEO, Guardrails AI
Similar ones you might find interesting
Handling fix results for streaming
How we handle fix results for streaming in Guardrails.
How we rewrote LLM Streaming to deal with validation failures
The new pipeline for LLM Streaming now includes ways to merge fixes across chunks after validating.
Latency and usability upgrades for ML-based validators
The numbers behind our validators