Portkey Brings AI Guardrails on Top of its Open Source AI Gateway
Portkey is filling a critical gap in productionizing AI apps – by bringing Guardrails on top of the AI Gateway, AI teams can now ship to production confidently.
LLMs are brittle – not just in API uptimes or their inexplicable 400/500 errors, but also in their core behavior. You can get a response with a 200 status code that completely errors out for your app’s pipeline due to mismatched output. We’ve long thought about the problem of fixing LLM outputs at Portkey, and wondered if we could bring some of the Guardrail abstractions towards the Gateway directly. With Portkey’s Guardrails, we now complete the loop on building robust & reliable AI apps that behave EXACTLY as you want, every time.
There was a crucial piece missing when it came to productionizing AI systems – orchestrating the LLM’s behavior based on guardrail verdicts on the inputs & outputs. We now solve it with this update!”— Rohit Agarwal
As Chip Huyen (VP of AI & OSS at Voltron Data) shared recently, Guardrails around a model Gateway solve 2 problems at once —
1. Identify/Fix faulty LLM outputs and
2. Orchestrate the request based on Guardrails results to actually make the app work.
Also Read: Extreme Networks and Intel Join Forces to Drive AI-Centric Product Innovation
We are teaming up with multiple AI Guardrail leaders in the industry to do exactly this – bring their SOTA AI Guardrails on top of Portkey’s open source AI Gateway, and make it incredibly easy for developers to use them in production. The Gateway orchestrates your LLM requests based on Guardrail results, and makes your app behave exactly as it should.
Comments are closed.