[bsfp-cryptocurrency style=”widget-18″ align=”marquee” columns=”6″ coins=”selected” coins-count=”6″ coins-selected=”BTC,ETH,XRP,LTC,EOS,ADA,XLM,NEO,LTC,EOS,XEM,DASH,USDT,BNB,QTUM,XVG,ONT,ZEC,STEEM” currency=”USD” title=”Cryptocurrency Widget” show_title=”0″ icon=”” scheme=”light” bs-show-desktop=”1″ bs-show-tablet=”1″ bs-show-phone=”1″ custom-css-class=”” custom-id=”” css=”.vc_custom_1523079266073{margin-bottom: 0px !important;padding-top: 0px !important;padding-bottom: 0px !important;}”]

CalypsoAI Launches Security Index; Provides First Comprehensive Safety Ranking of Major GenAI models

One Unified AI Security Platform

AI Security Leader’s Inference Red-Team Product Compromises World’s Largest Foundation Models

CalypsoAI, the leader in securing GenAI for enterprises, launched the CalypsoAI Security Leaderboard, the world’s first index of all the major AI models based on their security performance.

The CalypsoAI Security Leaderboard ranks all the major models on their ability to withstand advanced security attacks and presents a risk-to-performance (RTP) ratio as well as a valuable cost of security (CoS) metric. CalypsoAI compiled the Leaderboard after stress-testing AI models with its new Inference Red-Team solution, which combines Agentic Warfare with automated attacks.

Also Read: AiThority Interview with James Alger, Co-founder and COO of Qloo

“Our Inference Red-Team product has successfully broken all the world-class GenAI models that exist today,” said Donnchadh Casey, CEO of CalypsoAI. “Many organizations are adopting AI without understanding the risks to their business and clients; moving forward, the CalypsoAI Security Leaderboard provides a benchmark for business and technology leaders to integrate AI safely and at scale.”

CalypsoAI Inference Red-Team delivers automated, scalable assessments that run real-world attacks to proactively identify vulnerabilities and create a CalypsoAI Security Index-scored (CASI) AI inventory. By combining Agentic Warfare with a continuously updated library of extensive signature attacks and operational testing, Inference Red-Team empowers organizations to enhance governance, ensure compliance, and maintain proactive, secure and resilient AI systems.

Related Posts
1 of 42,016

“GenAI presents unparalleled opportunities for business transformation, yet security, governance and compliance risks remain significant barriers to adoption,” said Amit Levinstein, VP Security Architecture & CISO at CYE. “CalypsoAI’s breakthrough red teaming solution is a quantum leap in AI security and provides the hard evidence executives need, and confidence they desire, to deploy AI applications safely.”

With over 70 years of experience in security and AI among its engineering team, CalypsoAI recognized this need for clear, actionable reports identifying vulnerabilities to enable security teams to strengthen their AI systems and stay ahead of the latest threats.

Also Read: AiThority Interview with Jie Yang, Co-founder and CTO of Cybever

“CalypsoAI Inference Red-Team introduces Agentic Warfare as the latest method of finding security gaps in GenAI models,” said James White, President and CTO of CalypsoAI. “It signals the end of inefficient and inconsistent manual red teaming, which is still used by even the largest AI model developers, and with the CalypsoAI Security Leaderboard, closes a significant gap in publicly-available information on model security.”

The AI threat landscape is constantly evolving and most companies aren’t equipped to test their AI systems in the way attackers will; with its Agentic Warfare capability, CalypsoAI Inference Red-Team leverages AI-powered adversaries to engage dynamically, exposing hidden weaknesses that static tests miss. The agentic nature of this solution ensures teams can confidently select the safest models before deploying applications and as use cases evolve.

“CalypsoAI Red Team is a game-changer for businesses leading or even experimenting in AI,” said Jay Choi, CEO Typeform. “In many initiatives, we saw great opportunities to innovate by embedding Generative AI; but we constantly struggled with risk of security.  It was the unknown unknowns that were the most challenging. Calypso Inference Red-Teaming really addresses that risk so companies integrating AI-driven technology no longer need to choose between security and innovation.”

[To share your insights with us as part of editorial or sponsored content, please write to psen@itechseries.com]

Comments are closed.