Artificial Intelligence | News | Insights | AiThority
[bsfp-cryptocurrency style=”widget-18″ align=”marquee” columns=”6″ coins=”selected” coins-count=”6″ coins-selected=”BTC,ETH,XRP,LTC,EOS,ADA,XLM,NEO,LTC,EOS,XEM,DASH,USDT,BNB,QTUM,XVG,ONT,ZEC,STEEM” currency=”USD” title=”Cryptocurrency Widget” show_title=”0″ icon=”” scheme=”light” bs-show-desktop=”1″ bs-show-tablet=”1″ bs-show-phone=”1″ custom-css-class=”” custom-id=”” css=”.vc_custom_1523079266073{margin-bottom: 0px !important;padding-top: 0px !important;padding-bottom: 0px !important;}”]

Spectrum Labs Launches Content Moderation for Generative AI

Battle of the AIs: Spectrum Labs combats AI-generated hate speech, radicalization and grooming with industry-first AI-powered content moderation

Spectrum Labs, the leader in Text Analysis AI whose tools scale content moderation for games, apps and online platforms, announced today the launch of the world’s first AI content moderation solution that detects and prevents harmful and toxic behavior produced by Generative AI. With the rise of Generative AI, such as ChatGPT, Dall-E, Bard, Stable Diffusion and others, automatic content creation can now be used to produce racist images, propagate hate speech, radicalization, spam, scams, grooming and harassment quickly and on a massive scale with a low time investment by bad actors intent on misusing the new technology.

To begin to address this issue, Spectrum Labs has developed a first-of-its-kind moderation tool for Generative AI content that helps platforms auto-protect their communities from this highly-scalable adversarial content.

Recommended AI: How AI Can Help to Hit Outstanding Debts in the Booming BNPL Industry

“Platforms have already been struggling to sift through mountains of user-generated online content produced each day to identify and remove hateful, illegal and predatory content before Generative AI came along. Now, whether you are a spammer, a child groomer, a bully or recruiter for violent organizations, your job just got a lot easier,” said Justin Davis, CEO of Spectrum Labs. “Fortunately, our existing contextual AI content moderation tools can be adapted to address this new flood of content, because it was built to detect intent, not just a list of keywords or specific phrases, which Generative AI can easily avoid.”

Because Generative AI is designed to create plausible variations of human speech, traditional keyword-based moderation tools are unable to detect if the intent of content is hateful if it never uses specific racist words or phrases. (For example, a children’s story about why one race is superior to another without any racial slurs). Similarly, other existing contextual models that can detect sexual, threatening or toxic content but are unable to detect positive behaviors, such as encouragement, acknowledgment and rapport would redact Generative AI responses about sensitive topics even when the content intended to be helpful, supportive and assuring. (For example, if a user who has suffered from sexual abuse seeks help finding psychological support resources).

Related Posts
1 of 40,899

Even for image-based generative AIs such as Dall-E, automated detection and redaction of toxic human-generated prompts can prevent the creation of libraries of new AI-Generated image and video content that is hateful, threatening, radicalizing and more, while preserving the real-time latency that makes the user experience of generative AI seem so magical.

Recommended AI: Generative AI Will Amplify PR Pros’ Creativity

Future uses of multi-layer real-time AI moderation of Generative AI could include detection of copyright violations, detecting bias within AI-generated content to filter and eliminate biased and problematic training data sources as well as better analytics on what kinds of content people want to make and how it’s used. But right now the company is focused on quickly providing a basic set of tools to help protect users and platforms from a potential tidal wave of toxic content.

“At Spectrum Labs, we’re on a mission to make the internet a safer place for all. We know trust and safety workers are the unsung heroes in this fight, and we’re honored to support them. Together, we can build a safer digital world, one post at a time,” Davis added.

Recommended AI: Immersive Storytelling is the Next Generation of Consumer Experiences

[To share your insights with us, please write to sghosh@martechseries.com]

Comments are closed.