Intel Editorial Unlocking the Potential of Generative AI
Intel’s goal is to support all artificial intelligence models, including generative AI, with responsible perspectives and principles.
Generative artificial intelligence (AI) describes the algorithms used to create new data that can resemble human-generated content, including audio, code, images, text, simulations and videos. This technology is trained with existing content and data, creating the potential for applications like natural language processing, computer vision, metaverse and speech synthesis.
Generative AI is not new. It’s the tech that created voice assistants, infinitely evolving games, and chatbots.
Recommended AI News: Blotout Announces New Partnership with Fastly to Improve Meta Ad Spend with Blotout’s EdgeTag
Recently, several powerful AI tools, such as ChatGPT and DALL·E 2, have been used as generative AI, enabling people to build apps on them, and to experiment and post the results online. You may have seen one in action when social profiles were flooded with historically inspired photos from the MyHeritage AI Time Machine, which uses AI to generate hypothetical pictures of a person’s appearance if they lived in different eras. While some have shared concerns about the potential of generative AI to threaten jobs, there is a greater opportunity to responsibly use generative AI to improve people’s efficiency and creativity.
Intel’s Trusted Media team works to build generative AI applications with humans in mind. The team strives to create AI that improves people’s lives, limits harm and builds tools to make other technologies more natural. And it does it all with responsibility at each step of the process, not just the end.
Intel’s Approach to Generative AI
In the past few years, generative AI has become more powerful – and therefore more capable of doing problematic things in a more convincing and realistic manner.
For example, generative models for deepfakes aim to impersonate people. We defend against this in two ways. The first is with our deepfake detection algorithms integrated into our real-time platform. FakeCatcher, the core of the system, can detect fake videos with a 96% accuracy rate, enabling users to distinguish between real and fake content in real time. The second is through our responsible generators, one of which makes human puzzles. As opposed to creating images by training on real people’s faces, this generator mixes and matches regions (nose of person A, mouth of person B, eyes of person C, etc.) to create an entirely new face that does not already exist in a data set.
Recommended AI News: Complete Verkada Platform is Now Available Across the UK and Europe
We believe that AI should not only prevent harm but also enhance lives. To fulfill this vision, the team’s speech synthesis project aims to enable people who have lost their voices to talk again. This technology is used in Intel’s I Will Always Be Me digital storybook project in partnership with Dell Technologies, Rolls-Royce and the Motor Neuron Disease (MND) Association. The interactive website allows anyone diagnosed with MND or any disease expected to affect their speaking ability to record their voice to be used on an assistive speech device.
Finally, the Trusted Media research team is also working on using generative AI to make 3D experiences more realistic. For example, Intel’s CARLA is an open source urban driving simulator developed to support the development, training and validation of autonomous driving systems. Using generative AI, the scenes surrounding the driver would look more realistic and natural. Intel’s generative AI approaches also simplify 3D creation and rendering workflows, saving hours for 3D artists and making games run much more quickly.
Recommended AI News: Raising the Ceiling of Code-Free Programming for Robotics and Industrial Automation
[To share your insights with us, please write to firstname.lastname@example.org]
Comments are closed.