Artificial Intelligence | News | Insights | AiThority
[bsfp-cryptocurrency style=”widget-18″ align=”marquee” columns=”6″ coins=”selected” coins-count=”6″ coins-selected=”BTC,ETH,XRP,LTC,EOS,ADA,XLM,NEO,LTC,EOS,XEM,DASH,USDT,BNB,QTUM,XVG,ONT,ZEC,STEEM” currency=”USD” title=”Cryptocurrency Widget” show_title=”0″ icon=”” scheme=”light” bs-show-desktop=”1″ bs-show-tablet=”1″ bs-show-phone=”1″ custom-css-class=”” custom-id=”” css=”.vc_custom_1523079266073{margin-bottom: 0px !important;padding-top: 0px !important;padding-bottom: 0px !important;}”]

BrainChip Execs Talk Advantages of Putting AI on Sensors at the Edge as Part of Embedded Vision Summit

BrainChip Holdings Ltd., a leading provider of ultra-low power, high-performance AI technology, will present the Expert Bar session “Can You Put AI at the Sensor? at the Embedded Vision Summit May 27 at 11:30 a.m. PDT. The virtual presentation will be broadcast live as well as be available on-demand for attendees of the event.

“Can You Put AI at the Sensor? (Not the Edge of the Cloud!)”

The BrainChip team will help viewers better understand the requirements of sensors at the edge and how challenges associated with traditional machine learning make it difficult to properly enable these devices. Deploying a solution that leverages advanced neuromorphic computing as the engine for intelligent AI at the edge can be better used to solve critical problems such as privacy, security, latency, and low power requirements, while providing key features, such as one-shot learning and computing on the device itself, without dependency on the cloud.

Recommended AI News: Snyk to Provide Developer-First Security in Atlassian Bitbucket Cloud

Related Posts
1 of 41,061

“Cloud use for AI might be effective in a data center setting but relying on it for the millions of edge sensors being deployed in emerging ‘smart’ markets is a recipe for disaster,” said Anil Mankar, Chief Development Officer at BrainChip. “How do those devices overcome latency in communicating with the cloud? Next-generation AI needs a solution that will provide resources to edge devices. We look forward to sharing with attendees of the Embedded Vision Summit how our Akida Neural Processing Unit has been developed to address these concerns and provide true device intelligence without the need for the cloud.”

Recommended AI News: Sogou Launched World’s First AI Sign Language News Anchor

BrainChip is delivering on next-generation demands by achieving efficient, effective AI functionality. The company’s Akida neuromorphic processors are revolutionary advanced neural networking processors that bring artificial intelligence to the edge in a way that existing technologies are not capable. The solution is high-performance, small, ultra-low power and enables a wide array of edge capabilities. The Akida (NSoC) and intellectual property, can be used in applications including Smart Home, Smart Health, Smart City and Smart Transportation. These applications include, but are not limited to, home automation and remote controls, industrial IoT, robotics, security cameras, sensors, unmanned aircraft, autonomous vehicles, medical instruments, object detection, sound detection, odor and taste detection, gesture control and cybersecurity. The Akida NSoC is designed for use as a stand-alone embedded accelerator or as a co-processor, and includes interfaces for ADAS sensors, audio sensors, and other IoT sensors. Akida brings AI processing capability to edge devices for learning, enabling personalization of products without the need for retraining.

Recommended AI News: Seedtag to Offer Brands Precise Contextual Advertising During Euro 2020

Comments are closed.