nference Accelerates Self-Supervised Language Model Training With Cerebras CS-2 System
The ability to harness vast amounts of health data using advanced AI technology will lead to new discoveries and insights needed to improve patient care
-Cerebras Systems, the pioneer in high performance artificial intelligence (AI) compute, and nference, an AI-driven health technology company, announced a collaboration to accelerate natural language processing (NLP) for biomedical research and development by orders of magnitude with a Cerebras CS-2 system installed at the nference headquarters in Cambridge, Mass.
Download Our Top Whitepaper : Building Reliable and Secure Fintech Systems in 2022
“AI is driving an exponential increase in demand for compute”
The vast amounts of health data that lie within patient records, scientific papers, medical imagery, and genomic databases could be critical to advancing health outcomes. Unfortunately, this information is nearly impossible for data scientists and machine learning (ML) researchers to access, as it exists in unstructured, siloed, and incompatible forms, forcing researchers to sift through it manually. While data accessibility is a fundamental challenge in healthcare today, newer AI architectures such as transformer models can assist by processing data from various sources, de-identifying it, and converting it into structured, usable intelligence.
nference uses transformer AI models to employ self-supervised learning from large volumes of unstructured data without labels, translating vast amounts of health data into information that can be used to discover insights and drive research. However, training large models is complex, computationally intensive, and time-consuming, often requiring large clusters of conventional processors. A key feature of the Cerebras CS-2 architecture is the ability for data scientists and ML researchers to use data of longer sequence lengths than is practical using smaller, conventional processors – this is particularly relevant to research being conducted at nference.
“nference was founded to help solve complex medical problems and improve health outcomes by unlocking insights contained within biomedical data while protecting individual patient privacy,” said Ajit Rajasekharan, Chief Technology Officer, nference. “Our solution uses transformer models to help researchers and clinicians make sense of siloed and inaccessible health data, leading to new discoveries and findings that can impact patient outcomes. With Cerebras’s powerful CS-2 system, we can train transformer models with much longer sequence lengths than we could before, enabling us to iterate more rapidly and build better, more insightful models.”
Browse The Complete News About Aithority: Imperva Welcomes Dan Neault as SVP and GM, Data Security
“AI is driving an exponential increase in demand for compute,” said Andy Hock, Vice President of Product, Cerebras Systems. “As we have recently demonstrated across multiple customers and published work, the Cerebras CS-2 is orders of magnitude faster than legacy alternatives. This orders-of-magnitude performance advantage comes from the Cerebras Wafer Scale Engine (WSE-2), the world’s largest and most powerful AI processor. The WSE-2 is purpose built with 850,000 AI-optimized cores to accelerate the models of today and unlock future models not practical or possible on legacy infrastructure. The partnership and our work with nference is a great example of this, where their team – equipped with a CS-2 – is pushing the boundaries of AI to accelerate biomedical research and discovery to improve health outcomes.”
The Cerebras CS-2 system delivers the deep learning compute performance of hundreds of graphics processing units in a cluster, with the programming ease and efficiency of a single system. Powered by the largest and fastest processor ever built – the 2.6 trillion transistor second-generation Cerebras Wafer-Scale Engine (WSE-2) – the CS-2 delivers more AI-optimized compute cores, fast memory, and fabric bandwidth than any other deep learning processor in existence. A CS-2 delivers the wall-clock compute performance of many tens to hundreds of GPUs with the programming ease and efficiency of a single device.
Read More About Aithority News : Uber, IACLEA, and It’s On Us Partner to Share Spring Break Rideshare Safety Tips
[To share your insights with us, please write to sghosh@martechseries.com]
Comments are closed.