AI Collaborative Research Institute Launched

Security

A trio of companies is launching a new research institute whose intended purpose is to strengthen privacy and trust for decentralized artificial intelligence (AI). 

The Private AI Collaborative Research Institute, originally established by Intel‘s University Research & Collaboration Office (URC), is launching as a joint project involving digital security and privacy products vendor Avast and AI software-defined secure computing hardware services company Borsetta.

“As AI continues to grow in strength and scope, we have reached a point where action is necessary, not just talk,” said Michal Pechoucek, CTO at Avast.

“We’re delighted to be joining forces with Intel and Borsetta to unlock AI’s full potential for keeping individuals and their data secure.’’

By decentralizing AI, the companies aim to protect privacy and security, free inaccessible data from silos, and maintain efficiency. The trio said that centralized training can be easily attacked by modifying data anywhere between collection and the cloud. 

Another security issue surrounding contemporary AI stems from the limitations of Federated Machine Learning, a technique used to train an algorithm across multiple decentralized edge devices. 

While today’s federated AI can access data at the edge, the team behind the Institute said that this technique cannot simultaneously guarantee accuracy, privacy, and security.

“Research into responsible, secure, and private AI is crucial for its true potential to be realized,” said Richard Uhlig, Intel Senior Fellow, vice president and director of Intel Labs.

Borsetta said it was inspired to join the collaboration by its strong belief in driving a privacy-preserving framework to support the future hyperconnected world empowered by AI. 

“The mission of the Private AI Collaborative Institute is aligned with our vision for future proof security where data is provably protected with edge computing services that can be trusted,” said Pamela Norton, CEO of Borsetta.

“Trust will be the currency of the future, and we need to design AI embedded edge systems with trust, transparency, and security while advancing the human-driven values they were intended to reflect.”

A call for research proposals issued earlier this year has resulted in the selection of nine research projects at eight universities in Belgium, Canada, Germany, Singapore, and the United States to receive Institute support.

Products You May Like

Articles You May Like

A New Wiper Malware Was Behind Recent Cyberattack On Iranian Train System
New Apple technology will warn parents and children about sexually explicit photos in Messages
The pandemic effect is slowing
Evolve With XDR – The Modern Approach to SecOps
Keycloak tutorial: How to secure different application types

Leave a Reply

Your email address will not be published. Required fields are marked *