AI Threats Demand New Approach to Security Designs

Artificial Intelligence (AI) is advancing rapidly, raising concerns about potential threats it may pose. In response, a top US official emphasized the need to safeguard systems from the beginning rather than add protections later.

"We've normalized a world where technology products come off the line full of vulnerabilities and then consumers are expected to patch those vulnerabilities. We can't live in that world with AI," Jen Easterly, director of the US Cybersecurity and Infrastructure Security Agency, said.

"It is too powerful. It is moving too fast," she said in an interview after a meeting in Ottawa with Sami Khoury, head of Canada's Centre for Cyber Security.

Said Khoury, "We have to look at security throughout the lifecycle of that AI capability."

The Software and Information Industry Association (SIIA), a trade association comprised of over 380 global tech companies, recently created guidelines for the responsible development of artificial intelligence (AI) tools for education.

Now, the United States is among 18 countries to endorse new guidelines for AI cyber security. These guidelines, developed in Britain, emphasize secure AI design, development, deployment, and maintenance. As a result of this agreement, AI developers will collaborate with governments to establish a standardized process for identifying and mitigating risks.

Click over to Xchange and read more about last week's AI Safety Summit, including the 20-page document that provides general recommendations for companies developing and/or deploying AI systems.