Security News > 2023 > November > Released: AI security guidelines backed by 18 countries
The UK National Cyber Security Centre has published new guidelines that can help developers and providers of AI-powered systems "Build AI systems that function as intended, are available when needed, and work without revealing sensitive data to unauthorised parties."
The Guidelines for secure AI system development cover four key stages of the development lifecycle of machine learning applications.
Secure design hinges on all persons involved - system owners, developers, users - being aware of the unique security risks facing AI systems and being taught to avoid them.
"Model the threats to your system, and design your system for security as well as functionality and performance," the guidelines instruct.
"The new UK-led guidelines are the first of their kind to be agreed globally. They will help developers of any systems that use AI make informed cyber security decisions at every stage of the development process - whether those systems have been created from scratch or built on top of tools and service provided by others," the NCSC pointed out.
We urge all stakeholders to read these guidelines to help them make informed decisions about the design, deployment and operation of their machine learning AI systems.
News URL
https://www.helpnetsecurity.com/2023/11/27/ai-security-guidelines/
Related news
- Businesses turn to private AI for enhanced security and data management (source)
- CIOs want a platform that combines AI, networking, and security (source)
- Generative AI in Security: Risks and Mitigation Strategies (source)
- Unlocking the value of AI-powered identity security (source)
- Can Security Experts Leverage Generative AI Without Prompt Engineering Skills? (source)
- Eliminating AI Deepfake Threats: Is Your Identity Security AI-Proof? (source)
- Apple Opens PCC Source Code for Researchers to Identify Bugs in Cloud AI Security (source)
- Best AI Security Tools: Top Solutions, Features & Comparisons (source)
- How AI Is Changing the Cloud Security and Risk Equation (source)
- Google claims Big Sleep 'first' AI to spot freshly committed security bug that fuzzing missed (source)