- Get link
- X
- Other Apps
Short answer: Yes—AI and LLMs fit into all of these domains, but not as a separate CISSP domain. Instead, they’re treated as technologies you must secure and manage using the same principles.
Let’s map AI/LLMs (like Large Language Models) into each area so it clicks clearly:
1. Risk Management → defines strategy
AI introduces new types of risk.
AI-specific risks
- Data leakage (model reveals sensitive info)
- Bias and unfair decisions
- Hallucinations (incorrect outputs)
- Regulatory/legal issues
Example
A company uses an AI chatbot:
- Risk: it leaks customer data
- Mitigation: restrict training data + monitor outputs
👉 AI becomes part of your enterprise risk register
2. Asset Security → protects data
AI systems consume and generate data, so data protection is critical.
Focus areas
- Training data protection
- Sensitive prompts (user input)
- Output filtering
Example
- Don’t feed confidential data into public AI tools
- Encrypt datasets used to train models
3. Architecture & Engineering → build secure systems
AI systems must be securely designed like any other system.
Key concerns
- Secure model hosting (cloud/on-prem)
- API security (LLM endpoints)
- Model integrity (prevent tampering)
Example
- Protect your AI API with authentication + rate limiting
- Use zero trust architecture around AI services
4. Network Security → protect communication
AI often runs over APIs and cloud services.
Risks
- Intercepted API calls
- Data in transit exposure
Example
- Use HTTPS/TLS for all AI traffic
- Secure connections to AI providers
5. IAM → controls access
Who can use the AI? Who can train or modify it?
Controls
- Limit who can access models
- Restrict admin capabilities
- Track usage (logging)
Example
- Developers can deploy models
- Users can only query them
6. Testing → checks weaknesses
AI needs new types of testing beyond traditional systems.
AI-specific testing
- Prompt injection testing
- Adversarial attacks
- Output validation
Example
Try to “trick” the AI into leaking secrets → fix weaknesses
7. Operations → defends daily
AI systems require continuous monitoring.
Tasks
- Monitor outputs for abuse
- Detect anomalies
- Handle incidents (e.g., data leak)
Example
If AI starts generating harmful responses:
- Disable or retrain it
- Investigate logs
8. Development → prevents flaws early
AI must be built securely from the start.
Secure AI development
- Use trusted datasets
- Validate training data
- Apply secure coding practices to AI pipelines
Example
- Remove sensitive info before training
- Test models before deployment
Big Picture
AI doesn’t replace CISSP domains—it fits inside them:
- AI = new attack surface
- LLMs = new type of asset + system
- Security principles = still the same
Simple way to remember
Think of AI like this:
“It’s just another system—but one that can think, learn, and leak data in new ways.”
If you want, I can go deeper into:
- AI security risks companies are actually facing right now
- How to secure tools like ChatGPT in a workplace
- Or how AI shows up on the CISSP exam (very useful if you're studying)
Got it—you want real AI/LLM examples mapped to each CISSP domain so it’s concrete. Let’s plug in well-known systems like ChatGPT, Google Gemini, and Claude across the framework.
AI/LLMs Across CISSP Domains (with real names)
1. Risk Management → defines strategy
AI examples: ChatGPT, Google Gemini
What to manage
- Data leakage risk
- Legal/compliance (GDPR, HIPAA)
- AI hallucinations
Example
Company policy:
- Employees cannot paste sensitive data into ChatGPT or Gemini
2. Asset Security → protects data
AI examples: Claude, ChatGPT Enterprise
What to protect
- Training datasets
- Prompts (user inputs)
- AI-generated outputs
Example
- Use ChatGPT Enterprise instead of public tools to keep data private
- Encrypt datasets used for training Claude
3. Architecture & Engineering → build secure systems
AI examples: OpenAI API, Azure OpenAI Service
Focus
- Secure APIs
- Model hosting environments
- Zero trust architecture
Example
- Deploy models via Azure OpenAI Service with network isolation
- Protect OpenAI API keys
4. Communication & Network Security → protect data in transit
AI examples: OpenAI API, Google Vertex AI
Risks
- Intercepted API traffic
- Data exposure
Example
- All traffic to Vertex AI uses HTTPS/TLS encryption
- Secure VPN access for internal AI systems
5. Identity & Access Management (IAM) → controls access
AI examples: Azure Active Directory, Okta
Controls
- Who can use AI tools
- Who can train/deploy models
Example
- Use Azure Active Directory to restrict access to Azure OpenAI
- Developers vs regular users have different permissions
6. Security Assessment & Testing → checks weaknesses
AI examples: OpenAI Evals, IBM Adversarial Robustness Toolbox
Testing types
- Prompt injection testing
- Adversarial attacks
- Output validation
Example
- Use OpenAI Evals to test model behavior
- Use IBM tools to simulate attacks
7. Security Operations → defends daily
AI examples: Microsoft Sentinel, Splunk
Tasks
- Monitor AI usage
- Detect abuse or anomalies
Example
- Log all AI queries in Splunk
- Use Sentinel to detect suspicious prompt patterns
8. Software Development Security → prevents flaws early
AI examples: GitHub Copilot, Amazon CodeWhisperer
Risks
- AI-generated insecure code
- Vulnerable dependencies
Example
- Review code suggested by Copilot
- Scan outputs for vulnerabilities
Final takeaway
AI tools like ChatGPT, Claude, and Gemini are:
- Assets → need protection
- Systems → need secure architecture
- Users/tools → need IAM control
- Risks → must be managed
- Get link
- X
- Other Apps
Comments
Post a Comment