Course Outline

Day 1: Foundations and Core Threats

Module 1: Introduction to OWASP GenAI Security Project (1 hour)

Learning Objectives:

  • Understand the transition from the OWASP Top 10 to the specific security challenges of Generative AI.
  • Explore the ecosystem and resources provided by the OWASP GenAI Security Project.
  • Identify key differences between traditional application security and AI security.

Topics Covered:

  • Overview of the mission and scope of the OWASP GenAI Security Project.
  • Introduction to the Threat Defense COMPASS framework.
  • Understanding the AI security landscape and regulatory requirements for government.
  • Comparison of AI attack surfaces with traditional web application vulnerabilities.

Practical Exercise: Setting up the OWASP Threat Defense COMPASS tool and performing an initial threat assessment.

Module 2: OWASP Top 10 for LLMs - Part 1 (2.5 hours)

Learning Objectives:

  • Master the first five critical vulnerabilities in Large Language Models (LLMs).
  • Understand attack vectors and exploitation techniques.
  • Apply practical mitigation strategies.

Topics Covered:

LLM01: Prompt Injection

  • Techniques for direct and indirect prompt injection.
  • Hidden instruction attacks and cross-prompt contamination.
  • Practical examples of jailbreaking chatbots and bypassing safety measures.
  • Defense strategies: Input sanitization, prompt filtering, differential privacy.

LLM02: Sensitive Information Disclosure

  • Training data extraction and system prompt leakage.
  • Model behavior analysis for sensitive information exposure.
  • Privacy implications and regulatory compliance considerations.
  • Mitigation: Output filtering, access controls, data anonymization.

LLM03: Supply Chain Vulnerabilities

  • Security of third-party models and plugins.
  • Compromised training datasets and model poisoning.
  • Vendor risk assessment for AI components.
  • Secure model deployment and verification practices.

Practical Exercise: Hands-on lab demonstrating prompt injection attacks against vulnerable LLM applications and implementing defensive measures.

Module 3: OWASP Top 10 for LLMs - Part 2 (2 hours)

Topics Covered:

LLM04: Data and Model Poisoning

  • Techniques for manipulating training data.
  • Modifying model behavior through poisoned inputs.
  • Backdoor attacks and data integrity verification.
  • Prevention: Data validation pipelines, provenance tracking.

LLM05: Improper Output Handling

  • Insecure processing of LLM-generated content.
  • Code injection through AI-generated outputs.
  • Cross-site scripting via AI responses.
  • Output validation and sanitization frameworks.

Practical Exercise: Simulating data poisoning attacks and implementing robust output validation mechanisms.

Module 4: Advanced LLM Threats (1.5 hours)

Topics Covered:

LLM06: Excessive Agency

  • Risks associated with autonomous decision-making and boundary violations.
  • Management of agent authority and permissions.
  • Unintended system interactions and privilege escalation.
  • Implementing guardrails and human oversight controls.

LLM07: System Prompt Leakage

  • Vulnerabilities related to the exposure of system instructions.
  • Credential and logic disclosure through prompts.
  • Attack techniques for extracting system prompts.
  • Securing system instructions and external configuration.

Practical Exercise: Designing secure agent architectures with appropriate access controls and monitoring.

Day 2: Advanced Threats and Implementation

Module 5: Emerging AI Threats (2 hours)

Learning Objectives:

  • Understand cutting-edge AI security threats.
  • Implement advanced detection and prevention techniques.
  • Design resilient AI systems against sophisticated attacks for government.

Topics Covered:

LLM08: Vector and Embedding Weaknesses

  • Vulnerabilities in RAG systems and vector database security.
  • Embedding poisoning and similarity manipulation attacks.
  • Adversarial examples in semantic search.
  • Securing vector stores and implementing anomaly detection.

LLM09: Misinformation and Model Reliability

  • Detection and mitigation of hallucinations.
  • Bias amplification and fairness considerations.
  • Fact-checking and source verification mechanisms.
  • Content validation and human oversight integration.

LLM10: Unbounded Consumption

  • Resource exhaustion and denial-of-service attacks.
  • Rate limiting and resource management strategies.
  • Cost optimization and budget controls.
  • Performance monitoring and alerting systems.

Practical Exercise: Building a secure RAG pipeline with vector database protection and hallucination detection.

Module 6: Agentic AI Security (2 hours)

Learning Objectives:

  • Understand the unique security challenges of autonomous AI agents.
  • Apply the OWASP Agentic AI taxonomy to real-world systems for government.
  • Implement security controls for multi-agent environments.

Topics Covered:

  • Introduction to Agentic AI and autonomous systems.
  • OWASP Agentic AI Threat Taxonomy: Agent Design, Memory, Planning, Tool Use, Deployment.
  • Multi-agent system security and coordination risks.
  • Tool misuse, memory poisoning, and goal hijacking attacks.
  • Securing agent communication and decision-making processes.

Practical Exercise: Threat modeling exercise using OWASP Agentic AI taxonomy on a multi-agent customer service system for government.

Module 7: OWASP Threat Defense COMPASS Implementation (2 hours)

Learning Objectives:

  • Master the practical application of Threat Defense COMPASS for government.
  • Integrate AI threat assessment into organizational security programs.
  • Develop comprehensive AI risk management strategies.

Topics Covered:

  • Deep dive into Threat Defense COMPASS methodology.
  • OODA Loop integration: Observe, Orient, Decide, Act.
  • Mapping threats to MITRE ATT&CK and ATLAS frameworks.
  • Building AI Threat Resilience Strategy Dashboards.
  • Integration with existing security tools and processes.

Practical Exercise: Complete threat assessment using COMPASS for a Microsoft Copilot deployment scenario in government.

Module 8: Practical Implementation and Best Practices (2.5 hours)

Learning Objectives:

  • Design secure AI architectures from the ground up for government.
  • Implement monitoring and incident response for AI systems.
  • Create governance frameworks for AI security.

Topics Covered:

Secure AI Development Lifecycle:

  • Security-by-design principles for AI applications in government.
  • Code review practices for LLM integrations.
  • Testing methodologies and vulnerability scanning.
  • Deployment security and production hardening.

Monitoring and Detection:

  • AI-specific logging and monitoring requirements for government.
  • Anomaly detection for AI systems.
  • Incident response procedures for AI security events.
  • Forensics and investigation techniques.

Governance and Compliance:

  • AI risk management frameworks and policies for government.
  • Regulatory compliance considerations (GDPR, AI Act, etc.).
  • Third-party risk assessment for AI vendors.
  • Security awareness training for AI development teams.

Practical Exercise: Design a complete security architecture for an enterprise AI chatbot including monitoring, governance, and incident response procedures for government.

Module 9: Tools and Technologies (1 hour)

Learning Objectives:

  • Evaluate and implement AI security tools for government.
  • Understand the current AI security solutions landscape.
  • Build practical detection and prevention capabilities for government.

Topics Covered:

  • AI security tool ecosystem and vendor landscape for government.
  • Open-source security tools: Garak, PyRIT, Giskard.
  • Commercial solutions for AI security and monitoring.
  • Integration patterns and deployment strategies.
  • Tool selection criteria and evaluation frameworks.

Practical Exercise: Hands-on demonstration of AI security testing tools and implementation planning for government.

Module 10: Future Trends and Wrap-up (1 hour)

Learning Objectives:

  • Understand emerging threats and future security challenges for government.
  • Develop continuous learning and improvement strategies.
  • Create action plans for organizational AI security programs for government.

Topics Covered:

  • Emerging threats: Deepfakes, advanced prompt injection, model inversion.
  • Future OWASP GenAI project developments and roadmap.
  • Building AI security communities and knowledge sharing for government.
  • Continuous improvement and threat intelligence integration.

Action Planning Exercise: Develop a 90-day action plan for implementing OWASP GenAI security practices in participants' organizations for government.

Requirements

  • A general understanding of web application security principles
  • Basic familiarity with artificial intelligence and machine learning concepts
  • Experience with security frameworks or risk assessment methodologies is preferred

Audience for Government

  • Cybersecurity professionals
  • AI developers
  • System architects
  • Compliance officers
  • Security practitioners
 14 Hours

Number of participants


Price per participant

Testimonials (1)

Upcoming Courses

Related Categories