Get in Touch

Course Outline

Day 1
Anatomy of a Modern AI Agent

Beyond chatbots, agents are autonomous systems capable of reasoning and acting.

Paradigms include reactive, proactive, hybrid, and goal-directed agents.

Core components encompass perception, planning, memory, tool use, and action.

Single-agent versus multi-agent design tradeoffs are explored for government applications.

Agent Frameworks and the Modern Stack

Frameworks such as LangChain, LlamaIndex, AutoGen, and CrewAI are compared with classical frameworks like JADE and SPADE.

Choosing a framework based on production requirements for government use is discussed.

Tool Calling, Function Calling, and Structured Outputs

Hands-on session: scaffolding a single Python agent with tool calls.

Multi-Agent System Architectures

Designs for centralized, decentralized, hybrid, and layered multi-agent systems (MAS) are examined.

Communication protocols such as FIPA ACL and modern equivalents are covered.

Coordination patterns, including planning, negotiation, and synchronization, are discussed.

Emergent behavior and self-organization in agent populations are explored.

Decision-Making and Learning in Agents

Game theory for cooperative and competitive interactions among agents is introduced.

Reinforcement learning in multi-agent environments is covered.

Transfer learning and knowledge sharing across agents are discussed.

Conflict resolution and trust between coordinating agents are addressed.

Day 2
Multi-Modal Foundations for Agents

Multi-modal AI as a unified workflow across text, image, speech, and video is explored.

Leading multi-modal models such as GPT-4 Vision, Gemini, Claude, and Whisper are discussed.

Techniques for combining modalities within an agent's reasoning loop are covered.

Tradeoffs in latency, cost, and accuracy in multi-modal pipelines are examined.

Building the Perception Layer

Image processing techniques for classification, captioning, and object detection are introduced.

Speech recognition using Whisper ASR and streaming transcription is covered.

Text-to-speech synthesis and natural voice interaction are discussed.

Connecting perception outputs to LLM-driven reasoning and tool selection is explored.

Hands-On - Building a Multi-Modal Agent in Python

Defining the agent's task, context window, and tool inventory is covered.

Wiring up GPT-4 Vision and Whisper APIs end-to-end is demonstrated.

Implementing memory, state, and conversation management is addressed.

Safely adding tool calls that produce real-world side effects is discussed.

Hands-On - Orchestrating a Multi-Agent System

Composing specialized agents with AutoGen or CrewAI is demonstrated.

Defining roles, responsibilities, and inter-agent communication protocols is covered.

Resource allocation and coordination in a simulated environment are addressed.

Logging agent reasoning, tool calls, and decisions for inspection and audit is discussed.

Day 3
Threat Surface of Production AI Agents

Unique vulnerabilities of agentic AI compared to traditional software are explored.

Attack surfaces at data, model, prompt, tool, output, and interface layers are discussed.

Threat modeling for agent-based systems with autonomous tool use is covered.

Comparing AI cybersecurity practices to traditional cybersecurity is addressed.

Adversarial Attacks Hands-On

Methods such as FGSM, PGD, and DeepFool for adversarial examples are introduced.

White-box versus black-box attack scenarios are covered.

Model inversion and membership inference attacks are discussed.

Data poisoning and backdoor injection during training are addressed.

Prompt injection, jailbreaking, and tool misuse in LLM-based agents are explored.

Defensive Techniques and Model Hardening

Adversarial training and data augmentation strategies are introduced.

Defensive distillation and other robustness techniques are covered.

Input preprocessing, gradient masking, and regularization are discussed.

Differential privacy, noise injection, and privacy budgets are addressed.

Federated learning and secure aggregation for distributed training are explored.

Hands-On with the Adversarial Robustness Toolbox

Simulating attacks against the multi-modal agent built on Day 2 is demonstrated.

Measuring robustness under perturbation and quantifying degradation is covered.

Applying defenses iteratively and re-evaluating attack success rates is addressed.

Stress-testing tool-call pathways and prompt injection vectors is explored.

Day 4
Risk Management Frameworks for AI

The NIST AI Risk Management Framework (RMF) functions of govern, map, measure, and manage are introduced.

ISO/IEC 42001 and emerging AI-specific standards are discussed.

Mapping AI risk to existing enterprise GRC frameworks is covered.

AI accountability, auditability, and documentation requirements for government are addressed.

Regulatory Compliance for Agentic Systems

The EU AI Act's risk tiers, prohibited uses, and obligations for high-risk systems are introduced.

GDPR and CCPA implications for agent data pipelines are covered.

The U.S. Executive Order on Safe, Secure, and Trustworthy AI is addressed.

Sector-specific guidance for finance, healthcare, and public services is explored.

Third-party risk and supplier AI tool usage are discussed.

Ethics, Bias, and Explainability

Bias detection and mitigation across agent perception and reasoning are introduced.

Explainability and transparency as security-relevant properties are covered.

Fairness, downstream harm, and responsible deployment are addressed.

Designing inclusive, auditable agent behavior is explored.

Production Deployment, Monitoring, and Incident Response

Secure deployment patterns for single and multi-agent systems are introduced.

Continuous monitoring for drift, anomalies, and abuse is covered.

Logging, audit trails, and forensic readiness for agent actions are addressed.

AI security incident response playbooks and recovery are discussed.

Case studies of real-world AI breaches and lessons learned are explored.

Capstone and Synthesis

Reviewing the multi-modal multi-agent system built across the course is covered.

An end-to-end pipeline review, from design to deployment, is conducted.

Self-assessment of the system against NIST AI RMF functions is addressed.

A forward outlook on emerging trends in agentic AI and AI security is explored.

Summary and Next Steps

Requirements

Targeted Audience

This guidance is intended for AI engineers and architects developing agentic systems for production use in government and other sectors. It also targets cybersecurity, risk, and compliance professionals responsible for ensuring AI assurance in highly regulated industries such as finance, healthcare, and consulting. Additionally, senior developers and solution leads who are integrating multi-modal and multi-agent capabilities into enterprise platforms will find this information valuable.

 28 Hours

Number of participants


Price per participant

Testimonials (3)

Upcoming Courses

Related Categories