Learn AI: A Roadmap for Cloud Security Pros

From "I've used ChatGPT a few times" to "I can defend an agentic system in production." A practical, opinionated path for security people who want real AI fluency โ€” not buzzwords.

Abstract illustration of AI with silhouette head full of eyes, symbolizing observation
Photo by Tara Winstead on Pexels

ยท ยท Vendor-neutral ยท View source on GitHub

The honest version: Most security professionals don't need to train models from scratch. You need to understand how LLMs actually work, how to use them well, and how they fail โ€” because soon you will be asked to defend an agent that has access to your production cloud. This roadmap is the shortest path I'd recommend to a working CNAPP engineer, detection engineer, or cloud architect who has used ChatGPT but hasn't yet built or attacked anything with an LLM.

๐Ÿ“– The Path

  1. Why bother (and why now)
  2. Prerequisites
  3. Stage 1: Become a daily user (Weeks 0โ€“2)
  4. Stage 2: Mental model (Weeks 2โ€“6)
  5. Stage 3: Build something (Months 1.5โ€“3)
  6. Stage 4: AI/LLM security (Months 3โ€“6)
  7. Stage 5: Agents & MCP (Months 5โ€“8)
  8. Stage 6: Specialize
  9. Hands-on labs & CTFs
  10. Books & courses
  11. People & newsletters to follow
  12. Certifications worth your time
  13. Project ideas to actually build
  14. Frameworks & standards to know cold
  15. Stay current
  16. Common mistakes

Why bother (and why now)

Three reasons this is on a cloud-security site:

You don't need to become a machine-learning researcher. You need working AI literacy: enough to use these tools well, build small things, and reason about how they break.

Prerequisites

You'll have a much easier time if you already have:

That's it. You do not need linear algebra, calculus, or a PyTorch tattoo to be useful here.

Stage 1: Become a daily user (Weeks 0โ€“2)

Goal: Build instinct for what frontier models can and can't do. You can't defend a tool you've never used in anger.

1. Pick one frontier model and pay for it

Claude (Anthropic), ChatGPT (OpenAI), or Gemini (Google). Pay the $20/month โ€” the gap between the free and paid tier is enormous, and the gap between "occasional user" and "daily user" is bigger still. Don't try to evaluate three at once; you'll learn more by going deep on one.

2. Use it for everything for two weeks

Notice when it's confidently wrong. Notice when it saves you an hour. Notice how rephrasing a prompt changes the answer. That instinct is the foundation of everything below.

3. Try a coding agent

Use Claude Code, Cursor, GitHub Copilot, or Codex on a real task in a real repo. Write a small tool, debug a Lambda, refactor a script. The leap from "chat assistant" to "agent that edits files and runs commands" is exactly the shift you'll later need to defend.

Stage 1 milestone: You instinctively reach for an LLM for the right kinds of tasks and avoid it for the wrong ones. You can describe โ€” to a colleague โ€” three things this model does well and three it fails at.

Elegant 3D visualization of neural networks showcasing abstract connections in digital space
Photo by Google DeepMind on Pexels

Stage 2: Mental model (Weeks 2โ€“6)

Goal: Understand what's actually happening when you press send. No more magic.

1. The 30-minute mental model

You don't need to derive attention from scratch. You do need crisp answers to:

2. Recommended primers

3. Prompt engineering, briefly

Most "prompt engineering" content online is hype. The 80/20:

Stage 2 milestone: You can read a prompt-engineering blog post or a model evaluation paper and not get lost in the jargon. You stop calling LLMs "AI" in casual conversation because the precision matters.

A robot arm assists a professional with a book and coffee in a modern office
Photo by Pavel Danilyuk on Pexels
Most AI security work in 2026 is plumbing โ€” guardrails, secrets, observability โ€” not exotic adversarial ML. โ€” what the job actually looks like

Stage 3: Build something with the API (Months 1.5โ€“3)

Goal: Get past consumer-app polish and into raw model behavior.

1. The canonical first project: RAG over your own notes

Build a retrieval-augmented Q&A bot that answers questions about a corpus you control โ€” your blog posts, last year's incident write-ups, every CloudTrail finding from a CTF, your runbooks, whatever. You will:

It will work badly the first time. Fixing it is where the real learning lives.

2. Your second project: an evaluation harness

Pick a security task โ€” "extract IOCs from this incident report," "classify this CloudTrail event as benign or suspicious," "summarize this CVE in three bullets." Build a small dataset of ~30 examples with ground truth. Run the same prompt against it on every change. Watch your changes regress things. This is what real AI engineering feels like.

3. Recommended starting kits

Stage 3 milestone: You've shipped two things you wrote yourself, in code, that call a model API. You have an opinion about model providers based on real use, not Twitter takes.

Close-up of a vintage typewriter with 'AI ETHICS' typed on paper
Photo by Markus Winkler on Pexels

Stage 4: AI/LLM security (Months 3โ€“6)

Goal: Move from "I can build" to "I can break, and I can defend."

1. Read the canon

2. The attack classes you must understand

3. Defenses you'll build (or buy)

Stage 4 milestone: You can review an LLM-powered feature in your company's product and produce a credible threat model. You stop treating "the LLM is powerful" as a magic explanation.

Business professionals discussing graphs on a flipchart during a daylight meeting
Photo by Kaboompics on Pexels

Stage 5: Agents & MCP (Months 5โ€“8)

Goal: Internalize what changes when an LLM stops talking and starts doing.

Agents are LLMs in a loop with tools. Model Context Protocol (MCP) is the emerging open standard for plugging tools and data into them. Together they are the most consequential security shift since cloud itself, and most companies have no idea yet.

1. Build an agent end-to-end

2. Set up MCP locally

3. Read the agent-security canon

Stage 5 milestone: You can sketch the threat model for an agentic system on a whiteboard, identify the three highest-blast-radius tools, and explain to an exec why "but the system prompt forbids that" is not a control.

Stage 6: Specialize

Pick the lane your career and curiosity pull you toward:

Hands-on labs & CTFs

You can read for a year and learn less than you do in one weekend of CTFs. The full list lives in the CSOH CTF directory; here's what to start with:

For each lab, write up what you tried, what worked, and what didn't. Public write-ups are the highest-leverage portfolio item in this field right now.

Books & courses

The published-book layer is thinner than for cloud, but a few are worth it. The CSOH reading list tracks updates; here's a focused starter set:

Free courses worth your evenings:

People & newsletters to follow

The half-life of an AI blog post is about six weeks. Follow people, not blog posts:

Certifications worth your time

The AI security cert market is young. Most of the value is still in the work, not the badges. That said:

See also the CSOH cloud security certifications guide.

Project ideas to actually build

The portfolio that matters in 2026 is a small handful of running, public, AI-flavored projects. Pick one or two:

Publish the code. Write up what worked and what didn't. That artifact will outrun any cert in interviews.

Frameworks & standards to know cold

You'll be asked about all of these in interviews and audits. Skim them now, deep-read the ones your job touches:

Stay current

This field changes faster than cloud, which is saying something. A sustainable rhythm:

Common mistakes

Ready to start?