haotianblog
AI Security Engineering

AI Security Engineering

Topic hub

AI Security Engineering

A reproducible route through threat modeling, adversarial examples, poisoning, model privacy, and LLM/RAG/Agent security.

Built for engineers researching AI security threat modeling, robust evaluation, poisoning defense, membership inference, model extraction, and prompt-injection controls.

What you will build

You will use a safe toy lab and connect risks, metrics, boundaries, and engineering controls into one reviewable workflow.

  • AI security threat modeling
  • adversarial examples robust evaluation
  • data poisoning backdoor defense
  • model privacy membership inference
  • RAG prompt injection defense

Recommended reading order

Start with concepts, then move into runnable projects

AI Security Threat Modeling

Build a defense map with NIST adversarial ML, MITRE ATLAS, and OWASP LLM risks.

Level: Professional Reading time: 12 min
  • AI Security
  • Threat Modeling
  • NIST
  • MITRE ATLAS
  • OWASP

Adversarial Examples and Robust Evaluation

Evaluate clean and perturbed accuracy with an FGSM-style digits experiment.

Level: Professional Reading time: 11 min
  • Adversarial Examples
  • FGSM
  • Robust Evaluation
  • scikit-learn

Data Poisoning and Backdoor Defense

Study poison rate, trigger behavior, attack success rate, and training pipeline controls.

Level: Professional Reading time: 11 min
  • Data Poisoning
  • Backdoor Defense
  • Training Pipeline
  • scikit-learn

Model Privacy and Extraction Defense

Measure membership inference signal and surrogate fidelity against a local toy model.

Level: Professional Reading time: 12 min
  • Model Privacy
  • Membership Inference
  • Model Extraction
  • Prediction API

LLM, RAG, and Agent Security

Separate instructions from data and enforce tool permissions against indirect prompt injection.

Level: Professional Reading time: 12 min
  • LLM Security
  • RAG
  • Agent Tools
  • Prompt Injection

Resources and distribution assets

Code, data, diagrams, and share assets in one place

FAQ

Direct answers to common search questions

Does this hub provide steps for attacking real systems?

No. The lab uses only scikit-learn built-in data and synthetic toy data, with a focus on defensive evaluation, risk records, and engineering review.

Who is the intended reader?

It is for engineers who can read Python, understand basic ML workflows, and want to put AI systems into security review.