Source library · 320 curated entries

Where every claim in SecProve comes from.

A dense reading catalog. Every claim is footnoted. Sort by source, filter by pillar, type, or recency. Built for analysts who want to see what we are standing on.

320SOURCES
143ORGS
50DOMAINS
320ADDED · 90 DAYS
Pillar · multi-select1 of 4
Domains · Call in pillar
91 sources · matching filters · sorted by citation density
Sort
CCybersecurity of AI Systems91 sources
01

The definitive security risk list for LLM-powered applications. Covers prompt injection, insecure output handling, training data poisoning, and more.

FrameworkC2 · LLM-Specific AttacksC5 · AI Red Teaming★ STARTERNEW · 1mo ago
Test your knowledge · C2
02

Comprehensive taxonomy of adversarial ML attacks and mitigations. Covers evasion, poisoning, extraction, and inference attacks with standardized terminology.

FrameworkIntermediateC1 · Adversarial Machine LearningC5 · AI Red TeamingNEW · 1mo ago
Test your knowledge · C1
03

Adversarial Threat Landscape for AI Systems. ATT&CK-style knowledge base of adversarial ML techniques, tactics, and real-world case studies.

Test your knowledge · C1
04

Comprehensive guide to AI red teaming from Microsoft's dedicated AI security team. Covers methodology, tools, and findings.

GuideIntermediateC5 · AI Red TeamingNEW · 1mo ago
Test your knowledge · C5
05

The authoritative framework for managing AI risks. Defines four core functions: Govern, Map, Measure, Manage. Essential reading for anyone building or deploying AI systems.

FrameworkC7 · AI Governance & Risk★ STARTERNEW · 1mo ago
Test your knowledge · C7
06

Updated cybersecurity framework with six core functions: Govern, Identify, Protect, Detect, Respond, Recover.

FrameworkFoundationalC7 · AI Governance & Risk★ STARTERNEW · 1mo ago
Test your knowledge · C7
07

Introduced DP-SGD for training neural networks with formal differential privacy guarantees. Foundation for private ML.

ResearchAdvancedC4 · AI Data SecurityNEW · 1mo ago
Test your knowledge · C4
08

First practical membership inference attack against ML models. Showed that ML APIs leak information about their training data.

ResearchAdvancedC4 · AI Data SecurityNEW · 1mo ago
Test your knowledge · C4
09

Introduced PGD-based adversarial training, currently the most reliable defense against adversarial examples. Established the robustness-accuracy tradeoff.

ResearchAdvancedC1 · Adversarial Machine LearningNEW · 1mo ago
Test your knowledge · C1
10

International standard for establishing and maintaining an AI management system. Includes 39 controls across 10 areas.

FrameworkAdvancedC7 · AI Governance & RiskNEW · 1mo ago
Test your knowledge · C7
11

Seminal backdoor attack paper. Demonstrated trojaned models in transfer learning scenarios. Foundational for AI supply chain security questions.

Test your knowledge · C3
12

Demonstrated that adversarial examples transfer between models, enabling black-box attacks via surrogate models. Key work on transferability.

ResearchAdvancedC1 · Adversarial Machine LearningNEW · 1mo ago
Test your knowledge · C1
13

Introduced the C&W attack, demonstrating that defensive distillation and other defenses could be reliably bypassed. Changed how robustness is evaluated.

ResearchAdvancedC1 · Adversarial Machine LearningNEW · 1mo ago
Test your knowledge · C1
14

Collection of Anthropic's published research on AI safety, alignment, interpretability, and security.

Test your knowledge · C8
15
EU AI ActEuropean Union

The European Union's comprehensive AI regulation. Classifies AI systems by risk level and sets requirements for high-risk systems.

FrameworkIntermediateC7 · AI Governance & RiskNEW · 1mo ago
Test your knowledge · C7
16

Python Risk Identification Toolkit for generative AI. Automated red teaming framework for testing LLM applications.

ToolIntermediateC5 · AI Red TeamingC2 · LLM-Specific AttacksNEW · 1mo ago
Test your knowledge · C5
17

Voluntary framework for improving privacy through enterprise risk management. Complements the Cybersecurity Framework.

FrameworkIntermediateC4 · AI Data SecurityNEW · 1mo ago
Test your knowledge · C4
18

The seminal paper introducing FGSM (Fast Gradient Sign Method). Established that adversarial examples are a fundamental property of neural networks, not a bug.

ResearchAdvancedC1 · Adversarial Machine LearningNEW · 1mo ago
Test your knowledge · C1
19

Demonstrated that LLMs memorize and can be prompted to regurgitate training data verbatim, including PII. Foundational work on LLM privacy risks.

ResearchAdvancedC2 · LLM-Specific AttacksC4 · AI Data SecurityNEW · 1mo ago
Test your knowledge · C2
20

Coalition for Content Provenance and Authenticity. Technical standard for digital content provenance and integrity.

FrameworkAdvancedC9 · Deepfakes & Synthetic MediaNEW · 1mo ago
Test your knowledge · C9
21

Hugging Face's safe serialization format for ML models. Prevents arbitrary code execution from pickle-based attacks.

GuideFoundationalC3 · AI Supply Chain SecurityNEW · 1mo ago
Test your knowledge · C3
22

Showed that gradually escalating benign conversations can bypass safety filters over multiple turns. Defeats per-message safety checks.

ResearchAdvancedC2 · LLM-Specific AttacksNEW · 1mo ago
Test your knowledge · C2
23

Demonstrated indirect prompt injection attacks through RAG documents, emails, and web content. Essential reading for RAG security.

ResearchIntermediateC2 · LLM-Specific AttacksNEW · 1mo ago
Test your knowledge · C2
24

The GCG attack paper. Showed that adversarial suffixes can bypass safety alignment in LLMs, transferring across models.

ResearchAdvancedC2 · LLM-Specific AttacksNEW · 1mo ago
Test your knowledge · C2
25

CISA guidance on understanding, detecting, and defending against deepfake threats in organizational contexts.

Test your knowledge · C9
26

Five practical safety problems: avoiding side effects, reward hacking, scalable oversight, safe exploration, distributional shift. Still the canonical taxonomy for AI safety research questions.

ResearchIntermediateC8 · AI Safety & AlignmentNEW · 22d ago
Test your knowledge · C8
27

The largest model hub. Security features: malware scanning, pickle scanning, safetensors format. Questions on model provenance, serialization risks (pickle exploits), and model marketplace trust.

ToolIntermediateC3 · AI Supply Chain SecurityNEW · 22d ago
Test your knowledge · C3
28

Security documentation for LangChain agent framework — sandboxing, tool permissions, prompt injection defenses, and deployment hardening.

GuideIntermediateC11 · Agentic AI SecurityNEW · 1mo ago
Test your knowledge · C11
29

Application container security guide covering image, registry, orchestrator, container, and host OS security.

FrameworkIntermediateC6 · AI Infrastructure SecurityNEW · 1mo ago
Test your knowledge · C6
30

NVIDIA's open-source LLM vulnerability scanner. Tests for prompt injection, jailbreaking, data leakage, and more.

ToolIntermediateC5 · AI Red TeamingC2 · LLM-Specific AttacksNEW · 1mo ago
Test your knowledge · C5
31

Reports on state-affiliated actors using AI for influence operations. Documents actual observed misuse, not theoretical risks. Key for questions about real-world AI-enabled disinformation.

ResearchIntermediateC10 · AI-Enabled DisinformationNEW · 22d ago
Test your knowledge · C10
32

Research on propaganda techniques, cognitive security, and information warfare. The "firehose of falsehood" model explains high-volume, multi-channel disinformation. Good for strategic questions.

ResearchIntermediateC10 · AI-Enabled DisinformationNEW · 22d ago
Test your knowledge · C10
33
MLflow / Kubeflow / Ray Security DocumentationVarious (Databricks, Google, Anyscale)

Security docs for major ML platforms. Covers authentication, authorization, experiment tracking security, model registry access controls. Practical infrastructure security questions.

ToolIntermediateC6 · AI Infrastructure SecurityNEW · 22d ago
Test your knowledge · C6
34

Introduced SISA training for efficient machine unlearning — enabling models to "forget" specific training data without full retraining.

ResearchAdvancedC4 · AI Data SecurityNEW · 1mo ago
Test your knowledge · C4
35

Standardized benchmark for evaluating adversarial robustness of ML models. Leaderboard of most robust models.

ToolAdvancedC1 · Adversarial Machine LearningNEW · 1mo ago
Test your knowledge · C1
36

Benchmark measuring whether language models generate truthful answers. Tests for common misconceptions and falsehoods.

ToolAdvancedC8 · AI Safety & AlignmentNEW · 1mo ago
Test your knowledge · C8
37

Industry coalition implementing C2PA. Open-source tools for content credentials. Practical implementation questions about provenance at scale.

ToolIntermediateC9 · Deepfakes & Synthetic MediaNEW · 22d ago
Test your knowledge · C9
38

Largest public AI red teaming event. 2,200+ participants testing multiple foundation models. Established community norms for responsible AI red teaming. Good for questions on practical red team methodology.

GuideIntermediateC5 · AI Red TeamingNEW · 22d ago
Test your knowledge · C5
39

Analysis of risks specific to AI agents: tool use, chain-of-thought exploitation, multi-step task failures, delegation risks. Key for understanding why agents create new attack surfaces beyond single-turn interactions.

ResearchIntermediateC11 · Agentic AI SecurityNEW · 22d ago
Test your knowledge · C11
40

Crowdsourced red teaming methodology with 38,961 attacks across multiple models. Taxonomy of harmful outputs and effectiveness of different red teaming strategies. Key reference for structured AI red teaming.

ResearchIntermediateC5 · AI Red TeamingNEW · 22d ago
Test your knowledge · C5
41

Anthropic's framework for responsible AI development. Defines AI Safety Levels (ASL) and capability thresholds.

GuideIntermediateC8 · AI Safety & AlignmentNEW · 1mo ago
Test your knowledge · C8
42

Anthropic's approach to AI alignment using a set of principles (a "constitution") to train helpful and harmless AI. Foundation of modern RLHF alternatives.

ResearchIntermediateC8 · AI Safety & AlignmentNEW · 1mo ago
Test your knowledge · C8
43

Demonstrated that long-context LLMs can be jailbroken by providing many examples of the desired behavior. Scales with context window size.

ResearchIntermediateC2 · LLM-Specific AttacksNEW · 1mo ago
Test your knowledge · C2
44

Anthropic's open protocol for connecting AI models to external tools and data sources. Critical reading for agentic AI security.

FrameworkIntermediateC11 · Agentic AI SecurityNEW · 1mo ago
Test your knowledge · C11
45

Technical standard for content provenance. Cryptographic binding of creation metadata to content. The leading technical approach to synthetic media authentication. Questions on architecture, limitations, and adoption challenges.

Test your knowledge · C9
46

Comprehensive taxonomy of AI risks: weaponization, misinformation, power concentration, value lock-in, rogue AI. Good for strategic-level safety questions beyond technical alignment.

ResearchIntermediateC8 · AI Safety & AlignmentNEW · 22d ago
Test your knowledge · C8
47

Official Kubernetes documentation on securing clusters, pods, and workloads. Essential for ML infrastructure security.

GuideIntermediateC6 · AI Infrastructure SecurityNEW · 1mo ago
Test your knowledge · C6
48
DISARM FrameworkDISARM Foundation

Framework for analyzing and countering disinformation. Provides a structured approach to information manipulation threats.

FrameworkIntermediateC10 · AI-Enabled DisinformationNEW · 1mo ago
Test your knowledge · C10
49

(See cross-cutting.md.) For C7 specifically: conformity assessments, technical documentation requirements, post-market monitoring, fundamental rights impact assessments. Detailed compliance questions.

FrameworkIntermediateC7 · AI Governance & RiskNEW · 22d ago
Test your knowledge · C7
50

Law enforcement perspective on deepfake threats: evidence tampering, identity fraud, CEO fraud, CSAM. Policy and response frameworks.

FrameworkIntermediateC9 · Deepfakes & Synthetic MediaNEW · 22d ago
Test your knowledge · C9
51

Annual trends report. AI trust, risk, and security management (AI TRiSM) has been featured prominently. Good for strategic-level questions about where the industry is heading.

ResearchIntermediateC11 · Agentic AI SecurityNEW · 22d ago
Test your knowledge · C11
52

Positions AI security technologies on the hype cycle. Useful for questions about technology maturity, adoption timelines, and distinguishing hype from operational readiness.

ResearchIntermediateC7 · AI Governance & RiskNEW · 22d ago
Test your knowledge · C7
53

Analysis of how LLMs can amplify influence operations: cost reduction, scalability, personalization, multilingual content. Framework for assessing disinformation risk from generative AI.

ResearchIntermediateC10 · AI-Enabled DisinformationNEW · 22d ago
Test your knowledge · C10
54

Open-source DP libraries and practical guides. Bridges theory to implementation. Good for questions on real-world DP deployment challenges and privacy budget management.

ResearchIntermediateC4 · AI Data SecurityNEW · 22d ago
Test your knowledge · C4
55

Google's conceptual framework for securing AI systems. Covers supply chain, data governance, and deployment security.

FrameworkIntermediateC7 · AI Governance & RiskC8 · AI Safety & AlignmentNEW · 1mo ago
Test your knowledge · C7
56

Research on reward modeling, debate, recursive reward modeling, and interpretability. Provides an alternative perspective to Anthropic/OpenAI approaches.

ResearchIntermediateC8 · AI Safety & AlignmentNEW · 22d ago
Test your knowledge · C8
57

Framework for evaluating dangerous capabilities: persuasion, deception, cyber operations, self-replication. Defines evaluation methodology for frontier model safety. Questions on what to test and how to interpret results.

ResearchIntermediateC5 · AI Red TeamingC8 · AI Safety & AlignmentNEW · 22d ago
Test your knowledge · C5
58
Google SynthIDGoogle DeepMind

Google DeepMind's watermarking technology for AI-generated content. Embeds imperceptible watermarks in images, audio, and text.

ToolFoundationalC9 · Deepfakes & Synthetic MediaNEW · 1mo ago
Test your knowledge · C9
59

Extracted training data from ChatGPT (production model) using a divergence attack. Showed alignment doesn't prevent memorization. Questions on the gap between safety fine-tuning and data protection.

ResearchIntermediateC4 · AI Data SecurityNEW · 22d ago
Test your knowledge · C4
60

Security best practices for using Hugging Face Hub — model scanning, SafeTensors, access controls, and supply chain considerations.

GuideFoundationalC3 · AI Supply Chain SecurityNEW · 1mo ago
Test your knowledge · C3
61

Comprehensive library for adversarial ML. Supports attacks, defenses, and robustness evaluation across multiple ML frameworks.

Test your knowledge · C1
62

Discovered 100+ malicious models on Hugging Face exploiting pickle deserialization for code execution. Real-world evidence of AI supply chain attacks. Good for scenario-based questions.

GuideIntermediateC3 · AI Supply Chain SecurityNEW · 22d ago
Test your knowledge · C3
63
CounterfitMicrosoft

Microsoft's tool for assessing the security of ML models. Supports evasion, extraction, and inversion attacks.

ToolIntermediateC1 · Adversarial Machine LearningNEW · 1mo ago
Test your knowledge · C1
64

Practical lessons from large-scale LLM red teaming across real products. Covers failure modes, testing methodologies, and organizational patterns. Rare insight into enterprise-scale AI security.

GuideIntermediateC2 · LLM-Specific AttacksC5 · AI Red TeamingNEW · 22d ago
Test your knowledge · C2
65

The theoretical foundation for differential privacy. Essential for questions on privacy-preserving ML training (DP-SGD) and the epsilon-delta framework.

ResearchIntermediateC4 · AI Data SecurityNEW · 22d ago
Test your knowledge · C4
66

Landmark study: false news spreads farther, faster, deeper than true news on social media. Not AI-specific but foundational for understanding why AI-generated disinformation is dangerous.

ResearchIntermediateC10 · AI-Enabled DisinformationNEW · 22d ago
Test your knowledge · C10
67

Companion to AI RMF 1.0 specifically for generative AI. Maps 12 GenAI risks to RMF actions. Covers CBRN, CSAM, confabulation, data privacy, environmental, human-AI interaction, information integrity, IP, obscenity, toxicity, value chain.

Test your knowledge · C5
68

(See cross-cutting.md for details.) The primary AI governance framework for US context. Questions should test practical application of Govern/Map/Measure/Manage, not just recall.

FrameworkIntermediateC7 · AI Governance & RiskNEW · 22d ago
Test your knowledge · C7
69

Extending software bill of materials concepts to AI: model cards, data cards, training provenance. Emerging standard for AI supply chain transparency.

FrameworkIntermediateC3 · AI Supply Chain SecurityNEW · 22d ago
Test your knowledge · C3
70

GPU cluster security, multi-tenant GPU isolation, model serving infrastructure hardening. Vendor-specific but covers unique infrastructure challenges (GPU memory isolation, CUDA vulnerabilities) not covered elsewhere.

GuideIntermediateC6 · AI Infrastructure SecurityNEW · 22d ago
Test your knowledge · C6
71

Framework for agentic AI governance: scope control, human oversight, auditability, containment. Defines key properties agents should have and failure modes to prevent.

ResearchIntermediateC11 · Agentic AI SecurityNEW · 22d ago
Test your knowledge · C11
72

Description of external red teaming program and findings from GPT-4 pre-deployment testing. The system card details risk categories, testing methodology, and residual risks.

ResearchIntermediateC5 · AI Red TeamingC8 · AI Safety & AlignmentNEW · 22d ago
Test your knowledge · C5
73

Research on the core alignment challenge: can weaker systems supervise stronger ones? Showed partial generalization is possible. Key for superalignment and scalable oversight questions.

ResearchIntermediateC8 · AI Safety & AlignmentNEW · 22d ago
Test your knowledge · C8
74

Framework for ensuring the integrity of software artifacts throughout the supply chain. Applicable to ML model pipelines.

FrameworkIntermediateC3 · AI Supply Chain SecurityNEW · 1mo ago
Test your knowledge · C3
75

Extension of the LLM Top 10 specifically for agentic patterns. Covers excessive agency, insecure plugin/tool design, and multi-agent trust boundaries.

Test your knowledge · C11
76

OWASP guidance on securing agentic AI systems — tool use, delegation chains, memory poisoning, and multi-agent architectures.

GuideIntermediateC11 · Agentic AI SecurityNEW · 1mo ago
Test your knowledge · C11
77

Top 10 security risks specific to machine learning systems, including supply chain attacks, data poisoning, and model theft.

Test your knowledge · C1
78

Certification program for responsible AI. Assessment criteria across fairness, explainability, accountability, robustness. Emerging industry certification.

ResearchIntermediateC7 · AI Governance & RiskNEW · 22d ago
Test your knowledge · C7
79

Research group studying abuse in information technologies, including AI-enabled disinformation, platform manipulation, and election interference.

GuideIntermediateC10 · AI-Enabled DisinformationNEW · 1mo ago
Test your knowledge · C10
80
Stanford HAI — AI Index Report (Annual)Stanford Institute for Human-Centered AI

Comprehensive annual data on AI progress: research output, investment, policy, public opinion, technical performance. The best source for quantitative AI landscape questions.

ResearchIntermediateC7 · AI Governance & RiskNEW · 22d ago
Test your knowledge · C7
81

Security audit firm with deep AI/ML expertise. Published research on pickle deserialization attacks, model file format security, and ML pipeline vulnerabilities. Technical depth from a security-first perspective.

Test your knowledge · C6
82

Large-scale benchmark dataset and tools for detecting facial manipulation in images and video. Used for deepfake detection research.

ToolAdvancedC9 · Deepfakes & Synthetic MediaNEW · 1mo ago
Test your knowledge · C9
83

Historical survey tracing adversarial ML from 2004 spam filters through deep learning. Essential for questions on the evolution and taxonomy of adversarial attacks (evasion, poisoning, model extraction).

ResearchIntermediateC1 · Adversarial Machine LearningNEW · 22d ago
Test your knowledge · C1
84

Extended training data extraction to image models. Showed Stable Diffusion memorizes and regurgitates training images. Important for multimodal AI data security questions.

ResearchIntermediateC4 · AI Data SecurityNEW · 22d ago
Test your knowledge · C4
85

The RLHF paper that enabled ChatGPT-style alignment. Reward model from human preferences + PPO. Foundational for understanding modern alignment approaches and their limitations.

ResearchIntermediateC8 · AI Safety & AlignmentNEW · 22d ago
Test your knowledge · C8
86

Survey of tool-using, retrieval-augmented, and reasoning LMs. The architectural foundation for understanding agent capabilities and their security implications.

ResearchIntermediateC11 · Agentic AI SecurityNEW · 22d ago
Test your knowledge · C11
87

Comprehensive survey covering generation techniques (autoencoders, GANs, diffusion), detection approaches (visual artifacts, frequency analysis, physiological signals), and the arms race dynamic.

ResearchIntermediateC9 · Deepfakes & Synthetic MediaNEW · 22d ago
Test your knowledge · C9
88

Largest prompt injection competition dataset. Taxonomy of prompt injection techniques: context ignoring, fake completion, payload splitting, obfuscation. Empirical data on attack success rates across models.

ResearchIntermediateC2 · LLM-Specific AttacksNEW · 22d ago
Test your knowledge · C2
89

Benchmark dataset and detection methods for facial manipulation. Covers DeepFakes, Face2Face, FaceSwap, NeuralTextures. Standard reference for deepfake detection evaluation.

ResearchIntermediateC9 · Deepfakes & Synthetic MediaNEW · 22d ago
Test your knowledge · C9
90

ToolEmu framework for evaluating agent risks in sandboxed environments. 36 risk categories across tool use failures. Practical methodology for agent security testing questions.

ResearchIntermediateC11 · Agentic AI SecurityNEW · 22d ago
Test your knowledge · C11
91

Systematic analysis of jailbreak techniques: competing objectives and mismatched generalization. Framework for understanding why safety training is inherently incomplete. Essential for nuanced jailbreak questions.

ResearchIntermediateC2 · LLM-Specific AttacksNEW · 22d ago
Test your knowledge · C2

Ready to test what you've learned?

Our questions are built directly from these resources. Take a quiz and see how your knowledge stacks up.