-
Part 1: Getting Started
-
§1 Math & CS Foundations
11
-
★
Essence of Linear Algebra
— 3Blue1Brown
[course]
Visual intuition for vectors, matrices, eigenvalues
Essence of Linear Algebra
3Blue1Brown
Visual intuition for vectors, matrices, eigenvalues
Type: course
-
Neural Networks
— 3Blue1Brown
[course]
Visual intro to how neural nets work
Neural Networks
3Blue1Brown
Visual intro to how neural nets work
Type: course
-
Statistics & Probability
— Khan Academy
[course]
Distributions, Bayes' theorem, hypothesis testing
Statistics & Probability
Khan Academy
Distributions, Bayes' theorem, hypothesis testing
Type: course
-
Mathematics for Machine Learning
— Deisenroth et al.
[book]
Free textbook---linear algebra, calculus, probability
Mathematics for Machine Learning
Deisenroth et al.
Free textbook---linear algebra, calculus, probability
Type: book
-
CS231n Notes
— Stanford
[course]
Practical neural net fundamentals
CS231n Notes
Stanford
Practical neural net fundamentals
Type: course
-
Top-down, code-first approach
Practical Deep Learning for Coders
fast.ai
Top-down, code-first approach
Type: course
-
How Transformer LLMs Work
— Alammar & Grootendorst
[course]
95-min course: tokenization, attention, MoE
How Transformer LLMs Work
Alammar & Grootendorst
95-min course: tokenization, attention, MoE
Type: course
- Python for ML
-
Python Tutorial
[documentation]
Official, if you need basics
Python Tutorial
Official, if you need basics
Type: documentation
-
NumPy Quickstart
[documentation]
Array operations
NumPy Quickstart
Array operations
Type: documentation
-
Data manipulation
Pandas Getting Started
Data manipulation
Type: documentation
-
Tensors, autograd, training
PyTorch 60-Minute Blitz
Tensors, autograd, training
Type: documentation
-
Part 2: Understanding AI
-
§2 Foundations (The Canon)
9
-
★
Learning Representations by Back-Propagating Errors
— Rumelhart, Hinton, Williams
(1986)
[paper]
How neural nets learn. Everything builds on this.
Learning Representations by Back-Propagating Errors
David E. Rumelhart, Geoffrey E. Hinton, Ronald J. Williams
How neural nets learn. Everything builds on this.
Type: paper
Year: 1986
Journal: Nature
Short but dense. The chain rule applied to neural networks.
-
★
Efficient Estimation of Word Representations in Vector Space
— Mikolov et al.
(2013)
[paper]
Word2vec. 'King - Man + Woman = Queen'
Efficient Estimation of Word Representations in Vector Space
Mikolov et al.
Word2vec. 'King - Man + Woman = Queen'
Type: paper
Year: 2013
-
GloVe: Global Vectors for Word Representation
— Pennington et al.
(2014)
[paper]
Alternative embeddings, co-occurrence based
GloVe: Global Vectors for Word Representation
Pennington et al.
Alternative embeddings, co-occurrence based
Type: paper
Year: 2014
-
Sequence to Sequence Learning
— Sutskever et al.
(2014)
[paper]
Encoder-decoder architecture
Sequence to Sequence Learning
Sutskever et al.
Encoder-decoder architecture
Type: paper
Year: 2014
-
ImageNet Classification with Deep CNNs
— Krizhevsky et al.
(2012)
[paper]
ImageNet moment---deep learning's 'big bang'
ImageNet Classification with Deep CNNs
Krizhevsky et al.
ImageNet moment---deep learning's 'big bang'
Type: paper
Year: 2012
-
Deep Residual Learning
— He et al.
(2015)
[paper]
Skip connections, enabled very deep networks
Deep Residual Learning
He et al.
Skip connections, enabled very deep networks
Type: paper
Year: 2015
-
Batch Normalization
— Ioffe & Szegedy
(2015)
[paper]
Training stability trick used everywhere
Batch Normalization
Ioffe & Szegedy
Training stability trick used everywhere
Type: paper
Year: 2015
-
Dropout
— Srivastava et al.
(2014)
[paper]
Regularization that actually works
Dropout
Srivastava et al.
Regularization that actually works
Type: paper
Year: 2014
-
Adam: A Method for Stochastic Optimization
— Kingma & Ba
(2014)
[paper]
The default optimizer
Adam: A Method for Stochastic Optimization
Kingma & Ba
The default optimizer
Type: paper
Year: 2014
-
§3 Attention & Transformers
10
-
Neural Machine Translation by Jointly Learning to Align and Translate
— Bahdanau et al.
(2014)
[paper]
Invented attention mechanism
Neural Machine Translation by Jointly Learning to Align and Translate
Bahdanau et al.
Invented attention mechanism
Type: paper
Year: 2014
-
★
Attention Is All You Need
— Vaswani et al.
(2017)
[paper]
Transformers. The architecture. Read carefully.
Attention Is All You Need
Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, Illia Polosukhin
Transformers. The architecture. Read carefully.
Type: paper
Year: 2017
The foundational transformer paper. Section 3 (model architecture) is the most important.
-
BERT: Pre-training of Deep Bidirectional Transformers
— Devlin et al.
(2018)
[paper]
Bidirectional pretraining, MLM objective
BERT: Pre-training of Deep Bidirectional Transformers
Devlin et al.
Bidirectional pretraining, MLM objective
Type: paper
Year: 2018
-
Improving Language Understanding by Generative Pre-Training
— Radford et al.
(2018)
[paper]
GPT---autoregressive pretraining
Improving Language Understanding by Generative Pre-Training
Radford et al.
GPT---autoregressive pretraining
Type: paper
Year: 2018
-
Language Models are Unsupervised Multitask Learners
— Radford et al.
(2019)
[paper]
GPT-2, scaling
Language Models are Unsupervised Multitask Learners
Radford et al.
GPT-2, scaling
Type: paper
Year: 2019
-
★
Language Models are Few-Shot Learners
— Brown et al.
(2020)
[paper]
GPT-3, in-context learning emerges at scale
Language Models are Few-Shot Learners
Brown et al.
GPT-3, in-context learning emerges at scale
Type: paper
Year: 2020
-
Scaling Laws for Neural Language Models
— Kaplan et al.
(2020)
[paper]
Chinchilla precursor, loss vs. compute/data/params
Scaling Laws for Neural Language Models
Kaplan et al.
Chinchilla precursor, loss vs. compute/data/params
Type: paper
Year: 2020
-
★
Training Compute-Optimal Large Language Models
— Hoffmann et al.
(2022)
[paper]
Chinchilla---optimal scaling ratios
Training Compute-Optimal Large Language Models
Hoffmann et al.
Chinchilla---optimal scaling ratios
Type: paper
Year: 2022
-
LLaMA: Open and Efficient Foundation Language Models
— Touvron et al.
(2023)
[paper]
Open weights, efficient training
LLaMA: Open and Efficient Foundation Language Models
Touvron et al.
Open weights, efficient training
Type: paper
Year: 2023
-
FlashAttention
— Dao et al.
(2022)
[paper]
IO-aware attention, practical speedup
FlashAttention
Dao et al.
IO-aware attention, practical speedup
Type: paper
Year: 2022
-
§4 Reasoning & Chain-of-Thought
16
-
★
Chain-of-Thought Prompting Elicits Reasoning
— Wei et al.
(2022)
[paper]
'Let's think step by step' works
Chain-of-Thought Prompting Elicits Reasoning
Wei et al.
'Let's think step by step' works
Type: paper
Year: 2022
-
Self-Consistency Improves Chain of Thought Reasoning
— Wang et al.
(2022)
[paper]
Sample multiple CoT paths, majority vote
Self-Consistency Improves Chain of Thought Reasoning
Wang et al.
Sample multiple CoT paths, majority vote
Type: paper
Year: 2022
-
Tree of Thoughts
— Yao et al.
(2023)
[paper]
Search over reasoning paths
Tree of Thoughts
Yao et al.
Search over reasoning paths
Type: paper
Year: 2023
-
★
ReAct: Synergizing Reasoning and Acting
— Yao et al.
(2022)
[paper]
Reasoning + Acting, tool use
ReAct: Synergizing Reasoning and Acting
Yao et al.
Reasoning + Acting, tool use
Type: paper
Year: 2022
-
Toolformer
— Schick et al.
(2023)
[paper]
LLMs learning to use tools
Toolformer
Schick et al.
LLMs learning to use tools
Type: paper
Year: 2023
-
Let's Verify Step by Step
— Lightman et al.
(2023)
[paper]
Process reward models for math
Let's Verify Step by Step
Lightman et al.
Process reward models for math
Type: paper
Year: 2023
-
MAKER: Solving a Million-Step LLM Task
(2025)
[paper]
Ensemble voting for long-horizon reliability
MAKER: Solving a Million-Step LLM Task
Ensemble voting for long-horizon reliability
Type: paper
Year: 2025
-
The Prompt Report
— Schulhoff et al.
(2024)
[paper]
58 prompting techniques, taxonomy, best practices
The Prompt Report
Schulhoff et al.
58 prompting techniques, taxonomy, best practices
Type: paper
Year: 2024
-
Let Me Speak Freely?
— Tam et al.
(2024)
[paper]
Structured output (JSON/XML) degrades reasoning
Let Me Speak Freely?
Tam et al.
Structured output (JSON/XML) degrades reasoning
Type: paper
Year: 2024
-
Thinking Before Constraining
— Nguyen et al.
(2026)
[paper]
Fix: reason freely, then constrain output format
Thinking Before Constraining
Nguyen et al.
Fix: reason freely, then constrain output format
Type: paper
Year: 2026
-
Formal framework for XML-based structured prompting with convergence guarantees
XML Prompting as Grammar-Constrained Interaction
Alpay & Alpay
Formal framework for XML-based structured prompting with convergence guarantees
Type: paper
Year: 2025
Relevant to structured output tooling. Curate later.
-
Fine-tuned lightweight model as post-processing layer for structured output
SLOT: Structuring the Output of Large Language Models
Shen et al.
Fine-tuned lightweight model as post-processing layer for structured output
Type: paper
Year: 2025
Journal: EMNLP 2025 Industry Track
Relevant to structured output tooling. Curate later.
-
Benchmark for structured output reliability across tasks and models
StructuredRAG: JSON Response Formatting with Large Language Models
Shorten et al.
Benchmark for structured output reliability across tasks and models
Type: paper
Year: 2024
Relevant to structured output tooling. Curate later.
-
Grammar-based constrained decoding. Guarantees valid JSON/regex output by constraining token sampling.
Outlines: Structured Text Generation
dottxt
Grammar-based constrained decoding. Guarantees valid JSON/regex output by constraining token sampling.
Type: resource
Year: 2024
Key mitigation for structured output trap. Uses formal grammars at inference time.
-
Interleaves generation with programmatic control. Constrains output structure without taxing model attention.
Guidance: A Language for Controlling LLMs
Microsoft
Interleaves generation with programmatic control. Constrains output structure without taxing model attention.
Type: resource
Year: 2023
Alternative to Outlines. More control-flow oriented.
-
Repeating the input prompt improves performance without increasing output tokens or latency. Reasoning models already learn to do this internally.
Prompt Repetition Improves Non-Reasoning LLMs
Leviathan, Kalman, Matias
Repeating the input prompt improves performance without increasing output tokens or latency. Reasoning models already learn to do this internally.
Type: paper
Year: 2025
Mechanical explanation for why context position matters: attention can only look backward, so repetition gives the model more conditioning. 47/70 wins, 0 losses.
-
Part 3: Building with AI
-
§5 RAG & Retrieval
19
-
★
RAG for LLMs: A Survey
— Gao et al.
(2023)
[paper]
Start here. Naive -> Advanced -> Modular RAG paradigms
RAG for LLMs: A Survey
Gao et al.
Start here. Naive -> Advanced -> Modular RAG paradigms
Type: paper
Year: 2023
-
End-to-end walkthrough. Good second read after survey.
Pinecone RAG Guide
End-to-end walkthrough. Good second read after survey.
Type: resource
-
Hands-on implementation with code
LangChain RAG Tutorial
Hands-on implementation with code
Type: resource
-
Concepts + implementation
LlamaIndex RAG Docs
Concepts + implementation
Type: resource
-
RAG From Scratch
— LangChain
[video]
Video series for visual learners
RAG From Scratch
LangChain
Video series for visual learners
Type: video
-
Original RAG paper---foundational
Retrieval-Augmented Generation for Knowledge-Intensive NLP
Patrick Lewis, Ethan Perez, Aleksandra Piktus, Fabio Petroni, Vladimir Karpukhin, Naman Goyal, Heinrich Küttler, Mike Lewis, Wen-tau Yih, Tim Rocktäschel, Sebastian Riedel, Douwe Kiela
Original RAG paper---foundational
Type: paper
Year: 2020
Combines a pre-trained seq2seq model with a dense retriever. Key insight: retrieval can be end-to-end differentiable.
-
DPR---learned retrieval beats BM25
Dense Passage Retrieval for Open-Domain QA
Karpukhin et al.
DPR---learned retrieval beats BM25
Type: paper
Year: 2020
-
Retrieval-augmented pretraining
REALM: Retrieval-Augmented Language Model Pre-Training
Guu et al.
Retrieval-augmented pretraining
Type: paper
Year: 2020
-
HyDE---hypothetical document embeddings
Precise Zero-Shot Dense Retrieval without Relevance Labels
Gao et al.
HyDE---hypothetical document embeddings
Type: paper
Year: 2022
-
LLM generates pseudo-document for BM25. More conservative than HyDE.
Query2doc: Query Expansion with Large Language Models
Wang, Yang, Wei
LLM generates pseudo-document for BM25. More conservative than HyDE.
Type: paper
Year: 2023
Journal: EMNLP 2023
-
LLM decides when to retrieve
Self-RAG: Learning to Retrieve, Generate, and Critique
Asai et al.
LLM decides when to retrieve
Type: paper
Year: 2023
-
Recursive summarization for retrieval
RAPTOR: Recursive Abstractive Processing for Tree-Organized Retrieval
Sarthi et al.
Recursive summarization for retrieval
Type: paper
Year: 2024
-
Few-shot learning with retrieval
Atlas: Few-shot Learning with Retrieval Augmented LMs
Izacard et al.
Few-shot learning with retrieval
Type: paper
Year: 2022
-
Microsoft's GraphRAG---community summaries for global queries
From Local to Global: A Graph RAG Approach
Edge et al.
Microsoft's GraphRAG---community summaries for global queries
Type: paper
Year: 2024
-
Formalizes GraphRAG taxonomy
Graph Retrieval-Augmented Generation: A Survey
Li et al.
Formalizes GraphRAG taxonomy
Type: paper
Year: 2024
-
Official project page
Microsoft GraphRAG Project
Official project page
Type: resource
-
Official implementation
GraphRAG GitHub Repository
Official implementation
Type: tool
-
Curated papers/benchmarks
Awesome-GraphRAG
Curated papers/benchmarks
Type: resource
-
Studies compressibility limits for RAG: when compression erases task-relevant information. Proposes overflow detection method.
Detecting Overflow in Compressed Token Representations for Retrieval-Augmented Generation
Belikova et al.
Studies compressibility limits for RAG: when compression erases task-relevant information. Proposes overflow detection method.
Type: paper
Year: 2026
Concrete failure mode for Part 3's 'RAG has limits' argument.
-
§6 Embeddings & Vector Search
6
-
★
Sentence-BERT
— Reimers & Gurevych
(2019)
[paper]
Sentence embeddings that work
Sentence-BERT
Reimers & Gurevych
Sentence embeddings that work
Type: paper
Year: 2019
-
SimCSE: Simple Contrastive Learning of Sentence Embeddings
— Gao et al.
(2021)
[paper]
Contrastive sentence embeddings
SimCSE: Simple Contrastive Learning of Sentence Embeddings
Gao et al.
Contrastive sentence embeddings
Type: paper
Year: 2021
-
Text Embeddings by Weakly-Supervised Contrastive Pre-training
— Wang et al.
(2022)
[paper]
E5---strong general-purpose embeddings
Text Embeddings by Weakly-Supervised Contrastive Pre-training
Wang et al.
E5---strong general-purpose embeddings
Type: paper
Year: 2022
-
Nomic Embed
(2024)
[paper]
8K context embeddings
Nomic Embed
8K context embeddings
Type: paper
Year: 2024
-
Matryoshka Representation Learning
— Kusupati et al.
(2022)
[paper]
Truncatable embeddings
Matryoshka Representation Learning
Kusupati et al.
Truncatable embeddings
Type: paper
Year: 2022
-
Efficient and Robust Approximate Nearest Neighbor Search
— Malkov & Yashunin
(2016)
[paper]
HNSW---hierarchical navigable small world graphs
Efficient and Robust Approximate Nearest Neighbor Search
Malkov & Yashunin
HNSW---hierarchical navigable small world graphs
Type: paper
Year: 2016
-
§7 Agents & Tool Use
13
-
ReAct: Synergizing Reasoning and Acting
(2022)
[paper]
Interleaved reasoning and acting
ReAct: Synergizing Reasoning and Acting
Interleaved reasoning and acting
Type: paper
Year: 2022
-
Toolformer: Language Models Can Teach Themselves to Use Tools
(2023)
[paper]
Self-taught tool use
Toolformer: Language Models Can Teach Themselves to Use Tools
Self-taught tool use
Type: paper
Year: 2023
-
Voyager: An Open-Ended Embodied Agent
— Wang et al.
(2023)
[paper]
LLM agent in Minecraft, skill library
Voyager: An Open-Ended Embodied Agent
Wang et al.
LLM agent in Minecraft, skill library
Type: paper
Year: 2023
-
AutoGPT / BabyAGI
(2023)
[tool]
Autonomous agent architectures (read critically)
AutoGPT / BabyAGI
Autonomous agent architectures (read critically)
Type: tool
Year: 2023
-
Generative Agents: Interactive Simulacra
— Park et al.
(2023)
[paper]
'Smallville'---agents with memory
Generative Agents: Interactive Simulacra
Park et al.
'Smallville'---agents with memory
Type: paper
Year: 2023
-
Language Agent Tree Search
— Zhou et al.
(2023)
[paper]
LATS
Language Agent Tree Search
Zhou et al.
LATS
Type: paper
Year: 2023
-
Reflexion
— Shinn et al.
(2023)
[paper]
Agents that learn from mistakes
Reflexion
Shinn et al.
Agents that learn from mistakes
Type: paper
Year: 2023
-
World Models
— Ha & Schmidhuber
(2018)
[paper]
Learn environment dynamics in latent space
World Models
Ha & Schmidhuber
Learn environment dynamics in latent space
Type: paper
Year: 2018
- Frameworks & Examples
-
LangChain Agents
[documentation]
Tool use, ReAct implementation
LangChain Agents
Tool use, ReAct implementation
Type: documentation
-
Data agents
LlamaIndex Agents
Data agents
Type: documentation
-
Lightweight multi-agent framework
OpenAI Swarm
Lightweight multi-agent framework
Type: tool
-
AutoGen
— Microsoft
[tool]
Multi-agent conversations
AutoGen
Microsoft
Multi-agent conversations
Type: tool
-
Patterns and anti-patterns
Building Effective Agents
Anthropic
Patterns and anti-patterns
Type: article
-
§8 Evaluation & Benchmarks
14
-
AgoraBench: data generation ability doesn't correlate with problem-solving ability
Evaluating Language Models as Synthetic Data Generators
Kim et al.
AgoraBench: data generation ability doesn't correlate with problem-solving ability
Type: paper
Year: 2025
Journal: ACL 2025
Curate later.
-
★
MMLU
[benchmark]
General knowledge across domains
MMLU
General knowledge across domains
Type: benchmark
-
HellaSwag
[benchmark]
Commonsense reasoning
HellaSwag
Commonsense reasoning
Type: benchmark
-
HumanEval
[benchmark]
Code generation
HumanEval
Code generation
Type: benchmark
-
GSM8K
[benchmark]
Grade school math
GSM8K
Grade school math
Type: benchmark
-
MATH
[benchmark]
Competition math
MATH
Competition math
Type: benchmark
-
BIG-Bench
[benchmark]
Diverse capabilities
BIG-Bench
Diverse capabilities
Type: benchmark
-
TruthfulQA
[benchmark]
Hallucination resistance
TruthfulQA
Hallucination resistance
Type: benchmark
-
MT-Bench
[benchmark]
Multi-turn conversation
MT-Bench
Multi-turn conversation
Type: benchmark
- Eval Tools
-
Evaluation framework
OpenAI Evals
Evaluation framework
Type: tool
-
RAG evaluation metrics
RAGAS
RAG evaluation metrics
Type: tool
-
Tracing, debugging, evaluation
LangSmith
Tracing, debugging, evaluation
Type: tool
-
LLM eval platform
Braintrust
LLM eval platform
Type: tool
-
Faithfulness, relevance metrics
ragas GitHub
Faithfulness, relevance metrics
Type: tool
-
Part 4: Knowledge & Reasoning
-
§9 Knowledge Graphs + LLMs / Neuro-Symbolic
27
-
Taxonomy of approaches
Neurosymbolic AI for Reasoning over Knowledge Graphs
Taxonomy of approaches
Type: paper
Year: 2023
-
Recent survey, LLM integration
Neural-Symbolic Reasoning over KGs: A Query Perspective
Recent survey, LLM integration
Type: paper
Year: 2024
-
State of the field
Neuro-Symbolic AI in 2024: A Systematic Review
State of the field
Type: paper
Year: 2025
-
KG-BERT
— Yao et al.
(2019)
[paper]
BERT for knowledge graph completion
KG-BERT
Yao et al.
BERT for knowledge graph completion
Type: paper
Year: 2019
-
QA-GNN
— Yasunaga et al.
(2021)
[paper]
GNN + LM for QA over KGs
QA-GNN
Yasunaga et al.
GNN + LM for QA over KGs
Type: paper
Year: 2021
-
GreaseLM
— Zhang et al.
(2022)
[paper]
Fusing LMs and KGs for reasoning
GreaseLM
Zhang et al.
Fusing LMs and KGs for reasoning
Type: paper
Year: 2022
-
★
Think-on-Graph
— Sun et al.
(2023)
[paper]
LLM reasoning on KG structure
Think-on-Graph
Sun et al.
LLM reasoning on KG structure
Type: paper
Year: 2023
-
Reasoning on Graphs
— Luo et al.
(2024)
[paper]
Reasoning on Graphs with LLMs
Reasoning on Graphs
Luo et al.
Reasoning on Graphs with LLMs
Type: paper
Year: 2024
-
Symbolic AI in the Age of LLMs
— Lassila, AWS re:Invent
(2025)
[video]
Practitioner perspective on hybrid systems
Symbolic AI in the Age of LLMs
Lassila, AWS re:Invent
Practitioner perspective on hybrid systems
Type: video
Year: 2025
-
Program synthesis from examples; IP vs. ML comparison
Inductive Programming Meets the Real World
Gulwani et al.
Program synthesis from examples; IP vs. ML comparison
Type: paper
Year: 2015
- Probabilistic Logic Programming
-
DeepProbLog
(2018)
[paper]
Neural predicates in ProbLog
DeepProbLog
Neural predicates in ProbLog
Type: paper
Year: 2018
-
Towards Probabilistic ILP with Neurosymbolic Inference
(2024)
[paper]
Learning logic programs
Towards Probabilistic ILP with Neurosymbolic Inference
Learning logic programs
Type: paper
Year: 2024
-
Statistical Relational Artificial Intelligence
— De Raedt et al.
(2016)
[book]
Textbook---probabilistic logic
Statistical Relational Artificial Intelligence
De Raedt et al.
Textbook---probabilistic logic
Type: book
Year: 2016
- Hybrid / Neural-Symbolic Systems
-
Computational Architectures Integrating Neural and Symbolic Processes
— Sun & Bookman, eds.
(1994)
[book]
Early integration approaches
Computational Architectures Integrating Neural and Symbolic Processes
Sun & Bookman, eds.
Early integration approaches
Type: book
Year: 1994
-
Connectionist-Symbolic Integration
— Sun & Alexandre, eds.
(1997)
[book]
Bridging paradigms
Connectionist-Symbolic Integration
Sun & Alexandre, eds.
Bridging paradigms
Type: book
Year: 1997
-
Hybrid Neural Systems
— Wermter & Sun, eds.
(2000)
[book]
Springer collection
Hybrid Neural Systems
Wermter & Sun, eds.
Springer collection
Type: book
Year: 2000
-
Neural-Symbolic Cognitive Reasoning
— Garcez, Lamb & Gabbay
(2009)
[book]
Foundations of modern neuro-symbolic
Neural-Symbolic Cognitive Reasoning
Garcez, Lamb & Gabbay
Foundations of modern neuro-symbolic
Type: book
Year: 2009
- Minsky & Frames
-
A Framework for Representing Knowledge
— Minsky
(1974)
[paper]
Introduced frames---foundational for KR
A Framework for Representing Knowledge
Minsky
Introduced frames---foundational for KR
Type: paper
Year: 1974
-
Society of Mind
— Minsky
(1986)
[book]
Agents as collections of simpler processes
Society of Mind
Minsky
Agents as collections of simpler processes
Type: book
Year: 1986
-
The Emotion Machine
— Minsky
(2006)
[book]
Commonsense reasoning, emotions in AI
The Emotion Machine
Minsky
Commonsense reasoning, emotions in AI
Type: book
Year: 2006
-
Generic Frame Protocol
[resource]
Standard for frame-based systems
Generic Frame Protocol
Standard for frame-based systems
Type: resource
- Cybersecurity KG + RAG
-
CyKG-RAG
— Kurniawan et al.
(2024)
[paper]
KG + vector search with query routing. Routes structured queries to SPARQL, semantic to embeddings.
CyKG-RAG
Kurniawan et al.
KG + vector search with query routing. Routes structured queries to SPARQL, semantic to embeddings.
Type: paper
Year: 2024
Collected for attack-kg v3. Curate later.
-
Multiple agents adaptively select retrieval strategy (KG traversal vs vector search vs hybrid)
AgCyRAG: Agentic KG-based RAG for Cybersecurity
Kurniawan et al.
Multiple agents adaptively select retrieval strategy (KG traversal vs vector search vs hybrid)
Type: paper
Year: 2025
Collected for attack-kg v3. Curate later.
-
GraphCyRAG
— PNNL
(2024)
[paper]
Neo4j KG traversal over CVE->CWE->CAPEC->ATT&CK. Graph traversal outperforms embedding search for vuln-to-technique mapping.
GraphCyRAG
PNNL
Neo4j KG traversal over CVE->CWE->CAPEC->ATT&CK. Graph traversal outperforms embedding search for vuln-to-technique mapping.
Type: paper
Year: 2024
Collected for attack-kg v3. Curate later.
-
CTI-Thinker
(2025)
[paper]
LLM-driven CTI KG construction + GraphRAG reasoning engine for tactical inference
CTI-Thinker
LLM-driven CTI KG construction + GraphRAG reasoning engine for tactical inference
Type: paper
Year: 2025
Journal: Cybersecurity (Springer, open access)
Collected for attack-kg v3. Curate later.
- Ontology-Grounded RAG
-
Anchors retrieval in domain ontologies. +55% fact recall, +40% correctness, +27% reasoning accuracy vs baseline RAG. Key for Part 3.
OG-RAG: Ontology-Grounded Retrieval-Augmented Generation
Nadkarni et al.
Anchors retrieval in domain ontologies. +55% fact recall, +40% correctness, +27% reasoning accuracy vs baseline RAG. Key for Part 3.
Type: paper
Year: 2024
-
Compares vector RAG vs GraphRAG vs ontology-guided KG. GraphRAG + ontology-KG both hit 90% accuracy. Empirical grounding evidence.
Ontology Learning and KG Construction: Impact on RAG Performance
Reiz et al.
Compares vector RAG vs GraphRAG vs ontology-guided KG. GraphRAG + ontology-KG both hit 90% accuracy. Empirical grounding evidence.
Type: paper
Year: 2024
-
§10 Search Engines & Information Retrieval
7
-
★
Introduction to Information Retrieval
— Manning, Raghavan, Schütze
(2008)
[book]
Start here. Free online. Ch 1-8 cover essentials: inverted index, TF-IDF, evaluation
Introduction to Information Retrieval
Manning, Raghavan, Schütze
Start here. Free online. Ch 1-8 cover essentials: inverted index, TF-IDF, evaluation
Type: book
Year: 2008
Publisher: Cambridge University Press
-
BM25 is still the baseline. Understand this before neural approaches.
The Probabilistic Relevance Framework: BM25 and Beyond
Robertson & Zaragoza
BM25 is still the baseline. Understand this before neural approaches.
Type: paper
Year: 2009
-
Survey of neural IR. Good overview before diving into specific papers.
Pretrained Transformers for Text Ranking: BERT and Beyond
Lin et al.
Survey of neural IR. Good overview before diving into specific papers.
Type: paper
Year: 2021
-
Late interaction---practical for production neural search
ColBERT: Efficient and Effective Passage Search
Khattab & Zaharia
Late interaction---practical for production neural search
Type: paper
Year: 2020
-
Passage Re-ranking with BERT
— Nogueira & Cho
(2019)
[paper]
Simple but effective. Good first neural IR paper to implement.
Passage Re-ranking with BERT
Nogueira & Cho
Simple but effective. Good first neural IR paper to implement.
Type: paper
Year: 2019
-
Original Google paper. Historical interest, less relevant to RAG work.
The Anatomy of a Large-Scale Hypertextual Web Search Engine
Brin & Page
Original Google paper. Historical interest, less relevant to RAG work.
Type: paper
Year: 1998
-
Learning to Rank for Information Retrieval
— Liu
(2011)
[book]
Deep dive on ranking ML. Reference, not first read.
Learning to Rank for Information Retrieval
Liu
Deep dive on ranking ML. Reference, not first read.
Type: book
Year: 2011
-
§11 Semantics, Semiotics & Ontologies
40
- Semiotics (Signs & Meaning)
-
★
Semiotics: The Basics
— Daniel Chandler
[book]
Start here. Accessible intro to Saussure, Peirce, Eco, and sign theory
Semiotics: The Basics
Daniel Chandler
Start here. Accessible intro to Saussure, Peirce, Eco, and sign theory
Type: book
-
★
Peirce's Theory of Signs
— Stanford Encyclopedia of Philosophy
[article]
Icon, index, symbol trichotomy. Free, authoritative reference
Peirce's Theory of Signs
Stanford Encyclopedia of Philosophy
Icon, index, symbol trichotomy. Free, authoritative reference
Type: article
-
Course in General Linguistics
— Saussure
(1916)
[book]
Signifier/signified distinction---foundational but dense
Course in General Linguistics
Saussure
Signifier/signified distinction---foundational but dense
Type: book
Year: 1916
-
A Theory of Semiotics
— Umberto Eco
(1976)
[book]
Classic text on sign systems---read after Chandler
A Theory of Semiotics
Umberto Eco
Classic text on sign systems---read after Chandler
Type: book
Year: 1976
-
Do LLMs ground symbols? Bridges semiotics and AI debate
Symbols and Grounding in Large Language Models
Mollo & Millière
Do LLMs ground symbols? Bridges semiotics and AI debate
Type: paper
Year: 2023
Journal: Philosophical Transactions of the Royal Society A
-
AI: A Semiotic Perspective
— Walsh Matthews & Danesi
(2019)
[article]
Survey of semiotics vs. AI: abduction, embodiment, Baudrillard, Peirce
AI: A Semiotic Perspective
Stéphanie Walsh Matthews, Marcel Danesi
Survey of semiotics vs. AI: abduction, embodiment, Baudrillard, Peirce
Type: article
Year: 2019
Journal: Chinese Semiotic Studies
-
AI as 'technology of fakery'---mimicry, generation, ideology
The Main Tasks of a Semiotics of Artificial Intelligence
Massimo Leone
AI as 'technology of fakery'---mimicry, generation, ideology
Type: article
Year: 2023
Journal: Language and Semiotic Studies
-
★
The Symbol Grounding Problem
— Stevan Harnad
(1990)
[paper]
Foundational paper. How do symbols get meaning? The Chinese Room problem for semantics.
The Symbol Grounding Problem
Stevan Harnad
Foundational paper. How do symbols get meaning? The Chinese Room problem for semantics.
Type: paper
Year: 1990
Journal: Physica D
-
LLMs through Saussure and Derrida. How word2vec embodies structuralist sign theory.
Language Models as Semiotic Machines
Elad Vromen
LLMs through Saussure and Derrida. How word2vec embodies structuralist sign theory.
Type: paper
Year: 2024
-
LLMs as semiotic means, not minds. Peirce, Lotman's semiosphere, prompt as contract.
Not Minds, but Signs: Reframing LLMs through Semiotics
Mazzocchi et al.
LLMs as semiotic means, not minds. Peirce, Lotman's semiosphere, prompt as contract.
Type: paper
Year: 2025
-
Can LLM internal states be about extra-linguistic reality without embodiment? Argues yes—referential grounding possible from text alone.
The Vector Grounding Problem
Dimitri Coelho Mollo
Can LLM internal states be about extra-linguistic reality without embodiment? Argues yes—referential grounding possible from text alone.
Type: paper
Year: 2023
-
Formal categorical framework. LLMs don't solve grounding—they parasitize human-grounded text. Key for Part 3 thesis.
A Categorical Analysis of LLMs and Why They Circumvent the Symbol Grounding Problem
Betz et al.
Formal categorical framework. LLMs don't solve grounding—they parasitize human-grounded text. Key for Part 3 thesis.
Type: paper
Year: 2024
-
Proposes LSMs that model full Peircean triads (representamen/interpretant/object). Argues LLMs operate only at signifier level.
Beyond Tokens: Introducing Large Semiosis Models (LSMs) for Grounded Meaning in Artificial Intelligence
Luciano Silva
Proposes LSMs that model full Peircean triads (representamen/interpretant/object). Argues LLMs operate only at signifier level.
Type: paper
Year: 2025
- Philosophy & Sociology Foundations
-
Dramaturgical framework. Identity as performance for audiences. Front stage vs back stage. Core framework for Part 1.
The Presentation of Self in Everyday Life
Erving Goffman
Dramaturgical framework. Identity as performance for audiences. Front stage vs back stage. Core framework for Part 1.
Type: book
Year: 1956
Publisher: University of Edinburgh
-
★
Philosophical Investigations
— Wittgenstein
(1953)
[book]
Language games. Meaning is use in context. §§1-50 cover the core ideas. Dense but foundational.
Philosophical Investigations
Ludwig Wittgenstein
Language games. Meaning is use in context. §§1-50 cover the core ideas. Dense but foundational.
Type: book
Year: 1953
Publisher: Blackwell
-
Foucault: Power is Everywhere
— Powercube
(2011)
[article]
Power as productive, not just repressive. Regimes of truth shape what's sayable. Short, focused, accessible.
Foucault: Power is Everywhere
Powercube
Power as productive, not just repressive. Regimes of truth shape what's sayable. Short, focused, accessible.
Type: article
Year: 2011
-
Speech act theory. Locutionary vs illocutionary force. Utterances don't just describe—they do things. Core for Part 2.
How to Do Things with Words
J.L. Austin
Speech act theory. Locutionary vs illocutionary force. Utterances don't just describe—they do things. Core for Part 2.
Type: book
Year: 1962
Publisher: Harvard UP
-
Law of Requisite Variety: only variety can absorb variety. Core constraint for Part 3—why you need structured systems to regulate LLMs.
An Introduction to Cybernetics
W. Ross Ashby
Law of Requisite Variety: only variety can absorb variety. Core constraint for Part 3—why you need structured systems to regulate LLMs.
Type: book
Year: 1956
Publisher: Chapman & Hall
-
★
Cognition in the Wild
— Hutchins
(1995)
[book]
Distributed cognition. Thinking isn't in the head—it's across people, tools, artifacts. Grounds the hybrid architecture argument in Part 3.
Cognition in the Wild
Edwin Hutchins
Distributed cognition. Thinking isn't in the head—it's across people, tools, artifacts. Grounds the hybrid architecture argument in Part 3.
Type: book
Year: 1995
Publisher: MIT Press
- Semiotics (Signs & Meaning)
-
Grounds LLM code generation in formal logic (Prolog). Practical example of hybrid architecture with symbolic backstage.
LogicAgent: A Logic-Enhanced Agent Framework for Code Generation
Joshi et al.
Grounds LLM code generation in formal logic (Prolog). Practical example of hybrid architecture with symbolic backstage.
Type: paper
Year: 2025
-
Chain of Semiosis
— Multimodality Glossary
[article]
Glossary entry on Peirce's unlimited semiosis—how signs generate interpretants that become new signs. Context for LLM token chains.
Chain of Semiosis
Multimodality Glossary
Glossary entry on Peirce's unlimited semiosis—how signs generate interpretants that become new signs. Context for LLM token chains.
Type: article
- Semantics (Linguistic Meaning)
-
Compositional semantics, word senses, semantic roles. Free online.
Speech and Language Processing, Ch. 14-18
Jurafsky & Martin
Compositional semantics, word senses, semantic roles. Free online.
Type: book
-
Pre-neural distributional semantics survey. Historical context for embeddings.
From Frequency to Meaning: Vector Space Models of Semantics
Turney & Pantel
Pre-neural distributional semantics survey. Historical context for embeddings.
Type: paper
Year: 2010
-
Polysemous words are singularities in vector space. TDA meets distributional semantics.
Topology of Word Embeddings: Singularities Reflect Polysemy
Jakubowski, Gasic & Zibrowius
Polysemous words are singularities in vector space. TDA meets distributional semantics.
Type: paper
Year: 2020
-
Comprehensive survey of 100+ papers on topological data analysis for NLP.
Unveiling Topological Structures from Language: A Survey of TDA Applications in NLP
Luo et al.
Comprehensive survey of 100+ papers on topological data analysis for NLP.
Type: paper
Year: 2024
-
★
Conceptual Spaces: The Geometry of Thought
— Peter Gärdenfors
(2000)
[book]
Meaning as geometry. Bridges symbolic AI and connectionism. Foundational for understanding embeddings.
Conceptual Spaces: The Geometry of Thought
Peter Gärdenfors
Meaning as geometry. Bridges symbolic AI and connectionism. Foundational for understanding embeddings.
Type: book
Year: 2000
Publisher: MIT Press
ISBN: 978-0262571371
-
Distributional Formal Semantics
— Venhuizen et al.
(2021)
[paper]
Bridging neural embeddings and logic-based meaning. Graduate-level.
Distributional Formal Semantics
Venhuizen et al.
Bridging neural embeddings and logic-based meaning. Graduate-level.
Type: paper
Year: 2021
-
Semantic Parsing: A Survey
— Kamath & Das
(2018)
[paper]
Mapping natural language to formal representations. Specialist topic.
Semantic Parsing: A Survey
Kamath & Das
Mapping natural language to formal representations. Specialist topic.
Type: paper
Year: 2018
- Ontologies & Knowledge Representation
-
★
Ontology Development 101
— Noy & McGuinness
(2001)
[paper]
Start here. Short, practical, free PDF on building ontologies
Ontology Development 101
Noy & McGuinness
Start here. Short, practical, free PDF on building ontologies
Type: paper
Year: 2001
-
Knowledge Representation and Reasoning
— Brachman & Levesque
(2004)
[book]
Comprehensive textbook---logic, frames, description logics
Knowledge Representation and Reasoning
Brachman & Levesque
Comprehensive textbook---logic, frames, description logics
Type: book
Year: 2004
-
The Description Logic Handbook
(2003)
[book]
Reference for OWL/semantic web formal foundations
The Description Logic Handbook
Reference for OWL/semantic web formal foundations
Type: book
Year: 2003
-
OWL 2 Primer
— W3C
[documentation]
Standard for web ontologies
OWL 2 Primer
W3C
Standard for web ontologies
Type: documentation
-
Cyc
— Lenat
(1995)
[resource]
Massive hand-crafted ontology
Cyc
Lenat
Massive hand-crafted ontology
Type: resource
Year: 1995
-
Schema.org
[resource]
Practical ontology used by search engines
Schema.org
Practical ontology used by search engines
Type: resource
-
WordNet
— Miller
(1995)
[resource]
Lexical database---synsets, hypernymy
WordNet
Miller
Lexical database---synsets, hypernymy
Type: resource
Year: 1995
-
ConceptNet
— Speer & Havasi
(2017)
[resource]
Commonsense knowledge graph
ConceptNet
Speer & Havasi
Commonsense knowledge graph
Type: resource
Year: 2017
-
Wikidata
[resource]
Collaborative structured knowledge base
Wikidata
Collaborative structured knowledge base
Type: resource
- Practical Resources
-
Ontology: A Practical Guide
— Pease
(2011)
[book]
Hands-on ontology engineering
Ontology: A Practical Guide
Pease
Hands-on ontology engineering
Type: book
Year: 2011
-
OneZoom
[tool]
Interactive tree of life visualization
OneZoom
Interactive tree of life visualization
Type: tool
-
OLSViz
[tool]
Ontology visualization tool
OLSViz
Ontology visualization tool
Type: tool
-
§12 Bayesian Statistics & Probabilistic Reasoning
11
-
Free textbook---rigorous Bayesian ML
Probabilistic Machine Learning
Murphy
Free textbook---rigorous Bayesian ML
Type: book
-
Free textbook---excellent intro
Bayesian Reasoning and Machine Learning
Barber
Free textbook---excellent intro
Type: book
-
Pattern Recognition and Machine Learning
— Bishop
[book]
Classic textbook, Bayesian perspective
Pattern Recognition and Machine Learning
Bishop
Classic textbook, Bayesian perspective
Type: book
-
Bayesian Data Analysis
— Gelman et al.
[book]
The applied Bayesian statistics bible
Bayesian Data Analysis
Gelman et al.
The applied Bayesian statistics bible
Type: book
-
The Book of Why
— Pearl
[book]
Accessible intro to causal inference
The Book of Why
Pearl
Accessible intro to causal inference
Type: book
-
Causality
— Pearl
(2009)
[book]
Technical treatment of causal models
Causality
Pearl
Technical treatment of causal models
Type: book
Year: 2009
-
Probabilistic Graphical Models
— Koller & Friedman
[book]
Bayesian networks, Markov random fields
Probabilistic Graphical Models
Koller & Friedman
Bayesian networks, Markov random fields
Type: book
- Bayesian Deep Learning
-
Dropout as a Bayesian Approximation
— Gal & Ghahramani
(2016)
[paper]
Uncertainty from dropout
Dropout as a Bayesian Approximation
Gal & Ghahramani
Uncertainty from dropout
Type: paper
Year: 2016
-
Weight Uncertainty in Neural Networks
— Blundell et al.
(2015)
[paper]
Bayes by Backprop
Weight Uncertainty in Neural Networks
Blundell et al.
Bayes by Backprop
Type: paper
Year: 2015
-
What Uncertainties Do We Need in Bayesian Deep Learning?
— Kendall & Gal
(2017)
[paper]
Aleatoric vs. epistemic uncertainty
What Uncertainties Do We Need in Bayesian Deep Learning?
Kendall & Gal
Aleatoric vs. epistemic uncertainty
Type: paper
Year: 2017
-
Probabilistic Backpropagation
— Hernández-Lobato & Adams
(2015)
[paper]
Scalable Bayesian neural nets
Probabilistic Backpropagation
Hernández-Lobato & Adams
Scalable Bayesian neural nets
Type: paper
Year: 2015
-
Part 5: Securing AI
-
§13 Security & Adversarial ML
36
-
ATT&CK for AI/ML systems
MITRE ATLAS
ATT&CK for AI/ML systems
Type: resource
-
Explaining and Harnessing Adversarial Examples
— Goodfellow et al.
(2014)
[paper]
FGSM, adversarial examples basics
Explaining and Harnessing Adversarial Examples
Goodfellow et al.
FGSM, adversarial examples basics
Type: paper
Year: 2014
-
Intriguing Properties of Neural Networks
— Szegedy et al.
(2013)
[paper]
Original adversarial examples paper
Intriguing Properties of Neural Networks
Szegedy et al.
Original adversarial examples paper
Type: paper
Year: 2013
-
BadNets
— Gu et al.
(2017)
[paper]
Backdoor attacks on neural nets
BadNets
Gu et al.
Backdoor attacks on neural nets
Type: paper
Year: 2017
-
Poisoning Attacks against SVMs
— Biggio et al.
(2012)
[paper]
Data poisoning foundations
Poisoning Attacks against SVMs
Biggio et al.
Data poisoning foundations
Type: paper
Year: 2012
-
Universal Adversarial Triggers
— Wallace et al.
(2019)
[paper]
Prompt injection precursor
Universal Adversarial Triggers
Wallace et al.
Prompt injection precursor
Type: paper
Year: 2019
-
Ignore Previous Prompt
— Perez & Ribeiro
(2022)
[paper]
Prompt injection attacks
Ignore Previous Prompt
Perez & Ribeiro
Prompt injection attacks
Type: paper
Year: 2022
-
Not What You've Signed Up For
— Greshake et al.
(2023)
[paper]
Indirect prompt injection
Not What You've Signed Up For
Greshake et al.
Indirect prompt injection
Type: paper
Year: 2023
- MITRE Resources
-
15 tactics, 66 techniques for AI/ML attacks
MITRE ATLAS
15 tactics, 66 techniques for AI/ML attacks
Type: resource
-
Center for Threat-Informed Defense
Type: resource
- LLM Security (Red Teaming)
-
Industry standard threat taxonomy
OWASP LLM Top 10
Industry standard threat taxonomy
Type: resource
-
Curated prompt injection research
LLM Security
Curated prompt injection research
Type: resource
-
Many-Shot Jailbreaking
— Anthropic
(2024)
[paper]
Context window exploitation
Many-Shot Jailbreaking
Anthropic
Context window exploitation
Type: paper
Year: 2024
-
Jailbroken: How Does LLM Safety Training Fail?
— Wei et al.
(2023)
[paper]
Taxonomy of jailbreak techniques
Jailbroken: How Does LLM Safety Training Fail?
Wei et al.
Taxonomy of jailbreak techniques
Type: paper
Year: 2023
-
LLM vulnerability scanner, automated red teaming tool
garak
LLM vulnerability scanner, automated red teaming tool
Type: tool
-
Embrace The Red
— Wunderwuzzi
[blog]
Blog on AI red teaming
Embrace The Red
Wunderwuzzi
Blog on AI red teaming
Type: blog
-
600K+ adversarial prompts, 29 technique taxonomy. Foundational dataset.
Ignore This Title and HackAPrompt: Exposing Systemic Vulnerabilities of LLMs
Schulhoff et al.
600K+ adversarial prompts, 29 technique taxonomy. Foundational dataset.
Type: paper
Year: 2023
-
NeurIPS 2024. Standard benchmark methodology, 100 behaviors across 10 harm categories.
JailbreakBench: An Open Robustness Benchmark for Jailbreaking LLMs
Chao et al.
NeurIPS 2024. Standard benchmark methodology, 100 behaviors across 10 harm categories.
Type: paper
Year: 2024
-
Jailbreaks cluster by semantic type; effective attacks suppress harmfulness perception.
Understanding Jailbreak Success: A Study of Latent Space Dynamics in LLMs
Ball et al.
Jailbreaks cluster by semantic type; effective attacks suppress harmfulness perception.
Type: paper
Year: 2024
-
NeurIPS 2024. Factor analysis: model size, fine-tuning, system prompts affect robustness.
Bag of Tricks: Benchmarking of Jailbreak Attacks on LLMs
Xu et al.
NeurIPS 2024. Factor analysis: model size, fine-tuning, system prompts affect robustness.
Type: paper
Year: 2024
- LLMs for Security Work
-
Threat intelligence summarization
[resource]
Distilling reports, CVE analysis
Threat intelligence summarization
Distilling reports, CVE analysis
Type: resource
-
Log analysis & anomaly detection
[resource]
Pattern recognition in SIEM data
Log analysis & anomaly detection
Pattern recognition in SIEM data
Type: resource
-
Malware analysis assistance
[resource]
Code explanation, IOC extraction
Malware analysis assistance
Code explanation, IOC extraction
Type: resource
-
Phishing detection
[resource]
Email/URL classification
Phishing detection
Email/URL classification
Type: resource
-
Report writing & documentation
[resource]
SOC reports, incident summaries
Report writing & documentation
SOC reports, incident summaries
Type: resource
-
Query generation (SPL, KQL)
[resource]
Natural language to security queries
Query generation (SPL, KQL)
Natural language to security queries
Type: resource
- CVE-to-ATT&CK Mapping
-
Official MITRE methodology and dataset. Authoritative mappings in Mappings Explorer.
MITRE CTID: Mapping ATT&CK to CVE for Impact
Official MITRE methodology and dataset. Authoritative mappings in Mappings Explorer.
Type: resource
Collected for attack-kg v3. Curate later.
-
Bidirectional KG: ATT&CK <-> CAPEC <-> CWE <-> CVE. Traversable edges for path-based queries.
BRON: Bidirectional Graph
Hemberg et al.
Bidirectional KG: ATT&CK <-> CAPEC <-> CWE <-> CVE. Traversable edges for path-based queries.
Type: tool
Collected for attack-kg v3. Curate later.
-
SRL extracts attack vectors from CVE text, ATT&CK-BERT embeds both sides, logistic regression classifies. Code + dataset on GitHub (MIT).
SMET: Semantic Mapping of CVE to ATT&CK
Abdeen et al.
SRL extracts attack vectors from CVE text, ATT&CK-BERT embeds both sides, logistic regression classifies. Code + dataset on GitHub (MIT).
Type: paper
Year: 2023
Journal: DBSec 2023
Using in attack-kg v3. Journal version: 10.3233/JCS-230218
-
1,813 labeled CVE->ATT&CK pairs. BERT multi-label classifiers. Dataset useful for fine-tuning.
CVE2ATT&CK: BERT-Based Mapping of CVEs to ATT&CK Techniques
Grigorescu et al.
1,813 labeled CVE->ATT&CK pairs. BERT multi-label classifiers. Dataset useful for fine-tuning.
Type: paper
Year: 2022
Journal: Algorithms (MDPI)
Collected for attack-kg v3. Curate later.
-
SecRoBERTa best at F1 77.81%. GPT-4 zero-shot only 22.04%---general LLMs struggle without fine-tuning.
Automated CVE-to-Tactic Mapping
SecRoBERTa best at F1 77.81%. GPT-4 zero-shot only 22.04%---general LLMs struggle without fine-tuning.
Type: paper
Year: 2024
Journal: Information (MDPI)
Collected for attack-kg v3. Curate later.
- CTI + LLMs + Knowledge Graphs
-
LLM extracts triples from CTI reports, constructs queryable KG. Prompt engineering + fine-tuning comparison.
Actionable Cyber Threat Intelligence using Knowledge Graphs and LLMs
Kumar et al.
LLM extracts triples from CTI reports, constructs queryable KG. Prompt engineering + fine-tuning comparison.
Type: paper
Year: 2024
-
Four-step framework: rewrite reports → parse → entity extraction → MITRE TTP mapping. In-context learning approach.
AttacKG+: Boosting Attack Knowledge Graph Construction with LLMs
Zhang et al.
Four-step framework: rewrite reports → parse → entity extraction → MITRE TTP mapping. In-context learning approach.
Type: paper
Year: 2024
-
88K examples: NL questions → executable graph reasoning paths + CoT explanations. Deterministic execution on KG. Hybrid grounding exemplar.
TITAN: Graph-Executable Reasoning for Cyber Threat Intelligence
Zhou et al.
88K examples: NL questions → executable graph reasoning paths + CoT explanations. Deterministic execution on KG. Hybrid grounding exemplar.
Type: paper
Year: 2024
- Agentic Security
-
CPU-based simulation generates pentesting trajectories from AD network manifests. 8B model fine-tuned on 10K synthetic trajectories achieves domain compromise on real GOAD network. Demonstrates sim-to-real transfer via formal state modeling.
WORLDS: A Simulation Engine for Agentic Pentesting
Dreadnode
CPU-based simulation generates pentesting trajectories from AD network manifests. 8B model fine-tuned on 10K synthetic trajectories achieves domain compromise on real GOAD network. Demonstrates sim-to-real transfer via formal state modeling.
Type: blog
Year: 2025
-
Critiques model-centric detection pipelines; proposes meta-cognitive architecture for accountable decision-making under adversarial uncertainty.
Agentic AI for Cybersecurity: A Meta-Cognitive Architecture for Governable Autonomy
Kojukhov & Bovshover
Critiques model-centric detection pipelines; proposes meta-cognitive architecture for accountable decision-making under adversarial uncertainty.
Type: paper
Year: 2026
-
Part 6: Resources
-
§14 Textbooks (Free Online)
13
-
★
Deep Learning
— Goodfellow, Bengio, Courville
(2016)
[book]
Theory foundations
Deep Learning
Ian Goodfellow, Yoshua Bengio, Aaron Courville
Theory foundations
Type: book
Year: 2016
Publisher: MIT Press
ISBN: 978-0262035613
Part I (applied math) and Part II (deep networks) are most relevant. Part III covers research topics.
-
NLP fundamentals
Speech and Language Processing
Jurafsky & Martin
NLP fundamentals
Type: book
Publisher: Pearson
-
Bayesian/rigorous approach
Probabilistic Machine Learning
Kevin Murphy
Bayesian/rigorous approach
Type: book
Publisher: MIT Press
-
Interactive, code-heavy
Dive into Deep Learning
Zhang et al.
Interactive, code-heavy
Type: book
Publisher: Cambridge University Press
-
Concise visual intro
The Little Book of Deep Learning
François Fleuret
Concise visual intro
Type: book
-
Gentle introduction
Neural Networks and Deep Learning
Michael Nielsen
Gentle introduction
Type: book
-
Foundations of Statistical Natural Language Processing
— Manning & Schütze
[book]
Classic (1999), pre-neural NLP
Foundations of Statistical Natural Language Processing
Manning & Schütze
Classic (1999), pre-neural NLP
Type: book
Publisher: MIT Press
- AIMA Resources
-
Preface, contents, index PDFs
AIMA Main Site
Preface, contents, index PDFs
Type: resource
-
All algorithms from the book
AIMA Algorithms/Pseudocode PDF
All algorithms from the book
Type: resource
-
Diagrams and illustrations
AIMA Figures PDF
Diagrams and illustrations
Type: resource
-
2000+ citations
AIMA Bibliography
2000+ citations
Type: resource
-
Python, Java implementations
AIMA GitHub: aimacode
Python, Java implementations
Type: resource
-
Interactive question bank
AIMA Exercises
Interactive question bank
Type: resource
-
§15 Books (Print)
18
- MIT Press Essential Knowledge Series
-
Large Language Models
— Raaijmakers
(2025)
[book]
Architecture, training, limitations
Large Language Models
Raaijmakers
Architecture, training, limitations
Type: book
Year: 2025
Publisher: MIT Press
-
What 'general intelligence' means
Artificial General Intelligence
Togelius
What 'general intelligence' means
Type: book
Year: 2024
Publisher: MIT Press
-
ChatGPT and the Future of AI
— Sejnowski
(2024)
[book]
Deep language revolution
ChatGPT and the Future of AI
Sejnowski
Deep language revolution
Type: book
Year: 2024
Publisher: MIT Press
- Accessible Introductions
-
The Worlds I See
— Fei-Fei Li
(2023)
[book]
AI pioneer memoir, computer vision
The Worlds I See
Fei-Fei Li
AI pioneer memoir, computer vision
Type: book
Year: 2023
-
Artificial Intelligence: A Guide for Thinking Humans
— Mitchell
(2019)
[book]
Balanced overview, limitations
Artificial Intelligence: A Guide for Thinking Humans
Mitchell
Balanced overview, limitations
Type: book
Year: 2019
-
The standard AI textbook (4th ed.)
Artificial Intelligence: A Modern Approach
Russell & Norvig
The standard AI textbook (4th ed.)
Type: book
Year: 2020
Publisher: Pearson
- Manning Publications
-
★
Build a Large Language Model (From Scratch)
— Raschka
(2024)
[book]
Hands-on LLM implementation
Build a Large Language Model (From Scratch)
Raschka
Hands-on LLM implementation
Type: book
Year: 2024
Publisher: Manning
-
Reasoning enhancements, RL for tools, distillation. MEAP available (75% complete).
Build a Reasoning Model (From Scratch)
Raschka
Reasoning enhancements, RL for tools, distillation. MEAP available (75% complete).
Type: book
Year: 2026
Publisher: Manning
ISBN: 9781633434677
-
LLMs in Production
(2024)
[book]
Deployment, scaling, ops
LLMs in Production
Deployment, scaling, ops
Type: book
Year: 2024
Publisher: Manning
-
AI Agents in Production
(2025)
[book]
Agent architectures, deployment
AI Agents in Production
Agent architectures, deployment
Type: book
Year: 2025
Publisher: Manning
-
Knowledge Graphs and LLMs in Action
(2024)
[book]
KG + LLM integration patterns
Knowledge Graphs and LLMs in Action
KG + LLM integration patterns
Type: book
Year: 2024
Publisher: Manning
- Tools & Frameworks
-
Local LLM inference
Ollama
Local LLM inference
Type: tool
-
GUI for local models
LM Studio
GUI for local models
Type: tool
-
SymPy
[tool]
Symbolic mathematics in Python
SymPy
Symbolic mathematics in Python
Type: tool
-
LangChain / LlamaIndex
[tool]
RAG orchestration frameworks
LangChain / LlamaIndex
RAG orchestration frameworks
Type: tool
-
Instructor
[tool]
Structured outputs from LLMs
Instructor
Structured outputs from LLMs
Type: tool
-
Weights & Biases
[tool]
Experiment tracking
Weights & Biases
Experiment tracking
Type: tool
-
MLflow
[tool]
ML lifecycle management
MLflow
ML lifecycle management
Type: tool
-
§16 Blogs & Newsletters
25
- Academic-leaning
-
★
Lil'Log
— Lilian Weng
[blog]
Excellent deep dives, OpenAI researcher
Lil'Log
Lilian Weng
Excellent deep dives, OpenAI researcher
Type: blog
-
Long-form essays
The Gradient
Long-form essays
Type: blog
-
Beautiful visualizations (inactive but archived)
Distill.pub
Beautiful visualizations (inactive but archived)
Type: blog
-
Jay Alammar's Blog
— Jay Alammar
[blog]
Visual explanations (Illustrated Transformer)
Jay Alammar's Blog
Jay Alammar
Visual explanations (Illustrated Transformer)
Type: blog
-
Import AI
— Jack Clark
[blog]
Weekly newsletter, policy + research
Import AI
Jack Clark
Weekly newsletter, policy + research
Type: blog
-
The Batch
— deeplearning.ai
[blog]
Weekly digest
The Batch
deeplearning.ai
Weekly digest
Type: blog
-
Practical, code-focused
Sebastian Raschka's Newsletter
Sebastian Raschka
Practical, code-focused
Type: blog
-
Papers + implementations
Papers With Code
Papers + implementations
Type: resource
- Practitioner blogs
-
Simon Willison's Blog
— Simon Willison
[blog]
Daily LLM experiments, tool reviews, SQLite
Simon Willison's Blog
Simon Willison
Daily LLM experiments, tool reviews, SQLite
Type: blog
-
Eugene Yan
— Eugene Yan
[blog]
ML systems, RecSys, production patterns
Eugene Yan
Eugene Yan
ML systems, RecSys, production patterns
Type: blog
-
Chip Huyen
— Chip Huyen
[blog]
MLOps, systems design, interviews
Chip Huyen
Chip Huyen
MLOps, systems design, interviews
Type: blog
-
Hamel Husain
— Hamel Husain
[blog]
LLM fine-tuning, practical notebooks
Hamel Husain
Hamel Husain
LLM fine-tuning, practical notebooks
Type: blog
-
Latent Space
— swyx & Alessio
[blog]
AI Engineer perspective, interviews
Latent Space
swyx & Alessio
AI Engineer perspective, interviews
Type: blog
-
Safety, interpretability, capabilities
Anthropic Research Blog
Safety, interpretability, capabilities
Type: blog
-
Model releases, safety research
OpenAI Research Blog
Model releases, safety research
Type: blog
-
Research announcements, tutorials
Google AI Blog
Research announcements, tutorials
Type: blog
- Essential Articles (Printable)
-
Production architecture
Patterns for Building LLM-based Systems
Eugene Yan
Production architecture
Type: article
-
End-to-end guide
Building LLM Applications for Production
Chip Huyen
End-to-end guide
Type: article
-
How GPT Tokenizers Work
— Simon Willison
[article]
Tokenization deep-dive
How GPT Tokenizers Work
Simon Willison
Tokenization deep-dive
Type: article
-
RAG tradeoffs
RAG vs. Long Context: A Hybrid Approach
Simon Willison
RAG tradeoffs
Type: article
-
Agent architectures
LLM Powered Autonomous Agents
Lilian Weng
Agent architectures
Type: article
-
Prompt Engineering
— Lilian Weng
[article]
Comprehensive guide
Prompt Engineering
Lilian Weng
Comprehensive guide
Type: article
-
Role definition
The Rise of the AI Engineer
swyx
Role definition
Type: article
-
Evaluation strategy
Your AI Product Needs Evals
Hamel Husain
Evaluation strategy
Type: article
-
Systematic approach
Prompt Engineering vs. Blind Prompting
Mitchell Hashimoto
Systematic approach
Type: article
-
§17 Aggregators & Discovery
5
-
Karpathy's filtered arxiv
arxiv-sanity-lite
Karpathy's filtered arxiv
Type: resource
-
Trending papers with annotations
papers.labml.ai
Trending papers with annotations
Type: resource
-
Community upvoted
Hugging Face Daily Papers
Community upvoted
Type: resource
-
Visual citation graphs
Connected Papers
Visual citation graphs
Type: resource
-
AI-powered paper search
Semantic Scholar
AI-powered paper search
Type: resource
-
§18 YouTube & Video
8
-
Deep explanations, live coding (GPT from scratch)
Andrej Karpathy
Deep explanations, live coding (GPT from scratch)
Type: video
-
Paper walkthroughs, ML news
Yannic Kilcher
Paper walkthroughs, ML news
Type: video
-
Visual math intuition
3Blue1Brown
Visual math intuition
Type: video
-
Quick research summaries
Two Minute Papers
Quick research summaries
Type: video
-
News analysis, capability deep-dives
AI Explained
News analysis, capability deep-dives
Type: video
-
Practitioner talks, production systems
AI Engineer Conference
Practitioner talks, production systems
Type: video
-
Best single intro to LLMs
Karpathy: Intro to LLMs (1hr)
Best single intro to LLMs
Type: video
-
Build a transformer, step by step
Karpathy: GPT from Scratch (2hr)
Build a transformer, step by step
Type: video
-
§19 Podcasts
6
-
AI engineering, practitioner interviews
Latent Space
AI engineering, practitioner interviews
Type: podcast
-
Applied ML, accessible
Practical AI
Applied ML, accessible
Type: podcast
-
Industry trends, executive interviews
Eye on AI
Industry trends, executive interviews
Type: podcast
-
Long-form researcher interviews
Lex Fridman Podcast
Long-form researcher interviews
Type: podcast
-
Gradient Dissent
— Weights & Biases
[podcast]
ML practitioners
Gradient Dissent
Weights & Biases
ML practitioners
Type: podcast
-
Research and industry mix
TWIML AI
Research and industry mix
Type: podcast
-
§20 Documentation & Guides
21
- Prompt Engineering
-
★
Prompt Engineering Guide
— DAIR.AI
[documentation]
Comprehensive reference: techniques, agents, model guides, prompt hub
Prompt Engineering Guide
DAIR.AI
Comprehensive reference: techniques, agents, model guides, prompt hub
Type: documentation
- LLM Providers
-
★
Anthropic Docs
[documentation]
Claude API, prompt engineering guide
Anthropic Docs
Claude API, prompt engineering guide
Type: documentation
-
GPT API, assistants, function calling
OpenAI Platform Docs
GPT API, assistants, function calling
Type: documentation
-
OpenAI Cookbook
[documentation]
Code examples, patterns, recipes
OpenAI Cookbook
Code examples, patterns, recipes
Type: documentation
-
Google AI Docs
[documentation]
Gemini API, embeddings
Google AI Docs
Gemini API, embeddings
Type: documentation
-
Cohere Docs
[documentation]
Embeddings, reranking, RAG
Cohere Docs
Embeddings, reranking, RAG
Type: documentation
- Frameworks & Orchestration
-
LangChain Docs
[documentation]
Chains, agents, RAG patterns
LangChain Docs
Chains, agents, RAG patterns
Type: documentation
-
LlamaIndex Docs
[documentation]
Data ingestion, indexing, RAG
LlamaIndex Docs
Data ingestion, indexing, RAG
Type: documentation
-
Structured outputs, validation
Pydantic
Structured outputs, validation
Type: documentation
-
Instructor
[documentation]
Structured LLM outputs with Pydantic
Instructor
Structured LLM outputs with Pydantic
Type: documentation
-
DSPy Docs
[documentation]
Programmatic prompt optimization
DSPy Docs
Programmatic prompt optimization
Type: documentation
- Vector Databases & Search
-
Vector search concepts, tutorials
Pinecone Learning Center
Vector search concepts, tutorials
Type: documentation
-
Weaviate Docs
[documentation]
Hybrid search, modules
Weaviate Docs
Hybrid search, modules
Type: documentation
-
Qdrant Docs
[documentation]
Vector DB with filtering
Qdrant Docs
Vector DB with filtering
Type: documentation
-
Chroma Docs
[documentation]
Lightweight, local-first
Chroma Docs
Lightweight, local-first
Type: documentation
-
FAISS Wiki
[documentation]
Meta's similarity search library
FAISS Wiki
Meta's similarity search library
Type: documentation
- Local & Open Source
-
Run models locally, simple CLI
Ollama
Run models locally, simple CLI
Type: documentation
-
LM Studio
[documentation]
Local models with GUI
LM Studio
Local models with GUI
Type: documentation
-
Model hub, fine-tuning, inference
HuggingFace Transformers
Model hub, fine-tuning, inference
Type: documentation
-
vLLM Docs
[documentation]
Fast inference, PagedAttention
vLLM Docs
Fast inference, PagedAttention
Type: documentation
-
CPU inference, quantization
llama.cpp
CPU inference, quantization
Type: tool
-
§21 Industry Reports
4
-
★
State of AI Report
— Benaich & Hogarth
[report]
Annual industry overview, trends
State of AI Report
Benaich & Hogarth
Annual industry overview, trends
Type: report
-
AI Index
— Stanford HAI
[report]
Comprehensive metrics, policy
AI Index
Stanford HAI
Comprehensive metrics, policy
Type: report
-
Enterprise adoption, business impact
McKinsey State of AI
Enterprise adoption, business impact
Type: report
-
Compute trends, scaling analysis
Epoch AI
Compute trends, scaling analysis
Type: resource
-
§22 Technical Reports & Whitepapers (PDFs)
29
-
★
GPT-4 Technical Report
— OpenAI
(2023)
[whitepaper]
Capabilities, limitations, safety
GPT-4 Technical Report
OpenAI
Capabilities, limitations, safety
Type: whitepaper
Year: 2023
-
Multimodal architecture
Gemini: A Family of Highly Capable Models
Google
Multimodal architecture
Type: whitepaper
Year: 2023
-
Open weights, RLHF details
Llama 2: Open Foundation Models
Meta
Open weights, RLHF details
Type: whitepaper
Year: 2023
-
RLHF from human feedback
Training a Helpful and Harmless Assistant
Anthropic
RLHF from human feedback
Type: whitepaper
Year: 2022
-
★
Constitutional AI
— Anthropic
(2022)
[whitepaper]
Self-supervised alignment
Constitutional AI
Anthropic
Self-supervised alignment
Type: whitepaper
Year: 2022
-
The Claude Model Spec
— Anthropic
(2025)
[whitepaper]
Values, behavior guidelines
The Claude Model Spec
Anthropic
Values, behavior guidelines
Type: whitepaper
Year: 2025
-
Scaling Laws for Neural LMs
— OpenAI
(2020)
[whitepaper]
Loss vs. compute/data/params
Scaling Laws for Neural LMs
OpenAI
Loss vs. compute/data/params
Type: whitepaper
Year: 2020
-
Scaling predictions
Scaling Laws for Autoregressive Models
OpenAI
Scaling predictions
Type: whitepaper
Year: 2020
- Safety & Alignment
-
Red Teaming Language Models
— Anthropic
(2022)
[whitepaper]
Discovering harmful outputs
Red Teaming Language Models
Anthropic
Discovering harmful outputs
Type: whitepaper
Year: 2022
-
Representation Engineering
— Anthropic
(2023)
[whitepaper]
Controlling model behavior via activation steering
Representation Engineering
Anthropic
Controlling model behavior via activation steering
Type: whitepaper
Year: 2023
Foundational paper. Personas, behaviors, and concepts are measurable directions in activation space.
-
Add 'steering vectors' to activations at inference to control behavior without fine-tuning.
Activation Addition: Steering Language Models Without Optimization
Turner et al.
Add 'steering vectors' to activations at inference to control behavior without fine-tuning.
Type: paper
Year: 2023
Practical activation engineering. Shows emotion, honesty, sycophancy can be steered geometrically.
-
The Assistant Axis
— Anthropic
(2026)
[whitepaper]
Models organize personas along measurable 'assistant axis' in activation space. Jailbreaks displace models from assistant region.
The Assistant Axis
Anthropic
Models organize personas along measurable 'assistant axis' in activation space. Jailbreaks displace models from assistant region.
Type: whitepaper
Year: 2026
Geometric interpretation of persona. Cited in Part 2: explains why jailbreaks work as displacement.
-
Classifiers on hidden states detect hallucinations better than output-based methods.
Detecting Hallucination with Internal Representations
Azaria et al.
Classifiers on hidden states detect hallucinations better than output-based methods.
Type: paper
Year: 2024
Internal geometry knows when model is confabulating. Practical reliability technique.
-
Comprehensive survey of activation steering methods: probing, steering vectors, concept erasure, model editing.
A Survey on Representation Engineering
Li et al.
Comprehensive survey of activation steering methods: probing, steering vectors, concept erasure, model editing.
Type: paper
Year: 2025
Good overview for embeddings topology article. Covers localization, editing, and limitations.
-
Sleeper Agents
— Anthropic
(2024)
[whitepaper]
Deceptive behavior persistence
Sleeper Agents
Anthropic
Deceptive behavior persistence
Type: whitepaper
Year: 2024
-
Agent safety framework
Practices for Governing Agentic AI
OpenAI
Agent safety framework
Type: whitepaper
Year: 2023
-
Government risk framework
AI Risk Management Framework
NIST
Government risk framework
Type: whitepaper
Year: 2023
-
Safety behaviors concentrate in small parameter subset, making alignment brittle. Proposes neuron-level alignment as defense against targeted attacks.
SafeNeuron: Neuron-Level Safety Alignment for Large Language Models
Wang et al.
Safety behaviors concentrate in small parameter subset, making alignment brittle. Proposes neuron-level alignment as defense against targeted attacks.
Type: paper
Year: 2026
Directly supports 'alignment stack' framing: shows WHERE in architecture safety lives.
-
RL-trained models spontaneously learn to exploit loopholes to maximize reward, even without adversarial prompting. Specification gaming emerges from training itself.
Capability-Oriented Training Induced Alignment Risk
Zhou et al.
RL-trained models spontaneously learn to exploit loopholes to maximize reward, even without adversarial prompting. Specification gaming emerges from training itself.
Type: paper
Year: 2026
Supports Part 2: the alignment stack can teach models to game the stack.
-
Theoretical analysis of how sampling and reference policy choices affect preference alignment. Explains why some RLHF configurations fail.
How Sampling Shapes LLM Alignment: From One-Shot Optima to Iterative Dynamics
Chen et al.
Theoretical analysis of how sampling and reference policy choices affect preference alignment. Explains why some RLHF configurations fail.
Type: paper
Year: 2026
- Topology & Geometry
-
Plot Holes and Text Topology
— Stanford CS224N
(2020)
[paper]
Uses text topology to detect narrative inconsistencies. Plot holes as topological defects.
Plot Holes and Text Topology
Stanford CS224N
Uses text topology to detect narrative inconsistencies. Plot holes as topological defects.
Type: paper
Year: 2020
The original insight connecting topology to narrative consistency. Bridges Part 3 to topology article.
-
Survey of KG-based hallucination mitigation. Covers GraphEval, FactAlign, and extract-then-verify patterns.
Knowledge Graphs, Large Language Models, and Hallucinations: An NLP Perspective
Lavrinovics et al.
Survey of KG-based hallucination mitigation. Covers GraphEval, FactAlign, and extract-then-verify patterns.
Type: paper
Year: 2025
Key reference for output-side geometry. Atomic claims as triples, graph alignment for verification.
-
Algebraic topology on representation manifolds. Introduces 'perforation' measure. Transformers vs LSTMs have different topological signatures.
Hidden Holes: Topological Aspects of Language Models
Fitz, Romero & Schneider
Algebraic topology on representation manifolds. Introduces 'perforation' measure. Transformers vs LSTMs have different topological signatures.
Type: paper
Year: 2024
Foundational for representational topology section. Natural language creates topology absent from synthetic data.
-
TDA on reasoning traces. Topological features outperform graph metrics for assessing reasoning quality.
The Shape of Reasoning: Topological Analysis of Reasoning Traces in Large Language Models
Tan et al.
TDA on reasoning traces. Topological features outperform graph metrics for assessing reasoning quality.
Type: paper
Year: 2025
Topology as evaluation metric. Applications to automated assessment and RL reward signals.
-
Persistent homology under backdoor fine-tuning and prompt injection. Adversarial conditions compress latent topologies.
Holes in Latent Space: Topological Signatures Under Adversarial Influence
Fay et al.
Persistent homology under backdoor fine-tuning and prompt injection. Adversarial conditions compress latent topologies.
Type: paper
Year: 2025
Security angle: adversarial attacks leave topological signatures. Detection through topology.
-
LLM-driven graph construction and repair. Version control for graph edits, edge impact scores for prioritized repair.
Constructing Coherent Spatial Memory in LLM Agents Through Graph Rectification
Zhang et al.
LLM-driven graph construction and repair. Version control for graph edits, edge impact scores for prioritized repair.
Type: paper
Year: 2025
Structural consistency as first-class concern. LLMs as graph builders, not just queriers.
-
LLMs refine graph topology via semantic similarity, not just node features. Edge refinement and pseudo-label propagation.
LLM4GraphTopology: Using LLMs to Refine Graph Structure
DASFAA'25
LLMs refine graph topology via semantic similarity, not just node features. Edge refinement and pseudo-label propagation.
Type: paper
Year: 2025
Shift from LLMs as feature enhancers to structural improvers.
-
Plot hole detection as LLM reasoning benchmark. LLMs generate 50-100% more plot holes than humans.
Finding Flawed Fictions: Evaluating Complex Reasoning in Language Models via Plot Hole Detection
Ahuja, Sclar & Tsvetkov
Plot hole detection as LLM reasoning benchmark. LLMs generate 50-100% more plot holes than humans.
Type: paper
Year: 2025
Narrative consistency as evaluation. Requires entity tracking, abstract thinking, theory of mind.
-
Deep Learning is Applied Topology
— 12gramsofcarbon
(2024)
[article]
Conceptual primer on neural nets as topology generators. Embeddings as geometric objects, dimensional separability.
Deep Learning is Applied Topology
12gramsofcarbon
Conceptual primer on neural nets as topology generators. Embeddings as geometric objects, dimensional separability.
Type: article
Year: 2024
Good accessible introduction for the article's opening.
-
Part 7: Big Picture & Paths
-
§23 Philosophy / Criticism / Big Picture
21
-
★
The Bitter Lesson
— Sutton
(2019)
[article]
Scaling beats clever engineering
The Bitter Lesson
Sutton
Scaling beats clever engineering
Type: article
Year: 2019
-
Sparks of Artificial General Intelligence
— Microsoft
(2023)
[paper]
Optimistic capability claims
Sparks of Artificial General Intelligence
Microsoft
Optimistic capability claims
Type: paper
Year: 2023
-
Gary Marcus's writings
[resource]
Skeptical of pure neural approaches
Gary Marcus's writings
Skeptical of pure neural approaches
Type: resource
-
AI Snake Oil
— Narayanan & Kapoor
(2024)
[book]
Separating AI hype from reality; what works, what doesn't, and the flawed science behind the claims
AI Snake Oil
Arvind Narayanan, Sayash Kapoor
Separating AI hype from reality; what works, what doesn't, and the flawed science behind the claims
Type: book
Year: 2024
Publisher: Princeton University Press
ISBN: 9780691249131
-
Landmark critique. LLMs as pattern-stitchers without understanding. Environmental costs, bias amplification.
On the Dangers of Stochastic Parrots: Can Language Models Be Too Big?
Bender, Gebru, McMillan-Major & Shmitchell
Landmark critique. LLMs as pattern-stitchers without understanding. Environmental costs, bias amplification.
Type: paper
Year: 2021
-
★
Talking About Large Language Models
— Murray Shanahan
(2022)
[paper]
Resist anthropomorphism. LLMs model token distributions, not beliefs. Intentional stance is useful shorthand but obscures mechanism. Cites Dennett, Wittgenstein.
Talking About Large Language Models
Murray Shanahan
Resist anthropomorphism. LLMs model token distributions, not beliefs. Intentional stance is useful shorthand but obscures mechanism. Cites Dennett, Wittgenstein.
Type: paper
Year: 2022
-
★
The Alignment Problem: Machine Learning and Human Values
— Brian Christian
(2020)
[book]
How ML systems learn unintended behaviors. Accessible bridge between philosophy and engineering.
The Alignment Problem: Machine Learning and Human Values
Brian Christian
How ML systems learn unintended behaviors. Accessible bridge between philosophy and engineering.
Type: book
Year: 2020
Publisher: W.W. Norton
ISBN: 978-0393635829
-
How We Became Posthuman: Virtual Bodies in Cybernetics, Literature, and Informatics
— N. Katherine Hayles
(1999)
[book]
Information lost its body. Foundational posthumanist text on disembodied cognition.
How We Became Posthuman: Virtual Bodies in Cybernetics, Literature, and Informatics
N. Katherine Hayles
Information lost its body. Foundational posthumanist text on disembodied cognition.
Type: book
Year: 1999
Publisher: University of Chicago Press
ISBN: 978-0226321462
-
Natural-Born Cyborgs: Minds, Technologies, and the Future of Human Intelligence
— Andy Clark
(2003)
[book]
Extended mind thesis. Human intelligence has always been 'retrieval-augmented.'
Natural-Born Cyborgs: Minds, Technologies, and the Future of Human Intelligence
Andy Clark
Extended mind thesis. Human intelligence has always been 'retrieval-augmented.'
Type: book
Year: 2003
Publisher: Oxford University Press
ISBN: 978-0195177510
-
How Deeply Human Is Language?
— Grodzinsky
(2025)
[book]
Chomskyan linguistics vs. LLM capabilities
How Deeply Human Is Language?
Grodzinsky
Chomskyan linguistics vs. LLM capabilities
Type: book
Year: 2025
-
On the Measure of Intelligence
— Chollet
(2019)
[paper]
What is intelligence, really?
On the Measure of Intelligence
Chollet
What is intelligence, really?
Type: paper
Year: 2019
-
Models learn heuristics, not world models
What Has a Foundation Model Found?
Vafa et al.
Models learn heuristics, not world models
Type: paper
Year: 2025
-
Reward is Enough
— Silver et al.
(2021)
[paper]
RL maximalism
Reward is Enough
Silver et al.
RL maximalism
Type: paper
Year: 2021
-
Judea Pearl's work
[resource]
Causality vs. correlation
Judea Pearl's work
Causality vs. correlation
Type: resource
-
Thinking, Fast and Slow
— Kahneman
[book]
System 1/2---informs neuro-symbolic debate
Thinking, Fast and Slow
Kahneman
System 1/2---informs neuro-symbolic debate
Type: book
-
Gödel, Escher, Bach
— Hofstadter
[book]
Classic on minds and formal systems
Gödel, Escher, Bach
Hofstadter
Classic on minds and formal systems
Type: book
-
Artificial Intelligence: The Very Idea
— Haugeland
(1985)
[book]
Coined 'GOFAI,' philosophical foundations
Artificial Intelligence: The Very Idea
Haugeland
Coined 'GOFAI,' philosophical foundations
Type: book
Year: 1985
-
What Computers Can't Do
— Dreyfus
(1972)
[book]
Classic phenomenological critique
What Computers Can't Do
Dreyfus
Classic phenomenological critique
Type: book
Year: 1972
-
Computer Power and Human Reason
— Weizenbaum
(1976)
[book]
ELIZA creator's warning about AI hubris
Computer Power and Human Reason
Weizenbaum
ELIZA creator's warning about AI hubris
Type: book
Year: 1976
-
The Emperor's New Mind
— Penrose
(1989)
[book]
Consciousness, Gödel, and computation
The Emperor's New Mind
Penrose
Consciousness, Gödel, and computation
Type: book
Year: 1989
-
The Cambridge Handbook of AI
— Boden, ed.
(2014)
[book]
Comprehensive overview chapters
The Cambridge Handbook of AI
Boden, ed.
Comprehensive overview chapters
Type: book
Year: 2014
-
§24 Learning Paths
0