Large Language Models and Knowledge-aware AI (Seminar SS2026)
-
Seminar: Large Language Models and Knowledge-Aware AI
TU Dresden — Summer Term 2026
Overview
Large Language Models are the most powerful knowledge systems ever built — yet they remain, in many ways, black boxes. GPT-4, Llama, and their peers have silently absorbed vast swaths of human knowledge during training, encoding billions of facts across their parameters. But what exactly do they know? How much of it is right? And can we trust it?
This seminar tackles these questions across eight tracks:
Track 1 — Knowledge Auditing & Probing examines how we systematically audit what a model believes, from cloze-style queries to full-scale knowledge materialization, and how we benchmark what LLMs actually know versus what they merely appear to know.
Track 2 — Hallucination & Factual Reliability investigates how and why LLMs fabricate — from invented entities to confident but false claims — and whether detection or prevention is even possible, including recent theoretical arguments that hallucination may be an innate architectural limitation.
Track 3 — Knowledge Editing & Consistency explores whether targeted, surgical correction of LLM knowledge is feasible, or whether cascading inconsistencies and catastrophic forgetting make it a losing battle at scale.
Track 4 — Interpretability & Knowledge Mechanisms digs into how transformers store and recall facts — through feed-forward key-value memories, knowledge neurons, and the architectural constraints that produce phenomena like the Reversal Curse.
Track 5 — LLM Biases, Language Learning & Knowledge Construction addresses systematic distortions in what LLMs learn, including implicit biases introduced by persona assignment, and examines how structured knowledge can be constructed from or around LLMs.
Track 6 — LLM Limitations takes a broader look at fundamental shortcomings of current architectures — including failures in compositionality, long-context reasoning, counterfactual tasks, and self-correction — connecting these limitations back to the core challenge of reliable knowledge retrieval.
Track 7 — Responsible AI: Safety, Privacy & Misuse covers jailbreaking, training data extraction, memorization, and the structural reasons safety training can fail, examining both attacks and defenses.
Track 8 — Can We Afford the Perfect Prompt? examines the economics of prompting, asking whether state-of-the-art techniques like chain-of-thought are worth their computational cost, and what scaling laws apply to compound inference systems.
Logistics
Type Seminar (0/2/0) Instructors Simon Razniewski, Luca Giordano, Yujia Hu, Muhammed Saeed Registration: The number of participants is limited to 12, with priority given to Master students. To express interest, send an email to muhammed.saeed@tu-dresden.de, including a short motivation statement and your transcript. Places will be allocated based on background match (courses taken) and motivation.
Core Papers
These two papers form the backbone of the seminar. All participants should read them as shared reference points.
Paper Authors Venue GPTKB: Comprehensive General Knowledge from a Large Language Model Yujia Hu, Shrestha Mohanty, Manish Shrivastava, Simon Razniewski ACL 2025 Foundations of LLM Knowledge Materialization: Termination, Reproducibility, Robustness Luca Giordano, Simon Razniewski EACL Findings 2026 GPTKB materialized 101 million triples from GPT-4o-mini via recursive prompting — creating the largest LLM-derived knowledge base to date. It revealed that LLMs can serve as massive knowledge bases, but also exposed systemic issues: ~7% false triples, fabricated entities, and deep structural inconsistencies (e.g., only 8K of 318K spouse relations are symmetric). Foundations takes the next step, formally analyzing the theoretical properties of knowledge materialization — when does it terminate, how reproducible is it across runs, and how robust is it to perturbation?
Topics
Papers are organized by thematic track. Each participant selects one paper. Own topic suggestions are welcome.
Track 1: Knowledge Auditing & Probing
What do LLMs know — and how do we find out?
# Paper Venue 1 Language Models as Knowledge Bases? — Petroni et al. EMNLP 2019 2 How Can We Know What Language Models Know? — Jiang et al. TACL 2020 3 Head-to-Tail: How Knowledgeable are Large Language Models? — Sun et al. NAACL 2024 4 Do We Know What LLMs Don't Know? A Study of Consistency in Knowledge Probing EMNLP 2025 Findings 5 Evaluating Language Models for Knowledge Base Completion — Veseli et al. ESWC 2023 6 Unexpected Knowledge: Auditing Wikipedia and Grokipedia Search Recommendations — 7 Historical Perspective: As We May Think — Vannevar Bush The Atlantic, 1945 8 Enabling LLM Knowledge Analysis via Extensive Materialization ACL 2025 9 Foundations of LLM Knowledge Materialization: Termination, Reproducibility, Robustness EACL 2026 10 The Reversal Curse: LLMs Trained on "A is B" Fail to Learn "B is A" — Berglund et al. ICLR 2024
Track 2: Hallucination & Factual Reliability
LLMs fabricate. Can we detect it — and is it fixable?
# Paper Venue 11 HALoGEN: Fantastic LLM Hallucinations and Where to Find Them — 12 FActScore: Fine-grained Atomic Evaluation of Factual Precision — Min et al. EMNLP 2023 13 SelfCheckGPT: Zero-Resource Black-Box Hallucination Detection — Manakul et al. EMNLP 2023 14 Hallucination is Inevitable: An Innate Limitation of LLMs — Xu et al. arXiv 2024 15 Do Large Language Models Know What They Don't Know? — Yin et al. ACL 2023 Findings 16 Why Language Models Hallucinate arXiv 2025
Track 3: Knowledge Editing & Consistency
When LLMs are wrong, can we fix them?
# Paper Venue 17 Locating and Editing Factual Associations in GPT (ROME) — Meng et al. NeurIPS 2022 18 Mass-Editing Memory in a Transformer (MEMIT) — Meng et al. ICLR 2023 19 Evaluating the Ripple Effects of Knowledge Editing — Cohen et al. TACL / EMNLP 2024 20 Why Does New Knowledge Create Messy Ripple Effects in LLMs? — Qin et al. EMNLP 2024 21 WikiBigEdit: Understanding the Limits of Lifelong Knowledge Editing in LLMs — 22 Model Editing at Scale Leads to Gradual and Catastrophic Forgetting — Gupta et al. ACL 2024 Findings 23 WISE: Rethinking the Knowledge Memory for Lifelong Model Editing — Wang et al. NeurIPS 2024
Track 4: Interpretability & Knowledge Mechanisms
How do transformers store and recall facts — and what are the limits?
A good starting point for getting an overview of interpretability: https://thegradient.pub/explain-yourself/
# Paper Venue 24 Scaling Monosemanticity: Extracting Interpretable Features from Claude 3 Sonnet Anthropic 2024 25 Transformer Feed-Forward Layers Are Key-Value Memories — Geva et al. EMNLP 2021 26 Dissecting Recall of Factual Associations in Auto-Regressive LMs — Geva et al. EMNLP 2023 27 Knowledge Neurons in Pretrained Transformers — Dai et al. ACL 2022 28 Unveiling Factual Recall Behaviors of LLMs through Knowledge Neurons EMNLP 2024 29 The Reversal Curse: LLMs Trained on "A is B" Fail to Learn "B is A" — Berglund et al. ICLR 2024 30 Physics of Language Models (Storage Capacity) — Allen-Zhu & Li ICML 2024
Track 5: LLM Biases, Language Learning & Knowledge Construction
Broader questions about how LLMs learn, what they get wrong systematically, and how to build structured knowledge.
# Paper Venue 31 Bias Runs Deep: Implicit Reasoning Biases in Persona-Assigned LLMs — Gupta et al. ICLR 2024 32 Investigating Subtler Biases in LLMs: Ageism, Beauty, Institutional, and Nationality Bias — Kamruzzaman et al. EMNLP 2024 33 FoodTaxo: Generating Food Taxonomies with Large Language Models ACL Industry 2025 34 Extract, Define, Canonicalize: An LLM-Based Framework for Knowledge Graph Construction EMNLP 2024
Track 6: LLM Limitations
A broader look at the fundamental shortcomings of current LLM architectures — in compositionality, long-context reasoning, counterfactual tasks, and self-correction — and what these limitations mean for reliable knowledge use.
# Paper Venue 35 NYT-Connections: A Deceptively Simple Text Classification Task that Stumps System-1 Thinkers COLING 2025 36 Mission: Impossible Language Models ACL 2024 37 Lost in the Middle: How LLMs Use Long Contexts — Liu et al. TACL 2024 38 Faith and Fate: Limits of Transformers on Compositionality — Dziri et al. NeurIPS 2023 39 Dissociating Language and Thought in LLMs: A Cognitive Perspective — Mahowald et al. Trends in Cognitive Sciences 2024 40 Large Language Models Cannot Self-Correct Reasoning Yet — Huang et al. ICLR 2024 41 Reasoning or Reciting? Exploring the Capabilities and Limitations of LLMs Through Counterfactual Tasks — Wu et al. NAACL 2024
Track 7: Responsible AI — Safety, Privacy & Misuse
# Paper Venue 42 Jailbroken: How Does LLM Safety Training Fail? — Wei et al. NeurIPS 2023 43 Universal and Transferable Adversarial Attacks on Aligned LLMs — Zou et al. arXiv 2023 44 Extracting Training Data from Large Language Models — Carlini et al. USENIX Security 2021 45 Quantifying Memorization Across Neural Language Models — Carlini et al. ICLR 2023 46 Privacy in Large Language Models: Attacks, Defenses and Future Directions arXiv 2023 47 Do Anything Now: Characterizing and Evaluating In-The-Wild Jailbreak Prompts —
Track 8: Can We Afford the Perfect Prompt?
Inspired by the EPI paper (McDonald et al.), this track examines the economics and efficiency of prompting — asking whether state-of-the-art prompting techniques are worth their computational cost.
# Paper Venue 48 Can We Afford the Perfect Prompt? Balancing Cost and Accuracy with the Economical Prompting Index — McDonald et al. ACL 2025 49 Prompt Repetition Improves Non-Reasoning LLMs — Leviathan et al. arXiv 2024 50 Chain-of-Thought Prompting Elicits Reasoning in LLMs — Wei et al. NeurIPS 2022 51 Large Language Models are Zero-Shot Reasoners ("Think step by step") — Kojima et al. NeurIPS 2022 52 Are More LLM Calls All You Need? Towards Scaling Laws of Compound Inference Systems —
Background Reading
These papers provide excellent overviews for seminar preparation:
- A Review of Knowledge in Language Models — AlKhamissi et al. (arXiv 2022) — Comprehensive survey on how knowledge is stored, probed, and edited in LLMs.
- A Survey on Hallucination in Large Language Models — Huang et al. (arXiv 2023) — Taxonomy, challenges, and open questions.
- Editing Large Language Models: A Survey — Yao et al. (arXiv 2023) — Covers ROME, MEMIT, MEND, and more.
- Knowledge Mechanisms in Large Language Models: A Survey — Wang et al. (EMNLP 2024 Findings) — Theoretical grounding for storage, retrieval, and consistency.
- A Comprehensive Study of Knowledge Editing for LLMs (KnowEdit) — Zhang et al. (arXiv 2024) — Benchmark and survey for knowledge editing.
Grading
The final grade consists of:
- Report (33%): A written report (max. 4 pages, ACL-style)
- Presentation (33%): 20-minute presentation
- Q&A (33%): 15-minute Q&A session (5 min by peers, 10 min by course team). Each participant is assigned to ask questions to two peers, randomly assigned on the seminar days.
Tentative Timeline
Date Event Tue 7.4. Application deadline Fri 10.4. Notification of placement Wed 22.4., 09:20 "Introduction to KAAI" lecture — Location: S14-745 Wed 29.4., 09:20 "Seminar survival skills" lecture + topic assignment — Location: S14-745 May Meet with advisor Mon 22.6. Reports due Mon 29.6. Slides due Mon–Tue 6.–7.7. Block seminar (full day)
Material
- Slides 1st meeting
- Slides 2nd meeting
- Report template
This seminar discusses advanced topics at the interface of LLMs and KAAI.