Large Language Models and Knowledge-aware AI (Seminar)
TU Dresden | Sommersemester 2025
Large Language Models and Knowledge-aware AI (2025 Seminar)
Overview
This seminar discusses advanced topics at the interface of LLMs and KAAI. It is a block seminar and will take place on two consecutive days in the summer term 2025. There will also be two meetings at the beginning of the semester, for which participation is mandatory.
- Type: Seminar (0/2/0)
- Teacher: Simon Razniewski (lecturer)
- Modules: CMS-SEM, CMS-LM-ADV, CMS-LM-AI, INF-PM-FOR, INF-VERT2, INF-AQUA
Registration
- The number of participants is limited, with preference given to Master students
- To express interest, send an email to the lecturer (simon.razniewski@tu-dresden.de), including a short motivation statement and your transcript
- Places will be allocated based on background match (courses taken) and motivation
Topics
- Facts and hallucinations in LLMs
Do I Know This Entity? Knowledge Awareness and Hallucinations in Language Models
Javier Ferrando, Oscar Obeso, Senthooran Rajamanoharan, Neel Nanda
Locating and editing factual associations in gpt
K Meng, D Bau, A Andonian, Y Belinkov - Physical world knowledge
PhysBench: Benchmarking and Enhancing Vision-Language Models for Physical World Understanding
Wei Chow, Jiageng Mao, Boyi Li, Daniel Seita, Vitor Guizilini, Yue Wang - Knowledge representation in LLMs
The Reversal Curse: LLMs trained on "A is B" fail to learn "B is A"
Lukas Berglund et al
Supposedly Equivalent Facts That Aren't? Entity Frequency in Pre-training Induces Asymmetry in LLMs
Y He et al - Chain of thought reasoning
Chain-of-thought prompting elicits reasoning in large language models
J Wei et al
Whiteboard-of-Thought: Thinking Step-by-Step Across Modalities
Sachit Menon, Richard Zemel, Carl Vondrick - World model emergence in LLMs
Emergent World Representations: Exploring a Sequence Model Trained on a Synthetic Task
Kenneth Li, Aspen K. Hopkins, David Bau, Fernanda Viégas, Hanspeter Pfister, Martin Wattenberg
Connecting the Dots: LLMs can Infer and Verbalize Latent Structure from Disparate Training Data
Treutlein et al. - Cultural knowledge
Massively Multi-Cultural Knowledge Acquisition & LM Benchmarking
Yi Fung, Ruining Zhao, Jae Doo, Chenkai Sun, Heng Ji - Theory of mind
How FaR Are Large Language Models From Agents with Theory-of-Mind?
Pei Zhou et al. - Cognitive Psychology on LLMs
Using cognitive psychology to understand GPT-3
Marcel Binz, Eric Schulz - Historical perspective
As We May Think
Vannevar Bush - LLM versus human language learning
BabyLM challenge
Charpentier et al. - Taxonomy construction and refinement with LLMs
Refining wikidata taxonomy using large language modelsY Peng, T Bonald, M Alam
Towards ontology construction with language modelsM Funk, S Hosemann, JC Jung, C Lutz(own topic suggestions are welcome as well)
Deliverables
There are 5 deliverables. To pass the course, all have to be submitted on time. Percentages in brackets denote contribution to final grade.
- Outline of report (5%)
- Report 1st version (0%)*
- Reviews on two other reports (15%)
- Reports final version (40%)
- Presentation (40%)
- 1st revision is not graded, but the prime chance to obtain feedback from advisor and peers.
Tentative timeline
- Mon 7.4.: Application deadline
- Wed 9.4. Notification of accepted participants
- Wed 16.4., 10am-12: "Introduction to KAAI" lecture, location TBD
- Fr 25.4., 10am-12: "Seminar survival skills" lecture + topic assignment, location TBD
- Wed 7.5.: 1st deliverable due
- 12.-20.5.: Meetings with advisors
- Wed 11.6.: 2nd deliverable due + submit bids for reviewing other papers
- Thu 19.6.: 3rd deliverable due
- Thu 3.7.: 4th deliverable due
- Mon, Tu 14./15.7.: Block seminar presentations
Material
- Slides 1st meeting
- Slides 2nd meetings
- Report template
- Easychair for reviewing
This seminar discusses advanced topics at the interface of LLMs and KAAI.
Lade Bewertungsübersicht
Lade Übersicht