AI Forensics

Accountability through interpretability in AI systems

An interdisciplinary research project of HfG Karlsruhe, Universität Kassel, Durham University, Cambridge University, and UC Santa Barbara; funded by the VolkswagenStiftung, 2022–2025.

Event

Vector Media: Notes on the Epistemology of Machine Vision

Publication

Sparse Autoencoders Do Not Find Canonical Units of Analysis

Event

Final symposium, AI and the Society of the Future

Post

Feeling AI: ways of knowing beyond the rationalist perspective

Event

Tangible LLMs: Exploring Tangible Sense-Making For Trustworthy Large Language Models

Post

Algorithms + Slimes: Learning with Slime Moulds

Sociotechnical case study

Interpretability and Accountability of AI Systems in the Sciences

Sociotechnical case study

Scopic regimes of neural wiring

Forensics toolkit

Latent mechanistic interpretability

Sociotechnical case study

AI Design Interventions for Social Diversity

Sociotechnical case study

A Pedagogy of Machines: Technology in Education and Universities in Translation

Forensics toolkit

Exploratory machine learning interfaces

Sociotechnical case study

AI Interpretability and Accountability in the Humanitarian Sector

Toolkit + Case study

Exposing.ai – the production pipeline of facial recognition systems

Sociotechnical case study

AI Interpretability and Linear Cause-Effect Models in Medicine: Is Non-Linear Diagnosis Possible?