Ward Gauderis

prof_pic.jpg

FWO PhD Fellow

@ AILab @ VUB

Pleinlaan 9, 1050 Brussels, Belgium

Hello! I’m an FWO PhD Fellow in Brussels, supervised by Prof. Geraint Wiggins.

I refuse to treat neural networks as unstructured black boxes. Instead, I study compositionality as a mathematical foundation for deep learning. Using tensor networks and category theory, I ground the analysis of neural representations in the weight geometry of the architecture itself.

By viewing the model’s weights as a formal compositional system, emergent behaviour becomes a direct function of its algebraic wiring. This enables a divide-and-conquer approach that naturally links local mechanisms to global properties.

To this end, my work proposes tensor models as a unified framework for tractable neuro-symbolic integration and mechanistic analysis, ultimately aiming for truly compositional generalisation and interpretability.

Q-CHARM

How can compositional design improve compositional behaviour?

My FWO-funded PhD project, Q-CHARM, explores this question by distinguishing between a model’s architecture (its compositional design) and the structure that emerges during learning (its compositional behaviour).

The cure for black boxes is simply drawing better boxes.

I bridge two complementary paths: imposing explicit structure before training (neuro-symbolic design) and exposing implicit structure after training (mechanistic interpretability). Since deep networks cannot efficiently learn compositional functions from data alone, embedding domain structure is essential to guide learning toward representations that align with human understanding and generalise well.

To formalise this, I use string diagrams (life is too short for indices) to cast models as rigorous mathematical objects, cleanly separating high-level Syntax (symbolic rules and structure) from low-level Semantics (subsymbolic representations). As a devout Yoneda disciple, I believe this categorical grounding is the only way to formally reason about behaviour beyond the training set.

The practical blueprint for this vision lies in tensor models. They unify the expressivity of neural networks with the tractability of tensor networks. Because their weights possess a well-understood geometry, we can perform tractable analysis both pre- and post-training. While central to applied category theory and neuro-symbolic AI, they are rarely combined with interpretability; my work unites these perspectives.

Compositional Interpretability

Every good academic page needs a Venn diagram.

Current mechanistic interpretability lacks formal foundations, relying on post-hoc activation heuristics, effectively reading tea leaves in activation space. Because these methods often assume a layer-by-layer stratification, it is nearly impossible to tell if a feature is causally useful globally or just a local artifact.

The CompInterp framework shifts the focus from isolated features to their interactions as first-class citizens. To achieve measurable interpretability, we ground our analysis in formal decompositions rather than data-dependent heuristics:

  • Unified Algebra: By formulating weights, data, and subcircuit interactions via tensor contraction, standard matrix decompositions (SVD, ICA, etc.) naturally lift to complex architectures. Since the result remains a tensor model, we can trace discovered mechanisms back to the full architecture.
  • Weight-Based Analysis: Tensor networks capture higher-order relations between representation spaces. By analysing their polynomial coefficients directly within the weight geometry, dependence on input data becomes an explicit, optional choice, ensuring conclusions remain valid beyond the training distribution.
  • Disentangling Interactions: To filter out spurious correlations, decompositions must balance complexity and faithfulness. Because these properties are compositional, they can be traced throughout the model

We are currently scaling CompInterp methods to transformers and CNNs by leveraging their low-rank structure.

Why shouldn't math be beautiful?

Research Interests

If you want my full attention, just mention any of these…

  • Compositionality in AI: Category theory, string diagrams, geometric deep learning
  • Mechanistic Interpretability: Weight space analysis, parameter decompositions
  • Neuro-symbolic Architectures: Probabilistic circuits, tensor logic
  • Quantum-ish Mathematics: Tensor networks, Hilbert spaces, information geometry
  • Effective Theories of DL: Renormalisation, algebraic geometry, stochastic complexity
  • Models of Cognition & Creativity: Active inference, conceptual spaces

Hobbies

When I’m not agonising about model structure, I’m probably skating through the city, singing and playing piano, or falling down a philomathematical rabbit hole. I also love building FOSS, playing chess or other (board) games, and conversations that stretch the brain a little.

news

Dec 07, 2025 Our work on bilinear autoencoders got a spotlight at the Mechanistic interpretability workshop (NeurIPS 2025)! :partying_face:
Nov 01, 2025 I have been awarded a PhD Fellowship fundamental research from FWO to study the role of compositionality in deep learning! :mortar_board:
Oct 15, 2025 Come check out our deep dive on compositional interpretability at the Flanders AI Research Day 2025! :mag_right:
Sep 10, 2025 We are organizing the Choir in the Loop workshop (AIMC 2025), exploring human-AI collaboration in choirs. :musical_note:
Mar 04, 2025 Thomas Dooms and I are presenting our $\chi$-net poster at CoLoRAI (AAAI 2025)!

selected publications

  1. Finding Manifolds With Bilinear Autoencoders
    Thomas Dooms and Ward Gauderis
    In Mechanistic Interpretability Workshop: At the Thirty-Ninth Annual Conference on Neural Information Processing Systems, Oct 2025
  2. Compositionality Unlocks Deep Interpretable Models
    Thomas Dooms*, Ward Gauderis*, Geraint Wiggins, and 1 more author
    In Connecting Low-Rank Representations in AI: At the 39th Annual AAAI Conference on Artificial Intelligence, Nov 2024
  3. Quantum Theory in Knowledge Representation: A Novel Approach to Reasoning with a Quantum Model of Concepts
    Ward Gauderis and Geraint Wiggins
    Vrije Universiteit Brussel, Aug 2023