Gaël Gendron

Machine Learning Engineer / PhD Student in AI, Causality and Reasoning

Building reasoning systems that can be trusted

Profile

I am Gaël (pron. [gah-el]), a PhD student at the University of Auckland, New Zealand, creating trustworthy AI systems that can reason and understand the world around them. I focus on making safe, reliable and robust AI based on factuality and causality, with a particular interest in large reasoning models (LLMs/LRMs).

I showcased my work at top AI conferences and received the University of Auckland Best Student Published Paper in Computer Science award in 2023. As part of my work, I conducted the first evaluation of large language models (LLMs) on abstract reasoning, highlighting their brittleness and limitations. I developed ICLM, a novel modular language model architecture based on causal principles for out-of-distribution reasoning, and showed that causal models can improve the learning of interpretable, robust and domain-invariant mechanisms. My latest work, the Causal Cartographer is the first end-to-end framework for causal extraction and inference with LLM agents, enabling them to reason more reliably and efficiently in counterfactual environments (decreasing inference cost by up to 70%).

Latest Research

Causal Cartographer: From Mapping to Reasoning Over Counterfactual Worlds

Causal Cartographer: From Mapping to Reasoning Over Counterfactual Worlds

Gaël Gendron, Jože Rožanec, Michael Witbrock, Gillian Dobbie

We introduce a retrieval-augmented system for causal extraction and representating causal knowledge, and a methodology for provably estimating real-world counterfactuals. We show that causal-guided step-by-step LRMs can achieve competitive performance while greatly reducing LLMs' context and output length, decreasing inference cost up to 70%.

Large Language Models Causal Representation Learning Causal World Models Real-World Counterfactual Reasoning Natural Language Processing
Counterfactual Causal Inference in Natural Language

Counterfactual Causal Inference in Natural Language

Gaël Gendron, Jože Rožanec, Michael Witbrock, Gillian Dobbie

We build the first causal extraction and counterfactual causal inference system for natural language, and propose a new direction for model oversight and strategic foresight.

Large Language Models Causal Extraction Causal Inference Counterfactual Reasoning Natural Language Processing
Independent Causal Language Models

Independent Causal Language Models

Gaël Gendron, Bao Trung Nguyen, Alex Peng, Michael Witbrock, Gillian Dobbie

We develop a novel modular language model architecture sparating inference into independant causal modules, and show that it can be used to improve abstract reasoning performance and robustness for out-of-distribution settings.

Large Language Models Abstract Reasoning Independent Causal Mechanisms Out-Of-Distribution Generalization
Behaviour Modelling of Social Agents

Behaviour Modelling of Social Agents

Gaël Gendron, Yang Chen, Mitchell Rogers, Yiping Liu, Mihailo Azhar, Shahrokh Heidari, David Arturo Soriano Valdez, Kobe Knowles, Padriac O'Leary, Simon Eyre, Michael Witbrock, Gillian Dobbie, Jiamou Liu, Patrice Delmas

We model the behaviour of interacting social agents (e.g. meerkats) using a combination of causal inference and graph neural networks, and demonstrate increased efficiency and interpretability compared to existing architectures.

Graph Neural Networks Causal Structure Discovery Agent-Based Modelling
Large Language Models Are Not Strong Abstract Reasoners

Large Language Models Are Not Strong Abstract Reasoners

Gaël Gendron, Qiming Bao, Michael Witbrock, Gillian Dobbie

We evaluate the performance of large language models on abstract reasoning tasks and show that they fail to adapt to unseen reasoning chains, highlighting a lack of generalization and robustness.

Large Language Models Abstract Reasoning Evaluation Out-Of-Distribution Generalization