Yihong Chen
May 2025
Education:
UCL (Ph.D., CS)
Tsinghua (B.Eng., EE)
Experience:
Meta FAIR
Microsoft Research
Research:
Language Models
Knowledge Graphs
Continual Learning
Goal:
Robust, Steerable, Controllable AI
Structure
Explore how AI capture and acquire real world regularities
Destructure
Explore how AI break structures for adaptability and controllability
Applied Systems
Pretraining on graphs (AKBC 2021) and texts (NeurIPS 2023)
Theoretical Contribution
Unifying factorization models and GNNs (NeurIPS 2022)
\[ \begin{aligned} &\textbf{Theorem (Message Passing in FMs):} \\ &\text{The gradient descent operator } \text{GD} \text{ on the node embeddings of a DistMult model} \\ &\text{with the maximum likelihood objective and a multi-relational graph } \mathcal{T} \text{ over entities } \mathcal{E} \\ &\text{induces a message-passing operator whose composing functions are:} \end{aligned} \]
\[ q_{\mathrm{M}}(\phi[v], r, \phi[w]) = \begin{cases} \phi[w] \odot g(r) & \text{if } (r,w) \in \mathcal{N}_{+}^1[v], \\ (1 - P_\theta (v|w, r)) \phi[w] \odot g(r) & \text{if } (r, w) \in \mathcal{N}_-^1[v] \end{cases} \]
\[ q_{\mathrm{A}}(\{m[v, r, w] : (r,w) \in \mathcal{N}^1[v]\}) = \sum_{(r,w) \in \mathcal{N}^1[v]} m[v,r,w] \]
\[ q_{\mathrm{U}}(\phi[v], z[v]) = \phi[v] + \alpha z[v] - \beta n[v] \]
\[ n[v]= \frac{|\mathcal{N}_{+}^{1}[v]|}{|\mathcal{T}|} \mathbb{E}_{ P_{\mathcal{N}_+^{1}[v]} } \mathbb{E}_{ u \sim P_{\theta}(\cdot|v, r)} g(r) \odot \phi[u] + \frac{|\mathcal{T}^{-v}|}{|\mathcal{T}|} \mathbb{E}_{ P_{\mathcal{T}^{-v}} } P_\theta(v|s, r) g(r) \odot \phi[s] \]
where \( \mathcal{T}^{-v} = \{(s, r, o) \in \mathcal{T} : s \neq v \land o \neq v \} \), and \( P_{\mathcal{N}^{1}_+[v]} \), \( P_{\mathcal{T}^{-v}} \) are empirical probability distributions.
LLMs are excellent at structuring knowledge into neural weights, but poor at dismantling it. Unlike symbolic systems, they lack clearly addressable knowledge units.
Pistol: Benchmarking Structural Unlearning for LLMs (2024); unlearning difficulty increases as data inter-connectivity grows
The body is not general enough so requires lots of customization to the new language
Is there any way to make the body more general?
Is there any way to make the body more general?
Is there any way to make the body more general?
Is there any way to make the body more general?
Is there any way to make the body more general?
General v.s. Specific
Meta-learning with minimal intervention during pretraining
Meta-learning with minimal intervention during pretraining
Overfitting → Collapse: Rigid models fail under distributional shift;
Solution: Active Forgetting allows training more adaptive language models (NeurIPS 2023)
Fundamentally, the model does not know what it knows — or what it does not know.
Reformatting LLMs for making hidden model knowledge accessible.
Jet Expansions (2024): Extracting human-readable symbolic structures from LLM residual computation
Early steps learn meaningless bigrams like (yaml, Adam). As training progresses, the model picks up more sensible bigrams such as (its, own) and (make, sure).
RLHF improves ToxiGen scores, but LLMs like Llama-2-7B-Chat still retain toxic knowledge. With increasingly explicit prompts, their toxicity resurfaces: 84% for hard prompts. Jet bi-gram analysis (our method) confirms that RLHF mostly hides, rather than removes, toxic patterns.
🧱 Structure is both a solution and a problem
It empowers learning but can also trap learning due to outdated or inappropriate knowledge.
⚖️ Controllability means both building and breaking structure
Useful structure must be constructed — but also selectively remove to support adaptation, as in Active Forgetting.
🔍 To control models, we must reveal what they know
Neural models don't store knowledge in explicit units. Tools like Jet Expansions help us surface and inspect these hidden internal structures.
Questions and Discussion
Yihong Chen | yihong.chen@cs.ucl.ac.uk | github.com/yihong-chen
Relation Prediction for Multi-Relational Graph Representations
Chen, Yihong; Minervini, Pasquale; Riedel, Sebastian; Stenetorp, Pontus (AKBC 2021)
ReFactorGNNs: Revisiting Factorisation-Based Models
Chen, Yihong; Mishra, Pushkar; Franceschi, Luca; Minervini, Pasquale; Stenetorp, Pontus; Riedel, Sebastian (NeurIPS 2022)
Improving Language Plasticity via Pretraining with Active Forgetting
Chen, Yihong; Marchisio, Kelly; Raileanu, Roberta; Adelani, David Ifeoluwa; Stenetorp, Pontus; Riedel, Sebastian; Artetxe, Mikel (NeurIPS 2023)
PISTOL: Benchmarking Structural Unlearning
Qiu, Xinchi; Shen, William F.; Chen, Yihong; Cancedda, Nicola; Stenetorp, Pontus; Lane, Nicholas D.
arXiv:2406.16810, 2024
Jet Expansions of Residual Computation
Chen, Yihong; Xu, Xiangxiang; Lu, Yao; Stenetorp, Pontus; Franceschi, Luca
arXiv:2410.06024, 2024