Matthew Kowal

Member of Technical Staff

Matthew Kowal was a Member of Technical Staff at FAR.AI working on persuasion capabilities in LLMs. He specializes in interpreting the internal representations of multimodal models.

Matthew is a PhD candidate at York University, Toronto. His thesis focused on designing multi-layer interpretability methods with an emphasis on applications to understanding how spatiotemporal models process concepts across space and time. His previous research focused on interpreting the representations of CNNs with respect to specific concepts, such as measuring shape vs texture and position information.

Concept Influence: Leveraging Interpretability to Improve Performance and Efficiency in Training Data Attribution
February 19, 2026
concept-data-attribution-02-2026
Concept Influence: Leveraging Interpretability to Improve Performance and Efficiency in Training Data Attribution
concept-influence-leveraging-interpretability-to-improve-performance-and-efficiency-in-training-data-attribution
Revisiting Frontier LLMs’ Attempts to Persuade on Extreme Topics: GPT and Claude Improved, Gemini Worsened
February 11, 2026
revisiting-attempts-to-persuade
Revisiting Frontier LLMs’ Attempts to Persuade on Extreme Topics: GPT and Claude Improved, Gemini Worsened
revisiting-attempts-to-persuade
Frontier LLMs Attempt to Persuade into Harmful Topics
August 21, 2025
attempt-to-persuade-eval
It's the Thought that Counts: Evaluating the Attempts of Frontier LLMs to Persuade on Harmful Topics
its-the-thought-that-counts-evaluating-the-attempts-of-frontier-llms-to-persuade-on-harmful-topics
Concept Influence: Leveraging Interpretability to Improve Performance and Efficiency in Training Data Attribution
February 19, 2026
concept-influence-leveraging-interpretability-to-improve-performance-and-efficiency-in-training-data-attribution
Concept Influence: Leveraging Interpretability to Improve Performance and Efficiency in Training Data Attribution
concept-data-attribution-02-2026
Revisiting Frontier LLMs’ Attempts to Persuade on Extreme Topics: GPT and Claude Improved, Gemini Worsened
February 11, 2026
revisiting-attempts-to-persuade
Revisiting Frontier LLMs’ Attempts to Persuade on Extreme Topics: GPT and Claude Improved, Gemini Worsened
revisiting-attempts-to-persuade
Large language models can effectively convince people to believe conspiracies
January 9, 2026
large-language-models-can-effectively-convince-people-to-believe-conspiracies
Emergent Persuasion: Will LLMs Persuade Without Being Prompted?
October 21, 2025
emergent-persuasion-will-llms-persuade-without-being-prompted
It's the Thought that Counts: Evaluating the Attempts of Frontier LLMs to Persuade on Harmful Topics
July 20, 2025
its-the-thought-that-counts-evaluating-the-attempts-of-frontier-llms-to-persuade-on-harmful-topics
Frontier LLMs Attempt to Persuade into Harmful Topics
attempt-to-persuade-eval
Universal Sparse Autoencoders: Interpretable Cross-Model Concept Alignment
February 6, 2025
universal-sparse-autoencoders-interpretable-cross-model-concept-alignment