👨🏻💻 About Me
Hi, I’m Thomas Ferraz, I’m Research Scientist at NAVER LABS Europe, on the LLM Agents Team, and also PhD Candidate at Université Grenoble-Alpes, advised by Vassilina Nikoulina, Maxime Peyrard, and Eric Gaussier, exploring how large language models can reason and plan more effectively by combining efficient neural computation with interpretable latent structures.
My main research contributions rely on Efficient Learning for Language Models. Before working on LLM agents and reasoning, I contributed to Multilingual NLP and Low-Resource NLP, including works on multilingual speech recognition, distillation, and cross-lingual modeling.
I previously completed a Master’s in Applied Math & AI (MVA) at ENS Paris-Saclay and Institut Polytechnique de Paris, and an engineering degree from the University of São Paulo, where I graduated top of my class. I have also gained experience through research internships at Meta, Amazon, Apple, and NAVER LABS Europe.
🔬 Research Interests
- LLM Agents: Memory-augmented planning and reasoning for autonomous, multi-step agentic tasks.
- Efficient LLMs: Sparse, modular, and adaptive architectures enabling scalable and continual learning.
- Efficient Reasoning: Latent-space reasoning and cognitive-inspired mechanisms for faster, cheaper and robust deliberation.
- Interpretability: Neuro-symbolic methods and mechanistic analysis to expose and steer internal model computations.
📰 News
- Internships on LLM Agents Memory:
We are offering 2 internship positions in our LLM Agents team for 2026, focusing on Memory for LLM Agents. One position focuses on Prompt-based Memory and the other on Continuous Memory. We welcome applications from Master’s and PhD students. Interested in working with us next year? Apply here.
- I’M BACK TO THE ALPS! 🏔️⛷️🇫🇷:
I am excited to share that I am rejoining NAVER LABS Europe as a Research Scientist. I was intern there 2 years ago. I’ll be working on exciting topics around Memory and Reasoning for LLM Agents with Vassilina Nikoulina and Stéphane Clinchant.
- DeCRIM accepted at EMNLP 2024:
Our DeCRIM work on LLM self-correction was accepted to the EMNLP 2024 (see you in Miami 🇺🇸 in November!) and the System 2 Reasoning at Scale Workshop @ NeurIPS (see you in Vancouver 🇨🇦 in December!).
📝 Selected Publications
A selection of papers that reflect my main research focus and contributions.
-
LLM Self-Correction with DeCRIM: Decompose, Critique, and Refine for Enhanced Following of Instructions with Multiple Constraints
EMNLP, 2024 & Sys2Reasoning @ NeurIPS, 2024
TL;DR: Introducing RealInstruct to evaluate LLMs on real multi-constrained instructions, and DeCRIM self-correction that improves instruction following by decomposing requests and refining responses, enabling open LLMs to outperform GPT-4 with strong feedback.
-
Multilingual DistilWhisper: Efficient Distillation of Multi-task Speech Models via Language-Specific Experts
ICASSP, 2024
TL;DR: Proposes a lightweight adaptation method bridging the gap between small and large speech models on under-represented languages by leveraging language-specific experts and knowledge distillation, outperforming fine-tuning and LoRA with minimal overhead.
-
ZeroBERTo: Leveraging Zero-Shot Text Classification by Topic Modeling
PROPOR, 2022
TL;DR: ZeroBERTo combines topic modeling with language models for zero-shot text classification, overcoming input size limitations and reducing runtime, achieving a 12% better F1 score and 13x faster inference compared to XLM-R on a Portuguese benchmark.