I am a Research Scientist at Huawei’s Research Center in Paris, where I work with Milad Sefidgaran on (Unsupervised) Reinforcement Learning and Generalization. I obtained my PhD in Information Technology at Politecnico di Milano advised by Prof. Marcello Restelli at the RL^3 Group.

You can check out my (hopefully up-to-date) CV here, but for the freshest updates, my Scholar or BlueSky profiles might be more reliable. And if you’re curious about anything, feel free to drop me an email, I’m always happy to chat!

My research is in Reinforcement Learning (RL), and I am especially interested in getting around some of the usual pain points, like relying on massive amounts of data, training everything from scratch, or needing centralized schemes when multiple agents are involved. Real-world applications don’t always play by those rules, unfortunately.

Lately, I have been focusing on what we can do before the actual task is even defined, a field called unsupervised RL: things like pre-training models that make RL agents more general, more adaptable, and with more diverse behaviors.

To that end, I have dived into topics like partial observability, multi-agency, and decision-making under general utility functions. I have also worked with Siemens’s Cyber Physical Systems Team in Wien to bring scalable multi-agent RL into industrial production scheduling, and with Inephany on how to leverage RL for hyperparameter optimization in LLMs and large models in general.

News et al.

(2026)

(2025)

(2024)

(2023)

(2022)