I am a Research Scientist at Huawei’s Research Center in Paris, where I work with Milad Sefidgaran on (Unsupervised) Reinforcement Learning and Generalization. I obtained my PhD in Information Technology at Politecnico di Milano advised by Prof. Marcello Restelli at the RL^3 Group.

You can check out my (hopefully up-to-date) CV here, but for the freshest updates, my Scholar or BlueSky profiles might be more reliable. And if you’re curious about anything, feel free to drop me an email, I’m always happy to chat!

My research is in Reinforcement Learning (RL), and I am especially interested in getting around some of the usual pain points, as training everything from scratch. Real-world applications don’t always play by those rules, unfortunately.

Lately, I have been focusing on what we’d better do before the agent is released in the wild, a field called unsupervised RL: things like pre-training models (policies, policy spaces, representations etc.) that make RL agents more general, more adaptable, and with a more diverse set of skills.

To that end, I have also dived into topics like partial observability, multi-agency, and decision-making under general utility functions. Throughout my PhD, I have also worked with Siemens’s Cyber Physical Systems Team in Wien to bring scalable multi-agent RL into industrial production scheduling, and with Inephany’s great team on how to leverage RL for hyperparameter optimization for large models.

News et al.

(2026)

(2025)

(2024)

(2023)

(2022)