
Deep Learning Research Engineer
He specializes in building advanced models, including large language models (LLMs), capable of code refactoring, bug detection, and continual learning. João works extensively with PyTorch and deploys models on cloud platforms and high-performance computing systems.
Prior to ASML, he led research teams at GAIPS Lab, published in leading AI conferences, and secured competitive grants from the U.S. Air Force and FCT. He also taught AI courses, earning a Teaching Excellence Award for his contributions to education.
João’s key projects include advancing continual learning techniques, enabling AI to acquire new knowledge without forgetting previous tasks, and applying reinforcement learning to train models more efficiently with less data. He is passionate about making AI systems more effective, practical, and continually improving.


Engineering excellence
João’s overall performance in a 90-minute live technical assessment ranks in the top 5% of vetted Deep Learning Research Engineers at Proxify.
1This project investigates two hypothesis regarding the use of deep reinforcement learning in multiple tasks. The first hypothesis is driven by the question of whether a deep reinforcement learning algorithm, trained on two similar tasks, is able to outperform two single-task, individually trained algorithms, by more efficiently learning a new, similar task, that none of the three algorithms has encountered before. The second hypothesis is driven by the question of whether the same multi-task deep RL algorithm, trained on two similar tasks and augmented with elastic weight consolidation (EWC), is able to retain similar performance on the new task, as a similar algorithm without EWC, whilst being able to overcome catastrophic forgetting in the two previous tasks. We show that a multi-task Asynchronous Advantage Actor-Critic (GA3C) algorithm, trained on Space Invaders and Demon Attack, is in fact able to outperform two single-tasks GA3C versions, trained individually for each single-task, when evaluated on a new, third task—namely, Phoenix.
We also show that, when training two trained multi-task GA3C algorithms on the third task, if one is augmented with EWC, it is not only able to achieve similar performance on the new task, but also capable of overcoming a substantial amount of catastrophic forgetting on the two previous tasks.






Talk to an expert and get tailored matches from our network in just 2 days.
A network of over 6,000+ tech experts
Get matched with perfect-fit talent in 2 days on average
Hire quickly and easily with 94% match success