João M.

João M.

Deep Learning Research Engineer

Netherlands
Trusted member since 2025
10 years of experience

He specializes in building advanced models, including large language models (LLMs), capable of code refactoring, bug detection, and continual learning. João works extensively with PyTorch and deploys models on cloud platforms and high-performance computing systems.

Prior to ASML, he led research teams at GAIPS Lab, published in leading AI conferences, and secured competitive grants from the U.S. Air Force and FCT. He also taught AI courses, earning a Teaching Excellence Award for his contributions to education.

João’s key projects include advancing continual learning techniques, enabling AI to acquire new knowledge without forgetting previous tasks, and applying reinforcement learning to train models more efficiently with less data. He is passionate about making AI systems more effective, practical, and continually improving.

Main expertise

PythonPython10 years
Machine LearningMachine Learning10 years
Data Science10 years
Scikit-learnScikit-learn10 years
11+

Experience4

ASML

Deep Learning Research Engineer

ASML
Artificial Intelligence (AI)
Nov 2024 · 1y 4m
  • Led a research team on the "LLMs for Software Engineering" project, focusing on technical debt reduction, bug detection, and documentation analysis using Large Language Models.
  • Designed, implemented, trained, tested, and deployed LLMs for automatic code refactoring and bug detection.
  • Deployed models to cloud production environments and HPC distributed computing clusters.
  • Monitored the continual performance of deployed models using tools such as MLFlow, Sacred, and Weights & Biases.
  • Connected the company’s research department with academic partners at TU/e.
DockerDocker
JavaJava
FlaskFlask
PythonPython
C++C++
53+

Deep Learning Research Engineer

GAIPS Research
Artificial Intelligence (AI)
Feb 2019 - Oct 2024 · 5y 8m
  • Designed, implemented, trained, tested, and deployed state-of-the-art deep learning architectures, including Actor-Critics, DQNs, and LLMs, using convolutional, recurrent, and attention-based mechanisms for feature extraction across a wide range of tasks.
  • Deployed models to cloud production environments on platforms such as Google Cloud, Amazon AWS, and Slurm HPC distributed computing clusters.
  • Monitored the continual performance of deployed models using tools like MLFlow, Sacred, and Weights & Biases.
  • Assembled the company’s HPC Slurm cluster.
  • Led five research teams as first author, publishing a research paper for each in top-tier AI venues, including AAAI, IJCAI, ECAI, the Artificial Intelligence Journal, and PLoS One Journal.
  • Presented AI research at top-tier international conferences such as AAAI, IJCAI, and ECAI.
  • Secured two competitive funding grants, one from the U.S. Air Force Office of Scientific Research and another from the Portuguese Foundation for Science and Technology (FCT).
  • Received the Best Paper award for the project “Helping People On The Fly: Ad Hoc Teamwork for Human-Robot Teams.”
DockerDocker
PythonPython
Data Science
JoomlaJoomla
NumPyNumPy
30+
Thales

Software Engineer

Thales
Aerospace and Defense
May 2018 - Jan 2019 · 8m
  • Reduced technical debt and increased overall test coverage of the Top Sky Tower solution, a tool for air traffic controllers to manage electronic strips.
  • Implemented and tested critical security detection systems.
JavaJava
C++C++
C#C#
WPFWPF

Software Engineer

IST IT Department
Information Technology (IT) and Services
Mar 2017 - Mar 2018 · 1y
  • Trained a Convolutional Neural Network to classify valid identity card images.
  • Implemented software for automatic and periodic backups of the university’s records to the AWS cloud.
  • Re-implemented legacy software using modern technologies such as Scala and Kotlin.
JavaJava
PythonPython
AWS S3AWS S3
ScalaScala
KotlinKotlin
4+

Assessments

Engineering excellence

João’s overall performance in a 90-minute live technical assessment ranks in the top 5% of vetted Deep Learning Research Engineers at Proxify.

Portfolio

Highlighted by João

Multi-Task Learning & Catastrophic Forgetting in Continual Reinforcement Learning 1
Sep 2017
Multi-Task Learning & Catastrophic Forgetting in Continual Reinforcement Learning

This project investigates two hypothesis regarding the use of deep reinforcement learning in multiple tasks. The first hypothesis is driven by the question of whether a deep reinforcement learning algorithm, trained on two similar tasks, is able to outperform two single-task, individually trained algorithms, by more efficiently learning a new, similar task, that none of the three algorithms has encountered before. The second hypothesis is driven by the question of whether the same multi-task deep RL algorithm, trained on two similar tasks and augmented with elastic weight consolidation (EWC), is able to retain similar performance on the new task, as a similar algorithm without EWC, whilst being able to overcome catastrophic forgetting in the two previous tasks. We show that a multi-task Asynchronous Advantage Actor-Critic (GA3C) algorithm, trained on Space Invaders and Demon Attack, is in fact able to outperform two single-tasks GA3C versions, trained individually for each single-task, when evaluated on a new, third task—namely, Phoenix.

We also show that, when training two trained multi-task GA3C algorithms on the third task, if one is augmented with EWC, it is not only able to achieve similar performance on the new task, but also capable of overcoming a substantial amount of catastrophic forgetting on the two previous tasks.

Duration1y 1m

Other projects 3

PyTorch Encoder-Decoder Attention Model
odel-based Reinforcement Learning for Ad Hoc Teamwork
TopSky Tower

Education

Instituto Superior Técnico
Instituto Superior Técnico
Computer Science2019 - 2025
Instituto Superior Técnico
Instituto Superior Técnico
Information Systems and Computer Engineering2016 - 2018
Instituto Superior Técnico
Instituto Superior Técnico
Information Systems and Computer Engineering2012 - 2016

Stop browsing.
Get matched faster.