I am research scientist at Google DeepMind working on robotic manipulation. I earned my PhD at MIT working with Prof. Alberto Rodriguez.
I develop algorithms and solutions that enable robots to solve new tasks with high accuracy and dexterity.
My research was supported by
LaCaixa and Facebook fellowships.
My research focuses on developing algorithms for precise robotic generalization:
making robots capable of solving many tasks without compromising their performance and reliability.
By learning general AI models of perception and control,
we can provide robots with the right tools to thrive in diverse environments and task requirements.
In my work, I have studied how learning AI models allows precise control,
and how developing accurate visuo-tactile perception enables solving complex tasks,
such as grasping, localization, and precise placing without prior experience.
My goal is to continue developing algorithms that make robots dexterous and versatile at manipulating their environment.
Final PhD work: precise pick-and-placing of objects without prior experience! Why is this important? Currently, industry can't solve
this problem for a large variety of objects. Our system can, enabling robotics solutions in a wide variety of applications where flexibility is key.
The robot is capable to precisely pick-and-place objects that it had never interacted with before.
Latest News
June 2025 Invited talk at A Robot Touch of AI: London Summer School in Robotics & AI 2025.
Towards precise generalization of robot skills: accurate pick-and-place of novel objects" M. Bauza, T. Bronars, Y. Hou, N. Chavan-Dafle, A. Rodriguez
in progress , 2022
We learn in simulation how to accurate pick-and-place objects with visuo-tactile perception. Our solution transfers to the real world and succefully handles diferent types of objects shapes without requiring prior experience.
FingerSLAM: Closed-loop Unknown Object Localization and Reconstruction from Visuo-tactile Feedback
J. Zhao, M. Bauza, E. Adelson
under review, 2022
We address the problem of using visuo-tactile feedback for 6-DoF localization and 3D reconstruction of unknown in-hand objects.
We learn in simulation how to accurate localize objects with tactile. Our solution transfers to the real world, providing reliable pose distributions from the first touch.
Our technology is used by Magna and ABB and MERL. Our tactile sensor is Gelslim.
We optimize unsupervised losses for the current input. By optimizing where we act, we bypass generalization gaps and can impose a wide variety of inductive biases.
We learn to map functions to functions by combining graph networks and attention to build computational meshes and show this new framework can solve very diverse problems.
We propose a hybrid dynamics model, simulator-augmented interaction networks (SAIN), combining a physics engine with an object-based neural network for dynamics modeling.
We augment an analytical rigid-body simulator with a neural network that learns to model uncertainty as residuals. Best Paper Award on Cognitive Robotics at IROS 2018.