Osher Azulay

I’m a Fulbright postdoctoral researcher at the University of Michigan, working with Prof. Stella Yu. I work at the intersection of robotics, computer vision, tactile sensing, and machine learning, aiming to advance humanoid intelligence.

Previously, I earned my Ph.D. from Tel Aviv University in 2024, under the supervision of Dr. Avishai Sintov. My doctoral work focused on robotic in-hand manipulation—specifically, enabling robots to perceive and adapt to the physical world through vision, force sensing, and, most notably, tactile feedback.

I'm open to new collaborations and always enjoy discussing robotics and related research.

feel free to reach out!

Email  /  CV  /  Scholar  /  LinkedIn  /  Github

profile photo

News

  • July 2025 — Started my postdoc at UofM!
  • April 2025 — I gave a talk at ROAM-Lab Columbia University
  • Dec 2024 - I gave invited talks at Bar-Ilan University and Technion.
  • Nov 2024 - Recivied the Fulbright postdoctoral scholarship.

Show more

Selected Publications:

vt Embodiment-Agnostic Navigation Policy Trained with Visual Demonstrations
Nimrod Curtis*, Osher Azulay* and Avishai Sintov.
Under review.
paper / code / video

We proposed ViDEN, a framework using visual demonstrations for scalable, collision-free navigation.

vt Visuotactile-Based Learning for Insertion with Compliant Hands
Osher Azulay, Dhruv Metha Ramesh, Nimrod Curtis and Avishai Sintov.
IEEE RA-L & IROS, 2025.
website / paper / code

Sim2real learning of robust precision insertion polices with compliant hands.

ag Augmenting Tactile Simulators with Real-like and Zero-Shot Capabilities
Osher Azulay*, Alon Mizrahi*, Nimrod Curtis* and Avishai Sintov.
ICRA 2024.
paper / code

Tackling the sim-to-real problem for high resolution 3D round sensors using bi-directional Generative Adversarial Networks.

allsight AllSight: A Low-Cost and High-Resolution Round Tactile Sensor with Zero-Shot Learning Capability
Osher Azulay, Nimrod Curtis, Rotem Sokolovsky, Guy Levitski, Daniel Slomovik, Guy Lilling and Avishai Sintov.
IEEE RA-L & ICRA, 2024.
paper / video / code

Introducing AllSight, an optical tactile sensor with a round 3D structure designed for robotic inhand manipulation tasks

hapticrl Haptic-Based and SE(3)-Aware Object Insertion Using Compliant Hands
Osher Azulay, Max Monastirsky and Avishai Sintov.
IEEE RA-L & ICRA, 2023.
paper / video

Exploring complaint hands characteristics for object insertion using haptic-based residual RL.

throw Learning to Throw With a Handful of Samples Using Decision Transformers
Max Monastirsky, Osher Azulay and Avishai Sintov.
IEEE RA-L & IROS, 2023.
paper / video

Exploring the use of Decision Transformers for throwing and their ability for sim2real policy transfer.

hapticmpc Learning Haptic-based Object Pose Estimation for In-hand Manipulation Control with Underactuated Robotic Hands
Osher Azulay, Inbar Meir and Avishai Sintov.
IEEE Transactions on Haptics, 2022.
paper / video / code

In-hand object pose estimation and manipulation using Model Predictive Control.

ops Open-Sourcing Generative Models for Data-driven Robot Simulations
Eran Bamani, Osher Azulay, Anton Gurevich, and Avishai Sintov.
Data-Centric AI workshop, NeurIPS, 2021
paper / oral

Exploring the possibility of investing the recorded data in a generative model rather than directly to a regression model for real-robot applications.

komodo Wheel Loader Scooping Controller Using Deep Reinforcement Learning
Osher Azulay and Amir Shapiro.
IEEE Access, 2021
paper / video / code

A deep reinforcement learning-based controller for an unmanned ground vehicle with a custom-built scooping mechanism.


Template from source code.