Osher Azulay
I’m a Fulbright postdoctoral researcher at the University of Michigan, working with Prof. Stella Yu. I work at the intersection of robotics, computer vision, tactile sensing, and machine learning, aiming to advance humanoid intelligence.
Previously, I earned my Ph.D. from Tel Aviv University in 2024, under the supervision of Dr. Avishai Sintov. My doctoral work focused on robotic in-hand manipulation—specifically, enabling robots to perceive and adapt to the physical world through vision, force sensing, and, most notably, tactile feedback.
I'm open to new collaborations and always enjoy discussing robotics and related research.
feel free to reach out!
Email /
CV /
Scholar /
LinkedIn /
Github
|
|
News
- July 2025 — Started my postdoc at UofM!
- April 2025 — I gave a talk at ROAM-Lab Columbia University
- Dec 2024 - I gave invited talks at Bar-Ilan University and Technion.
- Nov 2024 - Recivied the Fulbright postdoctoral scholarship.
- Oct 2024 — Defened my PhD.
Show more
|
Selected Publications:
|
|
Embodiment-Agnostic Navigation Policy Trained with Visual Demonstrations
Nimrod Curtis*, Osher Azulay* and Avishai Sintov.
Under review.
paper
/
code
/
video
We proposed ViDEN, a framework using visual demonstrations for scalable, collision-free navigation.
|
|
Visuotactile-Based Learning for Insertion with Compliant Hands
Osher Azulay, Dhruv Metha Ramesh,
Nimrod Curtis and Avishai Sintov.
IEEE RA-L & IROS, 2025.
website
/
paper
/
code
Sim2real learning of robust precision insertion polices with compliant hands.
|
|
Augmenting Tactile Simulators with Real-like and Zero-Shot Capabilities
Osher Azulay*, Alon Mizrahi*,
Nimrod Curtis* and Avishai Sintov.
ICRA 2024.
paper
/
code
Tackling the sim-to-real problem for high resolution 3D round sensors using bi-directional Generative Adversarial Networks.
|
|
AllSight: A Low-Cost and High-Resolution Round Tactile Sensor with Zero-Shot Learning Capability
Osher Azulay,
Nimrod Curtis, Rotem Sokolovsky, Guy Levitski, Daniel Slomovik, Guy Lilling and
Avishai Sintov.
IEEE RA-L & ICRA, 2024.
paper
/
video
/
code
Introducing AllSight, an optical tactile sensor with a round 3D structure designed for robotic inhand manipulation tasks
|
|
Haptic-Based and SE(3)-Aware Object Insertion Using Compliant Hands
Osher Azulay, Max Monastirsky and Avishai Sintov.
IEEE RA-L & ICRA, 2023.
paper
/
video
Exploring complaint hands characteristics for object insertion using haptic-based residual RL.
|
|
Learning to Throw With a Handful of Samples Using Decision Transformers
Max Monastirsky, Osher Azulay and Avishai Sintov.
IEEE RA-L & IROS, 2023.
paper
/
video
Exploring the use of Decision Transformers for throwing and their ability for sim2real policy transfer.
|
|
Learning Haptic-based Object Pose Estimation for In-hand Manipulation Control with Underactuated Robotic Hands
Osher Azulay, Inbar Meir and Avishai Sintov.
IEEE Transactions on Haptics, 2022.
paper
/
video
/
code
In-hand object pose estimation and manipulation using Model Predictive Control.
|
|
Open-Sourcing Generative Models for Data-driven Robot Simulations
Eran Bamani, Osher Azulay, Anton Gurevich, and Avishai Sintov.
Data-Centric AI workshop, NeurIPS, 2021
paper
/
oral
Exploring the possibility of investing the recorded data in a generative model rather than directly to a regression model for real-robot applications.
|
|
Wheel Loader Scooping Controller Using Deep Reinforcement Learning
Osher Azulay and Amir Shapiro.
IEEE Access, 2021
paper
/
video
/
code
A deep reinforcement learning-based controller for an unmanned ground vehicle with a custom-built scooping mechanism.
|
|