FNU Abhimanyu

I am a recent graduate from Masters in Robotics(Research) from Carnegie Mellon University where I worked with Dr. Howie Choset and Dr. John Galeotti at the Biorobotics Lab, where I worked on robotics, computer vision and machine learning.

At CMU, as a master's student and as a research staff, I have developed CV/ML solutions for medical robotics as well as e-waste recycling. My master's thesis was on automating robotic ultrasound systems and improving ultrasound imaging especially for deformable subjects using deep learning methods. I have also worked on the topics of fine-grained image classification, 2D/3D semantic segmentation, 3D reconstruction, novel view synthesis using NeRF, 6D pose estimation using images, and robotic manipulation.

After CMU, I will be joining Kind Humanoid as Foundation Model and Machine Learning Engineer

Email  /  CV  /  Resume  /  Github  /  LinkedIn  /  Google Scholar

profile photo

Research

I'm interested in the area of computer vision, machine learning, robotics. Specifically, I aim to address the challenges in robotic manipulation of soft bodies, which are complex due to their dynamic nature and high degrees of freedom. My focus is on developing visual affordance models for handling deformable objects, moving beyond single-step manipulation to learn dense affordance for multi-step processes. I plan to utilize large-scale internet datasets for real-world tasks like cloth unfolding and rope handling, inspired by Dr. Deepak Pathak's work on leveraging human videos for robotic learning. Additionally, I'll explore integrating multi-modal data, including audio and images, using Foundation models to develop more generalized and robust manipulation policies.

Updates

  • August 2023: Presented paper on unsupevised registration in MHSRS'23.
  • August 2023: Defended my CMU MS in Robotics thesis on "Improving robotic-ultrasound AI using optical flow".
  • June 2023: Our paper on unsupervised deformable registration (U-RAFT) is accepted in IROS'23.
  • May 2023: Our paper on motion cancellation in ultrasound images is presented in ICRA'23 RAMI workshop.
  • October 2022: Presented my paper on multi-modal fine-grained image classification (RGB-X) for e-waste classification in IROS'22.
  • September 2022: My paper on e-waste recycling (RGB-X) featured on Apple Machine Learning Research.
  • May 2022: Our paper on autonomous robotics ultrasound scanning is presented in ICRA'22.
  • April 2022: Our CMU-Apple project on e-waste recycling featured on Apple's Environmental Progress Report'22.
  • November 2021: Our paper on performing real-time robotic surgery using robotic ultrasound scanning published in ISMR'21.
  • August 2021: I started as a master's student in the Robotics Institute at CMU
  • April 2020: Our (Vinit Sarode's and mine) work featured on OpenAI Gym leaderboard for BipedalWalker-v2 environment
  • December 2019: Our paper on modular robot for locomotion is published in Journal of Intelligent & Robotic Systems.
  • October 2019: Our paper on learning active spine behaviors for quadruped robots (lead Shounak Bhattacharya) is accepted in ROMAN'19.
  • October 2019: Our paper on trajectory based deep policy search (lead Shishir Kolathaya) is accepted in ROMAN'19.
  • August 2019: I started as a research staff in the Biorobotics Lab at CMU.
  • August 2019: Graduated from BITS Pilani, Goa Campus with B.E. in Mechanical Engineering!
  • July 2019: Presented my paper on analysis of modular robot designs in AIR'2019.
  • June 2019: Defended my B.Tech Thesis on control and simulation of quadrupedal robots!
  • June 2018: Started as a research intern at RBCCPS, IISc with Dr. Shishir Kolathaya.

Selected Projects

Physics inspired rope manipulation

This research focuses on developing a sophisticated dynamics model for rope manipulation, building upon insights from my previous work on unsupervised deformable registration. In this current project, we aim to decode the dynamics between specific actions (such as picking or placing) applied to a rope and the consequent shape changes of the rope.

Improving Robotic Ultrasound AI using Optical Flow
FNU Abhimanyu, Howie Choset
Masters Thesis, CMU, 2023  
Thesis / Slides

My thesis introduces a robotic ultrasound system with minimal human oversight, using an unsupervised method to estimate displacement fields in varying compression forces, aiding in realistic ultrasound image generation and enhanced vessel segmentation. This versatile approach proves effective in various medical scenarios, including respiratory motion cancellation and curved needle registration, demonstrating broad applicability and improved image quality.

U-RAFT: Unsupervised Deformable Ultrasound Image Registration and Its Application for Vessel Segmentation
FNU Abhimanyu, Andrew L. Orekhov, Ananya Bal, John Galeotti, Howie Choset
IROS, 2023   (Oral Presentation)
paper / video

Making RAFT training unsupervised and applying it to predict vessel deformation under forces in ultrasound images. We use this further to generate ultrasound images at multiple force values and improve segmentation by 12%

Phy-UGen: Simulating ultrasound image compression by using physics based NN
FNU Abhimanyu, John Galeotti, Howie Choset
slides

Uses a physics based neural network (PhyDNet) to simulate ultrasound image compression under the application of forces.

Unsupervised Deformable Image Registration for Respiratory Motion Compensation in Ultrasound Images
FNU Abhimanyu, Andrew L. Orekhov, John Galeotti, Howie Choset
Presented in ICRA, RAMI Workshop'23 in London,UK  
poster

Use Unsupervised RAFT (U-RAFT) to track pixel-movements in lung-ultrasound images to cancel the movement and reconstruct the fixed image. Reduce the pixel movement by 76%.

RGB-X Classification for Electronics Sorting (In collaboration with Apple Inc.)
FNU Abhimanyu, Tejas Zodage, Umesh Thillaivasan, Xinyue Lai, Rahul Chakwate, Javier Santillan, Emma Oti, Ming Zhao, Ralph Boirum Howie Choset, Matthew Travers
Presented in IROS'23, Kyoto, Japan. Featured on Apple Machine Learning Research.  
paper / video

RGB- X, a multi-modal image classification approach, that utilizes key features from external RGB images with those generated from X-ray images to accurately classify electronic objects. More specifically, this work develops Iterative Class Activation Mapping (iCAM), a novel network architecture that explicitly focuses on the finer-details in the multi-modal feature maps that are needed for accurate electronic object classification.

One shot object detection using data augmentation (In collaboration with Apple Inc.)
FNU Abhimanyu, Yigit Yakupoglu, Matt Travers, Howie Choset
video

We propose a one-shot image classification method featuring a novel data randomization approach. This method utilizes 3D point cloud reconstruction to render varied instances of smartphone models, enhancing dataset diversity. We employ cycle-GAN for style transfer between damaged and pristine phones. By using domain randomization, we expand our dataset 500-fold, encompassing diverse scenarios such as cracks, colors, backgrounds, and lighting. This approach significantly improves smartphone detection using YOLOv5.

Autonomous Ultrasound Scanning using Bayesian Optimization and Hybrid Force Control
Raghavv Goel*, FNU Abhimanyu*, Kirtan Patel, John Galeotti, Howie Choset
Presented in ICRA'22, Philadelphia, USA.  
paper / video

We propose an automated pipeline to use the Bayesian Optimization framework to guide the ultrasound probe to multiple points on the unknown surface. Furthermore, to scan the proposed region, a novel 6-axis hybrid force-position controller is presented to ensure acoustic coupling.

Modeling 3D Deformations under External Force from 2D Images
FNU Abhimanyu, Ananya Bal
Presented in 16-824, Visual Learning and Recognition by Prof. Deepak Pathak.  

Our pipeline learns deformation using 3D point clouds of the object deformation, material properties, force and its point of application and predicts a deformed version of the object. As we go from images to point clouds, our method uses 2D RGB images to learn 3D deformations.

Picking household items in simulation using DDPG
FNU Abhimanyu, Surya Kumar Singh
video

In this work, we setup a kitchen environment in Robosuite simulator and learn to pick household objects like soda cans using model-free reinforcment learning algorithm called DDPG.


Feel free to steal this website's source code. Do not scrape the HTML from this page itself, as it includes analytics tags that you do not want on your own website — use the github code instead. Also, consider using Leonid Keselman's Jekyll fork of this page.