At CMU, as a master's student and as a research staff, I have developed CV/ML solutions for medical robotics as well as
e-waste recycling. My master's thesis was on automating robotic ultrasound systems and
improving ultrasound imaging especially for deformable subjects using deep learning methods. I have also worked on the topics of fine-grained image classification, 2D/3D semantic segmentation, 3D reconstruction,
novel view synthesis using NeRF, 6D pose estimation using images, and robotic manipulation.
After CMU, I will be joining Kind Humanoid as Foundation Model and Machine Learning Engineer
I'm interested in the area of computer vision, machine learning, robotics.
Specifically, I aim to address the challenges in robotic manipulation of soft bodies,
which are complex due to their dynamic nature and high degrees of freedom.
My focus is on developing visual affordance models for handling
deformable objects, moving beyond single-step manipulation to learn dense affordance
for multi-step processes. I plan to utilize large-scale internet datasets for real-world
tasks like cloth unfolding and rope handling, inspired by Dr. Deepak Pathak's work
on leveraging human videos for robotic learning. Additionally,
I'll explore integrating multi-modal data, including audio and images,
using Foundation models to develop more generalized and robust manipulation policies.
Updates
August 2023: Presented paper on unsupevised registration in MHSRS'23.
August 2023: Defended my CMU MS in Robotics thesis on "Improving robotic-ultrasound AI using optical flow".
June 2023: Our paper on unsupervised deformable registration (U-RAFT) is accepted in IROS'23.
May 2023: Our paper on motion cancellation in ultrasound images is presented in ICRA'23 RAMI workshop.
October 2022: Presented my paper on multi-modal fine-grained image classification (RGB-X) for e-waste classification in IROS'22.
September 2022: My paper on e-waste recycling (RGB-X) featured on Apple Machine Learning Research.
May 2022: Our paper on autonomous robotics ultrasound scanning is presented in ICRA'22.
April 2022: Our CMU-Apple project on e-waste recycling featured on Apple's Environmental Progress Report'22.
November 2021: Our paper on performing real-time robotic surgery using robotic ultrasound scanning published in ISMR'21.
August 2021: I started as a master's student in the Robotics Institute at CMU
April 2020: Our (Vinit Sarode's and mine) work featured on OpenAI Gym leaderboard for BipedalWalker-v2 environment
December 2019: Our paper on modular robot for locomotion is published in Journal of Intelligent & Robotic Systems.
October 2019: Our paper on learning active spine behaviors for quadruped robots (lead Shounak Bhattacharya) is accepted in ROMAN'19.
October 2019: Our paper on trajectory based deep policy search (lead Shishir Kolathaya) is accepted in ROMAN'19.
August 2019: I started as a research staff in the Biorobotics Lab at CMU.
August 2019: Graduated from BITS Pilani, Goa Campus with B.E. in Mechanical Engineering!
July 2019: Presented my paper on analysis of modular robot designs in AIR'2019.
June 2019: Defended my B.Tech Thesis on control and simulation of quadrupedal robots!
This research focuses on developing a sophisticated dynamics model for
rope manipulation, building upon insights from my previous work on unsupervised
deformable registration. In this current project, we aim to decode the dynamics between
specific actions (such as picking or placing) applied to a rope and the consequent shape changes of the rope.
My thesis introduces a robotic ultrasound system with minimal human
oversight, using an unsupervised method to estimate displacement fields
in varying compression forces, aiding in realistic ultrasound image generation and
enhanced vessel segmentation. This versatile approach proves effective in
various medical scenarios, including respiratory motion cancellation and
curved needle registration, demonstrating broad applicability and improved image quality.
Making RAFT training unsupervised and applying it to predict vessel deformation under forces in ultrasound images.
We use this further to generate ultrasound images at multiple force values and improve segmentation by 12%
Phy-UGen: Simulating ultrasound image compression by using physics based NN FNU Abhimanyu,
John Galeotti,
Howie Choset
slides
Uses a physics based neural network (PhyDNet) to simulate ultrasound image compression under the application of forces.
Use Unsupervised RAFT (U-RAFT) to track pixel-movements in lung-ultrasound images to cancel the movement and reconstruct the fixed image.
Reduce the pixel movement by 76%.
RGB-
X, a multi-modal image classification approach, that utilizes
key features from external RGB images with those generated
from X-ray images to accurately classify electronic objects.
More specifically, this work develops Iterative Class Activation
Mapping (iCAM), a novel network architecture that explicitly
focuses on the finer-details in the multi-modal feature maps
that are needed for accurate electronic object classification.
We propose a one-shot image classification method featuring
a novel data randomization approach. This method utilizes 3D
point cloud reconstruction to render varied instances of
smartphone models, enhancing dataset diversity. We employ
cycle-GAN for style transfer between damaged and pristine
phones. By using domain randomization, we expand our dataset
500-fold, encompassing diverse scenarios such as cracks,
colors, backgrounds, and lighting. This approach significantly
improves smartphone detection using YOLOv5.
We propose an automated pipeline to use the Bayesian Optimization framework to guide the ultrasound probe to multiple points on the unknown surface.
Furthermore, to scan the proposed region, a novel 6-axis hybrid force-position controller is presented to ensure acoustic coupling.
Our pipeline learns deformation using 3D point clouds of the object deformation, material properties, force and its point of application and predicts a deformed version of the object.
As we go from images to point clouds, our method uses 2D RGB images to learn 3D deformations.
Picking household items in simulation using DDPG FNU Abhimanyu, Surya Kumar Singh video
In this work, we setup a kitchen environment in Robosuite simulator and learn to pick household objects like soda
cans using model-free reinforcment learning algorithm called DDPG.
Feel free to steal this website's source code. Do not scrape the HTML from this page itself, as it includes analytics tags that you do not want on your own website — use the github code instead. Also, consider using Leonid Keselman's Jekyll fork of this page.