hi, I'm Yatharth Ahuja,
and it's great to see you here!

profile photo

I am a robotic systems master's student at the Robotics Institute, Carnegie Mellon University. In my current program, we worked with Koppers Inc. on devising an industrial cleaning robot as a part of the sponsored capstone project, where I primarily led the perception, autonomy and planning subsystems. Presently, I am working on active perception in robotic manipulation tasks at the Momentum Robotics Lab under the guidance of Prof. Jeff Ichnowksi.

I like to engineer solutions - to bring betterments using automation. I am particularly interested in medicine, energy, and manufacturing applications. In the past I have led and worked on pertaining engineering projects, such as the Dual-Axis Solar Tracker - where we led the effort of creating an automated vending cart, as well as the RCT Volume Estimator - using morphological methods, developed for the dentists at MAIDS, Delhi to estimate the volume of the root canal treated teeth; among others.

My technical exposure has been across the stack, primarily around perception and controls. Always up for a chat on technology, history, sports or any other interesting thing happening around us!


Contact Links:  

 ~  Email  |  GitHub  |  LinkedIn |  Scholar |  X  ~ 


News

  • Feb '25 - Joined the JP Morgan Chase AI Maker Space, working on Visual SLAM.
  • Nov '24 - Deployed an industrial chemical cleaning robot as part of a capstone project sponsored by Koppers Inc.
  • Sep '24 - Joined Momentum Robotics Lab at CMU, focusing on active perception in robotic manipulation tasks.
  • Aug '24 - Completed Image Analysis Internship at Glimpse Engineering, optimizing AI pipelines for metal-ion cells.
  • Mar '23 - Accepted admission to Carnegie Mellon University for the Fall 2023 MRSD program.
  • Aug '22 - Graduated from Delhi Technological University with a B.Tech in Electrical Engineering, awarded First Class with Distinction.

Experience

Technical Lab Assistant at the JP Morgan Chase & Co. AI Makerspace

February 2025 - Present

Working on Visual Inertial SLAM for Pepper Robot, and integrating with navigation APIs. I also assist students and researchers to work with robots including Kinova, Fetch and Misty robots.

Research Associate at the Momentum Robotics Lab

September 2024 - Present

Working on adaptive actions with active perception for feedback in closed-loop robotic manipulation tasks. Implemented Koopman Operator Rollout for local, and a VLM for global planning for manipulation task recovery from 8 failure states with an avg. 73% success rate.

Image Analysis Intern at Glimpse Engineering

May 2024 - August 2024

Helped improve visual analysis and data processing of 3D CT scanning of metal-ion batteries for non-destructive testing. Aligned cylindrical Metal-Ion 2.5D cell scans using morphological operations and developed a parameter auto-tuner using heuristic search to automate the pipeline. Executed multiple CI/CD GitHub Actions development cycles to ensure real-time compute optimization with CUDA and production-ready code deployment.

Research and Development Engineer at CollabLens

January 2022 - December 2022

As a part of the initial founding team, juggled with deployments and experimentations to figure out valuable deliverables to clients. Architected and implemented data acquisition pipelines, enabling initial deployments and driving $200K+ in funding. Designed the systems for efficient compute allocation for client-server deployments, optimizing code for real-time performance.

Research and Development Engineer at the RuTAG Lab, IIT Delhi

April 2020 - September 2021

Led the development of a vending cart solution, fitted with a fridge and solar panel, to solve the problem of preserving farm produce for hawking vendors. Engineered the hybrid controller based dual-axis solar tracker and improved power acquisition efficiency by 25.1% over the static model, and with a runtime of >8 hrs in deployment. Presented the final system to the government officials, which led to a product roll-out by startup adoption.

Team Lead at Unmanned Ground Vehicle, DTU Delhi

December 2018 - April 2021

Lead of the university’s Unmanned Ground Vehicle Robotics Team, managing personnel and logistics. Led the development of the UGVs Minotaur (2019) and Centaur (2020), working on embedded, controls and perception systems. As a team, we secured 5th Position in Cyber Security Challenge and 9th overall at Intelligent Ground Vehicle Competition (IGVC) 2019 held at Oakland University, Michigan.

Publications

Full list can also be viewed here.

Depth Reconstruction Based Visual SLAM Using ORB Feature Extraction
Yatharth Ahuja, Tushar Nitharwal, Utkarsh Sundriyal, S. Indu, A. K. Mandpura
Journal of Images and Graphics (10th Edition) 2022

Effort to exploit visual information from a monocular input to derive depth information and combine it with the keypoints extracted from monocular input to be processed by ORB SLAM.

Development and deployment of a Hybrid Controller for a Dual-Axis Solar Tracker System
Yatharth Ahuja, Ghanvir Singh, Sushrut Singh, Suraj Bhat, S. K. Saha
International Conference on Electrical, Computer, Communications and Mechatronics Engineering (ICECCME) 2022

A system constructed for renewable energy harvesting techniques to a mobile street vending cart solution. The strategy of control deployed is that of a Hybrid Controller, which is a novel combination of Astronomical equations based on open-loop control and Differential Flatness based control over the tracking movement of the photovoltaic panel used.

Selective Lossy Image Compression for Autonomous Systems
Yatharth Ahuja, Shreyan Sood
XXIII Symposium on Image, Signal Processing and Artificial Vision (STSIVA) 2021

Object detection interfaced with selective lossy image compression techniques to improve the efficiency of subsequent image operations and reduce the memory requirement for storing the images in autonomous applications of self-driving vehicles.

Relevant Projects

Multimodal Odometry: Visual + LiDAR

September - November 2024

Developed an odometry model combining vision and LiDAR for better motion tracking. Improved accuracy on KITTI and TartanAir datasets.

Fruit Ninja Robot

January - May 2024

Built a robot to play Fruit Ninja using a 6-DOF arm and YoloV8 for object tracking. Achieved 94% accuracy and 0.6s reaction time.

mono3Dbox

January - May 2024

Enhanced 3D object detection on the KITTI dataset using a fine-tuned YoloV8, achieving better accuracy with optimized training.

Morphological RCT Drill Volume Estimation

July 2021

Designed a vision-based method to estimate tooth drill volumes for clinical use at MAIDS, New Delhi.





Find more details about my projects and experiences here.





Modified version of this website.