Jay Karhade

Generalizing Multi-Modal Perception

Hello! I'm an incoming Robotics PhD student and currently an M.S. Robotics student at CMU, where I'm advised by Dr. Sebastian Scherer at the AirLab.

My goal is to build generalized perception systems that learn to spatially and temporally reason about the environment through shared multi-sensory representations. I am interested in enhancing scene understanding and reconstruction by utilizing priors from world-models.

At present, my research focuses on using 2D, 3D and thermal data for multi-modal localization, and exploiting pre-trained representations for multi-sensor calibration and dense mapping.

Earlier at CMU, I had the opportunity to develop some amazing algorithms for multi-robot localization and mapping. In my undergrad, I explored Neural Point Cloud Rendering at ARC, NUS and AI for healthcare at BITS Pilani,Hyderabad.

I've been fortunate to collaborate with some wonderful researchers including Dr. Jonathon Luiten, Dr. Krishna Murthy, Dr. Sourav Garg, Prof. Deva Ramanan,and Prof. Rajesh Tripathy and Prof. Marcelo Ang.

Email  /  CV  /  Google Scholar  /  Twitter  /  Github

profile photo

Recent News

  • [February/2024]Accepted CMU RI PhD offer for Fall 2024!

  • [November/2023]SplaTAM is released!

  • [September/2023]Workshop co-organizer for Closing the Loop on Localization at IROS 2023! Also presenting 2 papers!

  • [August/2023]AnyLoc is released!

  • [June/2023]Reviewing for Field Robotics

  • [March/2023]Organizing Committee Member for ICCV'23 Workshop on Robot Learning and SLAM!

  • [March/2023]Reviewing for RSS 2023 and RAL

  • [Sept/2022] Started M.S.Robotics at CMU!

Selected Research

SplaTAM: Splat, Track & Map 3D Gaussians for Dense RGB-D SLAM
Nikhil Keetha, Jay Karhade, Krishna Murthy Jatavallabhulah, Gengshan Yang, Sebastian Scherer, Deva Ramanan and Jonathon Luiten
(CVPR 2024)
project page / video / Paper

Gaussian Splatting Meets Dense SLAM!

Robust Lidar Place Recognition with RoPE enhanced OverlapTransformer
Jay Karhade, Sebastian Scherer
Last-Mile Delivery Workshop, IROS, 2023
Short Paper

Modifying OverlapTransformer and adding Rotary Positional Embeddings to make Lidar Place Recognition robust under geometric aliasing.

AnyLoc: Towards universal Place Recognition
Nikhil Keetha*, Avneesh Mishra*, Jay Karhade*, Krishna Murthy Jatavallabhulah, Sebastian Scherer, K. Madhava Krishna, Sourav Garg,
RAL, 2023, ICRA 2024
project page / video / Paper

Combining self-supervised foundation model features from DINOv2 with unsupervised aggregation techniques to achieve zero-shot 400% improvement on Visual Place Recognition and emergence of semantic domains.

SubT-MRS: A Subterranean, Multi-Robot, Multi-Spectral and Multi-Degraded Dataset for Robust SLAM
(Under Review)
Paper

Sub-T and Multi-Robot Dataset for SLAM applications. Currently part of ICCV 2023 Challenge.

Time–frequency-domain deep learning framework for the automated detection of heart valve disorders using PCG signals.
Jay Karhade*, Shashwati Dash, Samit Kumar Ghosh, Dinesh Kumar Dash and Rajesh Kumar Tripathy
IEEE Transactions on Instrumentation and Measurement, 2022
Paper

Extract deep features time-domain and frequency-domain polynomial chirplet transform for automated detection of heart-valve disorders.

AFCNNet: Automated detection of AF using chirplet transform and deep convolutional bidirectional long short term memory network with ECG signals
Tejas Radhakrishnan*,Jay Karhade*, Samit Kumar Ghosh, Priya Ranjan Muduli, RK Tripathy, U Rajendra Acharya
Computers in Biology and Medicine, 2021
Paper

Extract deep features time-domain and frequency-domain polynomial chirplet transform for automated detection of heart-valve disorders.


Last updated on April 20th 2024. Website Template taken from Jon barron.