Ramya Hebbalaguppe

I'm a researcher in visual computing. My work spans: computer vision, reliable and robust machine learning, and computer graphics. I'm a Principal Scientist in Deep Learning and Artificial Intelligence Group (DLAI), TCS Research, IIT Delhi, India. My doctoral research advised by Prof. Chetan Arora at IIT Delhi is centered on proposing novel methods to enhance reliability in deep neural networks (DNNs)

Prior to this, I was fortunate to be working with Prof. Ramakrishna Kakarala at Nanyang Technological University on High Dynamic Range Imaging algorithms which formed a part of the image processing pipeline aimed at smartphone cameras. Our research was recognized with the Best Student Paper award at the 2012 SPIE conference in Burlingame, California. I completed my master's degree at DCUs School of Electronic Engineering and Computing in 2014, advised by Prof. Noel O'Connor and Prof. Alan Smeaton. I focused on reducing false alarms in surveillance camera networks. The result of this work, a portion of our research was licensed to Netwatch Systems.

In June 2015, I started working as a research scientist at TCS Research. Since the, I have been involved in various projects related to augmented reality. Specifically, I have focused on optimizing the layout of labels for immersive experiences and developing gestural interfaces for head-mounted devices and smartphones. As a team leader, I have overseen the development of a cost-effective industrial inspection framework. Recently, my team has been working on creative content generation (images, videos, 3D/4D data).

Outside work, I enjoy painting, traveling, cooking and baking, composting, planting tree saplings, and music.

> Email  /  Google Scholar  /  Twitter  /  Github

profile photo

Representative papers spanning the following themes reliable machine learning (out-of-distribution detection, uncertainty quantification, continual learning), 3D/4D/2D computer vision are highlighted.

LoMOE: Localized Multi-Object Editing via Multi-Diffusion
Goirik Chakrabarty, Aditya Chandrasekar, Ramya Hebbalaguppe, Prathosh AP,
ACM International Conference on Multimedia , 2024  

Recent developments in the field of diffusion models have demonstrated an exceptional capacity to generate high-quality prompt-conditioned image edits. Nevertheless, previous approaches have primarily relied on textual prompts for image editing, which tend to be less effective when making precise edits to specific objects or fine-grained regions within a scene containing single/multiple objects. We introduce a novel framework for zero-shot localized multi-object editing through a multi-diffusion process to overcome this challenge. This framework empowers users to perform various operations on objects within an image, such as adding, replacing, or editing many objects in a complex scene in one pass. Our approach leverages foreground masks and corresponding simple text prompts that exert localized influences on the target regions resulting in high-fidelity image editing. A combination of cross-attention and background preservation losses within the latent space ensures that the characteristics of the object being edited are preserved while simultaneously achieving a high-quality, seamless reconstruction of the background with fewer artifacts compared to the current methods. We also curate and release a dataset dedicated to multi-object editing, named LoMOE-Bench. Our experiments against existing state-of-the-art methods demonstrate the improved effectiveness of our approach in terms of both image editing quality and inference speed.

ReMOVE: A Reference-free Metric for Object Erasure
Aditya Chandrasekar, Goirik Chakrabarty, Jai Bardhan, Ramya Hebbalaguppe, Prathosh AP,
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPRW) The First Workshop on the Evaluation of Generative Foundation Models, 2024  
project page

We introduce ReMOVE, a novel reference-free metric for assessing object erasure efficacy in diffusion-based image editing models post-generation. Unlike existing measures such as LPIPS and CLIPScore, ReMOVE addresses the challenge of evaluating inpainting without a reference image, common in practical scenarios. ReMOVE effectively distinguishes between object removal and replacement, a key issue in diffusion models due to stochastic nature of image generation.

Transfer4D: A framework for frugal motion capture and deformation transfer
Shubh Maheshwari, Rahul Narain, Ramya Hebbalaguppe,
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2023  
project page

Animating a virtual character based on a real performance of an actor is a challenging task that currently requires expensive motion capture setups and additional effort by expert animators, rendering it accessible only to large production houses. The goal of our work is to democratize this task by developing a frugal alternative termed Transfer4D that uses only commodity depth sensors and further reduces animators' effort by automating the rigging and animation transfer process. Our approach can transfer motion from an incomplete, single-view depth video to a semantically similar target mesh, unlike prior works that make a stricter assumption on the source to be noise-free and watertight.

Calibrating Deep Neural Networks Using Explicit Regularisation and Dynamic Data Pruning
Rishabh Patra*, Ramya Hebbalaguppe*, Tirtharaj Dash, Gautam Shroff, Lovekesh Vig,
IEEE/CVF Winter Conference on Applications of Computer Vision, 2023   -- [Spotlight Presentation]
project page

We demonstrate state-of-the-art Deep Neural Network calibration performance via proposing a differentiable loss term that can be used effectively in gradient descent optimisation and dynamic data pruning strategy not only enhances legitimate high confidence samples to enhance trust in DNN classifiers but also reduce the training time for calibration.

A Novel Data Augmentation Technique for Out-of-Distribution Sample Detection using Compounded Corruptions
Ramya Hebbalaguppe, Soumya Suvra Ghosal, Jatin Prakash, Harshad Khadilkar, Chetan Arora,
European Conference on Machine Learning , 2022  
project page

We propose a novel Compounded Corruption(CnC) technique for the Out-of-Distribution data augmentation. One of the major advantages of CnC is that it does not require any hold-out data apart from the training set. Our extensive comparison with 20 methods from the major conferences in last 4 years show that a model trained using CnC based data augmentation, significantly outperforms SOTA, both in terms of OOD detection accuracy as well as inference time.

A stitch in time saves nine: A train-time regularizing loss for improved neural network calibration
Ramya Hebbalaguppe*, Jatin Prakash, Neelabh Madan*, Chetan Arora,
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) , 2022   -- [ORAL Presentation]
project page

We propose a novel auxiliary loss function: Multi-class Difference in Confidence and Accuracy (MDCA) for Deep Neural Network calibration. The loss can be combined with any application specific classification losses for image, NLP, Speech domains. We also demonstrate the utility of the loss in semantic segmentation tasks.

Current Research Team, Research and Innovation Park, IIT Delhi

  1. Tamoghno Kandar (IIT Bombay) Topic: Enhancing Trustworthiness in Foundational Models
  2. Jai Bardhan (IIIT Hyderabad) Topic: 3D Skeletonization/ Diffusion Models for creative content creation


(The list of researchers include full-time, pre-doctoral fellows, and research interns.)

  1. Adarsh Kappiyath (Spring'24) → Doctoral student at the University of Surrey, UK
  2. Meghal Dani → Doctoral student at IMPRS-IS, Max Planck School
  3. Surabhi Nath → Doctoral student at the Max Planck School of Cognition and the MPI for Biological Cybernetics
  4. Goirik Chakrabarty → Doctoral student (fall'24), University Göttingen Sinz lab
  5. Jatin Prakash → Doctoral student at New York University
  6. Neelabh Madan → Doctoral student at New York University
  7. Gaurav Gupta → Doctoral student at Rice University
  8. Apoorv Khattar → Doctoral student at University of Manchester, UK
  9. Neel Rakholia → Masters Student at Stanford
  10. Sharan Yalburgi → Visiting researcher at MIT proabilistic ML project
  11. Srinidhi Hegde → Masters Student at UMD
  12. Shubh Maheshwari → Graduate student at UCSD
  13. Pranay Gupta → Masters student at CMU
  14. Jitender Maurya → Researcher, Toshiba
  15. Archie Gupta → SDE, Microsoft
  16. Varun Jain → Masters student at CMU → Microsoft Fellow
  17. Additya Popli → SDE at Google
  18. Kshitijz Jain → Grad student at IITD
  19. Aravind Udupa → Grad student at IITD
  20. Soumya Suvra Ghosal → Masters Student at University of Wisconsin
  21. Gaurav Garg → Accenture
  22. Ramakrishna Perla → TTEC Digital

Academia: Thesis supervision

  1. Aditya C (IISc, Bangalore) - co-supervisor for M. Tech Thesis Topic: Metrics for image editing
  2. Shreyash Mohatta (BITS, Goa) - supervised M.Tech thesis on Egocentric Realtime Gesture Recognition with Dr. Ashwin Srinivasan → Masters student at NCSU
  3. Rishabh Patra (BITS, Goa)- supervised B.Tech thesis on uncertainty calibration with Dr. Tirtharaj Dash → SDE Amazon
  4. Ashwin Vaswani (BITS, Goa) - supervised BTP on Data-free Iterative Knowledge Distillation with Prof. Ashwin Srinivasan → Google Research → Masters student at CMU
  5. Het Shah (BITS, Goa) - supervised BTP on Knowledge Distillation, Pruning and Quantization with Prof. Ashwin Srinivasan → Research Associate at Google Research

Website inspired from Jon Barron's.