Decoding reality, one pixel at a time!

I am Omprakash Chakraborty, a Google PhD Scholar at Indian Institute of Technology Kharagpur (IIT KGP), India. My current field of research is Computer Vision and Machine Intelligence working under the supervision of Dr. Abir Das. I received by Masters from the Computer Science and Engineering Department at IIT Kharagpur. Before getting into IIT Kharagpur, I had done my Bachelors from Haldia Institute of Technology.

Omprakash Images


Was awarded the prestigious Google Fellowship.

Paper - "AnyDA: Anytime Domain Adaptation" accepted in ICLR 2023.


Organized CVPR 2022 Workshop on Dynamic Neural Networks Meets Computer Vision (DNetCV).


Paper - "Semi-Supervised Action Recognition with Temporal Contrastive Learning" accepted in CVPR 2021.

Organized CVPR 2021 Workshop on Dynamic Neural Networks Meets Computer Vision (DNetCV).



AnyDA: Anytime Domain Adaptation

Omprakash Chakraborty, Aadarsh Sahoo, Rameswar Panda, Abir Das; International Conference on Learning Representations (ICLR), 2023.

[Project] [Code]

Unsupervised domain adaptation is an open and challenging problem in computer vision. While existing research shows encouraging results in addressing cross-domain distribution shift on common benchmarks, they are often limited to testing under a specific target setting. This can limit their impact for many real-world applications that present different resource constraints. In this paper, we introduce a simple yet effective framework for anytime domain adaptation that is executable with dynamic resource constraints to achieve accuracy-efficiency trade-offs under domain-shifts. We achieve this by training a single shared network using both labeled source and unlabeled data, with switchable depth, width and input resolutions on the fly to enable testing under a wide range of computation budgets. Starting with a teacher network trained from a label-rich source domain, we utilize bootstrapped recursive knowledge distillation as a nexus between source and target domains to jointly train the student network with switchable subnetworks. Extensive experiments on several diverse benchmark datasets well demonstrate the superiority of our proposed approach over state-of-the-art methods.


Semi-Supervised Action Recognition with Temporal Contrastive Learning

O. Chakraborty, A. Singh, A. Varshney, R. Panda, R. Feris, K. Saenko, A. Das; Computer Vision and Pattern Recognition (CVPR), 2021.

[Project] [Code] [Poster] [video presentation]

Learning to recognize actions from only a handful of labeled videos is a challenging problem due to the scarcity of tediously collected activity labels. We approach this problem by learning a two-pathway temporal contrastive model using unlabeled videos at two different speeds leveraging the fact that changing video speed does not change an action. Specifically, we propose to maximize the similarity between encoded representations of the same video at two different speeds as well as minimize the similarity between different videos played at different speeds. This way we use the rich supervisory information in terms of ‘time’ that is present in otherwise unsupervised pool of videos. With this simple yet effective strategy of manipulating video playback rates, we considerably outperform video extensions of sophisticated state-of-the-art semi-supervised image recognition methods across multiple diverse benchmark datasets and network architectures. Interestingly, our proposed approach benefits from out-of-domain unlabeled videos showing generalization and robustness. We also perform rigorous ablations and analysis to validate our approach.