default search action
Makarand Tapaswi
Person information
Refine list
refinements active!
zoomed in on ?? of ?? records
view refined list in
export refined list as
2020 – today
- 2024
- [i36]Darshan Singh S, Zeeshan Khan, Makarand Tapaswi:
FiGCLIP: Fine-Grained CLIP Adaptation via Densely Annotated Videos. CoRR abs/2401.07669 (2024) - [i35]Yash Khandelwal, Mayur Arvind, Sriram Kumar, Ashish Gupta, Sachin Kumar Danisetty, Piyush Bagad, Anish Madan, Mayank Lunayach, Aditya Annavajjala, Abhishek Maiti, Sansiddh Jain, Aman Dalmia, Namrata Deka, Jerome White, Jigar Doshi, Angjoo Kanazawa, Rahul Panicker, Alpan Raval, Srinivas Rana, Makarand Tapaswi:
NurtureNet: A Multi-task Video-based Approach for Newborn Anthropometry. CoRR abs/2405.05530 (2024) - [i34]Haran Raajesh, Naveen Reddy Desanur, Zeeshan Khan, Makarand Tapaswi:
MICap: A Unified Model for Identity-aware Movie Descriptions. CoRR abs/2405.11483 (2024) - [i33]Aditya Kumar Singh, Dhruv Srivastava, Makarand Tapaswi:
"Previously on ..." From Recaps to Story Summarization. CoRR abs/2405.11487 (2024) - [i32]Darshana Saravanan, Darshan Singh S, Varun Gupta, Zeeshan Khan, Vineet Gandhi, Makarand Tapaswi:
VELOCITI: Can Video-Language Models Bind Semantic Concepts through Time? CoRR abs/2406.10889 (2024) - [i31]Kawshik Manikantan, Shubham Toshniwal, Makarand Tapaswi, Vineet Gandhi:
Major Entity Identification: A Generalizable Alternative to Coreference Resolution. CoRR abs/2406.14654 (2024) - 2023
- [c43]Piyush Bagad, Makarand Tapaswi, Cees G. M. Snoek:
Test of Time: Instilling Video-Language Models with a Sense of Time. CVPR 2023: 2503-2516 - [c42]Dhruv Srivastava, Aditya Kumar Singh, Makarand Tapaswi:
How You Feelin'? Learning Emotions and Mental States in Movie Scenes. CVPR 2023: 2517-2528 - [c41]Darshan Singh S, Anchit Gupta, C. V. Jawahar, Makarand Tapaswi:
Unsupervised Audio-Visual Lecture Segmentation. WACV 2023: 5221-5230 - [c40]Dhaval Taunk, Lakshya Khanna, Siri Venkata Pavan Kumar Kandru, Vasudeva Varma, Charu Sharma, Makarand Tapaswi:
GrapeQA: GRaph Augmentation and Pruning to Enhance Question-Answering. WWW (Companion Volume) 2023: 1138-1144 - [i30]Piyush Bagad, Makarand Tapaswi, Cees G. M. Snoek:
Test of Time: Instilling Video-Language Models with a Sense of Time. CoRR abs/2301.02074 (2023) - [i29]Dhaval Taunk, Lakshya Khanna, Pavan Kandru, Vasudeva Varma, Charu Sharma, Makarand Tapaswi:
GrapeQA: GRaph Augmentation and Pruning to Enhance Question-Answering. CoRR abs/2303.12320 (2023) - [i28]Dhruv Srivastava, Aditya Kumar Singh, Makarand Tapaswi:
How you feelin'? Learning Emotions and Mental States in Movie Scenes. CoRR abs/2304.05634 (2023) - [i27]Aroof Aimen, Arsh Verma, Makarand Tapaswi, Narayanan C. Krishnan:
Generalized Cross-domain Multi-label Few-shot Learning for Chest X-rays. CoRR abs/2309.04462 (2023) - [i26]Prajneya Kumar, Eshika Khandelwal, Makarand Tapaswi, Vishnu Sreekumar:
Eye vs. AI: Human Gaze and Model Attention in Video Memorability. CoRR abs/2311.16484 (2023) - 2022
- [c39]Pierre-Louis Guhur, Shizhe Chen, Ricardo Garcia Pinel, Makarand Tapaswi, Ivan Laptev, Cordelia Schmid:
Instruction-driven history-aware policies for robotic manipulations. CoRL 2022: 175-187 - [c38]Shizhe Chen, Pierre-Louis Guhur, Makarand Tapaswi, Cordelia Schmid, Ivan Laptev:
Think Global, Act Local: Dual-scale Graph Transformer for Vision-and-Language Navigation. CVPR 2022: 16516-16526 - [c37]Shizhe Chen, Pierre-Louis Guhur, Makarand Tapaswi, Cordelia Schmid, Ivan Laptev:
Learning from Unlabeled 3D Environments for Vision-and-Language Navigation. ECCV (39) 2022: 638-655 - [c36]Vladimír Petrík, Mohammad Nomaan Qureshi, Josef Sivic, Makarand Tapaswi:
Learning Object Manipulation Skills from Video via Approximate Differentiable Physics. IROS 2022: 7375-7382 - [c35]Jaidev Shriram, Makarand Tapaswi, Vinoo Alluri:
Sonus Texere! Automated Dense Soundtrack Construction for Books using Movie Adaptations. ISMIR 2022: 535-542 - [c34]Shizhe Chen, Pierre-Louis Guhur, Makarand Tapaswi, Cordelia Schmid, Ivan Laptev:
Language Conditioned Spatial Relation Reasoning for 3D Object Grounding. NeurIPS 2022 - [c33]Zeeshan Khan, C. V. Jawahar, Makarand Tapaswi:
Grounded Video Situation Recognition. NeurIPS 2022 - [i25]Shizhe Chen, Pierre-Louis Guhur, Makarand Tapaswi, Cordelia Schmid, Ivan Laptev:
Think Global, Act Local: Dual-scale Graph Transformer for Vision-and-Language Navigation. CoRR abs/2202.11742 (2022) - [i24]Vladimír Petrík, Mohammad Nomaan Qureshi, Josef Sivic, Makarand Tapaswi:
Learning Object Manipulation Skills from Video via Approximate Differentiable Physics. CoRR abs/2208.01960 (2022) - [i23]Shizhe Chen, Pierre-Louis Guhur, Makarand Tapaswi, Cordelia Schmid, Ivan Laptev:
Learning from Unlabeled 3D Environments for Vision-and-Language Navigation. CoRR abs/2208.11781 (2022) - [i22]Pierre-Louis Guhur, Shizhe Chen, Ricardo Garcia Pinel, Makarand Tapaswi, Ivan Laptev, Cordelia Schmid:
Instruction-driven history-aware policies for robotic manipulations. CoRR abs/2209.04899 (2022) - [i21]Zeeshan Khan, C. V. Jawahar, Makarand Tapaswi:
Grounded Video Situation Recognition. CoRR abs/2210.10828 (2022) - [i20]Darshan Singh S, Anchit Gupta, C. V. Jawahar, Makarand Tapaswi:
Unsupervised Audio-Visual Lecture Segmentation. CoRR abs/2210.16644 (2022) - [i19]Shizhe Chen, Pierre-Louis Guhur, Makarand Tapaswi, Cordelia Schmid, Ivan Laptev:
Language Conditioned Spatial Relation Reasoning for 3D Object Grounding. CoRR abs/2211.09646 (2022) - [i18]Arsh Verma, Makarand Tapaswi:
Can we Adopt Self-supervised Pretraining for Chest X-Rays? CoRR abs/2211.12931 (2022) - [i17]Jaidev Shriram, Makarand Tapaswi, Vinoo Alluri:
Sonus Texere! Automated Dense Soundtrack Construction for Books using Movie Adaptations. CoRR abs/2212.01033 (2022) - 2021
- [j3]Makarand Tapaswi, Vijay Kumar, Ivan Laptev:
Long term spatio-temporal modeling for action detection. Comput. Vis. Image Underst. 210: 103242 (2021) - [c32]Pierre-Louis Guhur, Makarand Tapaswi, Shizhe Chen, Ivan Laptev, Cordelia Schmid:
Airbert: In-domain Pretraining for Vision-and-Language Navigation. ICCV 2021: 1614-1623 - [c31]Rahul Vigneswaran, Marc T. Law, Vineeth N. Balasubramanian, Makarand Tapaswi:
Feature generation for long-tail classification. ICVGIP 2021: 41:1-41:9 - [i16]Pierre-Louis Guhur, Makarand Tapaswi, Shizhe Chen, Ivan Laptev, Cordelia Schmid:
Airbert: In-domain Pretraining for Vision-and-Language Navigation. CoRR abs/2108.09105 (2021) - [i15]Rahul Vigneswaran, Marc T. Law, Vineeth N. Balasubramanian, Makarand Tapaswi:
Feature Generation for Long-tail Classification. CoRR abs/2111.05956 (2021) - 2020
- [j2]Vivek Sharma, Makarand Tapaswi, M. Saquib Sarfraz, Rainer Stiefelhagen:
Video Face Clustering With Self-Supervised Representation Learning. IEEE Trans. Biom. Behav. Identity Sci. 2(2): 145-157 (2020) - [c30]Vladimír Petrík, Makarand Tapaswi, Ivan Laptev, Josef Sivic:
Learning Object Manipulation Skills via Approximate State Estimation from Real Videos. CoRL 2020: 296-312 - [c29]Anna Kukleva, Makarand Tapaswi, Ivan Laptev:
Learning Interactions and Relationships Between Movie Characters. CVPR 2020: 9846-9855 - [c28]Vivek Sharma, Makarand Tapaswi, M. Saquib Sarfraz, Rainer Stiefelhagen:
Clustering based Contrastive Learning for Improving Face Representations. FG 2020: 109-116 - [i14]Anna Kukleva, Makarand Tapaswi, Ivan Laptev:
Learning Interactions and Relationships between Movie Characters. CoRR abs/2003.13158 (2020) - [i13]Vivek Sharma, Makarand Tapaswi, M. Saquib Sarfraz, Rainer Stiefelhagen:
Clustering based Contrastive Learning for Improving Face Representations. CoRR abs/2004.02195 (2020) - [i12]Vivek Sharma, Makarand Tapaswi, Rainer Stiefelhagen:
Deep Multimodal Feature Encoding for Video Ordering. CoRR abs/2004.02205 (2020) - [i11]Vladimír Petrík, Makarand Tapaswi, Ivan Laptev, Josef Sivic:
Learning Object Manipulation Skills via Approximate State Estimation from Real Videos. CoRR abs/2011.06813 (2020)
2010 – 2019
- 2019
- [c27]Vivek Sharma, Makarand Tapaswi, M. Saquib Sarfraz, Rainer Stiefelhagen:
Self-Supervised Learning of Face Representations for Video Face Clustering. FG 2019: 1-8 - [c26]Antoine Miech, Dimitri Zhukov, Jean-Baptiste Alayrac, Makarand Tapaswi, Ivan Laptev, Josef Sivic:
HowTo100M: Learning a Text-Video Embedding by Watching Hundred Million Narrated Video Clips. ICCV 2019: 2630-2640 - [c25]Makarand Tapaswi, Marc T. Law, Sanja Fidler:
Video Face Clustering With Unknown Number of Clusters. ICCV 2019: 5026-5035 - [c24]Seung Wook Kim, Makarand Tapaswi, Sanja Fidler:
Visual Reasoning by Progressive Module Networks. ICLR (Poster) 2019 - [i10]Vivek Sharma, Makarand Tapaswi, M. Saquib Sarfraz, Rainer Stiefelhagen:
Self-Supervised Learning of Face Representations for Video Face Clustering. CoRR abs/1903.01000 (2019) - [i9]Antoine Miech, Dimitri Zhukov, Jean-Baptiste Alayrac, Makarand Tapaswi, Ivan Laptev, Josef Sivic:
HowTo100M: Learning a Text-Video Embedding by Watching Hundred Million Narrated Video Clips. CoRR abs/1906.03327 (2019) - [i8]Makarand Tapaswi, Marc T. Law, Sanja Fidler:
Video Face Clustering with Unknown Number of Clusters. CoRR abs/1908.03381 (2019) - [i7]Atef Chaudhury, Makarand Tapaswi, Seung Wook Kim, Sanja Fidler:
The Shmoop Corpus: A Dataset of Stories with Loosely Aligned Summaries. CoRR abs/1912.13082 (2019) - 2018
- [c23]Yuhao Zhou, Makarand Tapaswi, Sanja Fidler:
Now You Shake Me: Towards Automatic 4D Cinema. CVPR 2018: 7425-7434 - [c22]Paul Vicol, Makarand Tapaswi, Lluís Castrejón, Sanja Fidler:
MovieGraphs: Towards Understanding Human-Centric Situations From Videos. CVPR 2018: 8581-8590 - [i6]Seung Wook Kim, Makarand Tapaswi, Sanja Fidler:
Progressive Reasoning by Module Composition. CoRR abs/1806.02453 (2018) - 2017
- [c21]Ruiyu Li, Makarand Tapaswi, Renjie Liao, Jiaya Jia, Raquel Urtasun, Sanja Fidler:
Situation Recognition with Graph Neural Networks. ICCV 2017: 4183-4192 - [i5]Ruiyu Li, Makarand Tapaswi, Renjie Liao, Jiaya Jia, Raquel Urtasun, Sanja Fidler:
Situation Recognition with Graph Neural Networks. CoRR abs/1708.04320 (2017) - [i4]Paul Vicol, Makarand Tapaswi, Lluís Castrejón, Sanja Fidler:
MovieGraphs: Towards Understanding Human-Centric Situations from Videos. CoRR abs/1712.06761 (2017) - 2016
- [b1]Makarand Tapaswi:
Story Understanding through Semantic Analysis and Automatic Alignment of Text and Video. Karlsruhe University, Germany, 2016 - [c20]Makarand Tapaswi, Yukun Zhu, Rainer Stiefelhagen, Antonio Torralba, Raquel Urtasun, Sanja Fidler:
MovieQA: Understanding Stories in Movies through Question-Answering. CVPR 2016: 4631-4640 - [c19]Ziad Al-Halah, Makarand Tapaswi, Rainer Stiefelhagen:
Recovering the Missing Link: Predicting Class-Attribute Associations for Unsupervised Zero-Shot Learning. CVPR 2016: 5975-5984 - [c18]Monica-Laura Haurilet, Makarand Tapaswi, Ziad Al-Halah, Rainer Stiefelhagen:
Naming TV characters by watching and analyzing dialogs. WACV 2016: 1-9 - [i3]Ziad Al-Halah, Makarand Tapaswi, Rainer Stiefelhagen:
Recovering the Missing Link: Predicting Class-Attribute Associations for Unsupervised Zero-Shot Learning. CoRR abs/1610.04787 (2016) - [i2]Manuel Martínez, Monica-Laura Haurilet, Ziad Al-Halah, Makarand Tapaswi, Rainer Stiefelhagen:
Relaxed Earth Mover's Distances for Chain- and Tree-connected Spaces and their use as a Loss Function in Deep Learning. CoRR abs/1611.07573 (2016) - 2015
- [j1]Makarand Tapaswi, Martin Bäuml, Rainer Stiefelhagen:
Aligning plot synopses to videos for story-based retrieval. Int. J. Multim. Inf. Retr. 4(1): 3-16 (2015) - [c17]Makarand Tapaswi, Martin Bäuml, Rainer Stiefelhagen:
Book2Movie: Aligning video scenes with book chapters. CVPR 2015: 1827-1835 - [c16]Makarand Tapaswi, Martin Bäuml, Rainer Stiefelhagen:
Improved weak labels using contextual cues for person identification in videos. FG 2015: 1-8 - [c15]Marin Vlastelica P., Sergey Hayrapetyan, Makarand Tapaswi, Rainer Stiefelhagen:
KIT at MediaEval 2015 - Evaluating Visual Cues for Affective Impact of Movies Task. MediaEval 2015 - [c14]Esam Ghaleb, Makarand Tapaswi, Ziad Al-Halah, Hazim Kemal Ekenel, Rainer Stiefelhagen:
Accio: A Data Set for Face Track Retrieval in Movies Across Age. ICMR 2015: 455-458 - [i1]Makarand Tapaswi, Yukun Zhu, Rainer Stiefelhagen, Antonio Torralba, Raquel Urtasun, Sanja Fidler:
MovieQA: Understanding Stories in Movies through Question-Answering. CoRR abs/1512.02902 (2015) - 2014
- [c13]Martin Bäuml, Makarand Tapaswi, Rainer Stiefelhagen:
A time pooled track kernel for person identification. AVSS 2014: 7-12 - [c12]Makarand Tapaswi, Martin Bäuml, Rainer Stiefelhagen:
StoryGraphs: Visualizing Character Interactions as a Timeline. CVPR 2014: 827-834 - [c11]Makarand Tapaswi, Cemal Cagn Corez, Martin Bäuml, Hazim Kemal Ekenel, Rainer Stiefelhagen:
Cleaning up after a face tracker: False positive removal. ICIP 2014: 253-257 - [c10]Makarand Tapaswi, Omkar M. Parkhi, Esa Rahtu, Eric Sommerlade, Rainer Stiefelhagen, Andrew Zisserman:
Total Cluster: A person agnostic clustering method for broadcast videos. ICVGIP 2014: 7:1-7:8 - [c9]Makarand Tapaswi, Martin Bäuml, Rainer Stiefelhagen:
Story-based Video Retrieval in TV series using Plot Synopses. ICMR 2014: 137 - 2013
- [c8]Martin Bäuml, Makarand Tapaswi, Rainer Stiefelhagen:
Semi-supervised Learning with Constraints for Person Identification in Multimedia Data. CVPR 2013: 3602-3609 - [c7]Hervé Bredin, Johann Poignant, Guillaume Fortier, Makarand Tapaswi, Viet Bac Le, Anindya Roy, Claude Barras, Sophie Rosset, Achintya Kumar Sarkar, Qian Yang, Hua Gao, Alexis Mignon, Jakob Verbeek, Laurent Besacier, Georges Quénot, Hazim Kemal Ekenel, Rainer Stiefelhagen:
QCompere @ REPERE 2013. SLAM@INTERSPEECH 2013: 49-54 - 2012
- [c6]Martin Bäuml, Makarand Tapaswi, Arne Schumann, Rainer Stiefelhagen:
Contextual Constraints for Person Retrieval in Camera Networks. AVSS 2012: 221-227 - [c5]Makarand Tapaswi, Martin Bäuml, Rainer Stiefelhagen:
"Knock! Knock! Who is it?" probabilistic person identification in TV-series. CVPR 2012: 2658-2665 - [c4]Hervé Bredin, Johann Poignant, Makarand Tapaswi, Guillaume Fortier, Viet Bac Le, Thibault Napoléon, Hua Gao, Claude Barras, Sophie Rosset, Laurent Besacier, Jakob Verbeek, Georges Quénot, Frédéric Jurie, Hazim Kemal Ekenel:
Fusion of Speech, Faces and Text for Person Identification in TV Broadcast. ECCV Workshops (3) 2012: 385-394 - [c3]Tomas Semela, Makarand Tapaswi, Hazim Kemal Ekenel, Rainer Stiefelhagen:
KIT at MediaEval 2012 - Content - based Genre Classification with Visual Cues. MediaEval 2012
2000 – 2009
- 2008
- [c2]Amitava Das, Ohil K. Manyam, Makarand Tapaswi:
Audio-Visual Person Authentication with Multiple Visualized-Speech Features and Multiple Face Profiles. ICVGIP 2008: 39-46 - [c1]Amitava Das, Ohil K. Manyam, Makarand Tapaswi, Veeresh Taranalli:
Multilingual spoken-password based user authentication in emerging economies using cellular phone networks. SLT 2008: 5-8
Coauthor Index
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.
Unpaywalled article links
Add open access links from to the list of external document links (if available).
Privacy notice: By enabling the option above, your browser will contact the API of unpaywall.org to load hyperlinks to open access articles. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Unpaywall privacy policy.
Archived links via Wayback Machine
For web page which are no longer available, try to retrieve content from the of the Internet Archive (if available).
Privacy notice: By enabling the option above, your browser will contact the API of archive.org to check for archived content of web pages that are no longer available. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Internet Archive privacy policy.
Reference lists
Add a list of references from , , and to record detail pages.
load references from crossref.org and opencitations.net
Privacy notice: By enabling the option above, your browser will contact the APIs of crossref.org, opencitations.net, and semanticscholar.org to load article reference information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Crossref privacy policy and the OpenCitations privacy policy, as well as the AI2 Privacy Policy covering Semantic Scholar.
Citation data
Add a list of citing articles from and to record detail pages.
load citations from opencitations.net
Privacy notice: By enabling the option above, your browser will contact the API of opencitations.net and semanticscholar.org to load citation information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the OpenCitations privacy policy as well as the AI2 Privacy Policy covering Semantic Scholar.
OpenAlex data
Load additional information about publications from .
Privacy notice: By enabling the option above, your browser will contact the API of openalex.org to load additional information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the information given by OpenAlex.
last updated on 2024-08-02 20:23 CEST by the dblp team
all metadata released as open data under CC0 1.0 license
see also: Terms of Use | Privacy Policy | Imprint