default search action
Mohamed Elhoseiny
Person information
Refine list
refinements active!
zoomed in on ?? of ?? records
view refined list in
export refined list as
showing all ?? records
2020 – today
- 2024
- [j4]Deyao Zhu, Jun Chen, Kilichbek Haydarov, Xiaoqian Shen, Wenxuan Zhang, Mohamed Elhoseiny:
ChatGPT Asks, BLIP-2 Answers: Automatic Questioning Towards Enriched Visual Descriptions. Trans. Mach. Learn. Res. 2024 (2024) - [c83]Eslam Abdelrahman, Pengzhan Sun, Li Erran Li, Mohamed Elhoseiny:
ImageCaptioner2: Image Captioner for Image Captioning Bias Amplification Assessment. AAAI 2024: 20902-20911 - [c82]Xiaoqian Shen, Faizan Farooq Khan, Mohamed Elhoseiny:
EmoTalker: Audio Driven Emotion Aware Talking Head Generation. ACCV (5) 2024: 131-147 - [c81]Kilichbek Haydarov, Aashiq Muhamed, Xiaoqian Shen, Jovana Lazarevic, Ivan Skorokhodov, Chamuditha Jayanga Galappaththige, Mohamed Elhoseiny:
Adversarial Text to Continuous Image Generation. CVPR 2024: 6316-6326 - [c80]Faizan Farooq Khan, Diana Kim, Divyansh Jha, Youssef Mohamed, Hanna H. Chang, Ahmed Elgammal, Luba Elliott, Mohamed Elhoseiny:
AI Art Neural Constellation: Revealing the Collective and Contrastive State of AI-Generated and Human Art. CVPR Workshops 2024: 7470-7478 - [c79]Habib Slim, Mohamed Elhoseiny:
ShapeWalk: Compositional Shape Editing Through Language-Guided Chains. CVPR 2024: 22574-22583 - [c78]Wenxuan Zhang, Paul Janson, Rahaf Aljundi, Mohamed Elhoseiny:
Overcoming Generic Knowledge Loss with Selective Parameter Update. CVPR 2024: 24046-24056 - [c77]Kilichbek Haydarov, Xiaoqian Shen, Avinash Madasu, Mahmoud Salem, Li-Jia Li, Gamaleldin Elsayed, Mohamed Elhoseiny:
Affective Visual Dialog: A Large-Scale Benchmark for Emotional Reasoning Based on Visually Grounded Conversations. ECCV (75) 2024: 18-36 - [c76]Sanjoy Chowdhury, Sayan Nag, Subhrajyoti Dasgupta, Jun Chen, Mohamed Elhoseiny, Ruohan Gao, Dinesh Manocha:
MEERKAT: Audio-Visual Large Language Model for Grounding in Space and Time. ECCV (64) 2024: 52-70 - [c75]Xiang Li, Jian Ding, Zhaoyang Chen, Mohamed Elhoseiny:
Uni3DL: A Unified Model for 3D Vision-Language Understanding. ECCV (23) 2024: 74-92 - [c74]Kirolos Ataallah, Xiaoqian Shen, Eslam Abdelrahman, Essam Sleiman, Mingchen Zhuge, Jian Ding, Deyao Zhu, Jürgen Schmidhuber, Mohamed Elhoseiny:
Goldfish: Vision-Language Understanding of Arbitrarily Long Videos. ECCV (29) 2024: 251-267 - [c73]Youssef Mohamed, Runjia Li, Ibrahim Said Ahmad, Kilichbek Haydarov, Philip Torr, Kenneth Church, Mohamed Elhoseiny:
No Culture Left Behind: ArtELingo-28, a Benchmark of WikiArt with Captions in 28 Languages. EMNLP 2024: 20939-20962 - [c72]Eslam Mohamed Bakr, Mohamed Ayman, Mahmoud Ahmed, Habib Slim, Mohamed Elhoseiny:
CoT3DRef: Chain-of-Thoughts Data-Efficient 3D Visual Grounding. ICLR 2024 - [c71]Wenxuan Zhang, Youssef Mohamed, Bernard Ghanem, Philip Torr, Adel Bibi, Mohamed Elhoseiny:
Continual Learning on a Diet: Learning from Sparsely Labeled Streams Under Constrained Computation. ICLR 2024 - [c70]Deyao Zhu, Jun Chen, Xiaoqian Shen, Xiang Li, Mohamed Elhoseiny:
MiniGPT-4: Enhancing Vision-Language Understanding with Advanced Large Language Models. ICLR 2024 - [c69]Xinliang Zhu, Arnab Dhua, Douglas Gray, I. Zeki Yalniz, Tan Yu, Mohamed Elhoseiny, Bryan A. Plummer:
Multimodal Representation and Retrieval [MRR 2024]. SIGIR 2024: 3047-3050 - [c68]Salman Khan, Izzeddin Teeti, Andrew Bradley, Mohamed Elhoseiny, Fabio Cuzzolin:
A Hybrid Graph Network for Complex Activity Detection in Video. WACV 2024: 6748-6758 - [i90]Faizan Farooq Khan, Diana Kim, Divyansh Jha, Youssef Mohamed, Hanna H. Chang, Ahmed Elgammal, Luba Elliott, Mohamed Elhoseiny:
AI Art Neural Constellation: Revealing the Collective and Contrastive State of AI-Generated and Human Art. CoRR abs/2402.02453 (2024) - [i89]Kirolos Ataallah, Xiaoqian Shen, Eslam Abdelrahman, Essam Sleiman, Deyao Zhu, Jian Ding, Mohamed Elhoseiny:
MiniGPT4-Video: Advancing Multimodal LLMs for Video Understanding with Interleaved Visual-Textual Tokens. CoRR abs/2404.03413 (2024) - [i88]Wenxuan Zhang, Youssef Mohamed, Bernard Ghanem, Philip H. S. Torr, Adel Bibi, Mohamed Elhoseiny:
Continual Learning on a Diet: Learning from Sparsely Labeled Streams Under Constrained Computation. CoRR abs/2404.12766 (2024) - [i87]Junjie Fei, Mahmoud Ahmed, Jian Ding, Eslam Mohamed Bakr, Mohamed Elhoseiny:
Kestrel: Point Grounding Multimodal LLM for Part-Aware 3D Vision-Language Understanding. CoRR abs/2405.18937 (2024) - [i86]Abdulwahab Felemban, Eslam Mohamed Bakr, Xiaoqian Shen, Jian Ding, Abduallah A. Mohamed, Mohamed Elhoseiny:
iMotion-LLM: Motion Prediction Instruction Tuning. CoRR abs/2406.06211 (2024) - [i85]Xiang Li, Jian Ding, Mohamed Elhoseiny:
VRSBench: A Versatile Vision-Language Benchmark Dataset for Remote Sensing Image Understanding. CoRR abs/2406.12384 (2024) - [i84]Kirolos Ataallah, Chenhui Gou, Eslam Abdelrahman, Khushbu Pahwa, Jian Ding, Mohamed Elhoseiny:
InfiniBench: A Comprehensive Benchmark for Large Multimodal Models in Very Long Video Understanding. CoRR abs/2406.19875 (2024) - [i83]Sanjoy Chowdhury, Sayan Nag, Subhrajyoti Dasgupta, Jun Chen, Mohamed Elhoseiny, Ruohan Gao, Dinesh Manocha:
Meerkat: Audio-Visual Large Language Model for Grounding in Space and Time. CoRR abs/2407.01851 (2024) - [i82]Asma Alkhaldi, Raneem Alnajim, Layan Alabdullatef, Rawan Alyahya, Jun Chen, Deyao Zhu, Ahmed Alsinan, Mohamed Elhoseiny:
MiniGPT-Med: Large Language Model as a General Interface for Radiology Diagnosis. CoRR abs/2407.04106 (2024) - [i81]Kirolos Ataallah, Xiaoqian Shen, Eslam Abdelrahman, Essam Sleiman, Mingchen Zhuge, Jian Ding, Deyao Zhu, Jürgen Schmidhuber, Mohamed Elhoseiny:
Goldfish: Vision-Language Understanding of Arbitrarily Long Videos. CoRR abs/2407.12679 (2024) - [i80]Zilyu Ye, Jinxiu Liu, Ruotian Peng, Jinjin Cao, Zhiyang Chen, Yiyang Zhang, Ziwei Xuan, Mingyuan Zhou, Xiaoqian Shen, Mohamed Elhoseiny, Qi Liu, Guo-Jun Qi:
Openstory++: A Large-scale Dataset and Benchmark for Instance-aware Open-domain Visual Storytelling. CoRR abs/2408.03695 (2024) - [i79]Chenhui Gou, Abdulwahab Felemban, Faizan Farooq Khan, Deyao Zhu, Jianfei Cai, Hamid Rezatofighi, Mohamed Elhoseiny:
How Well Can Vision Language Models See Image Details? CoRR abs/2408.03940 (2024) - [i78]Wenxuan Zhang, Philip H. S. Torr, Mohamed Elhoseiny, Adel Bibi:
Bi-Factorial Preference Optimization: Balancing Safety-Helpfulness in Language Models. CoRR abs/2408.15313 (2024) - [i77]Xiaoqian Shen, Yunyang Xiong, Changsheng Zhao, Lemeng Wu, Jun Chen, Chenchen Zhu, Zechun Liu, Fanyi Xiao, Balakrishnan Varadarajan, Florian Bordes, Zhuang Liu, Hu Xu, Hyunwoo J. Kim, Bilge Soran, Raghuraman Krishnamoorthi, Mohamed Elhoseiny, Vikas Chandra:
LongVU: Spatiotemporal Adaptive Compression for Long Video-Language Understanding. CoRR abs/2410.17434 (2024) - [i76]Han Bao, Yue Huang, Yanbo Wang, Jiayi Ye, Xiangqi Wang, Xiuying Chen, Mohamed Elhoseiny, Xiangliang Zhang:
AutoBench-V: Can Large Vision-Language Models Benchmark Themselves? CoRR abs/2410.21259 (2024) - [i75]Youssef Mohamed, Runjia Li, Ibrahim Said Ahmad, Kilichbek Haydarov, Philip Torr, Kenneth Ward Church, Mohamed Elhoseiny:
No Culture Left Behind: ArtELingo-28, a Benchmark of WikiArt with Captions in 28 Languages. CoRR abs/2411.03769 (2024) - [i74]Jun Chen, Dannong Xu, Junjie Fei, Chun-Mei Feng, Mohamed Elhoseiny:
Document Haystacks: Vision-Language Reasoning Over Piles of 1000+ Documents. CoRR abs/2411.16740 (2024) - 2023
- [c67]Xiaoqian Shen, Xiang Li, Mohamed Elhoseiny:
MoStGAN-V: Video Generation with Temporal Motion Styles. CVPR 2023: 5652-5661 - [c66]Jun Chen, Ming Hu, Darren J. Coker, Michael L. Berumen, Blair R. Costelloe, Sara Beery, Anna Rohrbach, Mohamed Elhoseiny:
MammalNet: A Large-Scale Video Benchmark for Mammal Recognition and Behavior Understanding. CVPR 2023: 13052-13061 - [c65]Jun Chen, Deyao Zhu, Guocheng Qian, Bernard Ghanem, Zhicheng Yan, Chenchen Zhu, Fanyi Xiao, Sean Chang Culatana, Mohamed Elhoseiny:
Exploring Open-Vocabulary Semantic Segmentation from CLIP Vision Encoder Distillation Only. ICCV 2023: 699-710 - [c64]Wenxuan Zhang, Paul Janson, Kai Yi, Ivan Skorokhodov, Mohamed Elhoseiny:
Continual Zero-Shot Learning through Semantically Guided Generative Random Walks. ICCV 2023: 11540-11551 - [c63]Eslam Mohamed Bakr, Pengzhan Sun, Xiaoqian Shen, Faizan Farooq Khan, Li Erran Li, Mohamed Elhoseiny:
HRS-Bench: Holistic, Reliable and Scalable Benchmark for Text-to-Image Models. ICCV 2023: 19984-19996 - [c62]Runjia Li, Shuyang Sun, Mohamed Elhoseiny, Philip H. S. Torr:
OxfordTVG-HIC: Can Machine Make Humorous Captions from Images? ICCV 2023: 20236-20246 - [c61]Faizan Farooq Khan, Xiang Li, Andrew J. Temple, Mohamed Elhoseiny:
FishNet: A Large-scale Dataset and Benchmark for Fish Recognition, Detection, and Functional Trait Prediction. ICCV 2023: 20439-20449 - [c60]Deyao Zhu, Li Erran Li, Mohamed Elhoseiny:
Value Memory Graph: A Graph-Structured World Model for Offline Reinforcement Learning. ICLR 2023 - [c59]Hang Xu, Wenxuan Zhang, Jiawei Fei, Yuzhe Wu, Tingwen Xie, Jun Huang, Yuchen Xie, Mohamed Elhoseiny, Panos Kalnis:
SLAMB: Accelerated Large Batch Training with Sparse Communication. ICML 2023: 38801-38825 - [i73]Deyao Zhu, Yuhui Wang, Jürgen Schmidhuber, Mohamed Elhoseiny:
Guiding Online Reinforcement Learning with Action-Free Offline Pretraining. CoRR abs/2301.12876 (2023) - [i72]Deyao Zhu, Jun Chen, Kilichbek Haydarov, Xiaoqian Shen, Wenxuan Zhang, Mohamed Elhoseiny:
ChatGPT Asks, BLIP-2 Answers: Automatic Questioning Towards Enriched Visual Descriptions. CoRR abs/2303.06594 (2023) - [i71]Xiaoqian Shen, Xiang Li, Mohamed Elhoseiny:
MoStGAN-V: Video Generation with Temporal Motion Styles. CoRR abs/2304.02777 (2023) - [i70]Jun Chen, Deyao Zhu, Kilichbek Haydarov, Xiang Li, Mohamed Elhoseiny:
Video ChatCaptioner: Towards Enriched Spatiotemporal Descriptions. CoRR abs/2304.04227 (2023) - [i69]Eslam Mohamed Bakr, Pengzhan Sun, Li Erran Li, Mohamed Elhoseiny:
ImageCaptioner2: Image Captioner for Image Captioning Bias Amplification Assessment. CoRR abs/2304.04874 (2023) - [i68]Eslam Mohamed Bakr, Pengzhan Sun, Xiaoqian Shen, Faizan Farooq Khan, Li Erran Li, Mohamed Elhoseiny:
HRS-Bench: Holistic, Reliable and Scalable Benchmark for Text-to-Image Models. CoRR abs/2304.05390 (2023) - [i67]Deyao Zhu, Jun Chen, Xiaoqian Shen, Xiang Li, Mohamed Elhoseiny:
MiniGPT-4: Enhancing Vision-Language Understanding with Advanced Large Language Models. CoRR abs/2304.10592 (2023) - [i66]Jun Chen, Deyao Zhu, Guocheng Qian, Bernard Ghanem, Zhicheng Yan, Chenchen Zhu, Fanyi Xiao, Mohamed Elhoseiny, Sean Chang Culatana:
Exploring Open-Vocabulary Semantic Segmentation without Human Labels. CoRR abs/2306.00450 (2023) - [i65]Jun Chen, Ming Hu, Darren J. Coker, Michael L. Berumen, Blair R. Costelloe, Sara M. Beery, Anna Rohrbach, Mohamed Elhoseiny:
MammalNet: A Large-scale Video Benchmark for Mammal Recognition and Behavior Understanding. CoRR abs/2306.00576 (2023) - [i64]Runjia Li, Shuyang Sun, Mohamed Elhoseiny, Philip H. S. Torr:
OxfordTVG-HIC: Can Machine Make Humorous Captions from Images? CoRR abs/2307.11636 (2023) - [i63]Wenxuan Zhang, Paul Janson, Kai Yi, Ivan Skorokhodov, Mohamed Elhoseiny:
Continual Zero-Shot Learning through Semantically Guided Generative Random Walks. CoRR abs/2308.12366 (2023) - [i62]Wenxuan Zhang, Paul Janson, Rahaf Aljundi, Mohamed Elhoseiny:
Overcoming General Knowledge Loss with Selective Parameter Finetuning. CoRR abs/2308.12462 (2023) - [i61]Kilichbek Haydarov, Xiaoqian Shen, Avinash Madasu, Mahmoud Salem, Jia Li, Gamaleldin Elsayed, Mohamed Elhoseiny:
Affective Visual Dialog: A Large-Scale Benchmark for Emotional Reasoning Based on Visually Grounded Conversations. CoRR abs/2308.16349 (2023) - [i60]Eslam Mohamed Bakr, Mohamed Ayman, Mahmoud Ahmed, Habib Slim, Mohamed Elhoseiny:
CoT3DRef: Chain-of-Thoughts Data-Efficient 3D Visual Grounding. CoRR abs/2310.06214 (2023) - [i59]Jun Chen, Deyao Zhu, Xiaoqian Shen, Xiang Li, Zechun Liu, Pengchuan Zhang, Raghuraman Krishnamoorthi, Vikas Chandra, Yunyang Xiong, Mohamed Elhoseiny:
MiniGPT-v2: large language model as a unified interface for vision-language multi-task learning. CoRR abs/2310.09478 (2023) - [i58]Salman Khan, Izzeddin Teeti, Andrew Bradley, Mohamed Elhoseiny, Fabio Cuzzolin:
A Hybrid Graph Network for Complex Activity Detection in Video. CoRR abs/2310.17493 (2023) - [i57]Habib Slim, Xiang Li, Yuchen Li, Mahmoud Ahmed, Mohamed Ayman, Ujjwal Upadhyay, Ahmed Abdelreheem, Arpit Prajapati, Suhail Pothigara, Peter Wonka, Mohamed Elhoseiny:
3DCoMPaT++: An improved Large-scale 3D Vision Dataset for Compositional Recognition. CoRR abs/2310.18511 (2023) - [i56]Eslam Mohamed Bakr, Liangbing Zhao, Vincent Tao Hu, Matthieu Cord, Patrick Pérez, Mohamed Elhoseiny:
ToddlerDiffusion: Flash Interpretable Controllable Diffusion Model. CoRR abs/2311.14542 (2023) - [i55]Botos Csaba, Wenxuan Zhang, Matthias Müller, Ser-Nam Lim, Mohamed Elhoseiny, Philip H. S. Torr, Adel Bibi:
Label Delay in Continual Learning. CoRR abs/2312.00923 (2023) - [i54]Xiaoqian Shen, Mohamed Elhoseiny:
StoryGPT-V: Large Language Models as Consistent Story Visualizers. CoRR abs/2312.02252 (2023) - [i53]Xiang Li, Jian Ding, Zhaoyang Chen, Mohamed Elhoseiny:
Uni3DL: Unified Model for 3D and Language Understanding. CoRR abs/2312.03026 (2023) - 2022
- [c58]Ivan Skorokhodov, Sergey Tulyakov, Mohamed Elhoseiny:
StyleGAN-V: A Continuous Video Generator with the Price, Image Quality and Perks of StyleGAN2. CVPR 2022: 3616-3626 - [c57]Jun Chen, Han Guo, Kai Yi, Boyang Li, Mohamed Elhoseiny:
VisualGPT: Data-efficient Adaptation of Pretrained Language Models for Image Captioning. CVPR 2022: 18009-18019 - [c56]Jun Chen, Aniket Agarwal, Sherif Abdelkarim, Deyao Zhu, Mohamed Elhoseiny:
RelTransformer: A Transformer-Based Long-Tail Visual Relationship Recognition. CVPR 2022: 19485-19495 - [c55]Youssef Mohamed, Faizan Farooq Khan, Kilichbek Haydarov, Mohamed Elhoseiny:
It is Okay to Not Be Okay: Overcoming Emotional Bias in Affective Image Captioning by Contrastive Data Collection. CVPR 2022: 21231-21240 - [c54]Yuchen Li, Ujjwal Upadhyay, Habib Slim, Ahmed Abdelreheem, Arpit Prajapati, Suhail Pothigara, Peter Wonka, Mohamed Elhoseiny:
3D CoMPaT: Composition of Materials on Parts of 3D Things. ECCV (8) 2022: 110-127 - [c53]Kai Yi, Xiaoqian Shen, Yunhao Gou, Mohamed Elhoseiny:
Exploring Hierarchical Graph Representation for Large-Scale Zero-Shot Image Classification. ECCV (20) 2022: 116-132 - [c52]Abduallah A. Mohamed, Deyao Zhu, Warren Vu, Mohamed Elhoseiny, Christian G. Claudel:
Social-Implicit: Rethinking Trajectory Prediction Evaluation and The Effectiveness of Implicit Maximum Likelihood Estimation. ECCV (22) 2022: 463-479 - [c51]Youssef Mohamed, Mohamed Abdelfattah, Shyma Alhuwaider, Feifan Li, Xiangliang Zhang, Kenneth Church, Mohamed Elhoseiny:
ArtELingo: A Million Emotion Annotations of WikiArt with Emphasis on Diversity over Language and Culture. EMNLP 2022: 8770-8785 - [c50]Divyansh Jha, Kai Yi, Ivan Skorokhodov, Mohamed Elhoseiny:
Creative Walk Adversarial Networks: Novel Art Generation with Probabilistic Random Walk Deviation from Style Norms. ICCC 2022: 195-204 - [c49]Eslam Mohamed Bakr, Yasmeen Alsaedy, Mohamed Elhoseiny:
Look Around and Refer: 2D Synthetic Semantics Knowledge Distillation for 3D Visual Grounding. NeurIPS 2022 - [c48]Guocheng Qian, Yuchen Li, Houwen Peng, Jinjie Mai, Hasan Hammoud, Mohamed Elhoseiny, Bernard Ghanem:
PointNeXt: Revisiting PointNet++ with Improved Training and Scaling Strategies. NeurIPS 2022 - [c47]Ahmed Abdelreheem, Ujjwal Upadhyay, Ivan Skorokhodov, Rawan Al Yahya, Jun Chen, Mohamed Elhoseiny:
3DRefTransformer: Fine-Grained Object Identification in Real-World Scenes Using Natural Language. WACV 2022: 607-616 - [i52]Yuanpeng Li, Joel Hestness, Mohamed Elhoseiny, Liang Zhao, Kenneth Church:
Efficiently Disentangle Causal Representations. CoRR abs/2201.01942 (2022) - [i51]Kai Yi, Xiaoqian Shen, Yunhao Gou, Mohamed Elhoseiny:
Exploring Hierarchical Graph Representation for Large-Scale Zero-Shot Image Classification. CoRR abs/2203.01386 (2022) - [i50]Abduallah A. Mohamed, Deyao Zhu, Warren Vu, Mohamed Elhoseiny, Christian G. Claudel:
Social-Implicit: Rethinking Trajectory Prediction Evaluation and The Effectiveness of Implicit Maximum Likelihood Estimation. CoRR abs/2203.03057 (2022) - [i49]Youssef Mohamed, Faizan Farooq Khan, Kilichbek Haydarov, Mohamed Elhoseiny:
It is Okay to Not Be Okay: Overcoming Emotional Bias in Affective Image Captioning by Contrastive Data Collection. CoRR abs/2204.07660 (2022) - [i48]Jun Chen, Ming Hu, Boyang Li, Mohamed Elhoseiny:
Efficient Self-supervised Vision Pretraining with Local Masked Reconstruction. CoRR abs/2206.00790 (2022) - [i47]Deyao Zhu, Li Erran Li, Mohamed Elhoseiny:
Value Memory Graph: A Graph-Structured World Model for Offline Reinforcement Learning. CoRR abs/2206.04384 (2022) - [i46]Guocheng Qian, Yuchen Li, Houwen Peng, Jinjie Mai, Hasan Abed Al Kader Hammoud, Mohamed Elhoseiny, Bernard Ghanem:
PointNeXt: Revisiting PointNet++ with Improved Training and Scaling Strategies. CoRR abs/2206.04670 (2022) - [i45]Paul Janson, Wenxuan Zhang, Rahaf Aljundi, Mohamed Elhoseiny:
A Simple Baseline that Questions the Use of Pretrained-Models in Continual Learning. CoRR abs/2210.04428 (2022) - [i44]Youssef Mohamed, Mohamed Abdelfattah, Shyma Alhuwaider, Feifan Li, Xiangliang Zhang, Kenneth Ward Church, Mohamed Elhoseiny:
ArtELingo: A Million Emotion Annotations of WikiArt with Emphasis on Diversity over Language and Culture. CoRR abs/2211.10780 (2022) - [i43]Eslam Mohamed Bakr, Yasmeen Alsaedy, Mohamed Elhoseiny:
Look Around and Refer: 2D Synthetic Semantics Knowledge Distillation for 3D Visual Grounding. CoRR abs/2211.14241 (2022) - 2021
- [c46]Ahmed Ayyad, Yuchen Li, Raden Muaz, Shadi Albarqouni, Mohamed Elhoseiny:
Semi-Supervised Few-Shot Learning with Prototypical Random Walks. MetaDL@AAAI 2021: 45-57 - [c45]Deyao Zhu, Mohamed Zahran, Li Erran Li, Mohamed Elhoseiny:
Motion Forecasting with Unlikelihood Training in Continuous Space. CoRL 2021: 1003-1012 - [c44]Ivan Skorokhodov, Savva Ignatyev, Mohamed Elhoseiny:
Adversarial Generation of Continuous Images. CVPR 2021: 10753-10764 - [c43]Panos Achlioptas, Maks Ovsjanikov, Kilichbek Haydarov, Mohamed Elhoseiny, Leonidas J. Guibas:
ArtEmis: Affective Language for Visual Art. CVPR 2021: 11569-11579 - [c42]Divyansh Jha, Hanna H. Chang, Mohamed Elhoseiny:
Wölfflin's Affective Generative Analysis for Visual Art. ICCC 2021: 429-433 - [c41]Ivan Skorokhodov, Grigorii Sotnikov, Mohamed Elhoseiny:
Aligning Latent and Image Spaces to Connect the Unconnectable. ICCV 2021: 14124-14133 - [c40]Sherif Abdelkarim, Aniket Agarwal, Panos Achlioptas, Jun Chen, Jiaji Huang, Boyang Li, Kenneth Church, Mohamed Elhoseiny:
Exploring Long Tail Visual Relationship Recognition with Large Vocabulary. ICCV 2021: 15901-15910 - [c39]Ivan Skorokhodov, Mohamed Elhoseiny:
Class Normalization for (Continual)? Generalized Zero-Shot Learning. ICLR 2021 - [c38]Deyao Zhu, Mohamed Zahran, Li Erran Li, Mohamed Elhoseiny:
HalentNet: Multimodal Trajectory Forecasting with Hallucinative Intents. ICLR 2021 - [i42]Mohamed Elhoseiny, Kai Yi, Mohamed Elfeki:
CIZSL++: Creativity Inspired Generative Zero-Shot Learning. CoRR abs/2101.00173 (2021) - [i41]Panos Achlioptas, Maks Ovsjanikov, Kilichbek Haydarov, Mohamed Elhoseiny, Leonidas J. Guibas:
ArtEmis: Affective Language for Visual Art. CoRR abs/2101.07396 (2021) - [i40]Jun Chen, Han Guo, Kai Yi, Boyang Li, Mohamed Elhoseiny:
VisualGPT: Data-efficient Image Captioning by Balancing Visual Input and Linguistic Knowledge from Pretraining. CoRR abs/2102.10407 (2021) - [i39]Ivan Skorokhodov, Grigorii Sotnikov, Mohamed Elhoseiny:
Aligning Latent and Image Spaces to Connect the Unconnectable. CoRR abs/2104.06954 (2021) - [i38]Mohamed Elhoseiny, Divyansh Jha, Kai Yi, Ivan Skorokhodov:
Imaginative Walks: Generative Random Walk Deviation Loss for Improved Unseen Learning Representation. CoRR abs/2104.09757 (2021) - [i37]Jun Chen, Aniket Agarwal, Sherif Abdelkarim, Deyao Zhu, Mohamed Elhoseiny:
RelTransformer: Balancing the Visual Relationship Detection from Local Context, Scene and Memory. CoRR abs/2104.11934 (2021) - [i36]Kai Yi, Mohamed Elhoseiny:
Domain-Aware Continual Zero-Shot Learning. CoRR abs/2112.12989 (2021) - [i35]Ivan Skorokhodov, Sergey Tulyakov, Mohamed Elhoseiny:
StyleGAN-V: A Continuous Video Generator with the Price, Image Quality and Perks of StyleGAN2. CoRR abs/2112.14683 (2021) - 2020
- [c37]Abduallah A. Mohamed, Kun Qian, Mohamed Elhoseiny, Christian G. Claudel:
Social-STGCNN: A Social Spatio-Temporal Graph Convolutional Neural Network for Human Trajectory Prediction. CVPR 2020: 14412-14420 - [c36]Panos Achlioptas, Ahmed Abdelreheem, Fei Xia, Mohamed Elhoseiny, Leonidas J. Guibas:
ReferIt3D: Neural Listeners for Fine-Grained 3D Object Identification in Real-World Scenes. ECCV (1) 2020: 422-440 - [c35]Sayna Ebrahimi, Mohamed Elhoseiny, Trevor Darrell, Marcus Rohrbach:
Uncertainty-guided Continual Learning with Bayesian Neural Networks. ICLR 2020 - [c34]Yuanpeng Li, Liang Zhao, Kenneth Church, Mohamed Elhoseiny:
Compositional Language Continual Learning. ICLR 2020 - [c33]Uchenna Akujuobi, Jun Chen, Mohamed Elhoseiny, Michael Spranger, Xiangliang Zhang:
Temporal Positive-unlabeled Learning for Biomedical Hypothesis Generation via Risk Estimation. NeurIPS 2020 - [i34]Abduallah A. Mohamed, Kun Qian, Mohamed Elhoseiny, Christian G. Claudel:
Social-STGCNN: A Social Spatio-Temporal Graph Convolutional Neural Network for Human Trajectory Prediction. CoRR abs/2002.11927 (2020) - [i33]Sherif Abdelkarim, Panos Achlioptas, Jiaji Huang, Boyang Li, Kenneth Church, Mohamed Elhoseiny:
Long-tail Visual Relationship Recognition with a Visiolinguistic Hubless Loss. CoRR abs/2004.00436 (2020) - [i32]Jun Chen, Robert Hoehndorf, Mohamed Elhoseiny, Xiangliang Zhang:
Efficient long-distance relation extraction with DG-SpanBERT. CoRR abs/2004.03636 (2020) - [i31]