default search action
Search dblp
Full-text search
- > Home
Please enter a search query
- case-insensitive prefix search: default
e.g., sig matches "SIGIR" as well as "signal" - exact word search: append dollar sign ($) to word
e.g., graph$ matches "graph", but not "graphics" - boolean and: separate words by space
e.g., codd model - boolean or: connect words by pipe symbol (|)
e.g., graph|network
Update May 7, 2017: Please note that we had to disable the phrase search operator (.) and the boolean not operator (-) due to technical problems. For the time being, phrase search queries will yield regular prefix search result, and search terms preceded by a minus will be interpreted as regular (positive) search terms.
Author search results
no matches
Refine list
refine by author
- no options
- temporarily not available
refine by venue
- no options
- temporarily not available
refine by type
- no options
- temporarily not available
refine by access
- no options
- temporarily not available
refine by year
- no options
- temporarily not available
Publication search results
found 735 matches
- 2024
- Greeshma Arya, Mohammad Kamrul Hasan, Ashish Bagwari, Nurhizam Safie, Shayla Islam, Fatima Rayan Awad Ahmed, Aaishani De, Muhammad Attique Khan, Taher M. Ghazal:
Multimodal Hate Speech Detection in Memes Using Contrastive Language-Image Pre-Training. IEEE Access 12: 22359-22375 (2024) - Eesun Moon, A. S. M. Sharifuzzaman Sagar, Hyung Seok Kim:
Multimodal Daily-Life Emotional Recognition Using Heart Rate and Speech Data From Wearables. IEEE Access 12: 96635-96648 (2024) - Jianjun Lei, Jing Wang, Ying Wang:
Multi-level attention fusion network assisted by relative entropy alignment for multimodal speech emotion recognition. Appl. Intell. 54(17-18): 8478-8490 (2024) - Michael Neumann, Hardik Kothare, Vikram Ramanarayanan:
Multimodal speech biomarkers for remote monitoring of ALS disease progression. Comput. Biol. Medicine 180: 108949 (2024) - Ronghao Pan, José Antonio García-Díaz, Miguel Ángel Rodríguez-García, Rafael Valencia-García:
Spanish MEACorpus 2023: A multimodal speech-text corpus for emotion analysis in Spanish from natural environments. Comput. Stand. Interfaces 90: 103856 (2024) - Yujie Wan, Yuzhong Chen, Jiali Lin, Jiayuan Zhong, Chen Dong:
A knowledge-augmented heterogeneous graph convolutional network for aspect-level multimodal sentiment analysis. Comput. Speech Lang. 85: 101587 (2024) - Qi Zhou, Wannapon Suraworachet, Mutlu Cukurova:
Detecting non-verbal speech and gaze behaviours with multimodal data and computer vision to interpret effective collaborative learning interactions. Educ. Inf. Technol. 29(1): 1071-1098 (2024) - Mustaqeem Khan, Wail Gueaieb, Abdulmotaleb El-Saddik, Soonil Kwon:
MSER: Multimodal speech emotion recognition using cross-attention with deep fusion. Expert Syst. Appl. 245: 122946 (2024) - Gang Zhao, Yinan Zhang, Jie Chu:
A multimodal teacher speech emotion recognition method in the smart classroom. Internet Things 25: 101069 (2024) - Eniafe Festus Ayetiran, Özlem Özgöbek:
An inter-modal attention-based deep learning framework using unified modality for multimodal fake news, hate speech and offensive language detection. Inf. Syst. 123: 102378 (2024) - Puneet Kumar, Sarthak Malik, Balasubramanian Raman:
Interpretable multimodal emotion recognition using hybrid fusion of speech and image data. Multim. Tools Appl. 83(10): 28373-28394 (2024) - Maximilian Rosilius, Martin Spiertz, Benedikt Wirsing, Manuel Geuen, Volker Bräutigam, Bernd Ludwig:
Impact of Industrial Noise on Speech Interaction Performance and User Acceptance when Using the MS HoloLens 2. Multimodal Technol. Interact. 8(2): 8 (2024) - Samir Sadok, Simon Leglaive, Laurent Girin, Xavier Alameda-Pineda, Renaud Séguier:
A multimodal dynamical variational autoencoder for audiovisual speech representation learning. Neural Networks 172: 106120 (2024) - Auriane Boudin, Roxane Bertrand, Stéphane Rauzy, Magalie Ochs, Philippe Blache:
A multimodal model for predicting feedback position and type during conversation. Speech Commun. 159: 103066 (2024) - Zhe Chen, Hongcheng Liu, Yu Wang:
DialogMCF: Multimodal Context Flow for Audio Visual Scene-Aware Dialog. IEEE ACM Trans. Audio Speech Lang. Process. 32: 753-764 (2024) - Shiyao Cui, Jiangxia Cao, Xin Cong, Jiawei Sheng, Quangang Li, Tingwen Liu, Jinqiao Shi:
Enhancing Multimodal Entity and Relation Extraction With Variational Information Bottleneck. IEEE ACM Trans. Audio Speech Lang. Process. 32: 1274-1285 (2024) - Changkai Lin, Hongju Cheng, Qiang Rao, Yang Yang:
M${3}$SA: Multimodal Sentiment Analysis Based on Multi-Scale Feature Extraction and Multi-Task Learning. IEEE ACM Trans. Audio Speech Lang. Process. 32: 1416-1429 (2024) - Xinhao Mei, Chutong Meng, Haohe Liu, Qiuqiang Kong, Tom Ko, Chengqi Zhao, Mark D. Plumbley, Yuexian Zou, Wenwu Wang:
WavCaps: A ChatGPT-Assisted Weakly-Labelled Audio Captioning Dataset for Audio-Language Multimodal Research. IEEE ACM Trans. Audio Speech Lang. Process. 32: 3339-3354 (2024) - Alejandro Santorum Varela, Svetlana Stoyanchev, Simon Keizer, Rama Doddipatla, Kate M. Knill:
Entity Resolution in Situated Dialog With Unimodal and Multimodal Transformers. IEEE ACM Trans. Audio Speech Lang. Process. 32: 704-713 (2024) - Ziqi Yuan, Jingliang Fang, Hua Xu, Kai Gao:
Multimodal Consistency-Based Teacher for Semi-Supervised Multimodal Sentiment Analysis. IEEE ACM Trans. Audio Speech Lang. Process. 32: 3669-3683 (2024) - Yash Khurana, Swamita Gupta, R. Sathyaraj, S. P. Raja:
RobinNet: A Multimodal Speech Emotion Recognition System With Speaker Recognition for Social Interactions. IEEE Trans. Comput. Soc. Syst. 11(1): 478-487 (2024) - Ze-Yuan Huang, Qiang He, Kevin T. Maher, Xiaoming Deng, Yu-Kun Lai, Cuixia Ma, Sheng Feng Qin, Yong-Jin Liu, Hongan Wang:
SpeechMirror: A Multimodal Visual Analytics System for Personalized Reflection of Online Public Speaking Effectiveness. IEEE Trans. Vis. Comput. Graph. 30(1): 606-616 (2024) - Jordan Voas, David Harwath, Raymond Mooney:
Multimodal Contextualized Semantic Parsing from Speech. ACL (1) 2024: 7354-7369 - Ahmed El-Sayed, Omar Nasr:
AAST-NLP at Multimodal Hate Speech Event Detection 2024 : A Multimodal Approach for Classification of Text-Embedded Images Based on CLIP and BERT-Based Models. CASE 2024: 139-144 - Amrita Ganguly, Al Nahian Bin Emran, Sadiya Sayara Chowdhury Puspo, Md. Nishat Raihan, Dhiman Goswami, Marcos Zampieri:
MasonPerplexity at Multimodal Hate Speech Event Detection 2024: Hate Speech and Target Detection Using Transformer Ensembles. CASE 2024: 125-131 - Surendrabikram Thapa, Kritesh Rauniyar, Farhan Jafri, Hariram Veeramani, Raghav Jain, Sandesh Jain, Francielle Vargas, Ali Hürriyetoglu, Usman Naseem:
Extended Multimodal Hate Speech Event Detection During Russia-Ukraine Crisis - Shared Task at CASE 2024. CASE 2024: 221-228 - Yeshan Wang, Ilia Markov:
CLTL@Multimodal Hate Speech Event Detection 2024: The Winning Approach to Detecting Multimodal Hate Speech and Its Targets. CASE 2024: 73-78 - Yosuke Yamagishi:
YYama@Multimodal Hate Speech Event Detection 2024: Simpler Prompts, Better Results - Enhancing Zero-shot Detection with a Large Multimodal Model. CASE 2024: 60-66 - Cécile Macaire, Chloé Dion, Jordan Arrigo, Claire Lemaire, Emmanuelle Esperança-Rodier, Benjamin Lecouteux, Didier Schwab:
A Multimodal French Corpus of Aligned Speech, Text, and Pictogram Sequences for Speech-to-Pictogram Machine Translation. LREC/COLING 2024: 839-849 - Sepideh Kalateh, Luis Alberto Estrada-Jimenez, Sanaz Nikghadam-Hojjati, José Barata:
Multimodal Creativity State Detection from Speech and Voice. DoCEIS 2024: 123-136
skipping 705 more matches
loading more results
failed to load more results, please try again later
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.
Unpaywalled article links
Add open access links from to the list of external document links (if available).
Privacy notice: By enabling the option above, your browser will contact the API of unpaywall.org to load hyperlinks to open access articles. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Unpaywall privacy policy.
Archived links via Wayback Machine
For web page which are no longer available, try to retrieve content from the of the Internet Archive (if available).
Privacy notice: By enabling the option above, your browser will contact the API of archive.org to check for archived content of web pages that are no longer available. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Internet Archive privacy policy.
Reference lists
Add a list of references from , , and to record detail pages.
load references from crossref.org and opencitations.net
Privacy notice: By enabling the option above, your browser will contact the APIs of crossref.org, opencitations.net, and semanticscholar.org to load article reference information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Crossref privacy policy and the OpenCitations privacy policy, as well as the AI2 Privacy Policy covering Semantic Scholar.
Citation data
Add a list of citing articles from and to record detail pages.
load citations from opencitations.net
Privacy notice: By enabling the option above, your browser will contact the API of opencitations.net and semanticscholar.org to load citation information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the OpenCitations privacy policy as well as the AI2 Privacy Policy covering Semantic Scholar.
OpenAlex data
Load additional information about publications from .
Privacy notice: By enabling the option above, your browser will contact the API of openalex.org to load additional information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the information given by OpenAlex.
retrieved on 2024-09-24 09:12 CEST from data curated by the dblp team
all metadata released as open data under CC0 1.0 license
see also: Terms of Use | Privacy Policy | Imprint