Matches in SemOpenAlex for { <https://semopenalex.org/work/W4313349512> ?p ?o ?g. }
Showing items 1 to 86 of
86
with 100 items per page.
- W4313349512 endingPage "326" @default.
- W4313349512 startingPage "315" @default.
- W4313349512 abstract "Audio-visual emotion recognition aims to integrate audio and visual information for accurate emotion prediction, which is widely used in real application scenarios. However, most existing methods lack fully exploiting complementary information within modalities to obtain rich feature representations related to emotions. Recently, Transformer and CNN-based models achieve remarkable results in the field of automatic speech recognition. Motivated by this, we propose a novel audio-visual fusion network based on 3D-CNN and Convolution-augmented Transformer (Conformer) for multimodal emotion recognition. Firstly, the 3D-CNN is employed to process face sequences extracted from the video, and the 1D-CNN is used to process MFCC features of audio signals. Secondly, the visual and audio features are fed into a feature fusion module, which contains a set of convolutional layers for extracting local features and the self-attention mechanism for capturing global interactions of multimodal information. Finally, the fused features are input into linear layers to obtain the prediction results. To verify the effectiveness of the proposed method, experiments are performed on RAVDESS and a newly collected dataset named PKU-ER. The experimental results show that the proposed model achieves state-of-the-art performance in audio-only, video-only, and audio-visual fusion experiments." @default.
- W4313349512 created "2023-01-06" @default.
- W4313349512 creator A5042348823 @default.
- W4313349512 creator A5057287134 @default.
- W4313349512 creator A5071411382 @default.
- W4313349512 creator A5085600464 @default.
- W4313349512 date "2022-01-01" @default.
- W4313349512 modified "2023-09-26" @default.
- W4313349512 title "Audio-Visual Fusion Network Based on Conformer for Multimodal Emotion Recognition" @default.
- W4313349512 cites W2314395941 @default.
- W4313349512 cites W2472490454 @default.
- W4313349512 cites W2803193013 @default.
- W4313349512 cites W2807126412 @default.
- W4313349512 cites W2910191085 @default.
- W4313349512 cites W2963654155 @default.
- W4313349512 cites W3015558147 @default.
- W4313349512 cites W3095622232 @default.
- W4313349512 cites W3097777922 @default.
- W4313349512 cites W3160207687 @default.
- W4313349512 cites W3163107903 @default.
- W4313349512 cites W3199527474 @default.
- W4313349512 cites W3209279902 @default.
- W4313349512 cites W3211666705 @default.
- W4313349512 cites W4200633562 @default.
- W4313349512 cites W4221139382 @default.
- W4313349512 cites W4225959162 @default.
- W4313349512 cites W4286582832 @default.
- W4313349512 doi "https://doi.org/10.1007/978-3-031-20500-2_26" @default.
- W4313349512 hasPublicationYear "2022" @default.
- W4313349512 type Work @default.
- W4313349512 citedByCount "0" @default.
- W4313349512 crossrefType "book-chapter" @default.
- W4313349512 hasAuthorship W4313349512A5042348823 @default.
- W4313349512 hasAuthorship W4313349512A5057287134 @default.
- W4313349512 hasAuthorship W4313349512A5071411382 @default.
- W4313349512 hasAuthorship W4313349512A5085600464 @default.
- W4313349512 hasConcept C103038307 @default.
- W4313349512 hasConcept C138885662 @default.
- W4313349512 hasConcept C13895895 @default.
- W4313349512 hasConcept C153180895 @default.
- W4313349512 hasConcept C154945302 @default.
- W4313349512 hasConcept C158525013 @default.
- W4313349512 hasConcept C173414695 @default.
- W4313349512 hasConcept C2776401178 @default.
- W4313349512 hasConcept C28490314 @default.
- W4313349512 hasConcept C3017588708 @default.
- W4313349512 hasConcept C41008148 @default.
- W4313349512 hasConcept C41895202 @default.
- W4313349512 hasConcept C49774154 @default.
- W4313349512 hasConcept C52622490 @default.
- W4313349512 hasConcept C64922751 @default.
- W4313349512 hasConcept C81363708 @default.
- W4313349512 hasConceptScore W4313349512C103038307 @default.
- W4313349512 hasConceptScore W4313349512C138885662 @default.
- W4313349512 hasConceptScore W4313349512C13895895 @default.
- W4313349512 hasConceptScore W4313349512C153180895 @default.
- W4313349512 hasConceptScore W4313349512C154945302 @default.
- W4313349512 hasConceptScore W4313349512C158525013 @default.
- W4313349512 hasConceptScore W4313349512C173414695 @default.
- W4313349512 hasConceptScore W4313349512C2776401178 @default.
- W4313349512 hasConceptScore W4313349512C28490314 @default.
- W4313349512 hasConceptScore W4313349512C3017588708 @default.
- W4313349512 hasConceptScore W4313349512C41008148 @default.
- W4313349512 hasConceptScore W4313349512C41895202 @default.
- W4313349512 hasConceptScore W4313349512C49774154 @default.
- W4313349512 hasConceptScore W4313349512C52622490 @default.
- W4313349512 hasConceptScore W4313349512C64922751 @default.
- W4313349512 hasConceptScore W4313349512C81363708 @default.
- W4313349512 hasLocation W43133495121 @default.
- W4313349512 hasOpenAccess W4313349512 @default.
- W4313349512 hasPrimaryLocation W43133495121 @default.
- W4313349512 hasRelatedWork W2016461833 @default.
- W4313349512 hasRelatedWork W2059299633 @default.
- W4313349512 hasRelatedWork W2406522397 @default.
- W4313349512 hasRelatedWork W2518599539 @default.
- W4313349512 hasRelatedWork W2732542196 @default.
- W4313349512 hasRelatedWork W2760085659 @default.
- W4313349512 hasRelatedWork W2969680539 @default.
- W4313349512 hasRelatedWork W2977314777 @default.
- W4313349512 hasRelatedWork W2995914718 @default.
- W4313349512 hasRelatedWork W4307883119 @default.
- W4313349512 isParatext "false" @default.
- W4313349512 isRetracted "false" @default.
- W4313349512 workType "book-chapter" @default.