Matches in SemOpenAlex for { <https://semopenalex.org/work/W4361224703> ?p ?o ?g. }
- W4361224703 abstract "Automatic emotion recognition (AER) systems are burgeoning and systems based on either audio, video, text, or physiological signals have emerged. Multimodal systems, in turn, have shown to improve overall AER accuracy and to also provide some robustness against artifacts and missing data. Collecting multiple signal modalities, however, can be very intrusive, time consuming, and expensive. Recent advances in deep learning based speech-to-text and natural language processing systems, however, have enabled the development of reliable multimodal systems based on speech and text while only requiring the collection of audio data. Audio data, however, is extremely sensitive to environmental disturbances, such as additive noise, thus faces some challenges when deployed “in the wild.” To overcome this issue, speech enhancement algorithms have been deployed at the input signal level to improve testing accuracy in noisy conditions. Speech enhancement algorithms can come in different flavors and can be optimized for different tasks (e.g., for human perception vs. machine performance). Data augmentation, in turn, has also been deployed at the model level during training time to improve accuracy in noisy testing conditions. In this paper, we explore the combination of task-specific speech enhancement and data augmentation as a strategy to improve overall multimodal emotion recognition in noisy conditions. We show that AER accuracy under noisy conditions can be improved to levels close to those seen in clean conditions. When compared against a system without speech enhancement or data augmentation, an increase in AER accuracy of 40% was seen in a cross-corpus test, thus showing promising results for “in the wild” AER." @default.
- W4361224703 created "2023-03-31" @default.
- W4361224703 creator A5065418889 @default.
- W4361224703 creator A5068051036 @default.
- W4361224703 creator A5070243596 @default.
- W4361224703 date "2023-03-28" @default.
- W4361224703 modified "2023-10-16" @default.
- W4361224703 title "Task-specific speech enhancement and data augmentation for improved multimodal emotion recognition under noisy conditions" @default.
- W4361224703 cites W1534131679 @default.
- W4361224703 cites W1552007786 @default.
- W4361224703 cites W1603978816 @default.
- W4361224703 cites W1972921752 @default.
- W4361224703 cites W2055332436 @default.
- W4361224703 cites W2064641533 @default.
- W4361224703 cites W2090777335 @default.
- W4361224703 cites W2118479009 @default.
- W4361224703 cites W2125114513 @default.
- W4361224703 cites W2127707623 @default.
- W4361224703 cites W2133340843 @default.
- W4361224703 cites W2141998673 @default.
- W4361224703 cites W2154024118 @default.
- W4361224703 cites W2164471543 @default.
- W4361224703 cites W2168053878 @default.
- W4361224703 cites W2239141610 @default.
- W4361224703 cites W2250539671 @default.
- W4361224703 cites W2346454595 @default.
- W4361224703 cites W2394573230 @default.
- W4361224703 cites W2408520939 @default.
- W4361224703 cites W2534419903 @default.
- W4361224703 cites W2602034649 @default.
- W4361224703 cites W2610961739 @default.
- W4361224703 cites W2740550900 @default.
- W4361224703 cites W2767348466 @default.
- W4361224703 cites W2883496341 @default.
- W4361224703 cites W2889717020 @default.
- W4361224703 cites W2891138528 @default.
- W4361224703 cites W2918378401 @default.
- W4361224703 cites W2951442257 @default.
- W4361224703 cites W2955740278 @default.
- W4361224703 cites W2962843322 @default.
- W4361224703 cites W2962998773 @default.
- W4361224703 cites W2963341071 @default.
- W4361224703 cites W2963686995 @default.
- W4361224703 cites W2963873807 @default.
- W4361224703 cites W2973237404 @default.
- W4361224703 cites W2981677410 @default.
- W4361224703 cites W2991147398 @default.
- W4361224703 cites W2991489655 @default.
- W4361224703 cites W2994254596 @default.
- W4361224703 cites W2995515875 @default.
- W4361224703 cites W2997026866 @default.
- W4361224703 cites W3014475539 @default.
- W4361224703 cites W3047898105 @default.
- W4361224703 cites W3088631780 @default.
- W4361224703 cites W3151504414 @default.
- W4361224703 cites W3196937856 @default.
- W4361224703 cites W3197912330 @default.
- W4361224703 cites W3206776536 @default.
- W4361224703 cites W3212554192 @default.
- W4361224703 cites W3215560629 @default.
- W4361224703 cites W4205730971 @default.
- W4361224703 cites W4225325562 @default.
- W4361224703 cites W4232282348 @default.
- W4361224703 cites W4245744384 @default.
- W4361224703 cites W4285235436 @default.
- W4361224703 cites W4287643567 @default.
- W4361224703 cites W4288351520 @default.
- W4361224703 cites W4293661603 @default.
- W4361224703 cites W4299280181 @default.
- W4361224703 doi "https://doi.org/10.3389/fcomp.2023.1039261" @default.
- W4361224703 hasPublicationYear "2023" @default.
- W4361224703 type Work @default.
- W4361224703 citedByCount "0" @default.
- W4361224703 crossrefType "journal-article" @default.
- W4361224703 hasAuthorship W4361224703A5065418889 @default.
- W4361224703 hasAuthorship W4361224703A5068051036 @default.
- W4361224703 hasAuthorship W4361224703A5070243596 @default.
- W4361224703 hasBestOaLocation W43612247031 @default.
- W4361224703 hasConcept C104317684 @default.
- W4361224703 hasConcept C115961682 @default.
- W4361224703 hasConcept C144024400 @default.
- W4361224703 hasConcept C154945302 @default.
- W4361224703 hasConcept C162324750 @default.
- W4361224703 hasConcept C163294075 @default.
- W4361224703 hasConcept C185592680 @default.
- W4361224703 hasConcept C187736073 @default.
- W4361224703 hasConcept C204201278 @default.
- W4361224703 hasConcept C2776182073 @default.
- W4361224703 hasConcept C2779903281 @default.
- W4361224703 hasConcept C2780451532 @default.
- W4361224703 hasConcept C28490314 @default.
- W4361224703 hasConcept C36289849 @default.
- W4361224703 hasConcept C41008148 @default.
- W4361224703 hasConcept C55493867 @default.
- W4361224703 hasConcept C61328038 @default.
- W4361224703 hasConcept C63479239 @default.
- W4361224703 hasConcept C99498987 @default.
- W4361224703 hasConceptScore W4361224703C104317684 @default.
- W4361224703 hasConceptScore W4361224703C115961682 @default.
- W4361224703 hasConceptScore W4361224703C144024400 @default.