Item Infomation
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Nguyen, Dung | - |
dc.contributor.author | Nguyen, Duc Thanh | - |
dc.contributor.author | Sridha, Sridharan | - |
dc.date.accessioned | 2023-04-25T06:47:31Z | - |
dc.date.available | 2023-04-25T06:47:31Z | - |
dc.date.issued | 2023 | - |
dc.identifier.uri | https://link.springer.com/article/10.1007/s00521-023-08248-y | - |
dc.identifier.uri | https://dlib.phenikaa-uni.edu.vn/handle/PNK/8280 | - |
dc.description | CC BY | vi |
dc.description.abstract | Deep learning has been widely adopted in automatic emotion recognition and has lead to significant progress in the field. However, due to insufficient training data, pre-trained models are limited in their generalisation ability, leading to poor performance on novel test sets. To mitigate this challenge, transfer learning performed by fine-tuning pr-etrained models on novel domains has been applied. However, the fine-tuned knowledge may overwrite and/or discard important knowledge learnt in pre-trained models. In this paper, we address this issue by proposing a PathNet-based meta-transfer learning method that is able to (i) transfer emotional knowledge learnt from one visual/audio emotion domain to another domain and (ii) transfer emotional knowledge learnt from multiple audio emotion domains to one another to improve overall emotion recognition accuracy. | vi |
dc.language.iso | en | vi |
dc.publisher | Springer | vi |
dc.subject | Meta-transfer learning | vi |
dc.title | Meta-transfer learning for emotion recognition | vi |
dc.type | Book | vi |
Appears in Collections | ||
OER - Công nghệ thông tin |
Files in This Item: