TY - GEN
T1 - Zero-Shot Action Recognition with Knowledge Enhanced Generative Adversarial Networks
AU - Huang, Kaiqiang
AU - Miralles-Pechuan, Luis
AU - Mckeever, Susan
N1 - Publisher Copyright:
Copyright © 2021 by SCITEPRESS – Science and Technology Publications, Lda. All rights reserved.
PY - 2021
Y1 - 2021
N2 - Zero-Shot Action Recognition (ZSAR) aims to recognise action classes in videos that have never been seen during model training. In some approaches, ZSAR has been achieved by generating visual features for unseen classes based on the semantic information of the unseen class labels using generative adversarial networks (GANs). Therefore, the problem is converted to standard supervised learning since the unseen visual features are accessible. This approach alleviates the lack of labelled samples of unseen classes. In addition, objects appearing in the action instances could be used to create enriched semantics of action classes and therefore, increase the accuracy of ZSAR. In this paper, we consider using, in addition to the label, objects related to that action label. For example, the objects 'horse' and 'saddle' are highly related to the action 'Horse Riding' and these objects can bring additional semantic meaning. In this work, we aim to improve the GAN-based framework by incorporating object-based semantic information related to the class label with three approaches: replacing the class labels with objects, appending objects to the class, and averaging objects with the class. Then, we evaluate the performance using a subset of the popular dataset UCF101. Our experimental results demonstrate that our approach is valid since when including appropriate objects into the action classes, the baseline is improved by 4.93%.
AB - Zero-Shot Action Recognition (ZSAR) aims to recognise action classes in videos that have never been seen during model training. In some approaches, ZSAR has been achieved by generating visual features for unseen classes based on the semantic information of the unseen class labels using generative adversarial networks (GANs). Therefore, the problem is converted to standard supervised learning since the unseen visual features are accessible. This approach alleviates the lack of labelled samples of unseen classes. In addition, objects appearing in the action instances could be used to create enriched semantics of action classes and therefore, increase the accuracy of ZSAR. In this paper, we consider using, in addition to the label, objects related to that action label. For example, the objects 'horse' and 'saddle' are highly related to the action 'Horse Riding' and these objects can bring additional semantic meaning. In this work, we aim to improve the GAN-based framework by incorporating object-based semantic information related to the class label with three approaches: replacing the class labels with objects, appending objects to the class, and averaging objects with the class. Then, we evaluate the performance using a subset of the popular dataset UCF101. Our experimental results demonstrate that our approach is valid since when including appropriate objects into the action classes, the baseline is improved by 4.93%.
KW - Generative Adversarial Networks
KW - Human Action Recognition
KW - Zero-Shot Learning
UR - https://www.scopus.com/pages/publications/85146199484
U2 - 10.5220/0010717000003063
DO - 10.5220/0010717000003063
M3 - Conference contribution
AN - SCOPUS:85146199484
T3 - ICETE International Conference on E-Business and Telecommunication Networks (International Joint Conference on Computational Intelligence)
SP - 254
EP - 264
BT - IJCCI 2021 - Proceedings of the 13th International Joint Conference on Computational Intelligence
A2 - Back, Thomas
A2 - Wagner, Christian
A2 - Garibaldi, Jonathan
A2 - Lam, H. K.
A2 - Cottrell, Marie
A2 - Merelo, Juan Julian
A2 - Warwick, Kevin
PB - Science and Technology Publications, Lda
T2 - 13th International Joint Conference on Computational Intelligence, IJCCI 2021
Y2 - 25 October 2021 through 27 October 2021
ER -