Multi-task Sparse Learning with Beta Process Prior for Action Recognition

Chunfeng Yuan, Weiming Hu, Guodong Tian, Shuang Yang, Haoran Wang; The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2013, pp. 423-429

Abstract


In this paper, we formulate human action recognition as a novel Multi-Task Sparse Learning(MTSL) framework which aims to construct a test sample with multiple features from as few bases as possible. Learning the sparse representation under each feature modality is considered as a single task in MTSL. Since the tasks are generated from multiple features associated with the same visual input, they are not independent but inter-related. We introduce a Beta process(BP) prior to the hierarchical MTSL model, which efficiently learns a compact dictionary and infers the sparse structure shared across all the tasks. The MTSL model enforces the robustness in coefficient estimation compared with performing each task independently. Besides, the sparseness is achieved via the Beta process formulation rather than the computationally expensive l 1 norm penalty. In terms of non-informative gamma hyper-priors, the sparsity level is totally decided by the data. Finally, the learning problem is solved by Gibbs sampling inference which estimates the full posterior on the model parameters. Experimental results on the KTH and UCF sports datasets demonstrate the effectiveness of the proposed MTSL approach for action recognition.

Related Material


[pdf]
[bibtex]
@InProceedings{Yuan_2013_CVPR,
author = {Yuan, Chunfeng and Hu, Weiming and Tian, Guodong and Yang, Shuang and Wang, Haoran},
title = {Multi-task Sparse Learning with Beta Process Prior for Action Recognition},
booktitle = {The IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
month = {June},
year = {2013}
}