完整後設資料紀錄
DC 欄位 | 值 | 語言 |
---|---|---|
dc.contributor.author | Chou, Kuang-Pen | en_US |
dc.contributor.author | Prasad, Mukesh | en_US |
dc.contributor.author | Wu, Di | en_US |
dc.contributor.author | Sharma, Nabin | en_US |
dc.contributor.author | Li, Dong-Lin | en_US |
dc.contributor.author | Line, Yu-Feng | en_US |
dc.contributor.author | Blumenstein, Michael | en_US |
dc.contributor.author | Line, Wen-Chieh | en_US |
dc.contributor.author | Lin, Chin-Teng | en_US |
dc.date.accessioned | 2018-08-21T05:53:31Z | - |
dc.date.available | 2018-08-21T05:53:31Z | - |
dc.date.issued | 2018-01-01 | en_US |
dc.identifier.issn | 2169-3536 | en_US |
dc.identifier.uri | http://dx.doi.org/10.1109/ACCESS.2018.2809552 | en_US |
dc.identifier.uri | http://hdl.handle.net/11536/144800 | - |
dc.description.abstract | Automated human action recognition has the potential to play an important role in public security, for example, in relation to the multiview surveillance videos taken in public places, such as train stations or airports. This paper compares three practical, reliable, and generic systems for multiview video-based human action recognition, namely, the nearest neighbor classifier, Gaussian mixture model classifier, and the nearest mean classifier. To describe the different actions performed in different views, view-invariant features are proposed to address multiview action recognition. These features are obtained by extracting the holistic features from different temporal scales which are modeled as points of interest which represent the global spatial-temporal distribution. Experiments and cross-data testing are conducted on the KTH, WEIZMANN, and MuHAVi datasets. The system does not need to be retrained when scenarios are changed which means the trained database can be applied in a wide variety of environments, such as view angle or background changes. The experiment results show that the proposed approach outperforms the existing methods on the KTH and WEIZMANN datasets. | en_US |
dc.language.iso | en_US | en_US |
dc.subject | Multi-view video | en_US |
dc.subject | action recognition | en_US |
dc.subject | feature extraction | en_US |
dc.subject | background subtraction | en_US |
dc.subject | classification | en_US |
dc.subject | machine learning | en_US |
dc.title | Robust Feature-Based Automated Multi-View Human Action Recognition System | en_US |
dc.type | Article | en_US |
dc.identifier.doi | 10.1109/ACCESS.2018.2809552 | en_US |
dc.identifier.journal | IEEE ACCESS | en_US |
dc.citation.volume | 6 | en_US |
dc.citation.spage | 15283 | en_US |
dc.citation.epage | 15296 | en_US |
dc.contributor.department | 資訊工程學系 | zh_TW |
dc.contributor.department | 電機工程學系 | zh_TW |
dc.contributor.department | Department of Computer Science | en_US |
dc.contributor.department | Department of Electrical and Computer Engineering | en_US |
dc.identifier.wosnumber | WOS:000429258300001 | en_US |
顯示於類別: | 期刊論文 |