用户名: 密码: 验证码:
Discriminative Orderlet Mining for Real-Time Recognition of Human-Object Interaction
详细信息    查看全文
  • 作者:Gang Yu (17)
    Zicheng Liu (18)
    Junsong Yuan (17)

    17. School of Electrical and Electronic Engineering
    ; Nanyang Technological University ; Singapore ; Singapore
    18. Microsoft Research
    ; Redmond ; WA ; USA
  • 刊名:Lecture Notes in Computer Science
  • 出版年:2015
  • 出版时间:2015
  • 年:2015
  • 卷:9007
  • 期:1
  • 页码:50-65
  • 全文大小:733 KB
  • 参考文献:1. Shotton, J., Fitzgibbon, A., Cook, M., Sharp, T., Finocchio, M., Moore, R., Kipman, A., Blake, A.: Real-time human pose recognition in parts from single depth images. In: CVPR (2011)
    2. Chum, O., Philbin, J., Zisserman, A.: Near duplicate image detection: min-Hash and tf-idf weighting. In: BMVC (2008)
    3. Yagnik, J., Strelow, D., Ross, D., Lin, R.S.: The power of comparative reasoning. In: ICCV (2011)
    4. Wang, J., Liu, Z., Wu, Y., Yuan, J.: Mining actionlet ensemble for action recognition with depth cameras. In: CVPR (2012)
    5. Wang, J, Liu, Z, Chorowski, J, Chen, Z, Wu, Y Robust 3d action recognition with random occupancy patterns. In: Fitzgibbon, A, Lazebnik, S, Perona, P, Sato, Y, Schmid, C eds. (2012) Computer Vision 鈥?ECCV 2012. Springer, Heidelberg, pp. 872-885 CrossRef
    6. Schapire, R.: A brief introduction to boosting. In: IJCAI (1999)
    7. Tang, S, Wang, X, Lv, X, Han, TX, Keller, J, He, Z, Skubic, M, Lao, S Histogram of oriented normal vectors for object recognition with a depth sensor. In: Lee, KM, Matsushita, Y, Rehg, JM, Hu, Z eds. (2013) Computer Vision 鈥?ACCV 2012. Springer, Heidelberg, pp. 525-538 CrossRef
    8. Liu, J., Kuipers, B., Savarese, S.: Recognizing human actions by attributes. In: CVPR (2011)
    9. Schuldt, C., Laptev, I., Caputo, B.: Recognizing human actions: a local svm approach. In: ICPR (2004)
    10. Yu, G., Yuan, J., Liu, Z.: Unsupervised random forest indexing for fast action search. In: IEEE Conference on Computer Vision and Pattern Recognition (2011)
    11. Yu, G, Yuan, J, Liu, Z Propagative hough voting for human activity recognition. In: Fitzgibbon, A, Lazebnik, S, Perona, P, Sato, Y, Schmid, C eds. (2012) Computer Vision 鈥?ECCV 2012. Springer, Heidelberg, pp. 693-706 CrossRef
    12. Oreifej, O., Liu, Z.: HON4D: histogram of oriented 4D Normals for activity recognition from depth sequences. In: CVPR (2013)
    13. Laptev, I (2005) On space-time interest points. IJCV 64: pp. 107-123 CrossRef
    14. Dollar, P., Rabaud, V., Cottrell, G., Belongiel, S.: Behavior recognition via sparse spatio-temporal features. In: Visual Surveillance and Performance Evaluation of Tracking and Surveillance (2005)
    15. Wang, H., Klaser, A., Schmid, C., Liu, C.L.: Action recognition by dense trajectories. In: CVPR (2011)
    16. Jiang, Y-G, Dai, Q, Xue, X, Liu, W, Ngo, C-W Trajectory-based modeling of human actions with motion reference points. In: Fitzgibbon, A, Lazebnik, S, Perona, P, Sato, Y, Schmid, C eds. (2012) Computer Vision 鈥?ECCV 2012. Springer, Heidelberg, pp. 425-438 CrossRef
    17. Xia, L., Aggarwal, J.K.: Spatio-temporal depth cuboid similarity feature for activity recognition using depth camera. In: CVPR (2013)
    18. Ryoo, M.S.: Human activity prediction: early recognition of ongoing activities from streaming videos. In: ICCV (2011)
    19. Yang, X., Tian, Y.: EigenJoints-based action recognition using Naive-Bayes-Nearest-Neighbor. In: CVPRW (2012)
    20. Yang, X., Zhang, C., Tian, Y.: Recognizing actions using depth motion maps-based histograms of oriented gradients. In: ACM Multimedia (2012)
    21. Chen, H.S., Chen, H.T., Chen, Y.W., Lee, S.Y.: Human action recognition using star skeleton. In: ACM International Workshop on Video Surveillance and Sensor Networks (2006)
    22. Li, W., Zhang, Z., Liu, Z.: Action recognition based on a bag of 3d points. In: CVPRW (2010)
    23. Bentley, J (1984) Programming pearls: algorithm design techniques. Commun. ACM 27: pp. 865-873 CrossRef
    24. Zhu, Y., Chen, W., Guo, G.D.: Fusing spatiotemporal features and joints for 3D action recognition. In: CVPRW (2013)
    25. Hoai, M., DelaTorre, F.: Max-margin early event detectors. In: CVPR (2012)
    26. Zhou, B., Wang, X., Tang, X.: Understanding collective crowd behaviors: learning a mixture model of dynamic pedestrian-agents. In: CVPR (2012)
    27. Zanfir, M., Leordeanu, M., Sminchisescu, C.: The moving pose: an efficient 3d kinematics descriptor for low-latency action recognition and detection. In: ICCV (2013)
    28. Yu, G, Norberto, A, Yuan, J, Liu, Z (2011) Fast action detection via discriminative random forest voting and top-K subvolume search. IEEE Trans. Multimedia 13: pp. 507-517 CrossRef
    29. Sadanand, S., Corso, J.J.: Action bank: a high-level representation of activity in video. In: CVPR (2012)
    30. Chen, C.Y., Grauman, K.: Efficient activity detection with max-subgraph search. In: CVPR (2012)
    31. Gupta, A., Davis, L.S.: Objects in action: an approach for combining action understanding and object perception. In: CVPR (2007)
    32. Jain, A., Gupta, A., Rodriguez, M., Davis, L.S.: Representing videos using mid-level discriminative patches. In: CVPR (2013)
    33. Parikh, D., Grauman, K.: Relative attributes. In: ICCV (2011)
  • 作者单位:Computer Vision -- ACCV 2014
  • 丛书名:978-3-319-16813-5
  • 刊物类别:Computer Science
  • 刊物主题:Artificial Intelligence and Robotics
    Computer Communication Networks
    Software Engineering
    Data Encryption
    Database Management
    Computation by Abstract Devices
    Algorithm Analysis and Problem Complexity
  • 出版者:Springer Berlin / Heidelberg
  • ISSN:1611-3349
文摘
This paper presents a novel visual representation, called orderlets, for real-time human action recognition with depth sensors. An orderlet is a middle level feature that captures the ordinal pattern among a group of low level features. For skeletons, an orderlet captures specific spatial relationship among a group of joints. For a depth map, an orderlet characterizes a comparative relationship of the shape information among a group of subregions. The orderlet representation has two nice properties. First, it is insensitive to small noise since an orderlet only depends on the comparative relationship among individual features. Second, it is a frame-level representation thus suitable for real-time online action recognition. Experimental results demonstrate its superior performance on online action recognition and cross-environment action recognition.

© 2004-2018 中国地质图书馆版权所有 京ICP备05064691号 京公网安备11010802017129号

地址:北京市海淀区学院路29号 邮编:100083

电话:办公室:(+86 10)66554848;文献借阅、咨询服务、科技查新:66554700