687
Views
0
CrossRef citations to date
0
Altmetric
Research Article

Effective human–object interaction recognition for edge devices in intelligent space

ORCID Icon, ORCID Icon & ORCID Icon
Pages 1-9 | Received 27 Apr 2023, Accepted 04 Dec 2023, Published online: 23 Dec 2023

References

  • Lee J-H, Hashimoto H. Intelligent space–concept and contents. Adv Robot. 2002;16:265–280. doi:10.1163/156855302760121936.
  • Maegawa K, Shiotani T, Iwamoto K, et al. Ubiquitous display 2.0: development of new prototype and software modules for improvement. 10th International Conference on Ubiquitous Robots and Ambient Intelligence; 2013 Oct 30–Nov 2; Jeju, South Korea, p. 102–107. doi:10.1109/URAI.2013.6677483.
  • Park J, Lee J-H. Reconfigurable intelligent space, R+iSpace, and mobile module, MoMo. IEEE/RSJ International Conference on Intelligent Robots and Systems; 2012 Oct 7–12; Vilamoura-Algarve, Portugal, p. 3865–3866. doi:10.1109/IROS.2012.6386282.
  • Itsuka T, Song M, Kawamura A, et al. Development of ROS2 -TMS: new software platform for informationally structured environment. Robomech J. 2022;9:1. doi:10.1186/s40648-021-00216-2
  • Murakami K, Tsuji T, Hasegawa T, et al. Object tracking system by integrating multi-sensored data. IECON 2016 - 42nd Annual Conference of the IEEE Industrial Electronics Society; 2016 Oct 23–26; Florence, Italy, p. 747–754. doi:10.1109/IECON.2016.7793355
  • Pyo Y, Hasegawa T, Tanaka M, et al. Measurement and estimation of indoor human behavior of everyday life based on floor sensing with minimal invasion of privacy. 2013 IEEE International Conference on Robotics and Biomimetics; 2013 Dec 12–14; Shenzhen, China, p. 2170–2176. doi:10.1109/ROBIO.2013.6739791
  • Niitsuma M, Kobayashi H, Shiraishi A. Enhancement of spatial memory for applying to sequential activity. Acta Polytech Hung. 2012;9:121–137.
  • Irie K, Wakamura N, Umeda K. Construction of an intelligent room based on gesture recognition: operation of electric appliances with hand gestures. IEEE/RSJ International Conference on Intelligent Robots and Systems; 2004 Sept 28–Oct 2; Sendai, Japan, 193–198. doi:10.1109/IROS.2004.1389351
  • Ozaki H, Tran DT, Lee J-H. Human state recognition method in intelligent space. 2023 International Symposium on System Integration; 2023 Jan 17–20; Atlanta, GA, USA.
  • Gkioxari G, Girshick R, Dollár P, et al. Detecting and recognizing human–object interactions. 2018 IEEE/CVF Conference on Computer Vision and pattern Recognition; 2018 Jun 18–23; Salt Lake City, UT, USA, p. 8359–8367. doi:10.1109/CVPR.2018.00872
  • Chao Y-W, Liu Y, Liu X, et al. Learning to detect human–object interactions. 2018 IEEE Winter Conference on Applications of Computer Vision; 2018 Mar 12–15; Lake Tahoe, NV, USA, p. 381–389.
  • Wang T, Anwer RM, Khan MH, et al. Deep contextual attention for human–object interaction detection. In: 2019 IEEE/CVF International Conference on Computer Vision, 2019 Oct 27–Nov 2, Seoul, Korea, p. 5693–5701. doi:10.1109/ICCV.2019.00579
  • Li Y-L, Zhou S, Huang X, et al. Transferable interactiveness knowledge for human–object interaction detection. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition; 2019 Jun 15–20; Long Beach, CA, USA, p. 3580–3589. doi:10.1109/CVPR.2019.00370
  • Yan S, Xiong Y, Lin D. Spatial temporal graph convolutional networks for skeleton-based action recognition. The 32nd AAAI Conference on Artificial Intelligence; 2018 Feb 2–7; New Orleans, LA, USA, p. 1–9.
  • Qi S, Wang W, Jia B, et al. Learning human–object interactions by graph parsing neural networks. In: Ferrari V, Hebert M, Sminchisescu C, Weiss Y, editors. Computer vision- ECCV 2018. Lecture notes in computer science, Vol. 11213. Cham: Springer. doi:10.1007/978-3-030-01240-325
  • Duan H, Zhao Y, Chen K, et al. Revisiting skeleton-based action recognition. 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition; 2022 Jun 19–24; New Orleans, LA, USA, p. 2959–2968. doi:10.1109/CVPR52688.2022.00298
  • Kim B, Lee J, Kang J, et al. HOTR: End-to-end human–object interaction detection with transformers. 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition; 2021 Jun 20–25; Nashville, TN, USA, p. 74–83. doi:10.1109/CVPR46437.2021.00014
  • Pfister T, Charles J, Zisserman A. Flowing convnets for human pose estimation in videos. 2015 IEEE International Conference on Computer Vision (ICCV). IEEE; 2015, p. 1913–1921.
  • YOLOv5. [Cited 2023 Apr 10]. Available from: https://github.com/ultralytics/yolov5.
  • MediaPipe. [Cited 2023 Apr 10]. Available from: https://google.github.io/mediapipe/.
  • Lin T-Y, Maire M, Belongie S, et al. Microsoft COCO: common objects in context. In: Fleet D, Pajdla T, Schiele B, Tuytelaars T, editors. Computer vision- ECCV 2014: Lecture notes in computer science, vol 8693. Cham: Springer. doi:10.1007/978-3-319-10602-1
  • Bazarevsky V, Grishchenko I, Raveendran K, et al. BlazePose: On-device real-time body pose tracking. arXiv preprint. 2020. doi:10.48550/arXiv.2006.10204
  • Gupta A, Kembhavi A, Davis LS. Observing human–object interactions: using spatial and functional compatibility for recognition. IEEE Trans Pattern Anal Mach Intell. 2009;31:1775–1789. doi:10.1109/TPAMI.2009.83
  • Fan H, Zhuo T, Yu X, et al. Understanding atomic hand-object interaction with human intention. IEEE Trans Circuits Syst Video Technol. 2022;32:275–285. doi:10.1109/TCSVT.2021.3058688
  • Wang T, Yang T, Danelljan M, et al. Learning human-object interaction detection using interaction points. 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition; 2020 Jun 14–19, Online, p. 4115–4124. doi:10.1109/CVPR42600.2020.00417