Human-object maps for daily activity recognition

Haruya Ishikawa, Yuchi Ishikawa, Shuichi Akizuki, Yoshimitsu Aoki

研究成果: Conference contribution

抄録

In the field of action recognition, when and where an interaction between a human and an object happens has the potential to be valid information in enhancing action recognition accuracy. Especially, in daily life where each activities are performed in longer time frame, conventional short term action recognition may fail to generalize do to the variety of shorter actions that could take place during the activity. In this paper, we propose a novel representation of human object interaction called Human-Object Maps (HOMs) for recognition of long term daily activities. HOMs are 2D probability maps that represents spatio-temporal information of human object interaction in a given scene. We analyzed the effectiveness of HOMs as well as features relating to the time of the day in daily activity recognition. Since there are no publicly available daily activity dataset that depicts daily routines needed for our task, we have created a new dataset that contains long term activities. Using this dataset, we confirm that our method enhances the prediction accuracy of the conventional 3D ResNeXt action recognition method from 86.31% to 97.89%.

元の言語English
ホスト出版物のタイトルProceedings of the 16th International Conference on Machine Vision Applications, MVA 2019
出版者Institute of Electrical and Electronics Engineers Inc.
ISBN(電子版)9784901122184
DOI
出版物ステータスPublished - 2019 5 1
イベント16th International Conference on Machine Vision Applications, MVA 2019 - Tokyo, Japan
継続期間: 2019 5 272019 5 31

出版物シリーズ

名前Proceedings of the 16th International Conference on Machine Vision Applications, MVA 2019

Conference

Conference16th International Conference on Machine Vision Applications, MVA 2019
Japan
Tokyo
期間19/5/2719/5/31

ASJC Scopus subject areas

  • Computer Science Applications
  • Signal Processing
  • Computer Vision and Pattern Recognition

これを引用

Ishikawa, H., Ishikawa, Y., Akizuki, S., & Aoki, Y. (2019). Human-object maps for daily activity recognition. : Proceedings of the 16th International Conference on Machine Vision Applications, MVA 2019 [8757896] (Proceedings of the 16th International Conference on Machine Vision Applications, MVA 2019). Institute of Electrical and Electronics Engineers Inc.. https://doi.org/10.23919/MVA.2019.8757896

Human-object maps for daily activity recognition. / Ishikawa, Haruya; Ishikawa, Yuchi; Akizuki, Shuichi; Aoki, Yoshimitsu.

Proceedings of the 16th International Conference on Machine Vision Applications, MVA 2019. Institute of Electrical and Electronics Engineers Inc., 2019. 8757896 (Proceedings of the 16th International Conference on Machine Vision Applications, MVA 2019).

研究成果: Conference contribution

Ishikawa, H, Ishikawa, Y, Akizuki, S & Aoki, Y 2019, Human-object maps for daily activity recognition. : Proceedings of the 16th International Conference on Machine Vision Applications, MVA 2019., 8757896, Proceedings of the 16th International Conference on Machine Vision Applications, MVA 2019, Institute of Electrical and Electronics Engineers Inc., 16th International Conference on Machine Vision Applications, MVA 2019, Tokyo, Japan, 19/5/27. https://doi.org/10.23919/MVA.2019.8757896
Ishikawa H, Ishikawa Y, Akizuki S, Aoki Y. Human-object maps for daily activity recognition. : Proceedings of the 16th International Conference on Machine Vision Applications, MVA 2019. Institute of Electrical and Electronics Engineers Inc. 2019. 8757896. (Proceedings of the 16th International Conference on Machine Vision Applications, MVA 2019). https://doi.org/10.23919/MVA.2019.8757896
Ishikawa, Haruya ; Ishikawa, Yuchi ; Akizuki, Shuichi ; Aoki, Yoshimitsu. / Human-object maps for daily activity recognition. Proceedings of the 16th International Conference on Machine Vision Applications, MVA 2019. Institute of Electrical and Electronics Engineers Inc., 2019. (Proceedings of the 16th International Conference on Machine Vision Applications, MVA 2019).
@inproceedings{2b521e25dba84f9987b984d822cf10aa,
title = "Human-object maps for daily activity recognition",
abstract = "In the field of action recognition, when and where an interaction between a human and an object happens has the potential to be valid information in enhancing action recognition accuracy. Especially, in daily life where each activities are performed in longer time frame, conventional short term action recognition may fail to generalize do to the variety of shorter actions that could take place during the activity. In this paper, we propose a novel representation of human object interaction called Human-Object Maps (HOMs) for recognition of long term daily activities. HOMs are 2D probability maps that represents spatio-temporal information of human object interaction in a given scene. We analyzed the effectiveness of HOMs as well as features relating to the time of the day in daily activity recognition. Since there are no publicly available daily activity dataset that depicts daily routines needed for our task, we have created a new dataset that contains long term activities. Using this dataset, we confirm that our method enhances the prediction accuracy of the conventional 3D ResNeXt action recognition method from 86.31{\%} to 97.89{\%}.",
author = "Haruya Ishikawa and Yuchi Ishikawa and Shuichi Akizuki and Yoshimitsu Aoki",
year = "2019",
month = "5",
day = "1",
doi = "10.23919/MVA.2019.8757896",
language = "English",
series = "Proceedings of the 16th International Conference on Machine Vision Applications, MVA 2019",
publisher = "Institute of Electrical and Electronics Engineers Inc.",
booktitle = "Proceedings of the 16th International Conference on Machine Vision Applications, MVA 2019",

}

TY - GEN

T1 - Human-object maps for daily activity recognition

AU - Ishikawa, Haruya

AU - Ishikawa, Yuchi

AU - Akizuki, Shuichi

AU - Aoki, Yoshimitsu

PY - 2019/5/1

Y1 - 2019/5/1

N2 - In the field of action recognition, when and where an interaction between a human and an object happens has the potential to be valid information in enhancing action recognition accuracy. Especially, in daily life where each activities are performed in longer time frame, conventional short term action recognition may fail to generalize do to the variety of shorter actions that could take place during the activity. In this paper, we propose a novel representation of human object interaction called Human-Object Maps (HOMs) for recognition of long term daily activities. HOMs are 2D probability maps that represents spatio-temporal information of human object interaction in a given scene. We analyzed the effectiveness of HOMs as well as features relating to the time of the day in daily activity recognition. Since there are no publicly available daily activity dataset that depicts daily routines needed for our task, we have created a new dataset that contains long term activities. Using this dataset, we confirm that our method enhances the prediction accuracy of the conventional 3D ResNeXt action recognition method from 86.31% to 97.89%.

AB - In the field of action recognition, when and where an interaction between a human and an object happens has the potential to be valid information in enhancing action recognition accuracy. Especially, in daily life where each activities are performed in longer time frame, conventional short term action recognition may fail to generalize do to the variety of shorter actions that could take place during the activity. In this paper, we propose a novel representation of human object interaction called Human-Object Maps (HOMs) for recognition of long term daily activities. HOMs are 2D probability maps that represents spatio-temporal information of human object interaction in a given scene. We analyzed the effectiveness of HOMs as well as features relating to the time of the day in daily activity recognition. Since there are no publicly available daily activity dataset that depicts daily routines needed for our task, we have created a new dataset that contains long term activities. Using this dataset, we confirm that our method enhances the prediction accuracy of the conventional 3D ResNeXt action recognition method from 86.31% to 97.89%.

UR - http://www.scopus.com/inward/record.url?scp=85070482779&partnerID=8YFLogxK

UR - http://www.scopus.com/inward/citedby.url?scp=85070482779&partnerID=8YFLogxK

U2 - 10.23919/MVA.2019.8757896

DO - 10.23919/MVA.2019.8757896

M3 - Conference contribution

AN - SCOPUS:85070482779

T3 - Proceedings of the 16th International Conference on Machine Vision Applications, MVA 2019

BT - Proceedings of the 16th International Conference on Machine Vision Applications, MVA 2019

PB - Institute of Electrical and Electronics Engineers Inc.

ER -