Emotions recognition system for acoustic music data based on human perception features

Tatiana Endrjukaite, Yasushi Kiyoki

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Music plays an important role in the human's life. It is not only a set of sounds-music evokes emotions subjectively perceived by listeners. The growing amount of audio data wakes up a need for content-based searching. Traditionally, tunes information has been retrieved based on a reference information, for example, the title of a tune, the name of an artist, the genre and so on. When users would like to try to find music pieces in a specific mood such standard reference information of the tunes is not sufficiently effective. We need new methods and approaches to realize emotion-based search and tune content analysis. This paper proposes a new music-tune analysis approach to realize automatic emotion recognition by means of essential musical features. The innovativeness of this research is that it uses new musical features for tune's analysis, which are based on human's perception of the music. Most important distinction of the proposed approach is that it includes broader range of tunes genres, which is very significant for music emotion recognition system. Emotion description on continuous plane instead of categories results in more supported adjectives for emotion description which is also a great advantage.

Original languageEnglish
Title of host publicationInformation Modelling and Knowledge Bases XXVIII
PublisherIOS Press
Pages283-302
Number of pages20
Volume292
ISBN (Electronic)9781614997191
DOIs
Publication statusPublished - 2017

Publication series

NameFrontiers in Artificial Intelligence and Applications
Volume292
ISSN (Print)09226389

Fingerprint

Acoustics
Acoustic waves

Keywords

  • emotion recognition
  • instantaneous frequency spectrum
  • music analysis
  • music emotions
  • music repetitions
  • music similarity
  • tune internal homogeneity

ASJC Scopus subject areas

  • Artificial Intelligence

Cite this

Endrjukaite, T., & Kiyoki, Y. (2017). Emotions recognition system for acoustic music data based on human perception features. In Information Modelling and Knowledge Bases XXVIII (Vol. 292, pp. 283-302). (Frontiers in Artificial Intelligence and Applications; Vol. 292). IOS Press. https://doi.org/10.3233/978-1-61499-720-7-283

Emotions recognition system for acoustic music data based on human perception features. / Endrjukaite, Tatiana; Kiyoki, Yasushi.

Information Modelling and Knowledge Bases XXVIII. Vol. 292 IOS Press, 2017. p. 283-302 (Frontiers in Artificial Intelligence and Applications; Vol. 292).

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Endrjukaite, T & Kiyoki, Y 2017, Emotions recognition system for acoustic music data based on human perception features. in Information Modelling and Knowledge Bases XXVIII. vol. 292, Frontiers in Artificial Intelligence and Applications, vol. 292, IOS Press, pp. 283-302. https://doi.org/10.3233/978-1-61499-720-7-283
Endrjukaite T, Kiyoki Y. Emotions recognition system for acoustic music data based on human perception features. In Information Modelling and Knowledge Bases XXVIII. Vol. 292. IOS Press. 2017. p. 283-302. (Frontiers in Artificial Intelligence and Applications). https://doi.org/10.3233/978-1-61499-720-7-283
Endrjukaite, Tatiana ; Kiyoki, Yasushi. / Emotions recognition system for acoustic music data based on human perception features. Information Modelling and Knowledge Bases XXVIII. Vol. 292 IOS Press, 2017. pp. 283-302 (Frontiers in Artificial Intelligence and Applications).
@inproceedings{001ad6b3218a41e8806e84d7ceafd229,
title = "Emotions recognition system for acoustic music data based on human perception features",
abstract = "Music plays an important role in the human's life. It is not only a set of sounds-music evokes emotions subjectively perceived by listeners. The growing amount of audio data wakes up a need for content-based searching. Traditionally, tunes information has been retrieved based on a reference information, for example, the title of a tune, the name of an artist, the genre and so on. When users would like to try to find music pieces in a specific mood such standard reference information of the tunes is not sufficiently effective. We need new methods and approaches to realize emotion-based search and tune content analysis. This paper proposes a new music-tune analysis approach to realize automatic emotion recognition by means of essential musical features. The innovativeness of this research is that it uses new musical features for tune's analysis, which are based on human's perception of the music. Most important distinction of the proposed approach is that it includes broader range of tunes genres, which is very significant for music emotion recognition system. Emotion description on continuous plane instead of categories results in more supported adjectives for emotion description which is also a great advantage.",
keywords = "emotion recognition, instantaneous frequency spectrum, music analysis, music emotions, music repetitions, music similarity, tune internal homogeneity",
author = "Tatiana Endrjukaite and Yasushi Kiyoki",
year = "2017",
doi = "10.3233/978-1-61499-720-7-283",
language = "English",
volume = "292",
series = "Frontiers in Artificial Intelligence and Applications",
publisher = "IOS Press",
pages = "283--302",
booktitle = "Information Modelling and Knowledge Bases XXVIII",

}

TY - GEN

T1 - Emotions recognition system for acoustic music data based on human perception features

AU - Endrjukaite, Tatiana

AU - Kiyoki, Yasushi

PY - 2017

Y1 - 2017

N2 - Music plays an important role in the human's life. It is not only a set of sounds-music evokes emotions subjectively perceived by listeners. The growing amount of audio data wakes up a need for content-based searching. Traditionally, tunes information has been retrieved based on a reference information, for example, the title of a tune, the name of an artist, the genre and so on. When users would like to try to find music pieces in a specific mood such standard reference information of the tunes is not sufficiently effective. We need new methods and approaches to realize emotion-based search and tune content analysis. This paper proposes a new music-tune analysis approach to realize automatic emotion recognition by means of essential musical features. The innovativeness of this research is that it uses new musical features for tune's analysis, which are based on human's perception of the music. Most important distinction of the proposed approach is that it includes broader range of tunes genres, which is very significant for music emotion recognition system. Emotion description on continuous plane instead of categories results in more supported adjectives for emotion description which is also a great advantage.

AB - Music plays an important role in the human's life. It is not only a set of sounds-music evokes emotions subjectively perceived by listeners. The growing amount of audio data wakes up a need for content-based searching. Traditionally, tunes information has been retrieved based on a reference information, for example, the title of a tune, the name of an artist, the genre and so on. When users would like to try to find music pieces in a specific mood such standard reference information of the tunes is not sufficiently effective. We need new methods and approaches to realize emotion-based search and tune content analysis. This paper proposes a new music-tune analysis approach to realize automatic emotion recognition by means of essential musical features. The innovativeness of this research is that it uses new musical features for tune's analysis, which are based on human's perception of the music. Most important distinction of the proposed approach is that it includes broader range of tunes genres, which is very significant for music emotion recognition system. Emotion description on continuous plane instead of categories results in more supported adjectives for emotion description which is also a great advantage.

KW - emotion recognition

KW - instantaneous frequency spectrum

KW - music analysis

KW - music emotions

KW - music repetitions

KW - music similarity

KW - tune internal homogeneity

UR - http://www.scopus.com/inward/record.url?scp=85002637590&partnerID=8YFLogxK

UR - http://www.scopus.com/inward/citedby.url?scp=85002637590&partnerID=8YFLogxK

U2 - 10.3233/978-1-61499-720-7-283

DO - 10.3233/978-1-61499-720-7-283

M3 - Conference contribution

VL - 292

T3 - Frontiers in Artificial Intelligence and Applications

SP - 283

EP - 302

BT - Information Modelling and Knowledge Bases XXVIII

PB - IOS Press

ER -