Generating an integrated musical expression with a brain–computer interface

  • Takayuki Hamano
  • , Tomasz M. Rutkowski
  • , Hiroko Terasawa
  • , Kazuo Okanoya
  • , Kiyoshi Furukawa

Research output: Contribution to journalConference articlepeer-review

17 Scopus citations

Abstract

Electroencephalography (EEG) has been used to generate music for over 40 years, but the most recent developments in brain–computer interfaces (BCI) allow greater control and more flexible expression to use new musical instruments via EEG. We developed a real-time musical performance system using BCI technology and sonification techniques to generate chords with organically fluctuating timbre. We aimed to emulate the expressivity of traditional acoustic instruments by adding “non-coded” expressions that were not marked in the score. The BCI part of the system classifies patterns during neural activity while a performer imagines a chord. The sonification part of the system captures non-stationary changes in the brain waves and reflects them in the timbre by additive synthesis. In this paper, we discuss the conceptual design, system development, and the performance of this instrument.

Original languageEnglish
Pages (from-to)49-54
Number of pages6
JournalProceedings of the International Conference on New Interfaces for Musical Expression
StatePublished - 2013
Externally publishedYes
Event13th International conference on New Interfaces for Musical Expression, NIME 2013 - Daejeon, Korea, Republic of
Duration: 27 May 201330 May 2013

Keywords

  • Brain–computer interface (BCI)
  • Classification
  • Qualitative and quantitative information
  • Sonification

Fingerprint

Dive into the research topics of 'Generating an integrated musical expression with a brain–computer interface'. Together they form a unique fingerprint.

Cite this