Steffen Herff1, Garett Johnson2, Andrew Milne1, Christian Herff3, Jinsoo Kim4, Jerry Shih5, Dean Krusienski6
14:20 - 14:35 | Wed 12 Jul | Roentgen Hall | WeBT1.1
As brain-computer interface technology continues to advance toward practical applications, it can be extended to decoding and synthesizing music from mental imagery. Such a brain-actuated music synthesizer is envisioned to serve as a music composition and performance tool for musicians and non-musicians, as well as a potential communication and rehabilitation tool for the disabled. In order to decode brain activity during music listening, let alone generate new compositions directly from brain activity, a much more detailed knowledge of neural activations and pathways must be developed. To accomplish this, brain recordings with sufficient spatio-temporal resolution must be examined. This work presents preliminary results of signal characterization of electrocorticographic (ECoG) activity during perception of simple musical rhythms.