Title:
The sound within: Learning audio features from electroencephalogram recordings of music listening

Thumbnail Image
Author(s)
Vinay, Ashvala
Authors
Advisor(s)
Leslie, Grace
Advisor(s)
Editor(s)
Associated Organization(s)
Organizational Unit
Organizational Unit
Supplementary to
Abstract
We look at the intersection of music, machine Learning and neuroscience. Specifically, we are interested in understanding how we can predict audio onset events by using the electroencephalogram response of subjects listening to the same music segment. We present models and approaches to this problem using approaches derived by deep learning. We worked with a highly imbalanced dataset and present methods to solve it - tolerance windows and aggregations. Our presented methods are a feed-forward network, a convolutional neural network (CNN), a recurrent neural network (RNN) and a RNN with a custom unrolling method. Our results find that at a tolerance window of 40 ms, a feed-forward network performed well. We also found that an aggregation of 200 ms suggested promising results, with aggregations being a simple way to reduce model complexity.
Sponsor
Date Issued
2020-04-28
Extent
Resource Type
Text
Resource Subtype
Thesis
Rights Statement
Rights URI