Probing neural mechanisms of music perception, cognition, and performance using multivariate decoding
Psychomusicology: Music, Mind and Brain, 22(2):168–174, 2012
Recent neuroscience research has shown increasing use of multivariate decoding methods and machine learning. These methods, by uncovering the source and nature of informative variance in large data sets, invert the classical direction of inference that attempts to explain brain activity from mental state variables or stimulus features. However, these techniques are not yet commonly used among music researchers. In this position article, we introduce some key features of machine learning methods and review their use in the field of cognitive and behavioral neuroscience of music. We argue for the great potential of these methods in decoding multiple data types, specifically audio waveforms, electroen- cephalography, functional MRI, and motion capture data. By finding the most informative aspects of stimulus and performance data, hypotheses can be generated pertaining to how the brain processes incoming musical information and generates behavioral output, respectively. Importantly, these methods are also applicable to different neural and physiological data types such as magnetoencephalography, near-infrared spectroscopy, positron emission tomography, and electromyography.
Proceedings of the 12th International Conference on Music Perception and Cognition, page 943, Thessaloniki, Greece, July 2012. ICMPC/ESCOM. (abstract).
A software system, MediaMined, is described for the efficient analysis and classification of auditory signals. This system has been applied to the tasks of musical instrument identification, classifying musical genre, distinguishing between music and speech, and detection of the gender of human speakers. For each of these tasks, the same algorithm is applied, consisting of low-level signal analysis, statistical processing and perceptual modeling for feature extraction, and then supervised learning of sound classes. Given a ground truth dataset of audio examples, textual descriptive classification labels are then produced. Such labels are suitable for use in automating content interpretation (auditioning) and content retrieval, mixing and signal processing. A multidimensional feature vector is calculated from statistical and perceptual processing of low level signal analysis in the spectral and temporal domains. Machine learning techniques such as support vector machines are applied to produce classification labels given a selected taxonomy. The system is evaluated on large annotated ground truth datasets (n > 30000) and demonstrates success rates (F-measures) greater than 70% correct retrieval, depending on the task. Issues arising from labeling and balancing training sets are discussed. The performance of classification of audio using machine learning methods demonstrates the relative contribution of bottom-up signal derived features and data oriented classification processes to human cognition. Such demonstrations then sharpen the question as to the contribution of top-down, expectation based processes in human auditory cognition.
I'm a big fan of urban biking, having had the privilege of living in Amsterdam and Paris which are both very bike friendly cities for daily commuting. Returning back to NYC, I got a cast off commuter bike from a neighbour, that after considerable work, I got in working order. Given drivers in NYC are nowhere near as aware of cyclists as European drivers, lights are essential. I fitted the crappy old "Fairway Flyer" with a Dutch magneto (aka "dynamo" although that term is strictly incorrect) generator and 6VAC old school chrome light. In practice, the drag on wheel isn't enough to notice given the entire bike is hardly built for speed.
If you are a budding music or audio engineering undergraduate student, iZotope is hiring paid interns. The work mostly consists of auditioning our systems.
(Folks That Live On The Hill, by Jerome Kern)
The company I have been working with for about 18 months now, Imagine Research has now been acquired by iZotope Inc. based in Boston. This is a great opportunity for us to continue to develop our MediaMined technology, to broaden it's reach and incorporate it into iZotope's future products. For my local friends, I'll continue to be based in NYC.
Jill produced a beautiful interview that we made with my Dad, Peter, first shot after his diagnosis, and Jill rapidly edited it ready for playing at his wake. It was a deeply moving way for Dad to be there for all of the people that loved him.
This is the tribute I presented with my sister Gaye at my father, Peter John Smith's funeral:
I'd like to talk about Dad's life through some of our personal experiences and recollections of him.