Heart.FM, selected for ERC proof-of-concept funding, is an app-creation initiative to deliver tailored music therapy with physiological feedback in cardiovascular disease.
Daniel Bedoya, PhD candidate, designs citizen science experiments to help understand musical structures created in performance, and analyzes the perception of musical structures in performed music and physiological responses to these performed structures. He has an undergraduate degree in Sound Engineering (UDLA Quito-Ecuador) and a Master’s degree in Computer Science, Acoustics and Signal Processing Applied to Music (ATIAM – IRCAM-Sorbonne Université). Previously, he was a research assistant with Jean-Julien Aucouturier in the Perception and Sound Design (PDS) Team on the relationship of music and emotions in the ERC project CREAM and explored the influence of smiled speech in dyadic interactions in the REFLETS project.
Elaine Chew (PhD & SM, ORC, MIT; BAS, Math-CompSci & Music Perf, Stanford; FTCL & LTCL, Piano Perf, Trinity Coll Lon) is Principal Investigator of the ERC ADG project COSMOS. Her research centers on the mathematical and computational modeling of musical structures, with present focus on structures as they are communicated in performance and in ECG traces of cardiac arrhythmias. As a pianist, she has collaborated with composers to create and première new works, and she frequently designs and performs in concerts that present visualizations and compositions created in her research team. She is past recipient of PECASE/CAREER awards and Fellowships at the Radcliffe Institute for Advanced Studies|Harvard in the US. Her research has been supported by the ERC, EPSRC, AHRC, and NSF, and featured on BBC World Service/Radio 3, Smithsonian Magazine, Philadelphia Inquirer, Wired Blog, MIT Technology Review, and Los Angeles Philharmonic’s Inside the Music. She has also recorded on Albany and Neuma Records.
Lawrence Fyfe, Research Engineer, is creating web-based visualisation software and database infrastructure to harness volunteer thinking in the project’s citizen science modules. Lawrence received his PhD in Computational Media Design from the University of Calgary and a Master’s degree in Music, Science and Technology from the Centre for Computer Research in Music and Acoustics (CCRMA) at Stanford University. Before joining the COSMOS project, he worked on a binaural telepresence system for the Digiscope project at INRIA. The Digiscope project connected various visualisation labs around Paris via telepresence (audio and video conferencing) to facilitate collaboration. Before that, Lawrence developed a web site for listening to sonified EEG data, which was used to facilitate the diagnosis of epileptic seizures.
Corentin Guichaoua, Postdoctoral Researcher, is researching mathematical and computational techniques for automatic extraction of musical structure in performed music and cardiac signals. Previously, he was a postdoc with Moreno Andreatta at the University of Strasbourg in the SMIR project, where he implemented algebraic and topologic methods for systematic analysis and comparison of pieces of music. He holds PhD and Master’s degrees in Computer Science from the University of Rennes 1, and a concurrent Masters of Science and Engineering (Diplôme d’Ingénieur) from INSA (Institut national des sciences appliquées). His doctoral thesis, supervised by Frédéric Bimbot, focused on compressed descriptions of chord sequences from pieces of music using formal models, in order to extract information on their structure.
Gonzalo Romero, ATIAM (IRCAM’s Masters degree in Acoustics, Signal Professing, and Computer Science Applied to Music) intern, is developing scalable algorithms for the automatic transcription of rhythmic variations, and applying the computational techniques to create symbolic representations of long arrhythmia ECG sequences for structural analysis. He has received a Masters in Fundamental Mathematics at the Sorbonne University, and a Bachelor’s degree in Mathematics from Complutense University in Madrid, and a Bachelor’s degree in Composition from the Madrid Royal Conservatory (Real Conservatorio Superior de Música de Madrid). Gonzalo hails from a musical family and plays the violin and piano.
The Translational Electrophysiology Team led by Prof. Pier Lambiase—Professor of Cardiology at UCL’s Institute for Cardiovascular Science and Barts Heart Centre (BHC), Co-Director of Cardiovascular Research at Barts NHS Trust, and BHRS Committee Research Lead; member of the European Society of Cardiology and International Heart Rhythm Society clinical guideline committees—and BHC Electrophysiology Researchers, including Prof. Peter Taggart, Drs Ross Hunter (AF Lead) & Michele Orini (Engineer).
COSMOS: Computational Shaping and Modeling of Musical Structures (Principal Investigator: Elaine Chew) is a European Research Council Advanced Grant (AdG) project supported by the European Union’s Horizon 2020 research and innovation program under grant agreement No. 788960. COSMOS aims to use data science, optimization / data analytics, and citizen science to study musical structures as they are created in music performances and in unusual sources such as cardiac arrhythmias.
The project is hosted by the Centre National de la Recherche Scientifique (CNRS) at the Sciences et Technologies de la Musique et du Son (STMS) Laboratory, a joint research unit (UMR9912) of the CNRS, the Institut de Recherche et Coordination Acoustique/Musique (IRCAM), Sorbonne University, and the French Ministry of Culture. STMS is located at IRCAM, in the heart of Paris.
The project summary is given below and on the European Commission’s CORDIS website.
|Objective: Music performance is considered by many to be one of the most breath taking feats of human intelligence. That music performance is a creative act is no longer a disputed fact, but the very nature of this creative work remains illusive. Taking the view that the creative work of performance is the making and shaping of music structures, and that this creative thinking is a form of problem solving, COSMOS proposes an integrated programme of research to transform our understanding of the human experience of performed music, which is almost all music that we hear, and of the creativity of music performance, which addresses how music is made. The research themes are as follows: i) to find new ways to represent, explore, and talk about performance; ii) to harness volunteer thinking (citizen science) for music performance research by focussing on structures experienced and problem solving; iii) to create sandbox environments to experiment with making performed structures; iv) to create theoretical frameworks to discover the reasoning behind the structures perceived and made; and, v) to foster community engagement by training experts to provide feedback on structure solutions so as to increase public understanding of the creative work in music performance. Analysis of the perceived and designed structures will be based on a novel duality paradigm that turns conventional computational music structure analysis on its head to reverse engineer why a perceiver or a performer chooses a particular structure. Embedded in the approach is the use of computational thinking to optimise representations and theories to ensure accuracy, robustness, efficiency, and scalability. The PI is an established performer and a leading authority in music representation, music information research, and music perception and cognition. The project will have far reaching impact, reconfiguring expert and public views of music performance and time-varying music-like sequences such as cardiac arrhythmia.|