| Interacting With Digital 
        Music  Mark Sandler Queen Mary University of London 
         
 Date: Friday, April 21, 
        2006Place: Humanities and Social Sciences, 1173
 Time: 2:00 pm — 3:00 pm (Reception to 
        follow)
 
 Abstract:
 Although in a strict sense music went digital over two decades ago with the introduction of the CD, the term Digital Music has come to imply something more than a digital representation, often the inclusion of Internet technologies with music, and always the involvement of computers. By combining the latest Signal Processing techniques with Machine Learning and semantic processing, the latest and future generation Digital Music applications offer some really exciting ways to listen to, search for and interact with music.
 
 The talk will begin by looking at some widely available technologies and services for interacting with digital music, and then move on to a view of current research in the area with a look at some recent results from the Centre for Digital Music. Finally, the talk will take a view of the future, and get an idea how today's research informs future generation consumer products, with particular attention to the role of the Semantic Web.
 
    MARK SANDLER is Director 
        of the Centre for Digital Music and Professor of Signal Processing at 
        Queen Mary, University of London, where he moved in 2001 after 19 years 
        at King's College, also in the University of London. Mark received the 
        BSc and PhD degrees from University of Essex, UK, in 1978 and 1984, respectively. 
        Mark has published nearly 300 papers in journals and conferences. He is 
        a Senior Member of IEEE, a Fellow of IEE and a Fellow of the Audio Engineering 
        Society. He is two-times recipient of the IEE A.H.Reeves Premium Prize. 
        In 2003 he was General Chair of DAFx (6th International Workshop on Digital 
        Audio Effects) held at Queen Mary, and in 2005 was General Co-Chair of 
        ISMIR (6th International Conference on Music Information Retrieval) also 
        held at Queen Mary. He is founding Chair of the Audio Engineering Society's 
        Technical Committee on Semantic Audio Analysis, and is on the IEEE Technical 
        Committee of Audio and Electroacoustics. He was founding editor of the 
        EURASIP Journal of Applied Signal Processing, and is consulting editor 
        to Elsevier for books on Audio and Music Signal Processing. 
 He has worked 
        in Digital Signal Processing for Audio and Music for nearly 30 years on 
        a wide variety of topics including: Digital Power amplification; Drum 
        Synthesis; Chaos and Fractals for Analysis and Synthesis; non-linear dynamics; 
        Sigma-Delta Modulation & Direct Stream Digital technologies; Digital EQ; 
        Wavelet Audio compression; high quality audio compression; compression 
        domain processing; Internet Audio Streaming and Scalable coding; Automatic 
        Music Transcription and Musical Feature Extraction; Music Semantics and 
        Knowledge Representation; 3D sound reproduction; time stretching and audio 
        effects.
 
     Host: Dr. Xavier Amatriain, 
        Research Director of CREATE
 
 |