Prof. Simon Dixon
Contact Details
Position: Professor;
Director of the
UKRI Centre for Doctoral Training in Artificial
Intelligence and Music;
Deputy Director of the
Centre for Digital Music
Tel: Internal: [13] 7681
International: +44 20 7882 7681
Email: s.e.dixon@qmul.ac.uk
Office: ENG 406 (access from ENG403 via 4th floor of Graduate Centre)
Office Hours: Tue 3pm during semester.
Please mail me in advance to confirm availability or to request
an appointment at other times.
Research Group:
Centre for Digital Music
Google Scholar:
http://scholar.google.co.uk/citations?user=9GvyqXEAAAAJ
ORCID ID:
0000-0002-6098-481X
News & Events
- 6 Nov 2024: Yixiao Zhang passed his PhD viva on Improving Controllability and Editability for Pretrained Text-to-Music Generation Models
- 12 Jul 2024: the Best Paper Award at the Sound and Music Computing Conference (SMC 2024) was awarded to Alia Morsi (UPF), Huan Zhang (QMUL), Akira Maezawa (Yamaha), Simon Dixon (QMUL) and Xavier Serra (UPF) for our paper: Simulating Piano Performance Mistakes for Music Learning Contexts
- 26 Apr 2024: I will give a keynote presentation at the Northern Network for Empirical Music Research (NEMuR) at the University of Hull
- 7 Mar 2024: Brendan O'Connor passed his PhD viva on the topic of Analysis, Disentanglement, and Conversion of Singing Voice Attributes
- 26 Feb 2024: Applications close for a research assistent to work on the project Music Performance Assessment and Feedback
- 12 Jan 2024: Final workshop of JazzDAP (the Jazz Digital Archives Project, also known as New Directions in Digital Jazz Studies) at the Institute for Jazz Studies, Rutgers University, New Jersey.
- 11 Jan 2024: We (Adriana Cuervo, Pedro Cravinho, Simon Dixon, Haftor Medbøe, Gabriel Solis and Tillman Weyde) presented a panel on Artificial Intelligence and New Directions in Digital Jazz Studies at the Jazz Congress 2024, Lincoln Center, New York (the only time in my life I'll be on the same programme as Dee Dee Bridgewater, Marcus Miller, John Patitucci, etc.!)
- 28 Apr 2023: Carlos Lordelo passed his PhD viva on the topic of Deep Learning Methods for Instrument Separation and Recognition. Carlos was funded by the EU H2020 MIP-Frontiers project, and completed most of his PhD in Sweden with our industry partner DoReMIR Music Research, supervised by Emmanouil Benetos, Sven Ahlbäck from DoReMIR and me. His work is already in use in the ScoreCloud product range from DoReMIR. The examiners found his work on unsupervised domain adaptation particularly exciting, noting its potential beyond the immediate application areas of the thesis.
- 21 Nov 2022: I gave a Distinguished Lecture on Automatic Music Transcription and Music Understanding at McGill University, Centre for Interdisciplinary Research in Music Media and Technology (CIRMMT).
- 28 Apr 2022: Emir Demirel passed his PhD viva on the topic Deep Neural Networks for Automatic Lyrics Transcription, as part of the MIP-Frontiers project, in collaboration with DoReMIR Music Research.
- 12 Apr 2022: The New Directions in Digital Jazz Studies project, also known as JazzDAP (Jazz Digital Archive Project), is seeking a research assistant (postdoctoral or predoctoral) for a short-term (4-month) position analysing jazz recordings for main melody, chords and beats. Applications close on 5 May 2022. See advertisement here.
- 25 Mar 2022: Ruchit Agrawal passed his PhD viva on the topic Towards Context-Aware Neural Performance-Score Synchronisation, as part of the MIP-Frontiers project, in collaboration with Tido Music.
- 1 Dec 2021: At least 12 PhD studentships available for the Sept 2022 intake of the UKRI Centre for Doctoral Training in AI and Music. Application deadline: 30 Jan 2022.
- 1 Oct 2021: I was awarded Turing Fellow status by the Alan Turing Institute.
- 1 Apr 2021: The Journal of the American Musicological Society published a review of the Dig that Lick project by Chris Stover.
- 16 Dec 2020: Funding announced for a new project in collaboration with the University of Illinois, City University of London, Birmingham City University, the Institute of Jazz Studies at Rutgers University, and the Scottish Jazz Archive: New Directions in Digital Jazz Studies: Music Information Retrieval and AI Support for Jazz Scholarship in Digital Archives.
- 30 Nov 2020: At least 12 PhD studentships available for the Sept 2021 intake of the UKRI Centre for Doctoral Training in AI and Music. Application deadline: 27 Jan 2021.
- 14 Sep 2020: 15 students enrol as the second cohort of the UKRI Centre for Doctoral Training in AI and Music.
- 13 July 2020: Daniel Stoller successfully defends his PhD thesis Deep Learning for Music Information Retrieval in Limited Data Scenarios
- 20 Apr 2020: Daniel Stoller's paper Seq-U-Net: A One-Dimensional Causal U-Net for Efficient Sequence Modelling is accepted for presentation in the main track at IJCAI-PRICAI 2020.
- 20 Mar 2020: Francisco Rodríguez-Algarra successfully defends his PhD thesis A Critical Look at the Music Classification Experiment Pipeline: Using Interventions to Detect and Account for Confounding Effects
- 11 Feb 2020: Saumitra Mishra successfully defends his PhD thesis Interpretable Machine Learning for Machine Listening
- Jan 2020: funding announced for the Innovate UK Smart Grant (Project 43427) Hookslam: Making Hits with AI, 2020-2021
- Dec 2019: 12 PhD studentships available for the Sept 2020 intake of the UKRI Centre for Doctoral Training in AI and Music. Deadline: 31 Jan 2020.
- Sept 2019: First 12 students enrol in the UKRI CDT in AI and Music
- July 2019: Funding announced for the Innovate UK Smart Grant (Project 30863) Development of Next-Generation Music Recognition Algorithm for Content Monitoring, in collaboration with Breathe Music Ltd, Nov 2019 - May 2021
- April 2019: Dig That Lick project mentioned on Sandy Brown Jazz (scroll down / search for Weimar)
- 18 Mar 2019: 12 fully-funded PhD positions in Artificial Intelligence and Music starting in Sept 2019. Deadline for applications is 15 April. More details here
- 21 Feb 2019: Funding announced for the UKRI Centre for Doctoral Training in Artificial Intelligence and Music at Queen Mary University of London (2019-2027). This centre will fund over 70 PhD students spread over 5 cohorts starting each September from 2019. The AIM CDT involves over 20 industry partners and is open to expressions of interest from further prospective partners. PhD studentships for September 2019 start will be advertised in March.
- 15 Feb 2019: Jiajie Dai successfully defends her PhD thesis Modelling Intonation and Interaction in Vocal Ensembles
- 16 Jan 2019: Simon Dixon receives the Peter Claricoats Award for service as Director of Graduate Studies and for sustained high-quality research
- 4 Sep 2018: Launch of a new journal! The Transactions of the International Society for Music Information Retrieval (TISMIR) launched its first issue today. TISMIR is an open-access journal established to complement the widely cited ISMIR conference proceedings and provide a vehicle for the dissemination of the highest quality and most substantial scientific research in music information retrieval.
- 1 Apr 2018: EU H2020 MSCA ITN project New Frontiers in Music Information Processing (MIP-Frontiers) begins
- 26 Mar 2018: Adib Mehrabi successfully defends his PhD thesis Vocal Imitation for Query by Vocalisation
- 1 Mar 2018: Maria Panteli successfully defends her PhD thesis Computational Analysis of World Music Corpora
- 20 Dec 2017: Zulfadhli Mohamad successfully defends his PhD thesis Estimating Performance Parameters from Electric Guitar Recordings
- 30 Oct 2017: British Library Labs Research Award recognising "outstanding and innovative work that has been carried out using the British Library's digital collections and data" awarded to Maria Panteli, Emmanouil Benetos and Simon Dixon for the project A large-scale comparison of world music corpora with computational tools.
- 15 Sep 2017: Siying Wang successfully defends her thesis Computational Methods for the Alignment and Score-Informed Transcription of Piano Music, co-supervised by Sebastian Ewert
- 9 Aug 2017: Postdoctoral research position advertised here for the project Dig that Lick: Analysing large-scale data for melodic patterns in jazz performances . Deadline 15 September 2017.
- Jul 2017: Announcement of my promotion to Professor with effect from 1 October.
- 2 May 2017: Funding announced for New Frontiers in Music Information Processing, a Marie Sklodowska-Curie Innovative Training Network which will fund 15 PhD studentships between four universities (Queen Mary, Telecom ParisTech, Universitat Pompeu Fabra, Johannes Kepler University of Linz), three industry beneficiaries (Sony CSL, Roli Ltd, DoReMIR Music Research), and nine partner organisations (Native Instruments, Technicolor, Audionamix, Deezer, Karajan Institute, Wiener Staatsoper, Tido Music, BMAT Licensing, Jamendo).
- 29 March 2017: Funding announced by the Trans-Atlantic Program Digging into Data Challenge for the project: Dig that Lick: Analysing large-scale data for melodic patterns in jazz performances , a collaboration between Queen Mary University of London, UK (lead partner); Columbia University, USA; National Center for Scientific Research (CNRS), France; University of Music Franz Liszt, Weimar, Germany; University of Illinois Champaign Urbana, USA; and City University of London, UK. Job ads here.
- 11 November 2016: Siddharth Sigtia successfully defends his PhD thesis: Neural Networks for Analysing Music and Environmental Audio
- 28 October 2016: Tian Cheng successfully defends her PhD thesis: Exploiting Piano Acoustics in Automatic Transcription
- 10 August 2016: Maria Panteli wins Best Student Paper Award at the 17th International Society for Music Information Retrieval Conference, for Learning a Feature Space for World Music Style Similarity, by M. Panteli, E. Benetos and S. Dixon (2016).
- 27 July 2016: Magdalena Chudy successfully defends her PhD thesis: Discriminating music performers by timbre: On the relation between instrumental gesture, tone quality and perception in classical cello performance
- 15 Mar 2016: Robert Tubb successfully defends his PhD thesis: Creativity, Exploration and Control in Musical Parameter Spaces
- 14 Jan 2016: Shengchen Li successfully defends his PhD thesis: Expressive Timing Analysis in Classical Piano Performance by Mathematical Model Selection
- 5 Oct 2015: Yading Song successfully defends his PhD thesis: The Role of Emotion and Context in Musical Preference
- Jan 2015: Best Student Paper Award for the Audio and Acoustic Signal Processing track of the IEEE International Conference on Acoustics Speech and Signal Processing, 2015, awarded to Siying Wang for the paper: Compensating for Asynchronies between Musical Voices in Score-Performance Alignment, by S. Wang, S. Ewert and S. Dixon, to appear in ICASSP 2015
- 12 Jan 2015: Oxford University Press announces the most read article in the journal Early Music for 2014: The Temperament Police by Dan Tidhar, Simon Dixon, Emmanouil Benetos, and Tillman Weyde
- Jan 2015: Simon Dixon takes on role of Deputy Director of C4DM
- 8 Jan 2015: Peter Foster successfully defends his PhD thesis: Information-Theoretic Measures of Predictability for Music Content Analysis
- 4 Apr 2014: Amélie Anglade successfully defends her PhD thesis (no corrections required): Logic-based Modelling of Musical Harmony for Automatic Characterisation and Classification
- 24 Jan 2014: Holger Kirchhoff successfully defends his PhD thesis: A User-Assisted Approach to Multiple Instrument Music Transcription
- 27 Nov 2013: AHRC grant Digital Music Lab - Analysing Big Music Data awarded (with City University and the British Library)
- 15-18 Oct 2013: The MusicWest prize for the best poster paper at the 10th International Symposium on Computer Music Multidisciplinary Research (CMMR) was awarded to Holger Kirchhoff, Simon Dixon and Anssi Klapuri for their paper entitled Multiple Instrument Tracking based on Recontruction Error, Pitch Continuity and Instrument Activity
- 30 September 2013: Semantic Media mini-project, Computational Analysis of the Live Music Archive (CALMA), awarded funding (with Oxford University, Manchester University and the Internet Archive)
- 30 September 2013: Semantic Media mini-project, Metadata Used in Semantic Indexes and Charts (MUSIC), awarded funding (with Northampton University and Academic Press)
- 18 March 2013: Semantic Media mini-project, Large-Scale capture of Producer-Defined Musical Semantics, awarded funding (with Birmingham City University)
- 22 Feb 2013: Lesley Mearns successfully defends her PhD thesis: The Computational Analysis of Harmony in Western Art Music
- 22 Oct 2012: Emmanouil Benetos successfully defends his PhD thesis: Automatic Transcription of Polyphonic Music Exploiting Temporal Evolution
- 8 Oct 2012: Tutorial on Reusable software and reproducibility in music informatics research to be presented by Chris Cannam, Simon Dixon and Mark Plumbley at the 13th International Society for Music Information Retrieval Conference
- 19-22 June 2012: 9th International Symposium on Computer Music Modelling and Retrieval (CMMR): Music and Emotions was hosted at Queen Mary University of London, with over 150 attendees
- 17 May 2012: JISC awards funding for the project: Sound Data Management Training (SoDaMaT)
- 17 May 2012: JISC awards funding for the project: A shared open vocabulary for audio research and retrieval (SOVARR)
- 16 Jan 2012: Robert Macrae successfully defends his PhD thesis: Linking Music Metadata
- 27 Oct 2011: Simon Dixon elected as President Elect of the International Society for Music Information Retrieval
- 25 Oct 2011: Transcription Prelude by Emmanouil Benetos and Simon Dixon performed by Lika Pailodze at the ISMIR 2011 Concert, Miami, Florida
- Oct 2011: EU FP7 Project MIReS: Roadmap for Music Information ReSearch starts. Our partners for this 18 month project are: UPF (Barcelona), BMAT (Barcelona), OFAI (Vienna), IRCAM (Paris), INESC (Porto), Stromatalite (London)
- 14 Sep 2011: C4DM celebrates its 10-year anniversary, including a concert featuring the Non Hidden Markov Models and the Sparse Beats
- Sept 2011: Openings for a short-term research assistant and software developer (Oct 2011 - Mar 2012) for the JISC-funded project Sustainable Management of Digital Music Research Data. Deadline 19 Sept. Details: Research Assistant (Ad, Job Profile); Software Developer (Ad, Job Profile); Project Summary
- Sept 2011: JISC project Sustainable Management of Digital Music Research Data awarded
- Feb 2011: Opening for short-term research assistants (until July) to work on the Linked Music Metadata (LinkedBrainz) project. The positions have now been filled.
- 23-28 Jan 2011: Dagstuhl Seminar on Multimodal Music Processing, organised by Meinard Müller, Masataka Goto and Simon Dixon
- December 2010: Special issue of the Journal of New Music Research on Computational Musicology and the OMRAS-2 Project
- August 2010: Audio Chord Detection algorithm by Mauch and Dixon rated 1st in MIREX evaluation
- June 2010: JISC grant Linked Music Metadata awarded
- April 2010: EPSRC grant Musicology for the Masses awarded
- 4 March 2010: Matthias Mauch successfully defends his PhD thesis: Automatic Chord Transcription from Audio using Computational Models of Musical Context
- October 2009: Audio Chord Detection algorithm by Mauch and Dixon rated 1st in the MIREX evaluation
- October 2009: Music Structure Segmentation algorithm by Mauch, Noland and Dixon rated 1st in the MIREX evaluation
Press
Articles about some of the research projects I have worked on appeared in:- Scientific American, 18 March 2019
- MyScience.org, 31 March 2017
- New York Times, NYT Arts and Leisure, 5 June 2005
- Die Presse, 1 June 2004 (in German)
- FORMAT Science, 19 March 2004 (in German)
- Wired Magazine, Sept 2001, pp 100-112
- Der Spiegel, Nr.23/3.6.02, pp174-176 (in German)
Research
My research interests are in the fields of music informatics (music information retrieval), artificial intelligence, computer music, music signal processing, digital audio and music cognition. In recent years I have mainly worked on high-level music signal analysis, computational modelling of musical knowledge, and the study of musical performance, rhythm, harmony and intonation. For example, I have worked on note and chord transcription, beat tracking, multimodal music alignment, and singing intonation, using signal processing approaches, probabilistic models, and deep learning.
Projects
- Principal Investigator on UKRI Centre for Doctoral Training in Artificial Intelligence and Music, (EP/S022694/1), £6.24M, 2019-2027
- Principal Investigator on industry project Modelling Jazz Piano, sponsored by Yamaha Corporation, 2023-2025
- Principal Investigator on industry project Expressive Performance Rendering for Music Generation Systems, sponsored by DAACI Ltd, 2022-2026
- Co-Investigator on industry-funded project Deep Learning Technologies For Multi-instrument Automatic Music Transcription with Huawei Technologies, 2022-2023
- Co-Investigator on NEH/AHRC New Directions for Digital Scholarship in Cultural Institutions project AH/V009699/1, New Directions in Digital Jazz Studies: Music Information Retrieval and AI Support for Jazz Scholarship in Digital Archives (JazzDAP), £199685 total (£50k to QMUL), 2021-2023
- Co-Investigator on Innovate UK Smart Grant (Project 43427), Hookslam: Making Hits with AI, £250k total (£75k to QMUL), 2020-2021
- Principal Investigator on industry project Industry-Scale Machine Listening for Music and Audio Data, sponsored by Spotify Ltd, 2020-2024
- QMUL Principal Investigator on Innovate UK Smart Grant (Project 105516): Development of Next-Generation Music Recognition Algorithm for Content Monitoring, £350k total (£148k to QMUL), 2019-2021
- Principal Investigator on EU H2020 Innovative Training Network (Grant 765068) New Frontiers in Music Information Processing (MIP-Frontiers), €3937k total (~€1M to QMUL), 2018-2022
- Principal Investigator on Trans-Atlantic Programme Digging into Data Challenge project Dig that Lick: Analysing large-scale data for melodic patterns in jazz performances (ES/R004005/1), £646k total (£106219 to QMUL), 2017-2019
- Principal Investigator on industry-sponsored project Automatic Detection of Structural Boundaries in Musical Audio, with Rotor Videos, £13646, 2017
- Principal Investigator (for QMUL) on EU H2020 project Polifoni-X-Score, with DoReMIR Music Research AB, £77445, 2015-2016 (QMUL as 3rd party)
- Principal Investigator on industry-sponsored project Audio Alignment Algorithms, with Tido Music GmbH, £34474, 2014-2016; £21600, 2014-2015; and £7542, 2015
- Co-Investigator (2014) and Principal Investigator (2015) on AHRC project Digital Music Lab - Analysing Big Music Data (AH/L01016X/1), £564689 (£122940 to QMUL), 2014-2015
- Principal Investigator (2015) on Innovate UK Project Audio Data Exploration, £96804, 2014-15
- Principal Investigator (2015) on Technology Strategy Board (Innovate UK) Project Advanced Smart Microphone, £36000, 2014-15
- Principal Investigator on Semantic Media Mini-Project MUSIC, £12442 (£5940 to QMUL), 2014
- Principal Investigator on Semantic Media Mini-Project: CALMA, £23001 (£18820 to QMUL), 2013-2014
- Principal investigator on JISC project Sound Data Management Training (SoDaMaT), £57k, June 2012 - Feb 2013
- Principal investigator on JISC project A shared open vocabulary for audio research and retrieval (SOVARR), £69k, October 2012 - May 2013
- Principal investigator on JISC project Sustainable Management of Digital Music Research Data, £89k, Oct 2011 - May 2012
- Principal investigator (for QMUL) on EU FP7 project MIReS: Roadmap for Music Information ReSearch, €81k (to QMUL), Oct 2011 - Mar 2013
- Principal investigator on RCUK Digital Economy project EP/I001832/1 Musicology for the Masses, £250k, Sep 2010 - Feb 2012
- Principal investigator on JISC-funded Linked Music Metadata project, £95k, Aug 2010 - July 2011
- Co-investigator on EPSRC project EP/K009559/1 Platform Grant: Digital Music, £1161k, 2013-2018
- Co-investigator on TSB project 12033-76187, Making Musical Mood Metadata - M4, £145k, 2011-2013
- Co-investigator for EPSRC project EP/H043101/1 Sustainable Software for Digital Music and Audio Research, £947k, 2010-2014
- Co-investigator on WWTF-funded project AudioMiner: Mathematical Signal Analysis and Modeling for Manipulation of Sound Objects, €484k, 2010-2012
- Co-investigator on EPSRC-funded OMRAS-2 Project EP/E017614/1 Online Music Recognition And Searching: A Distributed Research Environment for Music Informatics and Computational Musicology, 2007-2010
- Co-investigator on Vienna Science and Technology Fund (WWTF) project Interfaces to Music, 2004-2007
- Researcher on EU 6th FP STREP Project Semantic Interaction with Music Audio Contents (SIMAC), 2004-2006
- Researcher on Gerhard Widmer's START Research Prize project Computer-Based Music Research: Artificial Intelligence Models of Musical Expression, 1999-2005
- Researcher on EU 5th FP RTN Project: Music Orchestration Systems in Algorithmic Research and Technology (MOSART), 2000-2003
Publications
This material is presented to ensure timely dissemination of scholarly and technical work. Copyright and all rights therein are retained by authors or by other copyright holders. All persons copying this information are expected to adhere to the terms and constraints invoked by each author's copyright. In most cases, these works may not be reposted without the explicit permission of the copyright holder.
Link to my page on Google Scholar.
Mary Pilataki, Matthias Mauch and Simon Dixon (2024),
Pitch-Aware Generative Pretraining Improves Multi-pitch
Estimation with Scarce Data.
ACM Multimedia Asia, to appear.
Yin-Jyun Luo, Kin Wai Cheuk, Woosung Choi, Wei-Hsiang Liao, Keisuke Toyama, Toshimitsu Uesaka, Koichi Saito, Chieh-Hsin Lai, Yuhta Takida, Simon Dixon, Yuki Mitsufuji (2024),
Disentangling Multi-instrument Music Audio for Source-Level Pitch and Timbre Manipulation.
Audio Imagination: NeurIPS 2024 Workshop, to appear.
Xavier Riley, Zixun Guo, Andrew Edwards and Simon Dixon (2024),
GAPS: A Large and Diverse Classical Guitar Dataset and Benchmark Transcription
Model.
International Society for Music Information Retrieval Conference, to appear.
Huan Zhang, Jinhua Liang and Simon Dixon (2024),
From Audio Encoders to Piano Judges: Benchmarking Performance Understanding for Solo Piano.
International Society for Music Information Retrieval Conference, to appear.
Andrew Edwards, Xavier Riley, Pedro Sarmento and Simon Dixon (2024),
MIDI-to-Tab: Guitar Tablature Inference via Masked Language Modeling.
International Society for Music Information Retrieval Conference, to appear.
S. Chang, E. Benetos, H. Kirchhoff and S. Dixon (2024),
YourMT3+: Multi-instrument Music Transcription with Enhanced Transformer Architectures and Cross-Dataset Stem Augmentation.
IEEE International Workshop on Machine Learning for Signal Processing (MLSP), to
appear.
arXiv preprint 2407.04822.
Huan Zhang, Shreyan Chowdhury, Carlos Eduardo Cancino-Chacón,
Jinhua Liang, Simon Dixon and Gerhard Widmer (2024),
DExter: Learning and Controlling Performance Expression with Diffusion Models.
Applied Sciences, 14, 6543.
DOI.
Yixiao Zhang, Yukara Ikemiya, Gus Xia, Naoki Murata,
Marco A. Martínez-Ramírez, Wei-Hsiang Liao, Yuki Mitsufuji, Simon Dixon (2024),
MusicMagus: Zero-Shot Text-to-Music Editing via Diffusion Models.
International Joint Conference on Artificial Intelligence, to appear.
Xavier Riley and Simon Dixon (2024),
Reconstructing the Charlie Parker Omnibook using an audio-to-score automatic transcription pipeline.
Sound and Music Computing Conference.
Alia Morsi, Huan Zhang, Akira Maezawa, Simon Dixon, and Xavier Serra (2024),
Simulating Piano Performance Mistakes for Music Learning.
Sound and Music Computing Conference (Best Paper Award).
Drew Edwards, Simon Dixon, Emmanouil Benetos, Akira Maezawa and Yuta Kusaka
(2024),
A Data-Driven Analysis of Robust Automatic Piano Transcription.
IEEE Signal Processing Letters, 31, pp 681-685.
DOI.
Müller, M., Dixon, S., Volk, A., Sturm, B. L. T., Rao, P., and Gotham, M. (2024),
Introducing the TISMIR Education Track: What, Why, How?.
Transactions of the International Society for Music Information Retrieval, 7(1), 85–98.
DOI.
Lucas Henry, Klaus Frieler, Gabriel Solis, Martin Pfleiderer, Simon Dixon, Frank Höger, Tillman Weyde and Hélène Camille Crayencour (2024),
Dig That Lick: Exploring Patterns in Jazz with Computational Methods.
Jazzforschung / Jazz Research, 50/51, to appear.
Xavier Riley, Drew Edwards and Simon Dixon (2024),
High Resolution Guitar Transcription via Domain Adaptation.
IEEE International Conference on Acoustics, Speech and Signal Processing
(ICASSP), to appear.
Yin-Jyun Luo and Simon Dixon (2024),
Posterior Variance-Parameterised Gaussian Dropout: Improving Disentangled Sequential Autoencoders for Zero-Shot Voice Conversion.
IEEE International Conference on Acoustics, Speech and Signal Processing
(ICASSP), to appear.
Yin-Jyun Luo, Sebastian Ewert and Simon Dixon (2024),
Unsupervised Pitch-Timbre Disentanglement of Musical Instruments Using a Jacobian Disentangled Sequential Autoencoder.
IEEE International Conference on Acoustics, Speech and Signal Processing
(ICASSP), to appear.
Drew Edwards, Simon Dixon and Emmanouil Benetos (2023),
PiJAMA: Piano Jazz with Automatic MIDI Annotations.
Transactions of the International Society for Music Information Retrieval, 6(1), pp. 89–102.
DOI.
Carey Bunks, Tillman Weyde, Simon Dixon and Bruno Di Giorgi (2023),
Modeling Harmonic Similarity for Jazz Using Co-occurrence Vectors and the Membrane Area.
International Society for Music Information Retrieval Conference, pp 757-764.
Xavier Riley and Simon Dixon (2023),
FiloBass: A Dataset and Corpus Based Study of Jazz Basslines.
International Society for Music Information Retrieval Conference, pp 500-507.
Huan Zhang, Emmanouil Karystinaios, Simon Dixon, Gerhard Widmer and Carlos
Eduardo Cancino-Chacón (2023),
Symbolic Music Representations for Classification Tasks: A Systematic Evaluation
.
International Society for Music Information Retrieval Conference, pp 848-858.
Ivan Shanin and Simon Dixon (2023),
Annotating Jazz Recordings using Lead Sheet Alignment with Deep Chroma Features
.
IEEE Workshop on Applications of Signal Processing to Audio and Acoustics
(WASPAA), DOI: 10.1109/WASPAA58266.2023.10248107.
Brendan O'Connor and Simon Dixon (2023),
A Comparative Analysis of Latent Regressor Losses for Singing Voice
Conversion.
Sound and Music Computing Conference, pp 289-295.
Xavier Riley and Simon Dixon (2023),
CREPE Notes: A New Method for Segmenting Pitch Contours into Discrete
Notes.
Sound and Music Computing Conference, pp 1-5.
Huan Zhang and Simon Dixon,
Disentangling the Horowitz factor: Learning content and style from expressive
piano performance.
IEEE International Conference on Acoustics, Speech and Signal Processing
(ICASSP), DOI: 10.1109/ICASSP49357.2023.10095009.
Xavier Riley, Drew Edwards and Simon Dixon (2023),
Automatic Guitar Transcription with a Composable Audio-to-MIDI-to-Tablature Architecture.
Digital Music Research Network One-Day Workshop, p 11.
Huan Zhang,
Jingjing Tang,
Syed Rafee,
Simon Dixon,
George Fazekas and
Geraint Wiggins (2022),
ATEPP: A Dataset of Automatically Transcribed Expressive Piano Performance.
International Society for Music Information Retrieval Conference, pp 446-453.
Yixiao Zhang, Junyan Jiang, Gus Xia and Simon Dixon (2022),
Interpreting Song Lyrics with an Audio-Informed Pre-trained Language Model.
International Society for Music Information Retrieval Conference, pp 19-26.
Yin-Jyun Luo, Sebastian Ewert and Simon Dixon (2022),
Towards Robust Unsupervised Disentanglement of Sequential Data — A Case Study Using Music Audio.
31st International Joint Conference on
Artificial Intelligence and the 25th European Conference on Artificial
Intelligence (IJCAI-ECAI 2022), pp 3299-3305.
Polina Proutskova,
Daniel Wolff,
György Fazekas,
Klaus Frieler,
Frank Höger,
Olga Velichkina,
Gabriel Solis,
Tillman Weyde,
Martin Pfleiderer,
Hélène Camille Crayencour,
Geoffroy Peeters,
and Simon Dixon (2022),
The Jazz Ontology:
A semantic model and large-scale RDF repositories for jazz.
Journal of Web Semantics, 74, 100735.
DOI.
Ruchit Agrawal, Daniel Wolff and Simon Dixon (2022),
A Convolutional-Attentional Neural Framework for Structure-Aware Performance-Score Synchronization.
IEEE Signal Processing Letters, 29, pp 344-348.
DOI.
C. Lordelo, E. Benetos, S. Dixon, S. Ahlbäck and P. Ohlsson (2021),
Adversarial Unsupervised Domain Adaptation for Harmonic-Percussive Source Separation.
IEEE Signal Processing Letters, 28, pp 81-85.
DOI.
Emir Demirel, Sven Ahlbäck and Simon Dixon (2021),
Computational Pronunciation Analysis in Sung Utterances.
29th European Signal Processing Conference (EUSIPCO), pp 186-190.
Emir Demirel, Sven Ahlbäck and Simon Dixon (2021),
MSTRE-NET: Multistreaming Acoustic Modeling for Automatic Lyrics Transcription.
22nd International Society for Music Information Retrieval Conference (ISMIR),
pp 151-158.
D. Foster and S. Dixon (2021),
Filosax: A Dataset of Annotated Jazz Saxophone Recordings.
22nd International Society for Music Information Retrieval Conference (ISMIR),
pp 205-212.
C. Lordelo, E. Benetos, S. Dixon and S. Ahlbäck (2021),
Pitch-Informed Instrument Assignment Using a Deep Convolutional Network with Multiple Kernel Shapes.
22nd International Society for Music Information Retrieval Conference (ISMIR),
pp 389-395.
K. O'Hanlon, E. Benetos and S. Dixon (2021),
Detecting cover songs with Key-invariant Pitch Class Networks.
2021 IEEE International Workshop on Machine Learning for Signal Processing
(MLSP), pp 1-6.
DOI.
B. O'Connor, S. Dixon and G. Fazekas (2021),
Zero-shot Singing Technique Conversion.
15th International Symposium on Computer Music Multidisciplinary Research (CMMR).
R. Agrawal, D. Wolff and S. Dixon (2021),
Structure-Aware Audio-to-Score Alignment Using Progressively
Dilated Convolutional Neural Networks.
IEEE International Conference on Acoustics, Speech and Signal Processing
(ICASSP), pp 571-575.
E. Demirel, S. Ahlbäck and S. Dixon (2021),
Low Resource Audio-to-Lyrics Alignment from Polyphonic Music Recordings.
IEEE International Conference on Acoustics, Speech and Signal Processing
(ICASSP), pp 586-590.
Y. Zhang, G. Xia, M. Levy and S. Dixon (2021),
COSMIC: A Conversational Interface for Human-AI Music Co-Creation.
International Conference on New Interfaces for Musical Expression (NIME).
J. Pauwels, S. Dixon and J.D. Reiss (2021),
A Frontend for Adaptive Online Listening Tests.
Web Audio Conference (WAC).
Yukun Li, Emir Demirel, Polina Proutskova and Simon Dixon (2021),
Phoneme-Informed Note Segmentation of Monophonic Vocal Music.
Second Workshop on NLP for Music and Spoken Audio (NLP4MuSA).
D. Stoller, M. Tian, S. Ewert and S. Dixon (2020),
Seq-U-Net: A One-Dimensional Causal U-Net for Efficient Sequence Modelling.
International Joint Conference on Artificial Intelligence, pp. 2893-2900.
D. Stoller, S. Ewert and S. Dixon (2020),
Training Generative Adversarial Networks from Incomplete Observations using
Factorised Discriminators.
International Conference on Learning Representations, April 2020.
B. O'Connor, S. Dixon and G. Fazekas (2020),
An Exploratory Study on Perceptual Spaces of the Singing Voice.
2020 Joint Conference on AI Music Creativity.
R. Agrawal and S. Dixon (2020),
Learning Frame Similarity using Siamese Networks for Audio-to-Score Alignment.
28th European Signal Processing Conference, pp 141-145.
E. Demirel, S. Ahlbäck and S. Dixon (2020),
Automatic Lyrics Transcription using Dilated Convolutional Neural Networks
with Self-Attention.
International Joint Conference on Neural Networks, July 2020.
S. Mishra, E. Benetos, B.L. Sturm and S. Dixon (2020),
Reliable Local Explanations for Machine Listening.
International Joint Conference on Neural Networks, July 2020.
F. Rodríguez-Algarra, B. L. Sturm and S. Dixon (2019),
Characterising Confounding Effects in Music Classification Experiments
through Interventions.
Transactions of the International Society for Music Information Retrieval,
2 (1), 52-66.
DOI
A. Mehrabi, S. Dixon and M. Sandler (2019),
Vocal imitation of percussion sounds:
On the perceptual similarity between imitations and imitated sounds.
PLoS ONE 14 (7): e0219955.
DOI
J. Dai and S. Dixon (2019),
Intonation Trajectories within Tones in Unaccompanied Soprano, Alto, Tenor,
Bass Quartet Singing.
Journal of the Acoustical Society of America, 146 (2): 1005-1014.
DOI
Copyright (2019) Acoustical Society of America. This article may be downloaded for personal use only. Any other use requires prior permission of the author and the Acoustical Society of America. The article appeared in the Journal of the Acoustical Society of America, 146 (2), 1005-1014 and may be found at http://scitation.aip.org/content/asa/journal/jasa/146/2/10.1121/1.5120483?aemail=author.
J. Dai and S. Dixon (2019),
Pitch Accuracy and Interaction in Unaccompanied Duet Singing.
Journal of the Acoustical Society of America, 145 (2), 663-675.
DOI
Copyright (2019) Acoustical Society of America. This article may be downloaded for personal use only. Any other use requires prior permission of the author and the Acoustical Society of America. The article appeared in the Journal of the Acoustical Society of America, 145 (2), 663-675 and may be found at http://scitation.aip.org/content/asa/journal/jasa/145/2/10.1121/1.5087817?aemail=author.
E. Benetos, S. Dixon, Z. Duan, and S. Ewert (2019),
Automatic Music Transcription: An Overview.
IEEE Signal Processing Magazine, 36 (1), 20-30.
DOI
C. Weiß, M. Mauch, S. Dixon and M. Müller (2019),
Investigating style evolution of Western classical music:
A computational approach.
Musicae Scientiae, 23 (4): 486-507.
DOI
C. Vianna Lordelo, E. Benetos, S. Dixon and S. Ahlbäck (2019),
Investigating kernel shapes and skip connections for deep learning-based harmonic-percussive separation.
IEEE Workshop on Applications of Signal Processing to Audio and Acoustics.
K. Frieler, D. Başaran, F. Höger, H.-C. Crayencour, G. Peeters and S. Dixon (2019),
Don't Hide in the Frames: Note- and Pattern-based Evaluation of Automated Melody Extraction Algorithms.
6th International Conference on Digital Libraries for Musicology, pp. 25-32.
DOI
R. Agrawal and S. Dixon (2019),
A Hybrid Approach to Audio-to-Score Alignment.
Machine Learning for Music Discovery Workshop at the
36th International Conference on Machine Learning.
E. Demirel, S. Ahlbäck and S. Dixon (2019),
Exploring Generalizability of Automatic Phoneme Recognition Models.
UK Speech 2019.
S. Mishra, D. Stoller, E. Benetos, B. Sturm and S. Dixon (2019),
GAN-Based Generation and Automatic Selection of Explanations for Neural
Networks.
Safe Machine Learning Workshop at the International Conference on Learning
Representations.
J. Dai and S. Dixon (2019),
Understanding Intonation Trajectories and Patterns of Vocal Notes.
25th International Conference on MultiMedia Modeling, pp 243-253.
DOI
M. Panteli, E. Benetos and S. Dixon (2018),
A review of manual and computational approaches for the study of world music
corpora.
Journal of New Music Research, 47 (2), 176-189.
DOI
K. Frieler, F. Höger, M. Pfleiderer and S. Dixon (2018),
Two Web Applications for Exploring Melodic Patterns in Jazz Solos.
19th International Society for Music Information Retrieval Conference,
pp 777-783.
Eita Nakamura, Ryo Nishikimi, Simon Dixon and Kazuyoshi Yoshii (2018),
Probabilistic Sequential Patterns for Singing Transcription.
Asia-Pacific Signal and Information Processing Association Annual Summit
and Conference.
DOI
S. Mishra, B. Sturm and S. Dixon (2018),
Understanding Deep Machine Listening Systems through Feature Inversion.
19th International Society for Music Information Retrieval Conference,
pp 755-762.
D. Stoller, S. Ewert and S. Dixon (2018),
Wave-U-Net:
A Multi-Scale Neural Network for End-to-End Audio Source Separation.
19th International Society for Music Information Retrieval Conference,
pp 334-340.
D. Stoller and S. Ewert and S. Dixon (2018),
Jointly Detecting and Separating Singing Voice: A Multi-Task Approach.
14th International Conference on Latent Variable Analysis and Signal Separation,
pp 329-339.
D. Stoller and V. Akkermans and S. Dixon (2018),
Detection of cut-points for automatic music rearrangement.
IEEE International Workshop on Machine Learning for Signal Processing.
S. Mishra, B. Sturm and S. Dixon (2018),
"What are You Listening to?"
Explaining Predictions of Deep Machine Listening Systems.
26th European Signal Processing Conference, pp 2260-2264.
Eita Nakamura, Emmanouil Benetos, Kazuyoshi Yoshii, Simon Dixon (2018),
Towards Complete Polyphonic Music Transcription:
Integrating Multi-Pitch Detection and Rhythm Quantization.
IEEE International Conference on Acoustics, Speech and Signal Processing
(ICASSP), pp 101-105.
Daniel Stoller, Sebastian Ewert and Simon Dixon (2018),
Adversarial Semi-Supervised Audio Source Separation applied to Singing
Voice Extraction.
IEEE International Conference on Acoustics, Speech and Signal Processing
(ICASSP), pp 2391-2395.
Adib Mehrabi, Kuenwoo Choi, Simon Dixon and Mark Sandler (2018),
Similarity Measures for Vocal-based Drum Sample Retrieval using Deep
Convolutional Auto-encoders.
IEEE International Conference on Acoustics, Speech and Signal Processing
(ICASSP), pp 356-360.
Shengchen Li, Simon Dixon and Mark Plumbley (2018),
A demonstration of hierarchical structure usage in expressive timing
analysis by model selection tests.
37th Chinese Control Conference, pp 3190-3195.
M. Panteli, E. Benetos and S. Dixon (2017),
A computational study on outliers in world music.
PLoS ONE 12 (12): e0189399
DOI
Z. Mohamad, S. Dixon and C. Harte (2017),
Pickup position and plucking point estimation on an electric guitar
via autocorrelation.
Journal of the Acoustical Society of America, 142 (6), 3530-3540.
DOI
Copyright (2017) Acoustical Society of America. This article may be downloaded for personal use only. Any other use requires prior permission of the author and the Acoustical Society of America. The article appeared in the Journal of the Acoustical Society of America, 142 (6), 3530-3540 and may be found at http://scitation.aip.org/content/asa/journal/jasa/142/6/10.1121/1.5016815?aemail=author.
B. Di Giorgi, S. Dixon, M. Zanoni, and A. Sarti (2017),
A Data-driven Model of Tonal Chord Sequence Complexity.
IEEE/ACM Transactions on Audio, Speech and Language Processing,
25 (11), 2237-2250.
DOI
S. Wang, S. Ewert and S. Dixon (2017),
Identifying Missing and Extra Notes in Piano Recordings Using Score-Informed
Dictionary Learning.
IEEE/ACM Transactions on Audio, Speech and Language Processing,
25 (10), 1877-1889.
DOI
E. Nakamura, S. Dixon and K. Yoshii (2017),
Note Value Recognition for Piano Transcription Using Markov Random Fields.
IEEE/ACM Transactions on Audio, Speech and Language Processing, 25 (9)
1846-1858.
DOI
A. Mehrabi, S. Dixon and M. Sandler (2017),
Vocal Imitation of Synthesised Sounds Varying in Pitch, Loudness and Spectral
Centroid.
Journal of the Acoustical Society of America, 141 (2), 783-796.
DOI
Copyright (2017) Acoustical Society of America. This article may be downloaded for personal use only. Any other use requires prior permission of the author and the Acoustical Society of America. The article appeared in the Journal of the Acoustical Society of America, 141 (2), pp 783-796 and may be found at http://scitation.aip.org/content/asa/journal/jasa/141/2/10.1121/1.4974825?aemail=author.
A. Mehrabi, C. Harte, C. Baume and S. Dixon (2017),
Music Thumbnailing for Radio Podcasts: A Listener Evaluation.
Journal of the Audio Engineering Society, 65 (6), 474-481.
J. Dai and S. Dixon (2017),
Analysis of Interactive Intonation in Unaccompanied SATB Ensembles.
18th International Society for Music Information Retrieval Conference
(ISMIR 2017), pp 599-605.
S. Li, S. Dixon and M. Plumbley (2017),
Clustering Expressive Timing with Regressed Polynomial Coefficients
Demonstrated by a Model Selection Test.
18th International Society for Music Information Retrieval Conference
(ISMIR 2017), pp 457-463.
S. Mishra, B. Sturm and S. Dixon (2017),
Local Interpretable Model-Agnostic Explanations for Music Content Analysis.
18th International Society for Music Information Retrieval Conference
(ISMIR 2017), pp 537-543.
V. Kedyte, M. Panteli, T. Weyde and S. Dixon (2017),
Geographical origin prediction of folk music recordings from the United Kingdom.
18th International Society for Music Information Retrieval Conference
(ISMIR 2017), pp 664-670.
Z. Mohamad, S. Dixon and C. Harte (2017),
Estimating Pickup and Plucking Positions of Guitar Tones and Chords
with Audio Effects.
20th International Conference on Digital Audio Effects (DAFx-17),
pp 420-427.
Z. Mohamad, S. Dixon and C. Harte (2017),
Pickup Position and Plucking Point Estimation on an Electric Guitar.
IEEE International Conference on Acoustics, Speech and Signal Processing
(ICASSP), pp 651-655.
M. Panteli, R. Bittner, J.P. Bello and S. Dixon (2017),
Towards the Characterization of Singing Styles in World Music.
IEEE International Conference on Acoustics, Speech and Signal Processing
(ICASSP), pp 636-640.
E. Quinton, K. O'Hanlon, S. Dixon and M. Sandler (2017),
Tracking Metrical Structure Changes with Sparse-NMF.
IEEE International Conference on Acoustics, Speech and Signal Processing
(ICASSP), pp 40-45.
S. Wang, S. Ewert and S. Dixon (2016),
Robust and Efficient Joint Alignment of Multiple Musical Performances.
IEEE/ACM Transactions on Audio, Speech and Language Processing,
24 (11), 2132-2145.
DOI
S. Sigtia, E. Benetos and S. Dixon (2016),
An End-to-End Neural Network for Polyphonic Piano Music Transcription.
IEEE/ACM Transactions on Audio, Speech and Language Processing,
24 (5), pp 927-939.
DOI
Y. Song, S. Dixon, M.T. Pearce and A.R. Halpern (2016),
Perceived and Induced Emotion Responses to Popular Music:
Categorical and Dimensional Models,
Music Perception, 33 (4), pp 472-492.
DOI
T. Cheng, M. Mauch, E. Benetos and S. Dixon (2016),
An Attack/Decay Model for Piano Transcription.
17th International Society for Music Information Retrieval Conference
(ISMIR 2016), pp 584-590.
J. Dai and S. Dixon (2016),
Analysis of Vocal Imitations of Pitch Trajectories.
17th International Society for Music Information Retrieval Conference
(ISMIR 2016), pp 87-93.
M. Panteli, E. Benetos and S. Dixon (2016),
Learning a Feature Space for World Music Style Similarity.
17th International Society for Music Information Retrieval Conference
(ISMIR 2016), pp 538-544 (Best Student Paper Award).
M. Panteli and S. Dixon (2016),
On the Evaluation of Rhythmic and Melodic Descriptors for Music Similarity.
17th International Society for Music Information Retrieval Conference
(ISMIR 2016), pp 468-474.
D. Stoller and S. Dixon (2016),
Analysis and Classification of Phonation Modes in Singing.
17th International Society for Music Information Retrieval Conference
(ISMIR 2016), pp 80-86.
E. Quinton, M. Sandler and S. Dixon (2016),
Estimation of the Reliability of Multiple Rhythm Features Extraction
from a Single Descriptor.
IEEE International Conference on Acoustics Speech and Signal Processing (ICASSP), pp 256-260.
M. Panteli, E. Benetos and S. Dixon (2016),
Automatic detection of outliers in world music collections.
Fourth International Conference on Analytical Approaches to World Music (AAWM).
Adib Mehrabi, Simon Dixon and Mark Sandler (2016),
Towards a Comprehensive Dataset of Vocal Imitations of Drum Sounds.
2nd AES Workshop on Intelligent Music Production, London, UK.
S. Li, S. Dixon, D. Black and M. Plumbley (2016),
A Model Selection Test for Factors Affecting the Choice of Expressive Timing Clusters for a Phrase.
Proceedings of the 13th Sound and Music Computing Conference (SMC), pp. 247-252.
P. Foster, S. Dixon and A. Klapuri (2015),
Identifying Cover Songs Using Information-Theoretic Measures of
Similarity,
IEEE/ACM Transactions on Audio, Speech and Language Processing,
23 (6), pp 993-1005.
DOI
J. Dai, M. Mauch and S. Dixon (2015),
Analysis of Intonation Trajectories in Solo Singing.
16th International Society for Music Information Retrieval Conference
(ISMIR 2015), pp 420-426.
S. Sigtia, N. Boulanger-Lewandowski and S. Dixon (2015),
Audio Chord Recognition with a Hybrid Recurrent Neural Network.
16th International Society for Music Information Retrieval Conference
(ISMIR 2015), pp 127-133.
J. Osmalskyj, J.-J. Embrechts, P. Foster and S. Dixon (2015),
Combining Features for Cover Song Identification.
16th International Society for Music Information Retrieval Conference
(ISMIR 2015), pp 462-468.
R. Tubb and S. Dixon (2015),
An Evaluation of Multidimensional Controllers for Sound Design Tasks,
Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing
Systems (CHI'15), pp 47-56.
Z. Mohamad, S. Dixon and C. Harte (2015),
Digitally Moving an Electric Guitar Pickup,
18th International Conference on Digital Audio Effects (DAFx-15), pp 284-291.
T. Cheng, S. Dixon and M. Mauch (2015),
Improving Piano Note Tracking by HMM Smoothing,
European Signal Processing Conference (EUSIPCO 2015), pp 2009-2013.
Y. Song and S. Dixon (2015),
How Well Can a Music Emotion Recognition System
Predict the Emotional Responses of Participants?
12th Sound and Music Computing Conference, pp 387-392.
S. Sigtia, E. Benetos, N. Boulanger-Lewandowski, T. Weyde, A.S. d'Avila Garcez
and S. Dixon (2015),
A Hybrid Recurrent Neural Network for Music Transcription,
IEEE International Conference on Acoustics Speech and Signal Processing (ICASSP), pp 2061-2065.
T. Cheng, S. Dixon and M. Mauch (2015),
Modelling the Decay of Piano Sounds,
IEEE International Conference on Acoustics Speech and Signal Processing (ICASSP), pp 594-598.
S. Wang, S. Ewert and S. Dixon (2015),
Compensating for Asynchronies between Musical Voices in Score-Performance
Alignment,
IEEE International Conference on Acoustics Speech and Signal Processing (ICASSP), pp 589-593 (Best Student Paper Award for the Audio and Acoustic
Signal Processing track).
M. Mauch, C. Cannam, R. Bittner, G. Fazekas, J. Salamon, J. Dai, J. Bello
and S. Dixon (2015),
Computer-aided Melody Note Transcription Using the Tony Software:
Accuracy and Efficiency.
First International Conference on Technologies for Music Notation and Representation (TENOR 2015), pp 23-30.
T. Wilmering, S. Bechhofer, S. Dixon, G. Fazekas and K. Page (2015),
Automating Annotation of Media with Linked Data Workflows,
Third International Workshop on Linked Media (LiME 2015), pp 737-738.
T. Wilmering, G. Fazekas, S. Dixon, K. Page and S. Bechhofer (2015),
Towards High Level Feature Extraction from Large Live Music Recording Archives,
Machine Learning for Music Discovery Workshop,
32nd International Conference on Machine Learning.
D. Tidhar, S. Dixon, E. Benetos and T. Weyde (2014),
The Temperament Police,
Early Music, 42 (4), pp 579-590.
DOI
P. Foster, M. Mauch and S. Dixon (2014),
Sequential Complexity as a Descriptor for Musical Similarity,
IEEE/ACM Transactions on Audio, Speech and Language Processing, 22 (12),
pp 1965-1977.
DOI
M. Mauch, K. Frieler and S. Dixon (2014),
Intonation in Unaccompanied Singing: Accuracy, Drift and a Model of Reference
Pitch Memory,
Journal of the Acoustical Society of America, 136 (1), pp 401-411.
DOI
Copyright (2014) Acoustical Society of America. This article may be downloaded for personal use only. Any other use requires prior permission of the author and the Acoustical Society of America. The article appeared in the Journal of the Acoustical Society of America, 136 (1), pp 401-411 and may be found at http://scitation.aip.org/content/asa/journal/jasa/136/1/10.1121/1.4881915?aemail=author.
R. Tubb and S. Dixon (2014),
A Zoomable Mapping of a Musical Parameter Space Using Hilbert Curves
,
Computer Music Journal, 38 (3), pp 23-33.
DOI
W. Goebl, S. Dixon and E. Schubert (2014),
Quantitative Methods: Motion Analysis, Audio Analysis, and Continuous Response
Techniques,
in Expressiveness in Music Performance:
Empirical Approaches across Styles and Cultures,
Ed. D. Fabian, R. Timmers and E. Schubert, Oxford University Press,
pp. 221-239.
D. Stowell and S. Dixon (2014),
Integration of informal music technologies in secondary school music lessons.
British Journal of Music Education, 31 (1), pp 19-39.
DOI
Copyright (2013) Cambridge University Press. The article appeared in the British Journal of Music Education.
T. Cheng, S. Dixon and M. Mauch (2014)
A Comparison of Extended Source-Filter Models for Musical Signal Reconstruction.
17th International Conference on Digital Audio Effects, pp 203-209.
H. Kirchhoff, R. Badeau and S. Dixon (2014),
Towards Complex Matrix Decomposition of Spectrograms based
on the Relative Phase Offsets of Harmonic Sounds.
IEEE International Conference on Acoustics Speech and Signal Processing (ICASSP), pp 1572-1576.
M. Mauch and S. Dixon (2014),
PYIN: a Fundamental Frequency Estimator Using Probabilistic Threshold
Distributions,
IEEE International Conference on Acoustics Speech and Signal Processing (ICASSP), 659-663.
DOI
S. Sigtia and S. Dixon (2014),
Improved Music Feature Learning With Deep Neural Networks,
IEEE International Conference on Acoustics Speech and Signal Processing (ICASSP), 6959-6963.
DOI
S. Sigtia, E. Benetos, S. Cherla, T. Weyde, A. Garcez and S. Dixon (2014),
An RNN-based Music Language Model for Improving Automatic Music Transcription,
15th International Society for Music Information Retrieval Conference,
pp 53-58.
L. Thompson, S. Dixon and M. Mauch (2014),
Drum Transcription via Classification of Bar-level Rhythmic Patterns,
15th International Society for Music Information Retrieval Conference,
pp 187-192.
S. Wang, S. Ewert and S. Dixon (2014),
Robust Joint Alignment of Multiple Versions of a Piece of Music,
15th International Society for Music Information Retrieval Conference,
pp 83-88.
R. Tubb and S. Dixon (2014),
The Divergent Interface: Supporting Creative Exploration of Parameter
Spaces,
14th International Conference on New Interfaces for Musical Expression.
R. Tubb and S. Dixon (2014),
A Four Strategy Model of Creative Parameter Space Interaction,
5th International Conference on Computational Creativity.
T. Weyde, S. Cottrell, E. Benetos, D. Wolff, D. Tidhar, J. Dykes, M. Plumbley, S. Dixon, M. Barthet, N. Gold, S. Abdallah, M. Mahey (2014),
Digital Music Lab - A Framework for Analysing Big Music Data,
European Conference on Data Analysis, Workshop on Statistical Musicology
(
abstract only).
T. Weyde, S. Cottrell, J. Dykes, E. Benetos, D. Wolff, D. Tidhar, N. Gold,
S. Abdallah, M. Plumbley, S. Dixon, M. Barthet, M. Mahey, A. Tovell,
A. Alancar-Brayner (2014),
Big Data for Musicology,
1st International Digital Libraries for Musicology Workshop.
C. Weiß, M. Mauch and S. Dixon (2014),
Timbre-invariant Audio Features for Style Analysis of Classical Music,
Joint 40th International Computer Music Conference and 11th Sound and Music Computing Conference, pp 1461-1468.
E. Benetos and S. Dixon (2013),
Multiple-instrument Polyphonic Music Transcription using a
Temporally-constrained Shift-invariant Model.
Journal of the Acoustical Society of America, 133 (3), pp 1727-1741.
DOI
Copyright (2013) Acoustical Society of America. This article may be downloaded for personal use only. Any other use requires prior permission of the author and the Acoustical Society of America. The article appeared in the Journal of the Acoustical Society of America, 131 (3), pp 1727-1741 and may be found at http://link.aip.org/link/?JAS/133/1727">
E. Benetos, S. Dixon, D. Giannoulis, H. Kirchhoff and A. Klapuri (2013),
Automatic Music Transcription: Challenges and Future Directions.
Journal of Intelligent Information Systems,
41 (3), pp 407-434.
DOI
X. Serra, M. Magas, E. Benetos, M. Chudy, S. Dixon, A. Flexer, E. Gómez, F. Gouyon, P. Herrera, S. Jorda, O. Paytuvi, G. Peeters, J. Schlüter, H. Vinet and G. Widmer (2013),
Roadmap for Music Information Research.
Final report of the EU FP7 MIReS
Project.
T. Cheng, S. Dixon and M. Mauch (2013),
A Deterministic Annealing EM Algorithm for Automatic Music Transcription.
In 14th International Society for Music Information Retrieval Conference,
pp 475-480.
Y. Song, S. Dixon, M. Pearce and A. Halpern (2013),
Do Online Social Tags Predict Perceived or Induced Emotional Responses to Music?
In 14th International Society for Music Information Retrieval Conference,
pp 89-94.
H. Kirchhoff, S. Dixon and A. Klapuri (2013),
Multiple instrument tracking based on reconstruction error, pitch continuity
and instrument activity.
In 10th International Symposium on Computer Music Multidisciplinary Research,
pp 894-903 (Best Poster Paper Award).
Y. Song, S. Dixon, M. Pearce and G. Fazekas (2013),
Using Tags to Select Stimuli in the Study of Music and Emotion.
Proceedings of the 3rd International Conference on Music and Emotion.
Identification of Cover Songs using Information Theoretic Measures of Similarity.
IEEE International Conference on Acoustics Speech and Signal Processing (ICASSP), pp 739-743.
H. Kirchhoff, S. Dixon and A. Klapuri (2013),
Missing Template Estimation for User-assisted Music Transcription.
IEEE International Conference on Acoustics Speech and Signal Processing
(ICASSP), pp 26-30.
M. Chudy, A. Perez Carrillo and S. Dixon (2013),
On the Relation between Gesture, Tone Production and Perception in Classical
Cello Performance.
Proceedings of Meetings on Acoustics 19 (1), 035017.
Presented at the 21st International Congress on Acoustics.
M. Chudy and S. Dixon (2013),
Recognising Cello Performers Using Timbre Models
(preprint).
In
B. Lausen, D. van den Poel, and A. Ultsch, editors, Joint Annual Conference
of the German Association for Pattern Recognition (DAGM) and the German
Classification Society (GfKl), and Symposium of the International Federation of
Classification Societies (IFCS): Algorithms from & for Nature and Life,
Studies in Classification, Data Analysis, and Knowledge Organization,
Heidelberg: Springer, pp 511-518.
A. Allik, G. Fazekas, S. Dixon, and M. Sandler (2013),
Facilitating Music Information Research with Shared Open Vocabularies.
Extended Semantic Web Conference (ESWC), pp 178-183.
(DOI)
A. Allik, G. Fazekas, S. Dixon, and M. Sandler (2013),
A Shared Vocabulary for Audio Features.
Extended Semantic Web Conference (ESWC), pp 285-286.
(DOI)
S. Dixon, M. Mauch and D. Tidhar (2012),
Estimation of Harpsichord Inharmonicity and Temperament from Musical
Recordings.
Journal of the Acoustical Society of America, 131 (1), pp 878-887.
Copyright (2012) Acoustical Society of America. This article may be downloaded for personal use only. Any other use requires prior permission of the author and the Acoustical Society of America. The article appeared in the Journal of the Acoustical Society of America, 131 (1), pp 878-887 and may be found at http://link.aip.org/link/?JAS/131/878
E. Benetos and S. Dixon (2012),
A shift-invariant latent variable model for automatic music transcription.
Computer Music Journal, 36 (4), pp 81-94.
DOI
A. Arzt, G. Widmer, and S. Dixon (2012),
Adaptive distance normalization for real-time music tracking.
In 20th European Signal Processing Conference, pp 2689-2693.
E. Benetos and S. Dixon (2012),
Temporally-constrained Convolutive Probabilistic Latent Component Analysis
for Multi-pitch Detection.
International Conference on Latent Variable Analysis and Signal Separation,
pp 364-371.
E. Benetos, A. Klapuri and S. Dixon (2012),
Score-informed transcription for automatic piano tutoring.
In 20th European Signal Processing Conference, pp 2153-2157.
Emmanouil Benetos, Mathieu Lagrange, and Simon Dixon (2012),
Characterisation of
acoustic scenes using a temporally-constrained shift-invariant model.
In 15th International Conference on Digital Audio Effects, pp 317-323.
Emmanouil Benetos, Simon Dixon, Dimitrios Giannoulis, Holger Kirchhoff, and
Anssi Klapuri (2012),
Automatic music transcription: Breaking the glass ceiling.
In 13th International Society for Music Information Retrieval Conference,
pp 379-384.
P. Foster, A. Klapuri, and S. Dixon (2012),
A method for identifying repetition structure in musical audio based on
time series prediction.
In 20th European Signal Processing Conference, pp 1299-1303.
J. Gansemann, P. Scheunders, and S. Dixon (2012),
Improving PLCA-based score-informed source separation with
invertible constant-Q transforms.
In 20th European Signal Processing Conference, pp 2634-2638.
H. Kirchhoff, S. Dixon and A. Klapuri (2012),
Shift-Variant Non-Negative Matrix Deconvolution for Music Transcription.
IEEE International Conference on Acoustics Speech and Signal Processing (ICASSP), pp 125-128.
Holger Kirchhoff, Simon Dixon, and Anssi Klapuri (2012),
Multi-template shift-variant non-negative matrix deconvolution for
semi-automatic music transcription.
In 13th International Society for Music Information Retrieval Conference,
pp 415-420.
R. Macrae and S. Dixon (2012),
Ranking lyrics for online search.
In 13th International Society for Music Information Retrieval Conference,
pp 361-366.
M. Mauch and S. Dixon (2012),
A corpus-based study of rhythm patterns.
In 13th International Society for Music Information Retrieval Conference,
pp 163-168.
M. Barthet, G. Fazekas, S. Dixon and M. Sandler (2012),
Social Media Retrieval for Music Education.
In Digital Futures 2012: The Third Annual Digital Economy All Hands Conference.
Y. Song, M. Pearce, and S. Dixon (2012),
A survey of music recommendation systems and future perspectives.
In 9th International Symposium on Computer Music
Modelling and Retrieval, pages 395-410, 2012.
Y. Song, M. Pearce, and S. Dixon (2012),
Evaluation of musical features for emotion classification.
In 13th International Society for Music Information Retrieval
Conference, pp 523-528.
M.J. Terrill, G. Fazekas, A.J.R. Simpson, J. Smith, and S. Dixon (2012),
Listening level changes music similarity.
In 13th International Society for Music Information
Retrieval Conference, pp 487-492.
R. Tubb, A. Klapuri, and S. Dixon (2012),
The Wablet: Scanned synthesis on a multi-touch interface.
In 15th International Conference on Digital Audio Effects,
pp 279-286.
S. Dixon, M. Mauch and A. Anglade (2011),
Probabilistic and Logic-Based Modelling of Harmony.
CMMR 2010 Post Symposium Proceedings, Ed. R. Kronland-Martinet, S. Ystad, K. Jensen and M. Aramaki, Springer LNCS 6684, Heidelberg, pp 1-19.
S. Dixon, D. Tidhar and E. Benetos (2011),
The Temperament Police:
The Truth, the Ground Truth and Nothing but the Truth.
In 12th International Society for Music Information Retrieval Conference,
pp 281-286.
R. Macrae and S. Dixon (2011),
Guitar Tab Mining, Analysis and Ranking.
In 12th International Society for Music Information Retrieval Conference,
pp 453-458.
D. Stowell and S. Dixon (2011),
MIR in School? Lessons from Ethnographic Observation of Secondary School
Music Classes.
In 12th International Society for Music Information Retrieval Conference,
pp 347-352.
M. Barthet and S. Dixon,
Ethnographic Observations of Musicologists at the British Library:
Implications for Music Information Retrieval.
In 12th International Society for Music Information Retrieval Conference,
pp 353-358.
E. Benetos and S. Dixon (2011),
Joint Multi-Pitch Detection using Spectral Envelope Estimation for Polyphonic Music Transcription.
Journal of Selected Topics in Signal Processing, 5 (6), pp 1111-1123.
E. Benetos and S. Dixon (2011),
A Temporally-Constrained Convolutive Probabilistic Model for Pitch
Detection.
IEEE Workshop on Applications of Signal Processing to Audio and Acoustics,
pp 133-136.
E. Benetos and S. Dixon (2011),
Polyphonic Music Transcription using Note Onset and Offset Detection.
IEEE International Conference on Acoustics, Speech and Signal Processing,
pp 37-40.
E. Benetos and S. Dixon (2011),
Multiple-Instrument Polyphonic Music Transcription using a Convolutive Probabilistic Model.
8th Sound and Music Computing Conference,
pp 19-24.
L.Mearns, E. Benetos and S. Dixon (2011),
Automatically Detecting Key Modulations in J.S. Bach Chorale Recordings.
8th Sound and Music Computing Conference,
pp 25-32.
R. Macrae, J. Neumann, X. Anguera, N. Oliver and S. Dixon (2011),
Real-Time Synchronisation of Multimedia Streams in a Mobile Device.
IEEE International Conference on Multimedia and Expo,
Workshop on Advances in Music Information Research (AdMIRe).
C. Mesnage, A. Rafiq, S. Dixon and R. Brixtel (2011),
Music Discovery with Social Networks.
Workshop on Music Recommendation and Discovery,
Chicago, USA, pp 1-6.
S. Dixon, M. Sandler, M. d’Inverno, and C. Rhodes (2010),
Towards a distributed research environment for music informatics and
computational musicology.
Journal of New Music Research, 39 (4): 291-294.
A. Anglade, E. Benetos, M. Mauch, and S. Dixon (2010),
Improving music genre classification using automatically induced harmony rules
.
Journal of New Music Research, 39 (4), 349-361.
E. Benetos and S. Dixon (2010),
Multiple-F0 estimation of piano sounds exploiting spectral structure and
temporal evolution.
In 2010 Workshop on Statistical and Perceptual Audition (SAPA), 2010.
M. Chudy and S. Dixon (2010),
Towards music performer recognition using timbre features.
In Third International Conference of Students of Systematic Musicology.
S. Dixon (2010),
Computational modelling of harmony.
In 7th International Symposium on Computer Music Modeling and Retrieval,
(CMMR 2010), Malaga, Spain.
K. Jacobson, S. Dixon, and M. Sandler (2010),
LinkedBrainz: Providing the MusicBrainz
next generation schema as linked data.
Late-breaking session, 11th International Society for Music Information
Retrieval Conference, Utrecht, Netherlands.
R. Macrae and S. Dixon (2010),
A guitar tablature score follower.
In IEEE International Conference on Multimedia and Expo.
R. Macrae and S. Dixon (2010),
Accurate real-time windowed time warping.
In 11th International Society for Music Information Retrieval Conference.
M. Mauch and S. Dixon (2010),
Simultaneous estimation of chords and musical context from audio.
IEEE Transactions on Audio, Speech and Language Processing, 18 (6), 1280-1289.
DOI: 10.1109/TASL.2009.2032947
M. Mauch and S. Dixon (2010),
Approximate Note Transcription for the Improved Identification of Difficult
Chords.
In 11th International Society for Music Information Retrieval Conference
(ISMIR 2010), Utrecht, Netherlands.
L. Mearns and S. Dixon (2010),
Characterisation of composer style using high level musical features.
Late-breaking session, 11th International Society for Music Information
Retrieval Conference, Utrecht, Netherlands.
L. Mearns, D. Tidhar, and S. Dixon (2010),
Characterisation of composer style using high-level musical features.
In 3rd ACM Workshop on Machine Learning and Music.
L. Mearns and S. Dixon (2010),
How do you characterise musical style?
In Third International Conference of Students of Systematic Musicology, 2010.
D. Tidhar, M. Mauch, and S. Dixon (2010),
High precision frequency estimation for harpsichord tuning classification.
In Proceedings of the IEEE International Conference on Acoustics, Speech and
Signal Processing, pages 61-64.
D. Tidhar, G. Fazekas, M. Mauch, and S. Dixon (2010),
TempEst: Harpsichord temperament estimation in a semantic-web environment.
Journal of New Music Research, 39 (4), 327-336.
D. Tidhar, G. Fazekas, M. Mauch, and S. Dixon (2010),
Temperament estimation as an MIR task.
Late-breaking session, 11th International Society for Music Information
Retrieval Conference, Utrecht, Netherlands.
A. Anglade and S. Dixon (2009),
First-order logic classification models
of musical genres based on harmony.
In 6th Sound and Music Computing Conference, pages 309-314.
A. Anglade, R. Ramirez, and S. Dixon (2009),
Genre classification using harmony
rules induced from automatic chord transcriptions.
In 10th International Society for
Music Information Retrieval Conference, pages 669-674.
M. Mauch, K. Noland, and S. Dixon (2009),
Using musical structure to enhance automatic chord transcription.
In 10th International Society for Music Information Retrieval Conference,
pages 231-236.
M. Mauch, C. Cannam, M. Davies, S. Dixon, C. Harte, S. Kolozali, and D. Tidhar
(2009),
OMRAS-2 metadata project 2009.
Late-breaking session, 10th International Society for Music Information
Retrieval Conference, Kobe, Japan.
A. Anglade, S. Dixon (2008),
Characterisation of Harmony with Inductive Logic Programming.
Proceedings of the
9th International Conference on Music Information Retrieval (ISMIR 2008),
Philadelphia, USA, pp 63-68.
A. Anglade, S. Dixon (2008),
Towards Logic-based Representations of Musical
Harmony for Classification, Retrieval and Knowledge Discovery.
Proceedings of the International Workshop on Machine Learning and Music
(MML2008 held in conjunction with ICML/COLT/UAI 2008), Helsinki, Finland,
July, 2008.
M. Mauch, S. Dixon (2008),
A Discrete Mixture Model for Chord Labelling.
Proceedings of the
9th International Conference on Music Information Retrieval (ISMIR 2008),
Philadelphia, USA.
M. Mauch, D. Müllensiefen, S. Dixon, G. Wiggins (2008),
Can Statistical Language Models be Used for the Analysis of Harmonic
Progressions?
Proceedings of the 10th International Conference on Music Perception and
Cognition, Sapporo, Japan.
M. Mauch, D. Müllensiefen, S. Dixon, G. Wiggins (2008),
Applying Tools from Natural Language Processing Analysis to Harmony:
A Bottom Up Approach.
Conference on Music, Language and the Mind.
R. Macrae, S. Dixon (2008),
From Toy to Tutor: Note-Scroller is a Game to Teach Music.
8th International Conference on New Interfaces for Musical Expression.
A. Arzt, G. Widmer, S. Dixon (2008),
Automatic Page Turning for Musicians via Real-Time Machine Listening.
Proceedings of the 18th European Conference on Artificial Intelligence, pp
241-245.
G. Widmer, S. Dixon, P. Knees, E. Pampalk, T. Pohle (2008),
From Sound to Sense via Feature Extraction and Machine Learning:
Deriving High-Level Descriptors for Characterising Music.
Sound to Sense - Sense to Sound: A State of the Art in Sound and Music
Computing, Ed. P. Polotti and D. Rocchesso, pp 161-194.
W. Goebl, S. Dixon, G. De Poli, A. Friberg, R. Bresin, G. Widmer (2008),
Sense in Expressive Music Performance: Data Acquisition, Computational Studies,
and Models.
Sound to Sense - Sense to Sound: A State of the Art in Sound and Music
Computing, Ed. P. Polotti and D. Rocchesso, pp 195-242.
S. Dixon (2007),
Evaluation of the Audio Beat Tracking System BeatRoot.
Journal of New Music Research, 36, 1, pp 39-50.
M. Mauch, S. Dixon, C. Harte, M. Casey, B. Fields
Discovering Chord Idioms Through Beatles and Real Book Songs.
Proceedings of the
8th International Conference on Music Information Retrieval (ISMIR 2007),
Vienna, Austria, September 2007, pp 111-114.
S. Dixon,
Tools for Analysis of Musical Expression.
Proceedings of the 19th International Congress on Acoustics,
Madrid, September 2007.
F. Gouyon, S. Dixon and G. Widmer,
Evaluating low-level features for beat classification and tracking.
Proceedings of the 2007 International Conference on
Acoustics, Speech and Signal Processing, Vol. IV, pp 1309-1312.
S. Dixon, W. Goebl and E. Cambouropoulos (2006),
Perceptual Smoothness of Tempo in Expressively Performed Music.
Music Perception, 23 (3), pp 195-214.
F. Gouyon, A. Klapuri, S. Dixon, M. Alonso,
G. Tzanetakis, C. Uhle and P. Cano (2006),
An Experimental Comparison of Audio Tempo Induction Algorithms.
IEEE Transactions on Audio, Speech and Language Processing, 14 (5), pp 1832-1844.
S. Dixon
Onset Detection Revisited
Proceedings of the
9th International Conference on Digital Audio Effects,
Montreal, September 2006, pp 133-137.
A. Flexer, F. Gouyon, S. Dixon and G. Widmer,
Probabilistic Combination of Features for Music Classification,
Proceedings of the
7th International Conference on Music Information Retrieval (ISMIR 2006),
Victoria BC, October 2006, pp 111-114.
S. Dixon,
Computer Analysis of Performance Timing,
9th International Conference on Music Perception and Cognition,
Bologna, August 2006.
F. Gouyon and S. Dixon,
A Review of Automatic Rhythm Description Systems.
Computer Music Journal, 29 (1), 2005, pp 34-54.
DOI
S. Dixon,
An On-Line Time Warping Algorithm for Tracking Musical Performances,
Proceedings of the
International Joint Conference on Artificial Intelligence,
Edinburgh, August 2005, pp 1727-1728.
S. Dixon,
Live Tracking of Musical Performances using On-Line Time Warping,
Proceedings of the 8th International Conference on Digital Audio Effects
(DAFx05), Madrid, Spain, September 2005, pp 92-97.
S. Dixon and G. Widmer,
MATCH: A Music Alignment Tool Chest,
6th International Conference on Music Information Retrieval (ISMIR 2005),
London, England, September 2005, pp 492-497. (Best Poster Award)
S. Dixon, W. Goebl and G. Widmer,
The 'Air Worm': An Interface for Real-Time Manipulation of Expressive Music
Performance,
Proceedings of the 2005 International Computer Music Conference (ICMC2005),
Barcelona, Spain, September 2005, pp 614-617.
G. Widmer, S. Dixon, A. Flexer, W. Goebl, P. Knees, S.T. Madsen, E. Pampalk,
T. Pohle, M. Schedl and A. Tobudic
The Maching Learning and Intelligent Music Processing Group at the Austrian
Research Institute for Artificial Intelligence (OFAI),
Proceedings of the 2005 International Computer Music Conference (ICMC2005),
Barcelona, Spain, September 2005, pp 169-172.
W. Goebl, S. Dixon, G. DePoli, A. Friberg, R. Bresin and G. Widmer
'Sense' in Expressive Music Performance:
Data Acquisition, Computational Studies, and Models,
S2S2 Summer School, Genova, 2005.
S. Dixon
Audio Analysis Applications for Music.
Encyclopedia of Information Science and Technology,
(ed. M.Khosrow-Pour), Idea Group, 2005, pp 188-196.
E. Pampalk and S. Dixon and G. Widmer,
Exploring Music Collections by Browsing Different Views.
Computer Music Journal, 28 (2), 2004, pp 49-62.
S. Dixon, F. Gouyon and G. Widmer,
Towards Characterisation of Music via Rhythmic Patterns,
5th International Conference on Music Information Retrieval (ISMIR 2004),
pp 509-516.
F. Gouyon and S. Dixon,
Dance Music Classification: A Tempo-Based Approach,
5th International Conference on Music Information Retrieval (ISMIR 2004),
pp 501-504.
F. Gouyon, N. Wack and S. Dixon,
An Open Source Tool for Semi-Automatic Rhythmic Annotation.
Proceedings of the 7th International Conference on Digital Audio Effects, 2004,
pp 193-196.
F. Gouyon, S. Dixon, E. Pampalk and G. Widmer,
Evaluating Rhythmic Descriptors for Musical Genre Classification.
Proceedings of the AES 25th International Conference,
London, June 2004, pp 196-204.
S. Dixon
Analysis of Musical Content in Digital Audio.
Computer Graphics and Multimedia: Applications, Problems, and Solutions
(ed. J. DiMarco), Idea Group, 2004, pp 214-235.
G. Widmer, S. Dixon, W. Goebl, E. Pampalk and A. Tobudic
In Search of the Horowitz Factor.
AI Magazine, 24 (3), 2003, pp 111-130.
S. Dixon, E. Pampalk and G. Widmer,
Classification of Dance Music by Periodicity Patterns.
4th International Conference on Music Information Retrieval (ISMIR 2003),
Baltimore MD, October 2003, pp 159 - 165.
E. Pampalk, S. Dixon and G. Widmer
Exploring Music Collections by Browsing Different Views.
4th International Conference on Music Information Retrieval (ISMIR 2003),
Baltimore MD, October 2003, pp 201 - 208.
E. Pampalk, S. Dixon and G. Widmer
On the Evaluation of Perceptual Similarity Measures for Music.
6th International Conference on Digital Audio Effects (DAFx-03),
London, UK, September 2003, pp 7-12.
S. Dixon
On the Analysis of Musical Expression in Audio Signals.
Storage and Retrieval for Media Databases 2003,
SPIE-IS&T Electronic Imaging, SPIE Vol. 5021, pp 122-132.
S. Dixon
(PDF)
Towards Automatic Analysis of Expressive Performance.
5th Triennial Conference of the European Society for the Cognitive Sciences of
Music (ESCOM5),
8 - 13 September 2003, Hanover, Germany, pp 107-110.
S. Dixon, W. Goebl and G. Widmer
The Performance Worm: Real Time Visualisation of
Expression Based on Langner's Tempo-Loudness Animation.
International Computer Music Conference,
16 - 21 September 2002, Göteborg, Sweden, pp 361-364.
S. Dixon, W. Goebl and G. Widmer
Real Time Tracking and Visualisation of Musical Expression.
II International Conference on Music and Artificial Intelligence,
12 - 15 September 2002, Edinburgh, Scotland, pp 58-68.
S. Dixon and W. Goebl
Pinpointing the Beat: Tapping to Expressive Performances.
7th International Conference on Music Perception and Cognition (ICMPC7),
17 - 21 July 2002, Sydney, Australia, pp 617-620.
S. Dixon.
Automatic Extraction of Tempo and Beat from Expressive Performances.
Journal of New Music Research, 30 (1), 2001, pp 39-58.
- Download BeatRoot software (requires Java 1.5 or higher)
- Audio beat tracking examples
- MIDI beat tracking examples
S. Dixon.
An Interactive Beat Tracking and Visualisation System.
Proceedings of the International Computer Music Conference,
Havana, Cuba, 2001, pp 215-218.
- Download BeatRoot software (requires Java 1.5 or higher)
- Screen shots: audio beat tracking, audio beat tracking (showing spectrogram), MIDI beat tracking
E. Cambouropoulos, S. Dixon, W. Goebl and G. Widmer.
Human Preferences for Tempo Smoothness.
VII International Symposium on Systematic and Comparative Musicology,
III International Conference on Cognitive Musicology,
16 - 19 August 2001, Jyväskylä, Finland, pp 18-26.
W. Goebl and S. Dixon.
Analysis of Tempo Classes in Performances of Mozart Sonatas.
VII International Symposium on Systematic and Comparative Musicology,
III International Conference on Cognitive Musicology,
16 - 19 August 2001, Jyväskylä, Finland, pp 65-76.
S. Dixon.
An Empirical Comparison of Tempo Trackers.
8th Brazilian Symposium on Computer Music,
31 July - 3 August 2001, Fortaleza, Brazil, pp 832-840.
S. Dixon, W. Goebl and E. Cambouropoulos.
Beat Extraction from Expressive Musical Performances.
2001 meeting of the Society for Music Perception and Cognition (SMPC2001),
9 - 11 August 2001, Kingston, Ontario, Canada, pp 62-63.
Technical report OFAI-TR-2001-22
S. Dixon.
Learning to Detect Onsets of Acoustic Piano Tones.
MOSART Workshop on Current Research Directions in Computer Music,
15 - 17 November 2001, Barcelona, Spain.
Empirical Music Performance Research: OeFAI's Position.
MOSART Workshop on Current Research Directions in Computer Music, 15 - 17 November 2001, Barcelona, Spain.
S. Dixon.
A Lightweight Multi-Agent Musical Beat Tracking System.
PRICAI 2000: Proceedings of the
Pacific Rim International Conference on Artificial Intelligence,
Melbourne, Australia, 2000, pp 778-788.
S. Dixon and E. Cambouropoulos.
Beat Tracking with Musical Knowledge.
ECAI 2000: Proceedings of the 14th European
Conference on Artificial Intelligence,
ed. W.Horn, IOS Press, Amsterdam, 2000, pp 626-630.
S. Dixon.
Extraction of Musical Performance Parameters from Audio Data.
IEEE Pacific-Rim Conference on Multimedia,
Sydney, Australia, December 2000, pp 42-45.
S. Dixon.
On the Computer Recognition of Solo Piano Music.
Mikropolyphonie, 6, 2000.
An earlier version appeared in:
Interfaces: Australasian Computer Music Conference,
Brisbane, Australia, 2000, pp 31-37.
Sound examples: MP3 files (~80kB) and WAV files (~1MB):
- Mozart Piano Sonata K.332 (2nd movt): Original (MP3) (WAV) Transcribed (MP3) (WAV)
- Mozart Piano Sonata K.279 (2nd movt): Original (MP3) (WAV) Transcribed (MP3) (WAV)
- Mozart Piano Sonata K.281 (3rd movt): Original (MP3) (WAV) Transcribed (MP3) (WAV)
S. Dixon.
A Beat Tracking System for Audio Signals.
Proceedings of the Conference on Mathematical and Computational Methods
in Music,
Vienna, Austria, December 1999, pp 101-110.
Sound examples: the beat tracking system inserts a cowbell sample at
each calculated beat time.
MP3 files (80kB each):
1,
2,
3.
WAV files (880kB each):
1,
2,
3.
S. Dixon.
Beat Induction and Rhythm Recognition.
Proceedings of the Australian Joint Conference on Artificial Intelligence,
Perth, Australia, 1997, pp 311-320.
S. Dixon.
A Dynamic Modelling Approach to Music Recognition.
Proceedings of the International Computer Music Conference,
Hong Kong, 1996, pp 83-86.
S. Dixon.
Multiphonic Note Identification.
Australian Computer Science Communications, 18, 1,
Proceedings of the 19th Australasian Computer Science Conference,
Melbourne, Australia, 1996, pp 318-323.
S. Dixon, D. M. W. Powers.
The Characterisation, Separation and Transcription of
Complex Acoustic Signals.
Proceedings of the Sixth Australian International Conference on
Speech Science and Technology,
Adelaide, Australia, 1996, pp 73-78.
D. M. W. Powers, C. R. Clark, S. Dixon, D. L. Weber.
Cocktails and Brainwaves:
Experiments with Complex and Subliminal Auditory Stimuli.
Proceedings of the Australian and New Zealand Conference on
Intelligent Information Systems,
Adelaide, Australia, 1996, pp 68-71.
S. Dixon, W. R. Wobcke.
A Nonmonotonic Reasoning Belief Revision System.
Australian Computer Science Communications, 17, 1,
Proceedings of the 18th Australasian Computer Science Conference,
Adelaide, Australia, 1995, pp 151-160.
S. Dixon.
Belief Revision: A Computational Approach.
(PDF)
(PS)
Unpublished PhD thesis, Basser Department of Computer Science,
University of Sydney, Australia, 1994.
S. Dixon, W. R. Wobcke.
The implementation of a first-order logic AGM belief revision system.
(PS)
Proceedings of 5th IEEE International Conference on Tools with AI,
Boston MA, pages 40--47.
IEEE Computer Society Press, 1993.
S. Dixon, N. Y. Foo.
Connections between the ATMS and AGM belief revision.
Proceedings of 13th International Joint Conference on
Artificial Intelligence,
Chambery, France, pages 534-539.
Morgan Kaufmann, 1993.
S. Dixon.
A finite base belief revision system.
Australian Computer Science Communications, 15(1):445-451.
Proceedings of ACSC-16: 16th Australian Computer Science Conference,
Brisbane, Australia, January 1993.
Queensland University of Technology, Brisbane, Australia, 1993.
S. Dixon, N. Y. Foo.
Beyond foundational reasoning.
Proceedings of AI'92:
5th Australian Joint Conference on Artificial Intelligence,
Hobart, Australia, pages 247-252.
World Scientific, Singapore, 1992.
S. Dixon, N. Y. Foo.
Encoding the ATMS in AGM logic (revised).
Technical Report 441, Basser Department of Computer Science,
University of Sydney, Australia, 1992.
N. Y. Foo, S. Dixon, D. Kuo.
Systems Dynamics via Theory Revision.
Open Distributed Processing Workshop Proceedings,
University of Sydney, Australia, 1990.
R. Colomb, S. Dixon.
Database Implementation of an Artificial Intelligence Programming Paradigm.
Proceedings of AI'90:
4th Australian Joint Conference on Artificial Intelligence,
Perth, Australia, pp 462-473,
World Scientific Press, Singapore, 1990.
Research Staff & Students
Sungkyun Chang
(Research assistant) |
Music Performance Assessment and Feedback |
Yukun Li
(PhD student, co-supervised by Polina Proutskova) |
Segmenting and Analysing Pitch Contours across Singing Styles |
Mary Pilataki-Manika
(PhD student on the AIM CDT, co-supervised by Matthias Mauch, in collaboration with Apple Music) |
Polyphonic Music Transcription using Deep Learning |
Dave Foster
(PhD student on the AIM CDT) |
Modelling the Creative Process of Jazz Improvisation |
Yin-Jyun Luo
(PhD student on the AIM CDT, co-supervised by Sebastian Ewert, in collaboration with Spotify) |
Industry-scale Machine Listening for Music and Audio Data |
Xavier Riley
(PhD student on the AIM CDT) |
Transcribing the Jazz Ensemble: Applying Deep Learning Models to Jazz Transcription |
Huan Zhang
(PhD student on the AIM CDT) |
Computational Modelling of Expressive Piano Performance |
Drew Edwards
(PhD student on the AIM CDT, co-supervised by Akira Maezawa, in collaboration with Yamaha) |
Modeling Jazz Piano: Symbolic Music Generation via Large-Scale Automatic Transcription |
Carey Bunks
(PhD student on the AIM CDT, co-supervised by Bruno di Giorgi, in collaboration with Apple) |
Cover Song Identification for Jazz |
Tyler McIntosh
(PhD student on the AIM CDT, co-supervised by Nadine Kroher, in collaboration with DAACI) |
Performance Rendering for Music Generation Systems |
Ningzhi Wang
(PhD student on the AIM CDT, in collaboration with Spotify) |
Generative Models for Music Audio Representation and Understanding |
Nicolas Guo
(PhD student on the AIM CDT) |
Towards Comprehensive Automatic Guitar Transcription via a Multi-modal Approach |
Ivan Shanin | Modeling Melodic Jazz Improvisation |
Completed PhDs
Former Research Assistants, Students and Visitors
Ivan Shanin
(Research Assistant on the JazzDAP (Jazz Digital Archive Project) project (Oct 2022 - Aug 2023). |
New Directions in Digital Jazz Studies: Music Information Retrieval and AI Support for Jazz Scholarship in Digital Archives |
Keitaro Tanaka Visiting PhD student (Mar 2023 - Sep 2023) |
Disentangled Representation Learning |
Ken O'Hanlon Postdoctoral Research Assistant on Innovate UK Project in collaboration with Breathe Music Ltd (Nov 2019 - Jun 2021) |
Development of Next-Generation Music Recognition Algorithm for Content Monitoring |
Polina Proutskova (Postdoctoral RA, Dec 2017 - Sep 2019) |
Dig that Lick: Analysing large-scale data for melodic patterns in jazz performances |
Wenming Gui (Visiting Fellow, Oct 2018 - Oct 2019) |
Deep Learning and Singing Voice Detection |
Jordan Smith (Postdoctoral RA, Mar - May 2019) |
Dig that Lick: Analysing large-scale data for melodic patterns in jazz performances |
Eita Nakamura (Visiting Postdoctoral Fellow, Nov 2016 - Oct 2017) |
Note Value Recognition for Rhythm Transcription |
Lasse Vetter
(Research student, co-supervised with George Fazekas) |
Polytimbral Instrument Activity Detection using Deep Recurrent Neural Networks |
Peter Foster (Postdoctoral Research Assistant, Nov 2015 - Oct 2016) |
Automatic Transcription of Music (with DoReMIR Music Research AB) |
Bruno Di Giorgi (Visiting PhD Student, Apr - Jul 2016) |
Modelling the Complexity of Tonal Chord Sequences |
Matthias Mauch (Royal Academy of Engineering Research Fellow, 2012-2015) |
Software Systems for Computer-Aided Music Understanding |
Emilia Gomez (Visiting Researcher, Jun - Sep 2015) |
Intonation in Flamenco Singing |
Julien Osmalskyj (Visiting PhD Student, Feb - Jun 2015) |
Cover Song Detection |
Peter Foster (RA Nov 2014 - Aug 2015) |
Audio Data Exploration (with Audio Analytic Ltd) |
Siddharth Sigtia (RA Nov 2014 - Aug 2015) |
Audio Data Exploration (with Audio Analytic Ltd) and Smart Microphone v2.0 (with Audio Analytic Ltd) |
Gabrielle Medeot (Visiting Research Student, Nov 2014 - Apr 2015) |
Memory-enhanced RNN-NADE networks for modelling polyphonic musical sequences |
Christof Weiss (Visiting PhD Student, Feb, Nov 2014) |
A musicologically motivated approach to automatic classification of music data |
Lucas Thompson (Vacation bursary Jun - Aug 2013) |
Analysis of Intonation in Unaccompanied Singing |
Alo Allik (RA Oct 2012 - May 2013) |
A Shared Open Vocabulary for Audio Research and Retrieval |
Stephen Welburn (RA Dec 2011 - Feb 2013) |
Sustainable Management of Digital Music Research Data (2011-12) and Sound Data Management Training (2012-13) |
Joachim Ganseman (Visiting PhD Student, November 2011 - September 2012) |
Sinusoidal modeling based on score-to-audio alignment |
Magdalena Chudy (RA Oct 2011 - Mar 2013) |
MIReS: Roadmap for Music Information ReSearch |
Marco Fabiani (RA October 2011 - May 2012) |
Sustainable Management of Digital Music Research Data |
Adam Stark (RA Dec 2011 - Mar 2012) |
Musicology for the Masses |
Xue Wen (RA Dec 2011 - Mar 2012) |
Musicology for the Masses |
Dan Stowell (RA Sept 2010 - March 2012) |
Musicology for the Masses |
Mathieu Barthet (RA Sept 2010 - Nov 2011) |
Musicology for the Masses |
Ivan Damnjanovic (RA Oct-Nov 2011) |
Sustainable Management of Digital Music Research Data |
Asma Rafiq (Research student) |
Modeling Users' Intentions for the Enhancement of Music Recommendation Systems |
Barry Norton (RA April-July 2011) |
Linked Music Metadata |
Thomas Gängler (RA July 2011) |
Linked Music Metadata |
Cedric Mesnage (RA March-July 2011) |
Linked Music Metadata |
Kurt Jacobson (Research Assistant, 2010) |
Linked Music Metadata |
Downloads
Beat tracking software
BeatRoot (requires: Java, any platform)
BeatRoot is also available as a
VAMP plugin for
Sonic Visualiser.
Audio alignment software
MATCH (requires: Java, any platform)
MATCH is also available as a
VAMP plugin for
Sonic Visualiser. There is a tutorial
here
.
Slides from the ISMIR 2015 Tutorial: Why Singing is Interesting by Simon Dixon, Masataka Goto and Matthias Mauch.
Slides from the IMA 2012 Tutorial on Music Signal Processing, by Mark Plumbley and Simon Dixon.
Slides from the ISMIR 2006 Tutorial on Computational Rhythm Description, by Fabien Gouyon and Simon Dixon.
Teaching
This year I am teaching (with Emmanouil Benetos):
- ECS7006: Music Informatics (for the MSc in Sound and Music Computing and the PhD programme in Artificial Intelligence and Music). This module introduces students to state-of-the-art methods for the analysis of music data, with a focus on music audio. It presents in-depth studies of general approaches to the low-level analysis of audio signals, and follows these with specialised methods for the high-level analysis of music signals, including the extraction of information related to the rhythm, melody, harmony, form and instrumentation of recorded music. This is followed by an examination of the most important methods of extracting high-level musical content, sound source separation, and on analysing multimodal music data.
- Artificial Intelligence
- Music Analysis and Synthesis
- Software Tools for Engineers
- Introduction to Multimedia
- Knowledge Representation and Reasoning
- Computational Logic
- Digital Audio and Computer Music
- Data Structures and Algorithms
- Cognitive Science
- Computer Programming
Biography
I completed my BSc(Hons) (1986-1989) and PhD (1990-1994) degrees in Computer Science at the University of Sydney, and also learnt Classical Guitar at the NSW Conservatorium of Music, obtaining the AMusA (1987) and LMusA (1988). I was a lecturer in Computer Science at Flinders University of South Australia from 1994 until mid 1999, and a research scientist in the Intelligent Music Processing Group at the Austrian Research Institute for Artificial Intelligence from 1999 to 2006. In December 2006 I joined QMUL's Centre for Digital Music as a Lecturer (now Professor), where I am now Deputy Director and lead the Music Informatics theme. I am Director of the UKRI Centre for Doctoral Training in Artificial Intelligence and Music, and founding editor of the journal Transactions of the International Society for Music Information Retrieval. I was Programme Co-chair for the 8th International Conference on Music Information Retrieval (2007) and the 14th International Society for Music Information Retrieval Conference (2013), and General Co-chair for the Dagstuhl Seminar on Multimodal Music Processing (2011) and the 9th International Symposium on Computer Music Modeling and Retrieval (CMMR 2012): Music and Emotions. I was President of the International Society for Music Information Retrieval (2014-15); and am on the Editorial Board of the Journal of New Music Research, and the Advisory Board of the Music Information Retrieval Evaluation Exchange (MIREX).