A vast amount of music information available on social media, web pages, online forums, and digital libraries, etc., is represented in natural language. Making sense of this information is challenging due to the unstructured nature of the data. Music and language data also share many similarities such as its sequential nature. With machine learning based natural language processing (NLP) technology, we attempt to tackle the rich complexity of human languages in order to extract useful insights for tasks such as music information retrieval (MIR) and audio AI. In this talk, I discuss the application of NLP in music information technology in the light of the latest transformations brought about by deep learning, enabling machines to make sense of the world through multimodal music and sound data. I conclude the talk by identifying emerging areas of interesting challenges and parallel trends at the intersection of these two exciting fields.
Shuo Zhang is a Senior Machine Learning Engineer (CED Applied Research) at Bose Corporation. His work encompasses natural language processing (NLP) and music information retrieval (MIR). Prior to Bose, he was a Researcher/Collaborator at the Music Technology Group (MTG), Department of Information and Communication Technologies (DTIC), Universitat Pompeu Fabra (UPF), Barcelona, Spain. Dr. Zhang serves on the program committees of conferences and workshops in communities such as ACL/NAACL/WWW/MASC-SLL. In 2016 he co-taught a tutorial on the application of NLP in MIR at the International Society of Music Information Retrieval (ISMIR) Conference at Columbia University in New York.