publications

publications by categories in reversed chronological order. generated by jekyll-scholar.

2025

  1. CrossMuSim: A cross-modal framework for music similarity retrieval with LLM-powered text description sourcing and mining
    Tristan Tsoi, Jiajun Deng, Yaolong Ju, Benno Weck, Holger Kirchhoff, and Simon Lui
    May 2025
    arXiv:2503.23128 [cs] Summary: This paper introduces a dual-source data acquisition approach combining online scraping and LLM-based prompting, where carefully designed prompts leverage LLMs’ comprehensive music knowledge to generate contextually rich descriptions.

2024

  1. End-to-End Automatic Singing Skill Evaluation Using Cross-Attention and Data Augmentation for Solo Singing and Singing With Accompaniment
    Yaolong Ju, Chun Yat Wu, Betty Cortiñas Lorenzo, Jing Yang, Jiajun Deng, Fan Fan, and Simon Lui
    In Proceedings of the 25th International Society for Music Information Retrieval Conference, May 2024
  2. Multi-View Midivae: Fusing Track- and Bar-View Representations for Long Multi-Track Symbolic Music Generation
    Zhiwei Lin, Jun Chen, Boshi Tang, Binzhu Sha, Jing Yang, Yaolong Ju, Fan Fan, Shiyin Kang, Zhiyong Wu, and Helen Meng
    In ICASSP 2024 - 2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Apr 2024
    Summary: Object and subjective experimental results demonstrate that, compared to the baseline, Multi-view MidiVAE exhibits significant improvements in terms of modeling long multi-track symbolic music.
  3. Cycle Frequency-Harmonic-Time Transformer for Note-Level Singing Voice Transcription
    Yulun Wu, Yaolong Ju, Simon Lui, Jing Yang, Fan Fan, and Xuhao Du
    In 2024 IEEE International Conference on Multimedia and Expo (ICME), Jul 2024
    Summary: A novel 3D Cycle Frequency-Harmonic-Time Transformer (CFT) is proposed to explicitly capture the harmonic series of singing voices, where a tokenization scheme is defined that captures harmonics across multiple octaves, then the harmonic features are aggregated into the frequency-harmonic-time representations via a cyclic architecture.
  4. Efficient adapter tuning for joint singing voice beat and downbeat tracking with self-supervised learning features
    Jiajun Deng, Yaolong Ju, Jing Yang, Simon Lui, and Xunying Liu
    In Proceedings of the 25th International Society for Music Information Retrieval Conference, Nov 2024
    Summary: A novel temporal convolutional network-based beat-tracking approach featuring self-supervised learning representations and adapter tuning is proposed to track the beat and downbeat of singing voices jointly.

2023

  1. Improving Automatic Singing Skill Evaluation with Timbral Features, Attention, and Singing Voice Separation
    Yaolong Ju, Chunyang Xu, Yichen Guo, Jinhu Li, and Simon Lui
    In 2023 IEEE International Conference on Multimedia and Expo (ICME), Jul 2023
    Summary: This paper proposes a more general ASSE model which applies to both solo singing and singing with accompaniment, and employs an existing singing voice separation tool for accompaniment removal and compares ASSE models trained with and without accompaniment.

2022

  1. VocEmb4SVS: Improving Singing Voice Separation with Vocal Embeddings
    Chenyi Li, Yi Li, Xuhao Du, Yaolong Ju, Shichao Hu, and Zhiyong Wu
    In 2022 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC), Nov 2022
    Summary: VocEmb4SVS is proposed, an SVS framework to utilize vocal embeddings of the singer as auxiliary knowledge for SVS conditioning and achieves state-of-the-art performance on the MUSDB18 dataset.
  2. AnimeTAB: A new guitar tablature dataset of anime and game music
    Yuecheng Zhou, Yaolong Ju, and Lingyun Xie
    Oct 2022
    arXiv:2210.03027 [cs] Summary: This paper presents AnimeTAB, a fingerstyle Guitar tablature dataset in MusicXML format, which provides more high-quality guitar tablature for both researchers and guitar players and an accompanying analysis toolkit, TABprocessor, is included to further facilitate its use.

2021

  1. Addressing ambiguity in supervised machine learning: A case study on automatic chord labelling
    Yaolong Ju
    McGill University, Oct 2021

2020

  1. Automatic Chord Labelling: A Figured Bass Approach
    Yaolong Ju, Sylvain Margot, Cory McKay, and Ichiro Fujinaga
    In Proceedings of the 7th International Conference on Digital Libraries for Musicology, Oct 2020
    Summary: This paper proposes a series of four rule-based algorithms that automatically generate chord labels for homorhythmic Baroque chorales based on both figured bass annotations and the musical surface, which are applied to the existing Bach Chorales Figured Bass dataset.
  2. Automatic Figured Bass Annotation Using the New Bach Chorales Figured Bass Dataset
    Yaolong Ju, Sylvain Margot, Cory McKay, Luke Dahn, and Ichiro Fujinaga
    In Proceedings of the 21th International Society for Music Information Retrieval Conference, Oct 2020
  3. Figured Bass Encodings for Bach Chorales in Various Symbolic Formats: A Case Study
    Yaolong Ju, Sylvain Margot, Cory McKay, and Ichiro Fujinaga
    In Proceedings of the Music Encoding Conference, Oct 2020
  4. Data Quality Matters: Iterative Corrections on a Corpus of Mendelssohn String Quartets and Implications for MIR Analysis
    Jacob Degroot-Maggetti, Timothy Reuse, Laurent Feisthauer, Samuel Howes, Yaolong Ju, Suzaka Kokubu, Sylvain Margot, Néstor Nápoles López, and Finn Upham
    In International Society for Music Information Retrieval Conference (ISMIR 2020), Oct 2020

2019

  1. An Interactive Workflow for Generating Chord Labels for Homorhythmic Music in Symbolic Formats
    Yaolong Ju, Samuel Howes, Cory McKay, Nathaniel Condit-Schultz, Jorge Calvo-Zaragoza, and Ichiro Fujinaga
    In Proceedings of the 20th International Society for Music Information Retrieval Conference, Oct 2019

2018

  1. A Flexible Approach to Automated Harmonic Analysis: Multiple Annotations of Chorales by Bach and Prætorius
    Nathaniel Condit-Schultz, Yaolong Ju, and Ichiro Fujinaga
    In Proceedings of the 19th International Society of Music Information Retrieval Conference, Oct 2018

2017

  1. Non-chord Tone Identification Using Deep Neural Networks
    Yaolong Ju, Nathaniel Condit-Schultz, Claire Arthur, and Ichiro Fujinaga
    In Proceedings of the 4th International Workshop on Digital Libraries for Musicology - DLfM ’17, Oct 2017
    Summary: The results suggest that DNNs offer an innovative and promising approach to tackling the problem of non-chord tone identification, as well as harmonic analysis.

2012

  1. K-means initial clustering center optimal algorithm based on Kruskal
    Lan Huang, Shixian Du, Yu Zhang, Yaolong Ju, and Zhuo Li
    J. Inf. Comput. Sci, Oct 2012