Let’s Jam: A Machine Learning Framework For Sample Pairing and Filtration Using Melodic and Tonal Movement Classification
DOI:
https://doi.org/10.58445/rars.3665Keywords:
Artificial Intelligence, Machine-Learning, Music Theory, MusicAbstract
The search engines of music sampling services are often inadequate when addressing samples that tonally and sonically pair well with each other. Using search filters and other functions within these services, one is able to find samples that are within the same key as the desired sample, and with the same sonic qualities. Current technology is either hyperspecific with the sample selections, or focuses too much on the instrumentation of the sample alone. Using a machine learning model (MLM), we plan to create a sample-sorting system that categorizes samples, not on the instrumentation or sonic similarity, but on the tonal qualities, like the chord structure the sample would fall on when assigned chords based on common practice chord assignments (I, II, III, and so on). This would allow the model to adequately categorize the samples by their movement (by breaking up the given sample into chord progressions such as I-IV-V). This delivers further classification advantages (such as filtering samples by what chords they would fall under) and allows sample libraries to provide sample pairings that are not sonically identical, but still fall under the same movement. Samples with similar chord structures are far more likely to sound harmonious without sounding exactly the same. This would lead to more diversity and efficiency with sample pairing and selection, while also allowing the user to exercise more tonal specificity when searching, improving workflow.
References
References
Avise, J. C. (2013). From perception to pleasure: Music and its neural substrates. National Library of Medicine. Retrieved December 27, 2025, from https://pmc.ncbi.nlm.nih.gov/articles/PMC3690607/
Bruin, J. (2021, August 22). Logistic Regression Analysis | Stata Annotated Output. OARC Stats. Retrieved February 10, 2026, from https://stats.oarc.ucla.edu/stata/output/logistic-regression-analysis/
Berg, L., King, B., Koenig, J., & McRoberts, R. L. (2022). Musician occupational and financial stress and mental health burden. Psychology of Music, 50(6), 1801-1815. https://doi.org/10.1177/03057356211064642 (Original work published 2022)
Bhattacharjee, Utpal & Mannala, Jyoti. (2019). An Experimental Analysis of Speech Features for Tone Speech Recognition. International Journal of Innovative Technology and Exploring Engineering. 2. 4355-4360. 10.35940/ijitee.B7748.129219.
Costa, L., & Los Angeles Film School. (2025). Music Sampling Explained: Creative Techniques and Copyright Laws Every Producer Should Know. The Los Angeles Film School. Retrieved December 27, 2025, from https://www.lafilm.edu/blog/music-sampling-explained/
Dambrin, D., & Schaack, D. (n.d.). Fruity Parametric EQ 2 - Effect Plugin. FL Studio. Retrieved December 28, 2025, from https://www.image-line.com/fl-studio-learning/fl-studio-online-manual/html/plugins/Fruity%20Parametric%20EQ%202.htm
de Haas, W. B., Robine, M., & Hanna, P. (2011). Comparing Approaches to the Similarity of Musical Chord Sequences. Hal Open Science. Retrieved December 28, 2025, from https://hal.science/hal-01006469v1/document
Eck, A. (2024). How Music Resonates in the Brain | Harvard Medicine Magazine. Harvard Medicine Magazine. Retrieved February 10, 2026, from https://magazine.hms.harvard.edu/articles/how-music-resonates-brain
Ellis, D. P.W., & LabROSA. (2007). CLASSIFYING MUSIC AUDIO WITH TIMBRAL AND CHROMA FEATURES. Columbia University. Retrieved December 28, 2025, from https://www.ee.columbia.edu/~dpwe/pubs/Ellis07-timbrechroma.pdf
FL Studio. (n.d.). Piano roll Basics. FL Studio. Retrieved December 28, 2025, from https://www.image-line.com/fl-studio-learning/fl-studio-online-manual/html/pianoroll.htm
Freeman J, Simoncelli EP. Metamers of the ventral stream. Nat Neurosci. 2011 Aug 14;14(9):1195-201. doi: 10.1038/nn.2889. PMID: 21841776; PMCID: PMC3164938.
Gagneja, N. (2020, October 8). Understanding phase and polarity in sound recording. Ux Collective. Retrieved December 17, 2025, from https://uxdesign.cc/understanding-phase-and-polarity-in-audio-101aa8cac2eb
Grey. (2025). Major Scale Chord Function: How Chords Behave. Hub Guitar. Retrieved December 27, 2025, from https://hubguitar.com/music-theory/chord-function
Hilsdorf, M. (2023, September 20). AI Music Source Separation: How it Works and Why It Is So Hard. Medium. Retrieved December 17, 2025, from https://medium.com/data-science/ai-music-source-separation-how-it-works-and-why-it-is-so-hard-187852e54752
Huss, P. J. (1983, 12). Vocal Pitch Range and Habitual Pitch Level: The Study of Normal College Age Speakers. Western Michigan University. Retrieved February 9, 2026, from https://scholarworks.wmich.edu/cgi/viewcontent.cgi?article=2616&context=masters_theses
Lenssen, Nathan & Needell, Deanna. (2014). An Introduction to Fourier Analysis with Applications to Music. Journal of Humanistic Mathematics. 4. 72-91. 10.5642/jhummath.201401.05.
Martin, T. (2025, June 5). Detailed Frequency Ranges of Instruments and Vocals. The Absolute Sound. Retrieved February 9, 2026, from https://www.theabsolutesound.com/freqchart/main_display.htm
Mencke, I., Omigie, D., & Wald-Fuhrmann, M. (2019, January 7). Atonal Music: Can Uncertainty Lead to Pleasure? Frontiers. Retrieved February 10, 2026, from https://www.frontiersin.org/journals/neuroscience/articles/10.3389/fnins.2018.00979/full
Moazeni, S. (2024, October 17). Logistic Regression: A Method for Classification | by S. Moazeni, PhD. Medium. Retrieved December 28, 2025, from https://medium.com/@AILearningHub/logistic-regression-a-method-for-classification-649ca3a68608
Music Theory Academy. (2025). 5 basic rules of Chord Progressions. Music Theory Academy. Retrieved December 28, 2025, from https://www.musictheoryacademy.com/understanding-music/chord-progressions/
Noble, J. (2024). What is Hierarchical Clustering? IBM. Retrieved December 28, 2025, from https://www.ibm.com/think/topics/hierarchical-clustering
Oudre, L., Grenier, Y., & IEEE. (2011). Chord Recognition by Fitting Rescaled Chroma Vectors to Chord Templates. Institut de Recherche en Informatique de Toulouse. Retrieved December 26, 2025, from https://www.irit.fr/~Cedric.Fevotte/publications/journals/ieee_asl_deterchord.pdf
Splice Records. (2022, December 1). Use Similar Sounds to find samples on Splice - Blog. Splice. Retrieved December 27, 2025, from https://splice.com/blog/introducing-similar-sounds/
Stryker, C. (2024). What Is Unsupervised Learning? IBM. Retrieved December 28, 2025, from https://www.ibm.com/think/topics/unsupervised-learning
Downloads
Posted
Categories
License
Copyright (c) 2026 Hritik Patel, Abhinav

This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.