Aarabi, Hadrien Foroughmand, and Geoffroy Peeters. 2018.
“Music Retiler: Using Nmf2d Source Separation for Audio Mosaicing.” In
Proceedings of the Audio Mostly 2018 on Sound in Immersion and Emotion, 27:1–7.
AM’18.
New York, NY, USA:
ACM.
https://doi.org/10.1145/3243274.3243299.
Alvarado, Pablo A., Mauricio A. Alvarez, and Dan Stowell. 2019.
“Sparse Gaussian Process Audio Source Separation Using Spectrum Priors in the Time-Domain.” In
ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 995–99.
https://doi.org/10.1109/ICASSP.2019.8683287.
Alvarado, Pablo A., and Dan Stowell. 2018.
“Efficient Learning of Harmonic Priors for Pitch Detection in Polyphonic Music.” November 16, 2018.
http://arxiv.org/abs/1705.07104.
Bertin, N., R. Badeau, and E. Vincent. 2010.
“Enforcing Harmonicity and Smoothness in Bayesian Non-Negative Matrix Factorization Applied to Polyphonic Music Transcription.” IEEE Transactions on Audio, Speech, and Language Processing 18 (3): 538–49.
https://doi.org/10.1109/TASL.2010.2041381.
Blaauw, Merlijn, and Jordi Bonada. 2017.
“A Neural Parametric Singing Synthesizer.” April 12, 2017.
http://arxiv.org/abs/1704.03809.
Blumensath, Thomas, and Mike Davies. 2006.
“Sparse and Shift-Invariant Representations of Music.” IEEE Transactions on Audio, Speech and Language Processing 14 (1): 50–57.
https://doi.org/10.1109/TSA.2005.860346.
Buch, Michael, Elio Quinton, and Bob L Sturm. 2017. “NichtnegativeMatrixFaktorisierungnutzendesKlangsynthesenSystem (NiMFKS): Extensions of NMF-Based Concatenative Sound Synthesis.” In Proceedings of the 20th International Conference on Digital Audio Effects, 7. Edinburgh.
Castro, Pablo de, and Tommaso Dorigo. 2019.
“INFERNO: Inference-Aware Neural Optimisation.” Computer Physics Communications 244 (November): 170–79.
https://doi.org/10.1016/j.cpc.2019.06.007.
Cichocki, A., R. Zdunek, and S. Amari. 2006.
“New Algorithms for Non-Negative Matrix Factorization in Applications to Blind Source Separation.” In
2006 IEEE International Conference on Acoustics Speech and Signal Processing Proceedings, 5:V–.
https://doi.org/10.1109/ICASSP.2006.1661352.
Driedger, Jonathan, Mathias Muller, and Sebastian Ewert. 2014.
“Improving Time-Scale Modification of Music Signals Using Harmonic-Percussive Separation.” IEEE Signal Processing Letters 21 (1, 1): 105–9.
https://doi.org/10.1109/LSP.2013.2294023.
Driedger, Jonathan, and Meinard Müller. 2016.
“A Review of Time-Scale Modification of Music Signals.” Applied Sciences 6 (2, 2): 57.
https://doi.org/10.3390/app6020057.
Driedger, Jonathan, Meinard Müller, and Sascha Disch. 2014.
“Extending Harmonic-Percussive Separation of Audio Signals.” In
ISMIR, 611–16.
http://www.terasoft.com.tw/conf/ismir2014/proceedings/T110_127_Paper.pdf.
Driedger, Jonathan, and Thomas Pratzlich. 2015.
“Let It Bee – Towards NMF-Inspired Audio Mosaicing.” In
Proceedings of ISMIR, 7.
Malaga.
http://ismir2015.uma.es/articles/13_Paper.pdf.
Elowsson, Anders, and Anders Friberg. 2017. “Long-Term Average Spectrum in Popular Music and Its Relation to the Level of the Percussion.” In Audio Engineering Society Convention 142, 13. Audio Engineering Society.
Févotte, Cédric, Nancy Bertin, and Jean-Louis Durrieu. 2008.
“Nonnegative Matrix Factorization with the Itakura-Saito Divergence: With Application to Music Analysis.” Neural Computation 21 (3): 793–830.
https://doi.org/10.1162/neco.2008.04-08-771.
Févotte, Cédric, and Jérôme Idier. 2011.
“Algorithms for Nonnegative Matrix Factorization with the β-Divergence.” Neural Computation 23 (9): 2421–56.
https://doi.org/10.1162/NECO_a_00168.
Fitzgerald, Derry. 2010.
“Harmonic/Percussive Separation Using Median Filtering.” http://arrow.dit.ie/argcon/67/.
FitzGerald, Derry, Antoine Liukus, Zafar Rafii, Bryan Pardo, and Laurent Daudet. 2013.
“Harmonic/Percussive Separation Using Kernel Additive Modelling.” In
Irish Signals & Systems Conference 2014 and 2014 China-Ireland International Conference on Information and Communications Technologies (ISSC 2014/CIICT 2014). 25th IET, 35–40.
IET.
http://ieeexplore.ieee.org/xpls/abs_all.jsp?arnumber=6912726.
Grais, Emad M., Dominic Ward, and Mark D. Plumbley. 2018.
“Raw Multi-Channel Audio Source Separation Using Multi-Resolution Convolutional Auto-Encoders.” March 1, 2018.
http://arxiv.org/abs/1803.00702.
Gribonval, R. 2003.
“Piecewise Linear Source Separation.” In
Proc. Soc. Photographic Instrumentation Eng., 5207:297–310.
San Diego, CA, USA.
https://doi.org/10.1117/12.504790.
Helén, M., and T. Virtanen. 2005.
“Separation of Drums from Polyphonic Music Using Non-Negative Matrix Factorization and Support Vector Machine.” In
Signal Processing Conference, 2005 13th European, 1–4.
http://www.cs.tut.fi/sgn/arg/music/tuomasv/EUSIPCO2005.pdf.
Hennequin, Romain, Anis Khlif, Felix Voituret, and Manuel Moussallam. 2019. “Spleeter: A Fast and State-of-the Art Music Source Separation Tool with Pre-Trained Models.” In, 2.
Hoffman, Matthew D, David M Blei, and Perry R Cook. 2010.
“Bayesian Nonparametric Matrix Factorization for Recorded Music.” In
International Conference on Machine Learning, 8.
http://soundlab.cs.princeton.edu/publications/2010_icml_gapnmf.pdf.
Hsieh, H., and J. Chien. 2011.
“Nonstationary and Temporally Correlated Source Separation Using Gaussian Process.” In
2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2120–23.
https://doi.org/10.1109/ICASSP.2011.5946745.
Jayaram, Vivek, and John Thickstun. 2020.
“Source Separation with Deep Generative Priors.” February 18, 2020.
http://arxiv.org/abs/2002.07942.
Klapuri, A., T. Virtanen, and T. Heittola. 2010.
“Sound Source Separation in Monaural Music Signals Using Excitation-Filter Model and Em Algorithm.” In
2010 IEEE International Conference on Acoustics, Speech and Signal Processing, 5510–13.
https://doi.org/10.1109/ICASSP.2010.5495216.
Lakatos, Stephen. 2000.
“A Common Perceptual Space for Harmonic and Percussive Timbres.” Perception & Psychophysics 62 (7): 1426–39.
https://doi.org/10.3758/BF03212144.
Laroche, Clément, Hélène Papadopoulos, Matthieu Kowalski, and Gaël Richard. 2017.
“Drum Extraction in Single Channel Audio Signals Using Multi-Layer Non Negative Matrix Factor Deconvolution.” In
ICASSP.
Nouvelle Orleans, United States.
https://hal.archives-ouvertes.fr/hal-01438851.
Leglaive, Simon, Roland Badeau, and Gaël Richard. 2017.
“Multichannel Audio Source Separation: Variational Inference of Time-Frequency Sources from Time-Domain Observations.” In
42nd International Conference on Acoustics, Speech and Signal Processing (ICASSP). Proc. 42nd
International Conference on
Acoustics,
Speech and
Signal Processing (
ICASSP).
La Nouvelle Orléans, LA, United States:
IEEE.
https://hal.archives-ouvertes.fr/hal-01416347.
Levin, David N. 2017.
“The Inner Structure of Time-Dependent Signals.” March 24, 2017.
http://arxiv.org/abs/1703.08596.
Liu, Yuzhou, Balaji Thoshkahna, Ali Milani, and Trausti Kristjansson. 2020.
“Voice and Accompaniment Separation in Music Using Self-Attention Convolutional Neural Network,” March.
https://arxiv.org/abs/2003.08954v1.
Liutkus, Antoine, Roland Badeau, and Gäel Richard. 2011.
“Gaussian Processes for Underdetermined Source Separation.” IEEE Transactions on Signal Processing 59 (7, 7): 3155–67.
https://doi.org/10.1109/TSP.2011.2119315.
Liutkus, Antoine, Zafar Rafii, Bryan Pardo, Derry Fitzgerald, and Laurent Daudet. 2014.
“Kernel Spectrogram Models for Source Separation.” In, 6–10.
IEEE.
https://doi.org/10.1109/HSCMA.2014.6843240.
Ma, Ning, Phil Green, Jon Barker, and André Coy. 2007.
“Exploiting Correlogram Structure for Robust Speech Recognition with Multiple Speech Sources.” Speech Communication 49 (12, 12): 874–91.
https://doi.org/10.1016/j.specom.2007.05.003.
Miron, Marius, Julio J. Carabias-Orti, Juan J. Bosch, Gó, Emilia Mez, and Jordi Janer. 2016.
“Score-Informed Source Separation for Multichannel Orchestral Recordings.” Journal of Electrical and Computer Engineering 2016 (December): e8363507.
https://doi.org/10.1155/2016/8363507.
Ono, N., K. Miyamoto, J. Le Roux, H. Kameoka, and S. Sagayama. 2008. “Separation of a Monaural Audio Signal into Harmonic/Percussive Components by Complementary Diffusion on Spectrogram.” In Signal Processing Conference, 2008 16th European, 1–4.
Ono, Nobutaka, Kenichi Miyamoto, Hirokazu Kameoka, and Shigeki Sagayama. 2008.
“A Real-Time Equalizer of Harmonic and Percussive Components in Music Signals.” In
ISMIR, 139–44.
https://books.google.ch/books?hl=en&lr=&id=OHp3sRnZD-oC&oi=fnd&pg=PA139&dq=harmonic+percussive+separation&ots=oEOOoIiBb6&sig=2J8Fr8t5CxlJWBz8NRpIUOW3nfc.
Ó Nuanáin, Cárthach, Sergi Jordà Puig, and Perfecto Herrera Boyer. 2016.
“An Interactive Software Instrument for Real-Time Rhythmic Concatenative Synthesis.” http://repositori.upf.edu/handle/10230/32951.
Park, S., and S. Choi. 2008.
“Gaussian Processes for Source Separation.” In
2008 IEEE International Conference on Acoustics, Speech and Signal Processing, 1909–12.
https://doi.org/10.1109/ICASSP.2008.4518008.
Pham, Dinh-Tuan, and Jean-François Cardoso. 2001.
“Blind Separation of Instantaneous Mixtures of Nonstationary Sources.” IEEE Transactions on Signal Processing 49 (9): 1837–48.
https://doi.org/10.1109/78.942614.
Prétet, Laure, Romain Hennequin, Jimena Royo-Letelier, and Andrea Vaglio. 2019.
“Singing Voice Separation: A Study on Training Data.” In
ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 506–10.
https://doi.org/10.1109/ICASSP.2019.8683555.
Qian, Kaizhi, Yang Zhang, Shiyu Chang, David Cox, and Mark Hasegawa-Johnson. 2020.
“Unsupervised Speech Decomposition via Triple Information Bottleneck.” August 11, 2020.
http://arxiv.org/abs/2004.11284.
Routtenberg, Tirza, and Joseph Tabrikian. 2010.
“Blind MIMO-AR System Identification and Source Separation with Finite-Alphabet.” IEEE Transactions on Signal Processing 58 (3): 990–1000.
https://doi.org/10.1109/TSP.2009.2036043.
Särelä, Jaakko, and Harri Valpola. 2005.
“Denoising Source Separation.” Journal of Machine Learning Research 6: 233–72.
http://www.jmlr.org/papers/v6/sarela05a.html.
Schlüter, J., and S. Böck. 2014.
“Improved Musical Onset Detection with Convolutional Neural Networks.” In
2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 6979–83.
https://doi.org/10.1109/ICASSP.2014.6854953.
Smaragdis, Paris. 2004.
“Non-Negative Matrix Factor Deconvolution; Extraction of Multiple Sound Sources from Monophonic Inputs.” In
Independent Component Analysis and Blind Signal Separation, edited by Carlos G. Puntonet and Alberto Prieto, 494–99. Lecture
Notes in
Computer Science.
Granada, Spain:
Springer Berlin Heidelberg.
https://doi.org/10.1007/978-3-540-30110-3_63.
Smaragdis, Paris, Cedric Fevotte, Gautham J. Mysore, Nasser Mohammadiha, and Matthew Hoffman. 2014.
“Static and Dynamic Source Separation Using Nonnegative Factorizations: A Unified View.” IEEE Signal Processing Magazine 31 (3): 66–75.
https://doi.org/10.1109/MSP.2013.2297715.
Smith, Jordan B L, Yuta Kawasaki, and Masataka Goto. 2019. “Unmixer: An Interface for Extracting and Remixing Loops.” In, 8.
Sprechmann, Pablo, Joan Bruna, and Yann LeCun. 2014.
“Audio Source Separation with Discriminative Scattering Networks.” December 22, 2014.
http://arxiv.org/abs/1412.7022.
Stöter, Fabian-Robert, Stefan Uhlich, Antoine Liutkus, and Yuki Mitsufuji. 2019.
“Open-Unmix - A Reference Implementation for Music Source Separation.” Journal of Open Source Software 4 (41): 1667.
https://doi.org/10.21105/joss.01667.
Tachibana, H., H. Kameoka, N. Ono, and S. Sagayama. 2012.
“Comparative Evaluations of Various Harmonic/Percussive Sound Separation Algorithms Based on Anisotropic Continuity of Spectrogram.” In
2012 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 465–68.
https://doi.org/10.1109/ICASSP.2012.6287917.
Tachibana, Hideyuki, Nobutaka Ono, and Shigeki Sagayama. 2014.
“Singing Voice Enhancement in Monaural Music Signals Based on Two-Stage Harmonic/Percussive Sound Separation on Multiple Resolution Spectrograms.” Audio, Speech, and Language Processing, IEEE/ACM Transactions on 22 (1): 228–37.
https://doi.org/10.1109/TASLP.2013.2287052.
Tenenbaum, J. B., and W. T. Freeman. 2000.
“Separating Style and Content with Bilinear Models.” Neural Computation 12 (6): 1247–83.
https://doi.org/10.1162/089976600300015349.
Turner, Richard E., and Maneesh Sahani. 2014.
“Time-Frequency Analysis as Probabilistic Inference.” IEEE Transactions on Signal Processing 62 (23): 6171–83.
https://doi.org/10.1109/TSP.2014.2362100.
Tzinis, Efthymios, Zhepei Wang, and Paris Smaragdis. 2020. “Sudo Rm -Rf: Efficient Networks for Universal Audio Source Separation.” In, 6.
Venkataramani, Shrikant, and Paris Smaragdis. 2017.
“End to End Source Separation with Adaptive Front-Ends.” May 6, 2017.
http://arxiv.org/abs/1705.02514.
Venkataramani, Shrikant, Y. Cem Subakan, and Paris Smaragdis. 2017.
“Neural Network Alternatives to Convolutive Audio Models for Source Separation.” September 20, 2017.
http://arxiv.org/abs/1709.07908.
Vincent, E., N. Bertin, and R. Badeau. 2008.
“Harmonic and Inharmonic Nonnegative Matrix Factorization for Polyphonic Pitch Transcription.” In
2008 IEEE International Conference on Acoustics, Speech and Signal Processing, 109–12.
https://doi.org/10.1109/ICASSP.2008.4517558.
Virtanen, T. 2007.
“Monaural Sound Source Separation by Nonnegative Matrix Factorization With Temporal Continuity and Sparseness Criteria.” IEEE Transactions on Audio, Speech, and Language Processing 15 (3): 1066–74.
https://doi.org/10.1109/TASL.2006.885253.
Virtanen, Tuomas. 2006.
“Unsupervised Learning Methods for Source Separation in Monaural Music Signals.” In
Signal Processing Methods for Music Transcription, 267–96.
Springer.
https://www.cs.tut.fi/sgn/arg/music/tuomasv/unsupervised_virtanen.pdf.
Yoshii, Kazuyoshi. 2013. “Beyond NMF: Time-Domain Audio Source Separation Without Phase Reconstruction,” 6.
———. 2018.
“Correlated Tensor Factorization for Audio Source Separation.” In
2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 731–35.
https://doi.org/10.1109/ICASSP.2018.8461434.
Yoshii, Kazuyoshi, Katsutoshi Itoyama, and Masataka Goto. 2016.
“Student’s T Nonnegative Matrix Factorization and Positive Semidefinite Tensor Factorization for Single-Channel Audio Source Separation.” In
2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 51–55.
https://doi.org/10.1109/ICASSP.2016.7471635.