[1]
|
Mesaros, A. and Virtanen, T. (2010) Recognition of Phonemes and Words in Singing. 2010 IEEE International Confer-ence on Acoustics, Speech and Signal Processing, Dallas, 14-19 March 2010, 2146-2149.
https://doi.org/10.1109/ICASSP.2010.5495585
|
[2]
|
Fujihara, H., Goto, M., Ogata, J. and Okuno, H.G. (2011) Lyric Synchronizer: Automatic Synchronization System between Musical Audio Signals and Lyrics. IEEE Journal of Se-lected Topics in Signal Processing, 5, 1252-1261.
https://doi.org/10.1109/JSTSP.2011.2159577
|
[3]
|
Sharma, B., Gupta, C., Li, H. and Wang, Y. (2019) Automatic Lyrics-to-Audio Alignment on Polyphonic Music Using Singing-Adapted Acoustic Models. 2019 IEEE International Conference on Acoustics, Speech and Signal Processing, Brighton, 12-17 May 2019, 396-400. https://doi.org/10.1109/ICASSP.2019.8682582
|
[4]
|
Gillet, O. and Richard, G. (2008) Transcription and Separation of Drum Signals from Polyphonic Music. The IEEE/ACM Transactions on Audio, Speech, and Language Processing, 3, 529-540.
https://doi.org/10.1109/TASL.2007.914120
|
[5]
|
Manilow, E., Seetharaman, P. and Pardo, B. (2020) Simultaneous Separation and Transcription of Mixtures with Multiple Polyphonic and Percussive Instruments. IEEE International Conference on Acoustics, Speech and Signal Processing, Barcelona, 4-8 May 2020, 771-775. https://doi.org/10.1109/ICASSP40776.2020.9054340
|
[6]
|
Gómez, J.S., Abeßer, J. and Cano, E. (2018) Jazz Solo Instrument Classification with Convolutional Neural Networks, Source Separation, and Transfer Learning. Proceedings of the 19th International Society for Music Information Retrieval Conference, ISMIR, Paris, 23-27 September 2018, 577-584.
|
[7]
|
Liu, J.-Y., Chen, Y.-H., Yeh, Y.-C. and Yang, Y.-H. (2019) Score and Lyrics-Free Singing Voice Gen-eration.
|
[8]
|
Jansson, A., Humphrey, E.J., Montecchio, N., Bittner, R.M., Kumar, A. and Weyde, T. (2017) Singing Voice Separation with Deep U-Net Convolutional Networks. 18th International Society for Music Information Retrieval Conference, Suzhou, 23-27 October 2017, 745-751.
|
[9]
|
Takahashi, N. and Mitsufuji, Y. (2017) Multi-Scale Mul-ti-Band DenseNets for Audio Source Separation. IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA), New Paltz, 15-18 October 2017, 261-265. https://doi.org/10.1109/WASPAA.2017.8169987
|
[10]
|
Takahashi, N., Goswami, N. and Mitsufuji, Y. (2018) Mmdenselstm: An Efficient Combination of Convolutional and Recurrent Neural Networks for Audio Source Separation. 2018 16th International Workshop on Acoustic Signal Enhancement (IWAENC), Tokyo, 17-20 September 2018, 106-110. https://doi.org/10.1109/IWAENC.2018.8521383
|
[11]
|
Lee, J.H., Choi, H.-S. and Lee, K. (2019) Audio Query-Based Music Source Separation. Proceedings of the 20th International Society for Music Information Retrieval Conference, ISMIR, Delft, 4-8 November 2019, 878-885.
|
[12]
|
Liu, J.-Y. and Yang, Y.-H. (2019) Dilated Convolution with Dilated GRU for Music Source Separation. Proceedings International Joint Conference on Artificial Intelligence (IJCAI), Macao, 10-16 August 2019, 4718-4724.
|
[13]
|
Luo, Y. and Mesgarani, N. (2019) Conv-TasNet: Surpassing Ideal Time-Frequency Magnitude Masking for Speech Separation. The IEEE/ACM Transactions on Audio, Speech, and Language Processing, 27, 1256-1266.
https://doi.org/10.1109/TASLP.2019.2915167
|
[14]
|
Madry, A., Makelov, A., Schmidt, L., et al. (2017) Towards Deep Learning Models Resistant to Adversarial Attacks.
|
[15]
|
Moosavi-Dezfooli, S.M., Fawzi, A. and Frossard, P. (2016) DeepFool: A Simple and Accurate Method to Fool Deep Neural Networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, 27-30 June 2016, 2574-2582. https://doi.org/10.1109/CVPR.2016.282
|
[16]
|
Goodfellow, I.J., Shlens, J. and Szegedy, C. (2014) Explaining and Harnessing Adversarial Examples.
|
[17]
|
Papernot, N., McDaniel, P., Jha, S., et al. (2016) The Limitations of Deep Learning in Adversarial Settings. 2016 IEEE European Symposium on Security and Privacy, Saarbruecken, 21-24 March 2016, 372-387.
https://doi.org/10.1109/EuroSP.2016.36
|
[18]
|
Szegedy, C., Zaremba, W., Sutskever, I., Bruna, J., Erhan, D., Good-fellow, I. and Fergus, R. (2014) Intriguing Properties of Neural Networks. https://arxiv.org/abs/1312.6199
|
[19]
|
Mimilakis, S.I., Drossos, K. and Schuller, G. (2020) Unsupervised Interpret-able Representation Learning for Singing Voice Separation. EUSIPCO 2020, Amsterdam, 24-28 August 2020, 1412-1416.
|
[20]
|
Rafii, Z., Liutkus, A., Stöter, F.-R., Mimilakis, S.I. and Bittner, R. (2017) MUSDB18—A Corpus for Music Separation. https://hal.inria.fr/hal-02190845
|
[21]
|
Vincent, E., Gribonval, R. and Févotte, C. (2006) Perfor-mance Measurement in Blind Audio Source Separation. IEEE Transactions on Audio, Speech, and Language Pro-cessing, 14, 1462-1469. https://doi.org/10.1109/TSA.2005.858005
|