An investigation of CNN-LSTM music recognition algorithm in ethnic vocal technique singing Online publication date: Mon, 09-Sep-2024
by Fang Dong
International Journal of Computational Science and Engineering (IJCSE), Vol. 27, No. 5, 2024
Abstract: A HPSS separation algorithm considering time and frequency features is proposed to address the issue of poor performance in music style recognition and classification. A CNN network structure was designed and the influence of different parameters in the network structure on recognition rate was studied. A deep hash learning method is proposed to address the issues of weak feature expression ability and high feature dimension in existing CNN, which is combined with LSTM networks to integrate temporal dimension information. The results showed that compared to other models such as GRU+LSTM, the double-layer LSTM model used in the study had higher recognition results, with a size of over 75%. This indicates that combining feature learning with hash encoding learning can achieve higher accuracy. Therefore, this model is more suitable for music style recognition technology, which helps in music information retrieval and improves the classification accuracy of music recognition.
Existing subscribers:
Go to Inderscience Online Journals to access the Full Text of this article.
If you are not a subscriber and you just want to read the full contents of this article, buy online access here.Complimentary Subscribers, Editors or Members of the Editorial Board of the International Journal of Computational Science and Engineering (IJCSE):
Login with your Inderscience username and password:
Want to subscribe?
A subscription gives you complete access to all articles in the current issue, as well as to all articles in the previous three years (where applicable). See our Orders page to subscribe.
If you still need assistance, please email subs@inderscience.com