Dense deep stochastic configuration network with hybrid training mechanism
by Weidong Zou; Yuanqing Xia; Weipeng Cao
International Journal of Computing Science and Mathematics (IJCSM), Vol. 15, No. 3, 2022

Abstract: Thanks to the supervised parameter generation strategy and non-iterative training mechanism, deep stochastic configuration network (DSCN) has achieved very efficient modelling efficiency in scenarios with relatively small problem complexity. However, the increasing number of hidden layers and the amount of training data have issued a challenge to the implementation of DSCN. To solve this problem, we propose a Dense DSCN with a Hybrid Training mechanism (HT-DDSCN), which extends the network structure of the DSCN to a dense connection type and combines three typical optimisation techniques and one universal control strategy to optimise the calculation process of the output weights. Extensive experiments on four benchmark regression problems show that HT-DDSCN can significantly improve the generalisation ability and the stability of DSCN.

Online publication date: Mon, 08-Aug-2022

The full text of this article is only available to individual subscribers or to users at subscribing institutions.

 
Existing subscribers:
Go to Inderscience Online Journals to access the Full Text of this article.

Pay per view:
If you are not a subscriber and you just want to read the full contents of this article, buy online access here.

Complimentary Subscribers, Editors or Members of the Editorial Board of the International Journal of Computing Science and Mathematics (IJCSM):
Login with your Inderscience username and password:

    Username:        Password:         

Forgotten your password?


Want to subscribe?
A subscription gives you complete access to all articles in the current issue, as well as to all articles in the previous three years (where applicable). See our Orders page to subscribe.

If you still need assistance, please email subs@inderscience.com