联系方式

  • QQ:99515681
  • 邮箱:99515681@qq.com
  • 工作时间:8:00-23:00
  • 微信:codehelp

您当前位置:首页 >> Java程序Java程序

日期:2021-06-08 10:24

IS71083A Financial Data Modelling
Coursework Assignment
Implementing an Incremental (online) Training Algorithm
for Density Neural Networks
Design and implement an incremental backpropagation algorithm for training density neural networks
(DNN). DNN networks compute not only the mean of the target distribution, but also the variance with a
second output node. That is why you have to develop the backprop algorithm to train all weights leading
to each of the two output nodes in DNN given in the figure below. Start the training process using
plausible randomly generated weights. After the design of the training algorithm, apply it to model and
forecast a time series, generated using the Mackey-Glass equation (available in the lecture handouts).
Use more lagged inputs and hidden units if necessary to achieve better predictions.
sum
Design a working prototype of the incremental backpropagation for DNN density networks in Matlab.
The prototype development should include data structures for the input-to-hidden and hidden-to-output
connections, and loops for the forward and backward pass.
References:
D.A.Nix and E.Weigend (1994). Estimating the mean and variance of the target probability distribution,
In: Proc. 1994 IEEE Int. Conf. on Neural Networks (ICNN'94), Orlando, FL, pp. 55-60.
Nikolay Nikolaev and Hitoshi Iba (2006).
"Adaptive Learning of Polynomial Networks: Genetic Programming, Backpropagation and Bayesian Methods",
Springer, New York, pp. 256 - 262.

版权所有:留学生编程辅导网 2021,All Rights Reserved 联系方式:QQ:99515681 电子信箱:99515681@qq.com
免责声明:本站部分内容从网络整理而来,只供参考!如有版权问题可联系本站删除。