文件名称:1
-
所属分类:
- 标签属性:
- 上传时间:2013-05-18
-
文件大小:4.28kb
-
已下载:0次
-
提 供 者:
-
相关连接:无下载说明:别用迅雷下载,失败请重下,重下不扣分!
介绍说明--下载内容来自于网络,使用问题请自行百度
bp_innerloop.m
Inner loop of the backpropagtion learning algorithm.
One hidden layer. Uses tanh as the transfer function.
Uses the following global variables for input and/or output:
Inputs1 - input patterns
Desired - desired output patterns
LearnRate - learning rate parameter
Momentum - momentum parameter
DerivIncr - increment to the derivative of the transfer function
(Fahlman s trick typical value 0.2)
Weights1 - first weight layer (updated by this routine)
Weights2 - second weight layer (updated by this routine)
deltaW1 - initialize to 0 before first call
deltaW2 - initialize to 0 before first call
TSS - total sum-squared error (set by this routine)
Recurrent state- bp_innerloop.m
Inner loop of the backpropagtion learning algorithm.
One hidden layer. Uses tanh as the transfer function.
Uses the following global variables for input and/or output:
Inputs1 - input patterns
Desired - desired output patterns
LearnRate - learning rate parameter
Momentum - momentum parameter
DerivIncr - increment to the derivative of the transfer function
(Fahlman s trick typical value 0.2)
Weights1 - first weight layer (updated by this routine)
Weights2 - second weight layer (updated by this routine)
deltaW1 - initialize to 0 before first call
deltaW2 - initialize to 0 before first call
TSS - total sum-squared error (set by this routine)
Recurrent state
Inner loop of the backpropagtion learning algorithm.
One hidden layer. Uses tanh as the transfer function.
Uses the following global variables for input and/or output:
Inputs1 - input patterns
Desired - desired output patterns
LearnRate - learning rate parameter
Momentum - momentum parameter
DerivIncr - increment to the derivative of the transfer function
(Fahlman s trick typical value 0.2)
Weights1 - first weight layer (updated by this routine)
Weights2 - second weight layer (updated by this routine)
deltaW1 - initialize to 0 before first call
deltaW2 - initialize to 0 before first call
TSS - total sum-squared error (set by this routine)
Recurrent state- bp_innerloop.m
Inner loop of the backpropagtion learning algorithm.
One hidden layer. Uses tanh as the transfer function.
Uses the following global variables for input and/or output:
Inputs1 - input patterns
Desired - desired output patterns
LearnRate - learning rate parameter
Momentum - momentum parameter
DerivIncr - increment to the derivative of the transfer function
(Fahlman s trick typical value 0.2)
Weights1 - first weight layer (updated by this routine)
Weights2 - second weight layer (updated by this routine)
deltaW1 - initialize to 0 before first call
deltaW2 - initialize to 0 before first call
TSS - total sum-squared error (set by this routine)
Recurrent state
(系统自动生成,下载前可以参看下载内容)
下载文件列表
elman_bp_innerloop.m
elman_bp_innerloop.m~
elman_test.m
elman_test.m~
elman_train.m
elman_train.m~
forwprop.m
elman_bp_innerloop.m~
elman_test.m
elman_test.m~
elman_train.m
elman_train.m~
forwprop.m
本网站为编程资源及源代码搜集、介绍的搜索网站,版权归原作者所有! 粤ICP备11031372号
1999-2046 搜珍网 All Rights Reserved.