Presentation is loading. Please wait.

Presentation is loading. Please wait.

前向人工神经网络敏感性研究 曾晓勤 河海大学计算机及信息工程学院 2003年10月.

Similar presentations


Presentation on theme: "前向人工神经网络敏感性研究 曾晓勤 河海大学计算机及信息工程学院 2003年10月."— Presentation transcript:

1 前向人工神经网络敏感性研究 曾晓勤 河海大学计算机及信息工程学院 2003年10月

2 一. 引言 1. 前向神经网络(FNN)介绍 ●神经元 – 离散型:自适应线性元(Adaline)
– 连续型:感知机(Perceptron) ●神经网络 – 离散型:多层自适应线性网(Madaline) – 连续型:多层感知机(BP网或MLP)

3 2. 研究提出 ●问题 – 硬件精度对权的影响 – 环境噪音对输入的影响 ●动机 – 参数的扰动对网络会产生怎样影响?
2. 研究提出 ●问题 – 硬件精度对权的影响 – 环境噪音对输入的影响 ●动机 – 参数的扰动对网络会产生怎样影响? – 如何衡量网络输出偏差的大小?

4 3. 研究内容 ●建立网络输出与网络参数扰动之间的关系 ●分析该关系,揭示网络的行为规律 ●量化网络输出偏差

5 4. 研究意义 ●度量网络性能,如容错和泛化能力 ●研究其它网络课题的基础,如网络结构的 裁剪和参数的挑选等
●指导网络设计,增强网络抗干扰能力 ●度量网络性能,如容错和泛化能力 ●研究其它网络课题的基础,如网络结构的 裁剪和参数的挑选等

6 二.研究纵览(典型方法和文献) Madaline的敏感性 ●n维几何模型(超球面)
M. Stevenson, R. Winter, and B. Widrow, “Sensitivity of Feedforward Neural Networks to Weight Errors,” IEEE Trans. on Neural, Networks, vol. 1, no. 1, 1990. ●统计模型(方差) S. W. Piché, “The Selection of Weight Accuracies for Madalines,” IEEE Trans. on Neural Networks, vol. 6, no. 2, 1995.

7 2. MLP的敏感性 ●分析方法(偏微分) S. Hashem, “Sensitivity Analysis for Feed- Forward Artificial Neural Networks with Differentiable Activation Functions”, Proc. of IJCNN, vol. 1, 1992. ●统计方法(标准差) J. Y. Choi & C. H. Choi, “Sensitivity Ana lysis of Multilayer Perceptron with Differ entiable Activation Functions,” IEEE Trans. on Neural Networks, vol. 3, no. 1, 1992.

8 3. 敏感性的应用 ●输入属性筛选 J. M. Zurada, A. Malinowski, S. Usui, “Perturbation Method for Deleting Redundant Inputs of Perceptron Networks”, Neurocomputing, vol. 14, ●网络结构裁减 A. P. Engelbrecht, “A New Pruning Heuristic Based on Variance Analysis of Sensitivity Information”, IEEE Trans. on Neural Networks, vol. 12, no. 6, 2001.

9 ●容错和泛化问题 J.L. Bernier et al, “A Quantitive Study of Fault Tolerance, Noise Immunity and Generalization Ability of MLPs,” Neural Computation, vol. 12, 2000.

10 三. 研究方法 1. 自底向上方法 2. 概率统计方法 3. n-维几何模型 ●单个神经元 ●整个网络 ●概率(离散型) ●均值(连续型)
●超矩形的顶点(离散型) ●超矩形体(连续型)

11 四.已获成果(代表性论文) ●敏感性分析: “Sensitivity Analysis of Multilayer Percep tron to Input and Weight Perturbations,” IEEE Trans. on Neural Networks, vol. 12, no.6, pp , Nov

12 ●敏感性量化: “A Quantified Sensitivity Measure for Multi- layer Perceptron to Input Perturbation,” Neural Computation, vol. 15, no. 1, pp , Jan

13 ●隐层节点的裁剪(敏感性应用): ●输入属性重要性的判定(敏感性应用):
“Hidden Neuron Pruning for Multilayer Perceptrons Using Sensitivity Measure,” Proc. of IEEE ICMLC2002, pp , Nov ●输入属性重要性的判定(敏感性应用): “Determining the Relevance of Input Features for Multilayer Perceptrons,” Proc. of IEEE SMC2003, Oct

14 五. 未来工作 ●进一步完善已有的结果,使之更加实用 – 放松限制条件 – 扩大分析范围 – 精确量化计算
●进一步应用所得的结果,解决实际问题 ●探索新方法,研究新类型的网络

15 结束 谢谢各位!

16

17

18

19

20

21 Effects of input & weight deviations on neurons’ sensitivity
Sensitivity increases with input and weigh deviations, but the increase has an upper bound.

22 Effects of input dimension on neurons’ sensitivity
There exists an optimal value for the dimension of input, which yields the highest sensitivity value.

23 Effects of input & weight deviations on MLPs’ sensitivity
Sensitivity of an MLP increases with the input and weight deviations.

24 Effects of the number of neurons in a layer
Sensitivity of MLPs: { n | 1n 10 } to the dimension of input.

25 Sensitivity of MLPs: { 2-n-2-1 | 1n 10 } to the number of neurons in the 1st layer.

26 Sensitivity of MLPs: { 2-2-n-1 | 1n 10 } to the number of neurons in the 2nd layer .
There exists an optimal value for the number of neurons in a layer, which yields the highest sensitivity value. The nearer a layer to the output layer is, The more effect the number of neurons in the layer has.

27 Effects of the number of layers
Sensitivity of MLPs:{2-1,2-2-1,.., } to the number of layers. Sensitivity decreases with the number increasing, and the decrease almost levels off when the number becomes large.

28 Sensitivity of the neurons with 2-dimensional input

29 Sensitivity of the neurons with 3-dimensional input

30 Sensitivity of the neurons with 4-dimensional input

31 Sensitivity of the neurons with 5-dimensional input

32 Sensitivity of the MLPs: 2-2-1, 2-3-1,2-2-2-1

33 Simulation 1 (Function Approximation)
Implement an MLP to approximate the function: where Implementation considerations The MLP architecture is restricted to 2-n-1. The convergence condition is MES-goal=0.01&Epoch105. The lowest trainable number of hidden neurons is n=5. The pruning processes start with MLPs of and stop at an architecture of The relevant data used by and resulted from the pruning process are listed in Table 1 and Table 2.

34 TABLE 1. Data for 3 MLPs with 5 hidden neurons to realize the function
2-5-1 Epoch MSE (training) MSE (testing) Trained weights and bias MSE-(goal=0.01 & epoch<=100000) Sensitivity Relevance 1 30586 [ ] [ ] [ ] [ ] [ ] [ ] bias=0 0.1733 0.0289 0.0184 0.3253 0.0158 2 65209 [ ] [ ] [ ] [ ] [ ] [ ] bias=0 0.0261 0.0072 0.0222 0.1888 0.3385 3 26094 [ ] [ ] [ ] [ ] [ ] [ ] bias=0 0.1194 0.1242 0.1791 0.4777 0.0243

35 TABLE 2. Data for the 3 pruned MLPs with 4 hidden neurons to realize the function
2-4-1 Epoch MSE (training) MSE (testing) Retrained weights and bias (goal=0.01 & epoch<=100000) Sensitivity Relevance 1 (Obtained by removing the 5th neuron from the MLP of 2-5-1) 2251 [ ] [ ] [ ] [ ] [ ] bias=4.2349 0.1541 0.0274 0.0193 0.3818 2 (Obtained by removing the 2nd neuron from the MLP of 2-5-1) 1945 [ ] [ ] [ ] [ ] [ ] bias= 0.0247 0.0230 0.1662 0.3914 3 13253 [ ] [ ] [ ] [ ] [ ] bias= 0.0259 0.0206 0.3737 0.1708

36 Simulation 2 (Classification)
Implement an MLP to solve the XOR problem: 1 Implementation considerations The MLP architecture is restricted to 2-n-1. The convergence condition is MES-goal=0.1&Epoch105. The pruning processes start with MLPs of and stop at an architecture of The relevant data used by and resulted from the pruning process are listed in Table 3 and Table 4.

37 TABLE 3. Data for 3 MLPs with 5 hidden neurons to realize the function
2-5-1 Epoch MSE (training) MSE (testing) Trained weights and bias (goal=0.1 & epoch<=100000) Sensitivity Relevance 1 44518 [ ] [ ] [ ] [ ] [ ] [ ] bias=0 0.6671 1.5725 0.8845 0.5348 2.1632 2 51098 [ ] [ ] [ ] [ ] [ ] [ ] bias=0 0.8997 0.3869 0.5480 0.6812 2.9828 3 33631 [ ] [ ] [ ] [ ] [ ] [ ] bias=0 1.4413 1.1773 0.7642 1.6469 0.9421

38 TABLE 4. Data for the 3 pruned MLPs with 4 hidden neurons to realize the function
2-4-1 Epoch MSE (training) MSE (testing) Retrained weights and bias (goal=0.1 & epoch<=100000) Sensitivity Relevance 1 (Obtained by removing the 4th neuron from the MLP of 2-5-1) 22611 [ ] [ ] [ ] [ ] [ ] bias=5.5570 0.9742 1.4699 1.0164 2.2865 2 (Obtained by removing the 2nd neuron from the MLP of 2-5-1) 14457 [ ] [ ] [ ] [ ] [ ] bias= 1.0768 0.9334 0.7517 3.1235 3 (Obtained by removing the 3rd neuron from the MLP of 2-5-1) 17501 [ ] [ ] [ ] [ ] [ ] bias=1.7474 1.7222 1.4662 1.7059 1.4967


Download ppt "前向人工神经网络敏感性研究 曾晓勤 河海大学计算机及信息工程学院 2003年10月."

Similar presentations


Ads by Google