1ST-ORDER AND 2ND-ORDER METHODS FOR LEARNING - BETWEEN STEEPEST DESCENT AND NEWTON METHOD

被引:813
作者
BATTITI, R [1 ]
机构
[1] UNIV TRENTO,DIPARTIMENTO MATEMAT,I-38050 TRENT,ITALY
关键词
D O I
10.1162/neco.1992.4.2.141
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
On-line first-order backpropagation is sufficiently fast and effective for many large-scale classification problems but for very high precision mappings, batch processing may be the method of choice. This paper reviews first- and second-order optimization methods for learning in feedforward neural networks. The viewpoint is that of optimization: many methods can be cast in the language of optimization techniques, allowing the transfer to neural nets of detailed results about computational complexity and safety procedures to ensure convergence and to avoid numerical problems. The review is not intended to deliver detailed prescriptions for the most appropriate methods in specific applications, but to illustrate the main characteristics of the different methods and their mutual relations.
引用
收藏
页码:141 / 166
页数:26
相关论文
共 53 条
[1]  
ALLRED LG, 1990, P INT JOINT C NEURAL, V1, P721
[2]  
[Anonymous], P INT NEUR NETW C PA
[3]  
BARNARD E, 1988, CSE89014 OREG GRAD C
[4]  
BATTITI R, 1991, PARALLEL COMPUTING P
[5]  
BATTITI R, 1990, P IFIP WORKSH SILICO
[6]  
Battiti R., 1989, COMPLEX SYSTEMS, V3, P331
[7]  
BECKER S, 1989, 1988 P CONN MOD SUMM
[8]  
BENGIO Y, 1989, P GNCB CNR SCH TRENT
[9]  
Bingham John A. C., 1988, THEORY PRACTICE MODE
[10]  
Broomhead D. S., 1988, Complex Systems, V2, P321