TY - GEN
T1 - Accelerating generalized iterative scaling using componentwise extrapolations for on-line conditional random fields
AU - Yang, Hee Deok
AU - Lee, Seong Whan
PY - 2010
Y1 - 2010
N2 - In this paper, a simple and globally convergent method based on penalized generalized iterative scaling (GIS) with staggered Aitken acceleration is proposed to efficiently estimate the parameters for an on-line conditional random field (CRF). The staggered Aitken acceleration method, which alternates between an acceleration step and a non-acceleration step, provides numerical stability and computational simplicity in analyzing the incompleteness of data. The proposed method is based on stochastic gradient descent (SGD) and it has the following advantages: (1) it can approximate parameters close to the empirical optimum in a single pass through the training examples; (2) it can reduce the computing time by eliminating computation of the inverse of the objective function's Hessian matrix with staggered Aitken acceleration. We show the convergence of penalized GIS based on the staggered Aitken acceleration method, compare its speed of convergence with that of other stochastic optimization methods, and also illustrate experimental results with public data sets.
AB - In this paper, a simple and globally convergent method based on penalized generalized iterative scaling (GIS) with staggered Aitken acceleration is proposed to efficiently estimate the parameters for an on-line conditional random field (CRF). The staggered Aitken acceleration method, which alternates between an acceleration step and a non-acceleration step, provides numerical stability and computational simplicity in analyzing the incompleteness of data. The proposed method is based on stochastic gradient descent (SGD) and it has the following advantages: (1) it can approximate parameters close to the empirical optimum in a single pass through the training examples; (2) it can reduce the computing time by eliminating computation of the inverse of the objective function's Hessian matrix with staggered Aitken acceleration. We show the convergence of penalized GIS based on the staggered Aitken acceleration method, compare its speed of convergence with that of other stochastic optimization methods, and also illustrate experimental results with public data sets.
KW - Aitken acceleration
KW - Conditional random field
KW - Incremental learning
KW - On-line learning
UR - http://www.scopus.com/inward/record.url?scp=78149341745&partnerID=8YFLogxK
U2 - 10.1109/ICMLC.2010.5580707
DO - 10.1109/ICMLC.2010.5580707
M3 - Conference contribution
AN - SCOPUS:78149341745
SN - 9781424465262
T3 - 2010 International Conference on Machine Learning and Cybernetics, ICMLC 2010
SP - 3169
EP - 3173
BT - 2010 International Conference on Machine Learning and Cybernetics, ICMLC 2010
T2 - 2010 International Conference on Machine Learning and Cybernetics, ICMLC 2010
Y2 - 11 July 2010 through 14 July 2010
ER -