#### Title

A Batch Variable Learning Rate Gradient Descent Algorithm with the Smoothing L1/2 Regularization for Takagi-Sugeno Models

#### Document Type

Article

#### Publication Title

IEEE Access

#### Abstract

A batch variable learning rate gradient descent algorithm is proposed to efficiently train a neuro-fuzzy network of zero-order Takagi-Sugeno inference systems. By using the advantages of regularization, the smoothing L_{1/2} regularization is utilized to find more appropriate sparse network. Combining the second-order information of the smoothing error function, a variable learning rate is chosen along the steep descent direction, which avoids line search procedure and may reduce the cost of computation. In order to appropriately adjust the Lipschitz constant of the smoothing error function in the learning rate, a new scheme is proposed by introducing a hyper-parameter. Also the article applies the modified secant equation for estimating the Lipschitz constant, which makes the algorithm greatly reduce the oscillating phenomenon and improve the robustness. Under appropriate assumptions, a convergent result of the proposed algorithm is also given. Simulation results for two identification and classification problems show that the proposed algorithm has better numerical performance and promotes the sparsity capability of the network, compared with the common batch gradient descent algorithm and a variable learning rate gradient-based algorithm.

#### First Page

100185

#### Last Page

100193

#### DOI

10.1109/ACCESS.2020.2997867

#### Publication Date

1-1-2020

#### Recommended Citation

Lu, Yunlong; Li, Wenyu; and Wang, Hongwei, "A Batch Variable Learning Rate Gradient Descent Algorithm with the Smoothing L1/2 Regularization for Takagi-Sugeno Models" (2020). *Mathematics & Physics Faculty Publications*. 36.

https://rio.tamiu.edu/math_physics_facpubs/36