摘要
转自:爱可可-爱生活
论文《A New Backpropagation Algorithm without Gradient Descent》摘要:
The backpropagation algorithm, which had been originally introduced in the 1970s, is the workhorse of learning in neural networks. This backpropagation algorithm makes use of the famous machine learning algorithm known as Gradient Descent, which is a first-order iterative optimization algorithm for finding the minimum of a function. To find a local minimum of a function using gradient descent, one takes steps proportional to the negative of the gradient (or of the approximate gradient) of the function at the current point. In this paper, we develop an alternative to the backpropagation without the use of the Gradient Descent Algorithm, but instead we are going to devise a new algorithm to find the error in the weights and biases of an artificial neuron using Moore-Penrose Pseudo Inverse. The numerical studies and the experiments performed on various datasets are used to verify the working of this alternative algorithm.
链接:
https://arxiv.org/abs/1802.00027
原文链接:
https://m.weibo.cn/1402400261/4205024235742492
领取专属 10元无门槛券
私享最新 技术干货