Abstract: Artificial Neural Networks (ANNs) play an important role
in machine learning nowadays. A typical ANN is traditionally trained using stochastic gradient descent (SGD) with backpropagation (BP). However, it is unlikely that a real biological neural network is trained similarly. Neuroscience theories, such as Hebbian Theory could inspire adaption from the traditional training method SGD to make ANNs more biologically plausible. Two mathematical descriptions of Hebbian theory will be suggested based on the limitations of the mathematical framework of Hebbian theory: A competitive Hebbian learning rule and an imply Hebbian learning rule. Finally, this research will propose a method, called the Hebbian Plasticity Term (HPT) method, that incorporates the mathematical description of Hebbian theory to modify the traditional training method SGD. Therefore two variants of the HPT method are proposed: HPT-Competitive and HPT-Imply. The influence of the hebbian plasticity term ϕ on SGD shows more biologically realistic plasticity of a synaptic connection between neurons in the ANN at the cost of performance.
Publication Year: 2020
Publication Date: 2020-01-01
Language: en
Type: dissertation
Access and Citation
Cited By Count: 1
AI Researcher Chatbot
Get quick answers to your questions about the article from our AI researcher chatbot