Development of a fast relu activation function algorithm for deep learning problems


Download 1.34 Mb.
bet1/6
Sana05.05.2023
Hajmi1.34 Mb.
#1431413
  1   2   3   4   5   6
Bog'liq
Article ReLU




DEVELOPMENT OF A FAST RELU ACTIVATION FUNCTION ALGORITHM FOR DEEP LEARNING PROBLEMS


Karimberdiyev Jahongir
Tashkent university of information technologies, Tashkent, Uzbekistan


Abstract

Artificial neural networks are inspired by the biological neurons within the human body which activate under certain circumstances resulting in a related action performed by the body in response. Artificial neural nets consist of various layers of interconnected artificial neurons powered by activation functions which help in switching them ON/OFF. Like traditional machine learning algorithms, here too, there are certain values that neural nets learn in the training phase.


Keywords: TensorFlow; Convolutional Neural Networks; Depth Learning; Activation Function, ReLU

Introduction


Briefly, each neuron receives a multiplied version of inputs and random weights which is then added with static bias value (unique to each neuron layer), this is then passed to an appropriate activation function which decides the final value to be given out of the neuron. There are various activation functions available as per the nature of input values. Once the output is generated from the final neural net layer, loss function (input vs output)is calculated and backpropagation is performed where the weights are adjusted to make the loss minimum. Finding optimal values of weights is what the overall operation is focusing around.


* Corresponding author. Tel.: +998946560444.
E-mail address: jahongirkarimberdiyev618@gmail.com



What is activation function?
As mentioned above, activation functions give out the final value given out from a neuron, but what is activation function and why do we need it? 
So, an activation function is basically just a simple function that transforms its inputs into outputs that have a certain range. There are various types of activation functions that perform this task in a different manner, For example, the sigmoid activation function takes input and maps the resulting values in between 0 to 1.
One of the reasons that this function is added into an artificial neural network in order to help the network learn complex patterns in the data. These functions introduce nonlinear real-world properties to artificial neural networks. Basically, in a simple neural network, x is defined as inputs, w weights, and we pass f (x) that is the value passed to the output of the network. This will then be the final output or the input of another layer.
If the activation function is not applied, the output signal becomes a simple linear function. A neural network without activation function will act as a linear regression with limited learning power. But we also want our neural network to learn non-linear states as we give it complex real-world information such as image, video, text, and sound.

Download 1.34 Mb.

Do'stlaringiz bilan baham:
  1   2   3   4   5   6




Ma'lumotlar bazasi mualliflik huquqi bilan himoyalangan ©fayllar.org 2024
ma'muriyatiga murojaat qiling