We deal with two sorts of parameters in machine learning: machine learnable parameters and hyper-parameters.

**Machine learnable parameters –**The parameters that the algorithms learn/estimate on their own during training for a particular dataset**Hyper-parameters**are variables that machine learning engineers or data scientists provide precise values to regulate how algorithms learn and modify the model’s performance.

The learning rate, denoted by the symbol α, is a hyper-parameter used to govern the pace at which an algorithm updates or learns the values of a parameter estimate. In other words,** the learning rate regulates the weights of our ****neural network**** concerning the loss gradient**>. It indicates how often the neural network refreshes the notions it has learned.

From examples in the training dataset, a neural network learns or approximates a function to optimally map inputs to outputs.

The rate of learning or speed at which the model learns is controlled by the hyperparameter. It regulates the amount of allocated error with which the model’s weights are updated each time they are updated, such as at the end of each batch of training instances.

The model will learn to best estimate the function given available resources – the number of layers and nodes per layer in a particular number of training epochs -passes through the training data if the learning rate is perfectly calibrated.

A desirable learning rate is low enough for the network to converge on something useful while yet being high enough to train in a reasonable length of time.

Smaller learning rates necessitate more training epochs because of the fewer changes. On the other hand, larger learning rates result in faster changes.

Moreover, larger learning rates frequently result in a suboptimal final set of weights.

An analytical method cannot be used to calculate the weights of a neural network. Instead, the weights must be discovered using stochastic gradient descent, an empirical optimization approach. In simpler terms, **the stochastic gradient descent algorithm is used to train deep ****learning rate ****neural networks**.

- Stochastic gradient descent is an optimization technique that uses instances from the training dataset to estimate the error gradient for the current state of the model and then uses backpropagation to update the model’s weights.

As a result, we should avoid using a learning rate that is either too high or too low. However, we must set up the model so that a decent enough set of weights is determined on average to approximate the mapping issue as represented by the training dataset.

It allows the training algorithm to **keep track of the model’s performance and automatically alter the learning rate** to achieve optimum results.

The learning rate increases or decreases in this method depending on the cost function’s gradient value.

- The learning rate will be reduced when the gradient value is higher, and larger when the gradient value is lower.

As a result, learning slows down and speeds up at steeper and shallower regions of the cost function curve, respectively.

The most basic model of this reduces the learning rate once the model’s performance reaches a plateau. The model accomplishes this by reducing the learning rate by a factor of two, or by an order of magnitude. If the performance does not improve, the learning rate might be increased again.

**Adaptive learning rates****frequently beat fixed****AI learning rates**in neural networks.

An adaptive learning rate in machine learning is commonly utilized when using stochastic gradient descent to build deep neural nets.

There are, however, various sorts of learning rate approaches:

**Decaying Learning Rate –**The learning rate drops as the number of epochs/iterations increases in this learning rate technique.**Scheduled Drop Learning rate –**The learning rate is lowered by a specified proportion at a specified frequency in the drop learning rate method, as opposed to the decay technique, where the learning rate declines repetitively.**Cycling learning rate –**The learning rate cyclically changes between a base rate and a maximum rate in this methodology. At a constant frequency, the learning rate varies in a triangular pattern between the maximum and base rates.**The Gradient Descent Method –**is a well-known optimization approach for estimating model parameters in machine learning. The value of each parameter is originally assumed or assigned random values when training a model. The cost function is generated using the initial values, and the parameter estimations are improved over time so that the cost function eventually assumes a minimum value.