*Last updated: 26th Nov, 2023*

In this post, you will learn about **gradient descent algorithm** and its importance in training machine learning models. For a data scientist, it is of utmost importance to get a good grasp on the concepts of gradient descent algorithm as it is widely used for optimizing / minimizing the **objective function** / **loss function** related to various machine learning models such as regression, neural network etc. in terms of learning optimal weights / parameters. This algorithm is essential because it underpins many machine learning models, enabling them to learn from data by optimizing their performance. By understanding gradient descent, one gains insight into how algorithms can be trained to make predictions, recognize patterns, and make decisions, making it a cornerstone concept in the field of AI and data science.

## Introduction to Gradient Descent Algorithm

The gradient descent algorithm is an **optimization technique** used to **minimize a function,** commonly referred to as the **objective function** in machine learning. This objective function, often termed as the **cost or loss function**, is a measure of the error, typically the **squared difference between actual values and predictions**. In machine learning, gradient descent aims to find the most optimal values of parameters or weights that minimize this loss function.

The term “gradient descent” is derived from two key concepts: “**gradient**” and “**descent**.”

**Gradient**: In mathematics and physics, the gradient of a function at any point is a**multi-dimensional vector that points in the direction of the steepest increase of that function**. The magnitude of the gradient vector at a point corresponds to the rate of increase in the function’s value. In the context of a loss function in machine learning, the gradient indicates the direction in which the function is rising most rapidly.**Descent**: The word “descent” refers to the method’s objective of**moving downwards to find the minimum of the function**. Since the gradient points towards the greatest increase, the opposite direction, which the algorithm moves in, will lead to the steepest decrease. By iteratively moving in the direction opposite to the gradient (hence “descending”), the algorithm seeks to find the point where the function attains its minimum value.

Thus, “gradient descent” is aptly named as it describes the process of descending (minimizing) a function by moving against (opposite to) its gradient. The following plot can be used to understand the gradient descent algorithm.

The plot visualizes the concept of gradient descent on a simple quadratic function *f*(*x*)=*x*2. The red dots represent the steps taken by the gradient descent algorithm starting from an initial point (here, *x*=9) and moving towards the minimum of the function at *x*=0.

Each step is determined by the gradient (slope) of the function at that point, and the algorithm iteratively moves in the direction that reduces the function’s value (i.e., descending). The red dashed line connects these steps, illustrating the path taken by the algorithm as it seeks the function’s minimum. This exemplifies how gradient descent navigates the function’s landscape to find the point of lowest value.

## How to calculate Gradient Descent?

Calculating gradient descent involves several steps aimed at iteratively finding the minimum of a function. Here’s an expanded explanation with examples for better understanding:

**Calculate the Gradient**: The gradient is the derivative of the function with respect to its variables. In a simple one-dimensional case, if you have a function*f*(*x*), you calculate its derivative*f*′(*x*). For instance, if $f(x) = x^2$, then*f*′(*x*)=2*x*. In multi-dimensional scenarios, such as functions with multiple parameters (e.g.,*f*(*x*,*y*)), you calculate the partial derivatives with respect to each parameter.**Calculate Descent Value**: The descent value is computed by multiplying the gradient by the learning rate (a small, positive number often less than 1) and -1. The learning rate determines the size of the steps the algorithm takes towards the minimum. For example, if the gradient at a point is 4 and the learning rate is 0.1, the descent value would be −0.1×4=−0.4.**Update the Parameters**: The parameters of the function are updated by subtracting the descent value from their current values. This step is repeated iteratively. Continuing with the example above, if the current value of a parameter is 5, it would be updated to 5−(−0.4)=5.4. This updating moves the parameter towards the function’s minimum. The following is the representation of updation of parameters:

In case of multiple parameters, the value of different parameters would need to be updated as given below:**Iterative Process**: The process of calculating the gradient, descent value, and updating parameters is repeated until the changes become negligibly small or for a set number of iterations. This iterative process ensures that the algorithm moves towards the function’s minimum. The following is the representation using formula:**Example in a Real-World Scenario**: Consider a machine learning model where you’re trying to minimize the error between predicted and actual values (the loss function). The parameters could be the weights in a linear regression model predicting house prices. The gradient descent algorithm would iteratively adjust these weights to minimize the difference (loss) between the predicted and actual house prices.**Visualization**: Think of this process as a ball rolling down a hill. The ball represents the parameters, and the hill represents the function to be minimized. The ball rolls in the direction that takes it to the lowest point of the hill, which represents the minimum of the function.

## Different Types of Gradient Descent Algorithms

Gradient descent algorithms could be implemented in the following two different ways:

### Batch Gradient Descent

In batch gradient descent, the update of model parameters or weights is calculated using the entire training dataset at once. This means that for each step in the training process, the algorithm calculates the gradient of the cost function for the entire dataset.

This algorithm is often used when the training dataset is relatively small and can fit into memory comfortably. It’s also suitable for less complex models where the cost of computing the gradient for the whole dataset is not prohibitively high. For example, consider a linear regression model where you’re predicting housing prices based on features like size and location. If your dataset is small, you could use batch gradient descent to train your model, updating the weights based on the error calculated from the entire dataset.

### Stochastic Gradient Descent (SGD)

Stochastic gradient descent updates the model’s parameters using only one training example or a small batch of examples at a time. This means the gradient and hence the parameter update is calculated and applied after each example or small batch, making the process more incremental.

SGD is particularly useful when dealing with large datasets that cannot fit into memory. It’s also favored when training complex models, as it can handle the high computational cost more efficiently. Additionally, the stochastic nature of SGD can help the model escape local minima, potentially leading to better solutions. For example, in a deep learning model for image classification with a massive dataset, using SGD would be practical. Here, the model parameters are updated incrementally as each image (or a small batch of images) is processed, enabling the model to learn progressively without the need to load the entire dataset into memory.

## Gradient Descent Python Code Example

The following code represents the python code for implementing and plotting gradient descent algorithm:

import numpy as np import matplotlib.pyplot as plt # Defining a simple quadratic function and its derivative f = lambda x: x**2 f_prime = lambda x: 2*x # Generating values x = np.linspace(-10, 10, 400) y = f(x) # Gradient descent settings learning_rate = 0.15 x_start = 9 # Starting point steps = [x_start] n_iterations = 15 # Gradient Descent Iteration for _ in range(n_iterations): x_start = x_start - learning_rate * f_prime(x_start) steps.append(x_start) # Plotting the function plt.figure(figsize=(10, 6)) plt.plot(x, y, lw=2, label="f(x) = x²") # Plotting the steps plt.scatter(steps, f(np.array(steps)), color='red', s=100, label='Steps') plt.plot(steps, f(np.array(steps)), color='red', linestyle='--', lw=1.5) # Annotations and labels plt.title('Gradient Descent on f(x) = x²') plt.xlabel('x') plt.ylabel('f(x)') plt.legend() plt.grid(True) plt.show()

## Top Youtube Videos on Gradient Descent Algorithm

Here is the list of top 5 Youtube Videos that could be viewed to get a good understanding of Gradient descent algorithm.

**Gradient descent, how neural networks learn**

2. **Gradient Descent, Step-by-Step**

3. **Khan Academy – Why the gradient is the direction of steepest ascent?**

## Conclusion

As a summary, you learned the concepts of **Gradient Descent **along with some of the following aspects:

- Gradient descent algorithm is an optimization algorithm which is used to minimise the objective function.
- The objective function that needs to be minimised is termed as cost function or loss function.
- Gradient of a function at any point can be calculated as the first-order derivative of that function at that point.
- In case of multiple dimension, gradient of function at any point can be calculated as the partial derivative of the function a that point against different dimensions.
- Gradient descent represents the opposite direction of gradient. Gradient of a function at any point represents direction of steepest ascent of the function at that point.
- Batch gradient descent is updating the weights after all the training examples are processed. Stochastic gradient descent is about updating the weights based on each training data or a small group of training data.

- Model Parallelism vs Data Parallelism: Examples - April 11, 2024
- Model Complexity & Overfitting in Machine Learning: How to Reduce - April 10, 2024
- 6 Game-Changing Features of ChatGPT’s Latest Upgrade - April 9, 2024

## Leave a Reply