0
0
ML Pythonml~5 mins

Gradient descent optimization in ML Python - Cheat Sheet & Quick Revision

Choose your learning style9 modes available
Recall & Review
beginner
What is gradient descent in simple terms?
Gradient descent is a way to find the lowest point of a hill by taking small steps downhill. In machine learning, it helps find the best model by reducing errors step by step.
Click to reveal answer
beginner
What role does the learning rate play in gradient descent?
The learning rate controls how big each step downhill is. If it's too big, you might jump past the lowest point. If it's too small, it takes a long time to reach the bottom.
Click to reveal answer
intermediate
Explain the difference between batch, stochastic, and mini-batch gradient descent.
Batch gradient descent uses all data to take one step, which is slow but stable. Stochastic uses one data point per step, which is fast but noisy. Mini-batch uses a small group of data points, balancing speed and stability.
Click to reveal answer
intermediate
Why can gradient descent get stuck in local minima?
Sometimes the error surface has many low points. Gradient descent might find a small dip (local minimum) instead of the lowest dip (global minimum), so it stops there and misses the best solution.
Click to reveal answer
advanced
What is the purpose of using momentum in gradient descent?
Momentum helps gradient descent keep moving in the right direction by remembering past steps. It can speed up learning and help jump over small bumps or flat areas.
Click to reveal answer
What does gradient descent aim to minimize?
AThe number of data points
BThe learning rate
CThe model size
DThe error or loss function
If the learning rate is too high, what can happen?
AThe model will always find the global minimum
BThe model converges too slowly
CThe model might overshoot the minimum and fail to converge
DThe model ignores the data
Which type of gradient descent updates the model after looking at one data point?
AStochastic gradient descent
BMini-batch gradient descent
CBatch gradient descent
DMomentum gradient descent
What problem does momentum help solve in gradient descent?
AAvoiding local minima
BSpeeding up convergence and smoothing updates
CIncreasing the learning rate automatically
DReducing the size of the dataset
Which of these is NOT a type of gradient descent?
ARandom gradient descent
BStochastic gradient descent
CBatch gradient descent
DMini-batch gradient descent
Describe how gradient descent helps a machine learning model learn from data.
Think about walking downhill to find the lowest point.
You got /4 concepts.
    Explain the differences and trade-offs between batch, stochastic, and mini-batch gradient descent.
    Consider how much data is used for each update.
    You got /3 concepts.