Complete the code to create an SGD optimizer for the model parameters with a learning rate of 0.01.
optimizer = torch.optim.[1](model.parameters(), lr=0.01)
The SGD optimizer is created using torch.optim.SGD. It updates parameters using stochastic gradient descent.
Complete the code to create an Adam optimizer with a learning rate of 0.001.
optimizer = torch.optim.[1](model.parameters(), lr=0.001)
The Adam optimizer is created using torch.optim.Adam. It adapts learning rates for each parameter.
Fix the error in the code to correctly update the optimizer after computing gradients.
loss.backward()
optimizer.[1]()After computing gradients with loss.backward(), call optimizer.step() to update model parameters.
Fill both blanks to reset gradients and then update parameters correctly.
optimizer.[1]() loss.backward() optimizer.[2]()
First, clear old gradients with optimizer.zero_grad(). Then compute gradients with loss.backward(). Finally, update parameters with optimizer.step().
Fill all three blanks to create an Adam optimizer with weight decay, reset gradients, and update parameters.
optimizer = torch.optim.Adam(model.parameters(), lr=[1], weight_decay=[2]) optimizer.[3]() loss.backward() optimizer.step()
The Adam optimizer is created with learning rate 0.001 and weight decay 0.0001. Before backward pass, call optimizer.zero_grad() to clear gradients.