Experiment - Activation functions (ReLU, Sigmoid, Softmax)
Problem:You are training a simple neural network to classify handwritten digits (0-9) using the MNIST dataset. The current model uses Sigmoid activation functions in all layers.
Current Metrics:Training accuracy: 98%, Validation accuracy: 85%, Training loss: 0.05, Validation loss: 0.35
Issue:The model shows signs of overfitting and slow convergence. Validation accuracy is much lower than training accuracy, and the model struggles to classify some digits correctly.