import tensorflow as tf
from tensorflow.keras import layers, models
# Load MNIST dataset
(X_train, y_train), (X_test, y_test) = tf.keras.datasets.mnist.load_data()
# Normalize pixel values
X_train, X_test = X_train / 255.0, X_test / 255.0
# Build model with dropout and smaller hidden layer
model = models.Sequential([
layers.Flatten(input_shape=(28, 28)),
layers.Dense(64, activation='relu'),
layers.Dropout(0.3),
layers.Dense(10, activation='softmax')
])
# Compile model with lower learning rate
model.compile(
optimizer=tf.keras.optimizers.Adam(learning_rate=0.001),
loss='sparse_categorical_crossentropy',
metrics=['accuracy']
)
# Use early stopping callback
early_stop = tf.keras.callbacks.EarlyStopping(monitor='val_loss', patience=3, restore_best_weights=True)
# Train model
history = model.fit(
X_train, y_train,
epochs=30,
batch_size=64,
validation_split=0.2,
callbacks=[early_stop],
verbose=0
)
# Evaluate on test data
test_loss, test_acc = model.evaluate(X_test, y_test, verbose=0)
print(f'Test accuracy: {test_acc * 100:.2f}%')