Why do machine learning researchers use benchmark datasets?
Think about why having a common dataset helps researchers.
Benchmark datasets provide a standard set of data so researchers can compare model performance fairly and consistently.
What is the output shape of the features when loading the Iris dataset using scikit-learn?
from sklearn.datasets import load_iris iris = load_iris() X = iris.data print(X.shape)
Check how many samples and features the Iris dataset has.
The Iris dataset has 150 samples and 4 features, so the shape of the data is (150, 4).
You want to classify handwritten digits from the MNIST dataset. Which model is best suited for this task?
Consider the type of data and the task (image classification).
CNNs are designed to work well with image data and are commonly used for digit classification tasks like MNIST.
Which metric is most appropriate to evaluate a classification model on the CIFAR-10 benchmark dataset?
Think about the type of task CIFAR-10 represents.
CIFAR-10 is a classification dataset, so accuracy is a suitable metric to measure how many images are correctly classified.
What error will this code raise when trying to load the Boston Housing dataset using scikit-learn?
from sklearn.datasets import load_boston boston = load_boston()
Check recent changes in scikit-learn about this dataset.
The Boston Housing dataset was removed from scikit-learn due to ethical concerns, so trying to import load_boston raises ImportError.