More training data means the model processes more information, which takes more energy. Simpler models or fewer epochs reduce energy use. CPUs are usually less efficient than GPUs for AI training, so switching to CPU often increases energy use.
Total CO2 = Energy (kWh) × Emission factor (kg CO2/kWh) = 500 × 0.4 = 200 kg CO2.
Fine-tuning a smaller pre-trained model requires less energy and time than training a large model from scratch, reducing environmental impact while keeping good accuracy.
Without early stopping, the model trains for all 100 epochs even if it stops improving early, wasting energy.
Deploying optimized smaller models near users (edge computing) reduces energy from data transfer and large centralized servers, balancing performance and environmental impact.