Basic image manipulation with arrays in NumPy - Time & Space Complexity
When working with images as arrays, it is important to understand how the time to process them grows as the image size increases.
We want to know how the time to manipulate image pixels changes when the image gets bigger.
Analyze the time complexity of the following code snippet.
import numpy as np
height, width = 100, 100 # Define height and width for the image
# Create a sample image as a 2D array (grayscale)
image = np.random.randint(0, 256, size=(height, width), dtype=np.uint8)
# Invert the image colors
inverted_image = 255 - image
# Increase brightness by 50, clip max value to 255
bright_image = np.clip(image + 50, 0, 255)
# Create a mask for pixels above 200
mask = image > 200
# Set those pixels to 255 (white)
image[mask] = 255
This code creates a grayscale image as a 2D array and performs simple pixel-wise operations like inversion, brightness increase, and masking.
Identify the loops, recursion, array traversals that repeat.
- Primary operation: Pixel-wise arithmetic and comparisons over the entire image array.
- How many times: Once for each pixel in the image (height x width times).
As the image size grows, the number of pixels to process grows too, so the time grows proportionally.
| Input Size (pixels) | Approx. Operations |
|---|---|
| 10 x 10 = 100 | About 100 pixel operations |
| 100 x 100 = 10,000 | About 10,000 pixel operations |
| 1000 x 1000 = 1,000,000 | About 1,000,000 pixel operations |
Pattern observation: The operations increase directly with the number of pixels, so doubling width and height quadruples the work.
Time Complexity: O(n)
This means the time to manipulate the image grows linearly with the total number of pixels.
[X] Wrong: "The operations only depend on width or height, not both."
[OK] Correct: Each pixel is identified by both row and column, so total pixels are width x height, not just one dimension.
Understanding how image size affects processing time helps you explain efficiency in real tasks like photo editing or computer vision.
"What if we changed the image from grayscale (2D array) to color (3D array with RGB channels)? How would the time complexity change?"