The exact value of the diffraction-limited aperture is often a contentious topic amongst photographers. Some might claim it's at f/11 for a given digital camera, while others will insist that it's closer to f/16, for example. While the precise f-stop doesn't really matter, it's good to clarify why there's so many opinions, and how these differences might translate into how your photograph actually appears. This article is intended as an addendum to the earlier tutorial on diffraction in photography.


Knowing the diffraction limit requires knowing how much detail a camera could resolve under ideal circumstances. With a perfect sensor, this would simply correlate with the size of the camera sensor's pixels. However, real-world sensors are a bit more complicated.

Bayer Array

Bayer Color Filter Arrays. Most of today's digital cameras only capture one of the primary colors at each pixel: green, red or blue (as depicted to the left). Furthermore, the color array or "mosaic" is arranged so that green is captured at twice as many pixel locations than is red or blue. Using sophisticated demosaicing algorithms, full color information at each pixel is then estimated based on the values of adjacent pixels.

Anti-Aliasing Filter & Microlenses. In front of the color filter array, most camera sensors also have another layer of material which aims to (i) minimize digital artifacts, such as aliasing and moiré, and to (ii) improve light-gathering ability. While this can further reduce resolution, with recent digital cameras resolution isn't too severely impacted (compared to what's already lost from the color filter array).

The end result is that the camera's resolution is better than one would expect if each 2x2 block containing all three colors represented a pixel, but not quite as good as each individual pixel. The exact resolution is really a matter of definition.


If it's not simply the pixel size, how does one define the resolution of a digital camera? Resolution is usually measured by taking a photo of progressively more closely spaced parallel lines. Once these lines can no longer be distinguished, the camera's resolution is said to have been surpassed. However, as with most things, it's not quite so clear cut.

Photograph of Converging Lines Used for Measuring Resolution

For the purpose of this tutorial, two resolution definitions are relevant:

Extinction Resolution. This describes the smallest features or edges that can be captured by your camera, although such detail may not necessarily be distinguished if it's too closely spaced, and such detail may also exhibit noticeable demosaicing artifacts. Beyond this resolution, progressively finer detail will appear almost completely smooth. In the above diagram, the extinction resolution will be near (or just beyond) the far right edge.

Artifact-Free Resolution. This describes the most closely spaced details which are still clearly distinguished from one another. Beyond this resolution, detail may exhibit visible demosaicing artifacts, and may not be depicted in the same way it appears in-person. The artifact-free resolution is typically what one is referring to when they speak of the resolution limit. In the above diagram, the resolution limit would be past roughly the first two-thirds of the image.

One could make the argument that any detail finer than the artifact-free resolution limit isn't actually "real detail," and therefore shouldn't matter, since it's largely the result of the bayer array and digital artifacts. However, one could also argue that such features do indeed improve apparent image detail, which is often helpful with objects such as fine leaves or grass. Edges will also appear more sharp and defined.

Technical Notes:
The real-world resolution limit of a Bayer array is typically around 1.5X as large as the individual pixels. For example, Canon's EOS 5D camera has 2912 vertical pixels, but can only resolve ~2000 horizontal lines. However, its extinction resolution is around ~2500 horizontal lines, so details as fine as 1.2-1.3X the pixel size may also appear. While these values will vary slightly based on the camera and RAW converter, they're a good rough estimate that's based on actual measurements.


Our human visual system has adapted to be most sensitive to the green region of the light spectrum (since this is where our sun emits the most light -- see the tutorial on human color perception). The result is that green light contributes much more to our perception of luminance, and is also why color filter arrays are designed to capture twice as much green light as either of the other two colors.

However, this human sensory trick isn't without trade-offs. It also means that color filter arrays can resolve differences in lightness far better than differences in color -- with important implications for the diffraction limit. Since the resolution of green/lightness is roughly twice as high as that for red or blue, a higher f-stop is required before diffraction begins to limit the resolution of strictly red/blue details.

<— Less Diffraction More Diffraction —>

Finally, regardless of the type of camera sensor, diffraction doesn't affect all colors equally. The longer its wavelength, the more diffraction will scatter a given color of light. In the visual rainbow -- red-orange-yellow-green-blue-indigo-violet -- the red end of the spectrum has the longest wavelength, and the wavelength gets progressively shorter as one moves toward the violet end of the spectrum. So why is this relevant? With camera color filters, blue light is much less susceptible to diffraction than red light, so it requires a higher f-stop before diffraction begins to limit its resolution.


The resolution limit of any optical system can be described based on how finely it's able to record a point light source. The shape at which this point light source is recorded is called the "airy disk." The width of an airy disk is controlled by the camera's f-stop; if this width increases too much, closely spaced points may no longer be resolved (as shown below).

3D view of airy disk Airy disks resolved

Airy Disk
(smallest point light source)
Closely Spaced Points
Barely Resolved Unresolved

At low f-stop values, the width of this airy disk is typically much smaller than the width of a camera pixel, so your pixel size is all that determines the camera's maximum resolution. However, at high f-stop values, the width of this airy disk can surpass* that of your camera's pixels -- in which case your camera's resolution is said to be "diffraction limited." Under such a scenario, the resolution of your camera is determined only by the width of this airy disk, and this resolution will always be less than the pixel resolution.

The key here is that airy disks begin overlapping even before they become unresolved, so diffraction will become visible (and potentially erase subtle tonal variations) even before standard black and white lines are no longer distinguishable. In other words, the diffraction limit isn't an all or nothing image quality cliff -- its onset is gradual.

*Strictly speaking, the width of the airy disk needs to be at least ~3X the pixel width in order for diffraction to limit artifact-free, grayscale resolution on a Bayer sensor, although it will likely still become visible when the airy disk width is near 2X the pixel width.


Use the following calculator to estimate when diffraction begins to influence an image.  This only applies for images viewed on-screen at 100%; whether this will be apparent in the final print also depends on viewing distance and print size. To calculate this as well, please visit: diffraction limits and photography.

Diffraction Limited Aperture Calculator
 Sensor Size
 Resolution Megapixels
Diffraction Likely Becomes Visible
Diffraction Limits Extinction Resolution
Diffraction Limits Standard Grayscale Resolution

The above values are only theoretical best case scenarios;
actual results will also depend on lens characteristics, demosaicing and subject detail.
Color resolution limits are not listed because these are highly dependent on demosaicing software, image content and color purity; expect these to be at 20-50% higher f-stops for Bayer sensors.

Most will find that the f-stop given in the "diffraction limits extinction resolution" field tends to correlate well with the f-stop values where one first starts to see fine detail being softened. All other pages of this website therefore use this as the criterion for determining the diffraction-limited aperture.

The purpose of the above calculator isn't so that you can scrutinize your images to see when blue details are impacted, for example. Instead, it's intended to do the opposite: to give an idea of how gradual and broad diffraction's onset can be, and how its "limit" depends on what you're using as the image quality criterion.


Thus far, you're probably thinking, "diffraction more easily limits resolution as the number of camera megapixels increases, so more megapixels must be bad, right?" No -- at least not as far as diffraction is concerned. Having more megapixels just provides more flexibility. Whenever your subject matter doesn't require a high f-stop, you have the ability to make a larger print, or to crop the image more aggressively. Alternatively, a 20MP camera that requires an f-stop beyond its diffraction limit could always downsize its image to produce the equivalent from a 10MP camera that uses the same f-stop (but isn't yet diffraction limited).

Regardless, the onset of diffraction is gradual, and its limiting f-stop shouldn't be treated as unsurpassable. Diffraction is just something to be aware of when choosing your exposure settings, similar to how one would balance other trade-offs such as noise (ISO) vs shutter speed. While calculations can be a helpful at-home guide, the best way to identify the optimal trade-off is to experiment -- using your particular lens and subject.

For more background reading on this topic, also visit:
Understanding Diffraction, Part 1: Pixel Size, Aperture & Airy Disks

-Back to Photography Tutorials-