In the field of industrial acoustic inspection, we encounter acoustic cameras in all shapes and sizes—some compact and portable with barely a dozen microphones, others resembling giant honeycombs densely packed with over a hundred sensors.
When consulting with clients, a straightforward question frequently arises: "The more microphones, the more accurate the localization, right? Should I just pick the model with the highest microphone count?"
The answer is more nuanced than a simple "yes" or "no." In reality, the relationship between microphone count and performance is not strictly linear. It is governed by three interdependent constraints: physical aperture, computational power, and application context.
To grasp this, we must first understand how an acoustic camera operates. It functions like a miniature phased-array radar: multiple microphones capture the same sound wave from slightly different positions, and beamforming algorithms compute the direction of arrival.
The critical metric determining how precisely the camera can "see" a sound source is called the main lobe width (3dB beamwidth). A narrower main lobe translates directly to a better ability to separate two adjacent noise sources.
For common linear or cross-shaped arrays, the main lobe width approximates the following relationship:
Main Lobe Width ≈ 0.891 × λ / (N × d)
λ: Wavelength of the sound (shorter for higher frequencies)
N: Number of microphones
d: Spacing between adjacent microphones
The product N × d represents the physical aperture of the array—essentially, the maximum span across which microphones are distributed. Therefore, it is the aperture that truly dictates localization resolution. This is why military sonar arrays are built dozens or even hundreds of meters long: to achieve an enormous aperture for detecting distant underwater targets.

From this principle, two obvious performance levers emerge:
Increase microphone count while maintaining spacing. For instance, with spacing fixed at 3 cm, increasing from 4 to 32 microphones dramatically sharpens the main lobe.
Increase spacing while maintaining microphone count. Spreading the same 16 microphones across a wider physical area yields a similarly significant performance boost.
At this point, one might wonder: "Why not just buy a larger housing and cram it full of microphones?"
The issue is that device dimensions are finite. If you purchase a handheld acoustic camera, its array disk diameter might be fixed at 10 cm. Within that fixed 10 cm physical space:
Microphone Spacing d ≈ Array Diameter D / Microphone Count N
In other words, inside a fixed-size enclosure, more microphones force tighter spacing. And excessively tight spacing triggers a cascade of engineering drawbacks:
Degraded High-Frequency Performance: Overly dense spacing causes the main lobe for high-frequency signals to become broad and flat, losing the ability to focus sharply. No matter how many extra microphones you add, the physical aperture ceiling caps the achievable resolution.
Elevated Sidelobes and Cluttered Imaging: Simulations demonstrate that for the same 10 cm diameter, a 16-microphone configuration and a 64-microphone configuration exhibit nearly identical main lobe widths. However, the 64-microphone version displays noticeably higher and more chaotic sidelobes. Imagine 30 people crowded into a room meant for 10—mutual interference produces noise, not clarity.
Exponential Computational Burden: Doubling the microphone count increases beamforming computational load geometrically, not linearly. In field applications requiring real-time imaging, an excessive number of microphones causes system lag, latency, and a degraded user experience.
Wasted Cost and Power: Every additional microphone channel demands its own high-precision analog-to-digital converter, preamplifier, and anti-aliasing filter. These expensive components drive up cost, power consumption, and potential failure points—without delivering a corresponding improvement in accuracy.
So, does a higher microphone count mean a better acoustic camera?
Our answer: Not entirely. It depends on the phase of development.
Growth Phase (32 → 64 → 128 Microphones): Before the device hits its physical size bottleneck and when array real estate remains ample, adding microphones effectively expands the physical aperture or enhances spatial sampling. In this stage, performance improvements are nearly proportional. Moving from 32 to 64, and from 64 to 128, you will observe clear gains in low-frequency resolution and dynamic range.
Maturity Phase (Beyond 128 Microphones): Once the array's physical dimensions are fixed (e.g., limited by a handheld device's 50 cm diameter) and the microphone count exceeds a threshold of approximately 128, marginal returns diminish sharply. Continuing to pack in additional microphones compresses the spacing to its physical limit. Further increases not only fail to narrow the main lobe but actively degrade high-frequency response due to excessive density, amplify sidelobe interference, and overwhelm the processing system.
True engineering wisdom lies in balance. For a fixed-size device, the optimal solution is not blindly maximizing microphone count, but rather finding the sweet spot among main lobe sharpness, sidelobe suppression, frequency coverage, and real-time computational efficiency. In real-world field conditions, a well-designed 64- or 128-microphone array will consistently outperform a 256-microphone "spec-stuffed" array plagued by cramped spacing.
Next time you evaluate an acoustic camera, consider asking: "What is this device's physical aperture?" instead of focusing solely on the number of microphones listed in the brochure.