There is a newer version of the record available.

Published January 24, 2025 | Version v1
Software Open

Sound of Interference: Electromagnetic Eavesdropping Attack on Digital Microphones Using Pulse Density Modulation

  • 1. The University of Electro-Communications
  • 2. ROR icon University of Florida

Contributors

  • 1. ROR icon University of Florida

Description

We introduce a novel electromagnetic (EM) side-channel attack that allows for acoustic eavesdropping on electronic devices. This method specifically targets modern digital microelectromechanical systems (MEMS) microphones, which transmit captured audio via pulse-density modulation (PDM), that translate the analog sound signal into the density of output pulses in the digital domain. We discover that each harmonic of these digital pulses retains acoustic information, allowing the original audio to be retrieved through simple FM demodulation using standard radio receivers. An attacker can exploit this phenomenon to capture what the victim microphone hears remotely without installing malicious software or tampering with the device. We verify the vulnerability presence by conducting real-world evaluation on several PDM microphones and electronic devices, including laptops and smart speakers.
For example, we demonstrate that the attack achieves up to 94.2% accuracy in recognizing spoken digits, up to 2 meters from a victim laptop located behind a 25 cm concrete wall. We also evaluate the attacker capability to eavesdrop on speech using popular speech-to-text APIs (e.g., OpenAI) not trained on EM traces, achieving a maximum of 14% transcription error rate in recovering the Harvard Sentences dataset. We further demonstrate that similar accuracy can be achieved with a cheap and stealthy antenna made out of copper tape. We finally discuss the limited effectiveness of current defenses such as resampling, and we propose a new hardware defense
based on clock randomization.

The work will be presented at Usenix Security 2025.

The repository contains the instructions and data to setup the environment and to replicate the results. Please see the README file (README.md) for the details.

The artifacts consist of:
1. Speech transcription results and training log files for digit and speaker classification models, along with the corresponding recorded audio files for the tested devices.
2. The scripts for training and fine-tuning the pipeline we used for our digit and speaker classification models and for testing HuBERT, Microsoft, and OpenAI transcription models.

Note that due to responsible disclosure, two of the evaluated devices ("Redacted laptop" and "Redacted" in the paper) are reported as Device_1 and Device_2 in the repository, respectively.

Files

Behind_the_wall.zip

Files (17.2 GB)

Name Size Download all
md5:a57f5dda120ff025c0ff7961db1e4ad4
4.5 GB Preview Download
md5:1d0db93dd216401eea3082060c64c4dc
3.4 GB Preview Download
md5:77cda2d38dc7e94719003a18205a6df6
2.2 MB Preview Download
md5:b8cad6b65fe9e23ed1af29e1c4ec7f72
3.5 GB Preview Download
md5:a3f3c3900bb9357d9bdacb811519e95b
628.8 MB Preview Download
md5:1c6be0d4765b16c2c754083c255c323b
1.2 GB Preview Download
md5:07bf874b37f43913d4749836908053f3
3.2 GB Preview Download
md5:296d7bccb6fdff28cf8ef7673f375721
5.3 kB Preview Download
md5:ad9ead3d6b6515bef749f91961eca185
841.5 MB Preview Download

Additional details

Dates

Accepted
2025
Data