Light Commands is a vulnerability of MEMS microphones that allows attackers to remotely inject inaudible and invisible commands into voice assistants, such as Google assistant, Amazon Alexa, Facebook Portal, and Apple Siri using light.
In our paper we demonstrate this effect, successfully using light to inject malicious commands into several voice controlled devices such as smart speakers, tablets, and phones across large distances and through glass windows.
The implications of injecting unauthorized voice commands vary in severity based on the type of commands that can be executed through voice. As an example, in our paper we show how an attacker can use light-injected voice commands to unlock the victim's smart-lock protected home doors, or even locate, unlock and start various vehicles.
Read the Paper Cite@inproceedings{Sugawara2020LightCommands,
author = {Sugawara, Takeshi and Cyr, Benjamin and Rampazzi, Sara and Genkin, Daniel and Fu, Kevin},
title = {Light Commands: Laser-Based Audio Injection on Voice-Controllable Systems},
year = {2019}
}
Light Commands were discovered by the following team of academic researchers:
Contact us at LightCommandsTeam@gmail.com
By shining the laser through the window at microphones inside smart speakers, tablets, or phones, a far away attacker can remotely send inaudible and potentially invisible commands which are then acted upon by Alexa, Portal, Google assistant or Siri.
Making things worse, once an attacker has gained control over a voice assistant, the attacker can use it to break other systems. For example, the attacker can:
Microphones convert sound into electrical signals. The main discovery behind light commands is that in addition to sound, microphones also react to light aimed directly at them. Thus, by modulating an electrical signal in the intensity of a light beam, attackers can trick microphones into producing electrical signals as if they are receiving genuine audio.
Voice assistants inherently rely on voice to interact with the user. By shining a laser on their microphones, an attacker can effectively hijack the voice assistant and send inaudible commands to the Alexa, Siri, Portal, or Google Assistant.
Light can easily travel long distances, limiting the attacker only in the ability to focus and aim the laser beam. We have demonstrated the attack in a 110 meter hallway, which is the longest hallway available to us at the time of writing.
Careful aiming and laser focusing is indeed required for light commands to work. To focus the laser across large distances one can use a commercially avalible telephoto lens. Aiming can be done using a geared tripod head, which greatly increases accuracy. An attacker can use a telescope or binocular in order to see the device's microphone ports at large distances.
In our experiments, we test our attack on the most popular voice recognition systems, namely Amazon Alexa, Apple Siri, Facebook Portal, and Google Assistant. We benchmark multiple devices such as smart speakers, phones, and tablets as well as third-party devices with built-in speech recognition.
| Device | Voice Recognition System |
Minimun Laser Power at 30 cm [mW] |
Max Distance at 60 mW [m]* |
Max Distance at 5 mW [m]** |
| Google Home | Google Assistant | 0.5 | 50+ | 110+ |
| Google Home mini | Google Assistant | 16 | 20 | - |
| Google NEST Cam IQ | Google Assistant | 9 | 50+ | - |
| Echo Plus 1st Generation | Amazon Alexa | 2.4 | 50+ | 110+ |
| Echo Plus 2nd Generation | Amazon Alexa | 2.9 | 50+ | 50 |
| Echo | Amazon Alexa | 25 | 50+ | - |
| Echo Dot 2nd Generation | Amazon Alexa | 7 | 50+ | - |
| Echo Dot 3rd Generation | Amazon Alexa | 9 | 50+ | - |
| Echo Show 5 | Amazon Alexa | 17 | 50+ | - |
| Echo Spot | Amazon Alexa | 29 | 50+ | - |
| Facebook Portal Mini | Alexa + Portal | 18 | 5 | - |
| Fire Cube TV | Amazon Alexa | 13 | 20 | - |
| EchoBee 4 | Amazon Alexa | 1.7 | 50+ | 70 |
| iPhone XR | Siri | 21 | 10 | - |
| iPad 6th Gen | Siri | 27 | 20 | - |
| Samsung Galaxy S9 | Google Assistant | 60 | 5 | - |
| Google Pixel 2 | Google Assistant | 46 | 5 | - |
While we do not claim that our list of tested devices is exhaustive, we do argue that it does provide some intuition about the vulnerability of popular voice recognition systems to Light Commands.
Note:
* Limited to a 50 m long corridor.
** Limited to a 110 m long corridor.
At the time of writing, speaker recognition is off by default for smart speakers and is only enabled by default for devices like phones and tablets. Thus, Light Commands can be used on these smart speakers without imitating the owner's voice. Moreover, even if enabled, speaker recognition only verifies that the wake-up words (e.g., "Ok Google" or "Alexa") are said in the owner's voice, and not the rest of the command. This means that one "OK Google" or "Alexa" spoken by the owner can be used to compromise all the commands. Finally, as we show in our work, speaker recognition for wake-up words is often weak and can be sometimes bypassed by an attacker using online text-to-speech synthesis tools for imitating the owner's voice.
The Light Commands attack can be mounted using a simple laser pointer ($13.99, $16.99, and $17.99 on Amazon), a laser driver (Wavelength Electronics LD5CHA, $339), and a sound amplifier (Neoteck NTK059, $27.99 on Amazon). A telephoto lens (Opteka 650-1300mm, $199.95 on Amazon) can be used to focus the laser for long range attacks.