New System Uses Smartphone Video to Accurately Locate Shooters

Smartphone Video and Audio Can Pinpoint Shooters

A system developed by Carnegie Mellon University uses video from smartphones to help pinpoint the location of shooters. Here, analysis of video from two smartphones during the 2017 mass shooting in Las Vegas indicates the shooter was in the north wing of the Mandalay Bay hotel. Credit: Carnegie Mellon University

New analytical tool could aid human rights, public safety workers.

Researchers at Carnegie Mellon University have developed a system that can accurately locate a shooter based on video recordings from as few as three smartphones.

When demonstrated using three video recordings from the 2017 mass shooting in Las Vegas that left 58 people dead and hundreds wounded, the system correctly estimated the shooter’s actual location — the north wing of the Mandalay Bay hotel. The estimate was based on three gunshots fired within the first minute of what would be a prolonged massacre.

Alexander Hauptmann, research professor in CMU’s Language Technologies Institute, said the system, called Video Event Reconstruction and Analysis (VERA), won’t necessarily replace the commercial microphone arrays for locating shooters that public safety officials already use, although it may be a useful supplement for public safety when commercial arrays aren’t available.

One key motivation for assembling VERA was to create a tool that could be used by human rights workers and journalists who investigate war crimes, terrorist acts, and human rights violations, Hauptmann said.

Researchers at Carnegie Mellon University have developed a system that can accurately locate a shooter based on video recordings from as few as three smartphones. Credit: Carnegie Mellon University

“Military and intelligence agencies are already developing these types of technologies,” said fellow researcher Jay D. Aronson, a professor of history at CMU and director of the Center for Human Rights Science. “We think it’s crucial for the human rights community to have the same types of tools. It provides a necessary check on state power.”

The researchers presented VERA and released it as open-source code last month at the Association for Computing Machinery’s International Conference on Multimedia in Nice, France.

Hauptmann said he has used his expertise in video analysis to help investigators analyze events such as the 2014 Maidan massacre in Ukraine, which left at least 50 antigovernment protesters dead. Inspired by that work — and the insight of ballistics experts and architecture colleagues from the firm SITU Research — Hauptmann, Aronson, and Junwei Liang, a Ph.D. student in language and information technology, have pulled together several technologies for processing video, while automating their use as much as possible.

VERA uses machine learning techniques to synchronize the video feeds and calculate the position of each camera based on what that camera is seeing. But it’s the audio from the video feeds that’s pivotal in localizing the source of the gunshots, Hauptmann said. Specifically, the system looks at the time delay between the crack caused by a supersonic bullet’s shock wave and the muzzle blast, which travels at the speed of sound. It also uses audio to identify the type of gun used, which determines bullet speed. VERA can then calculate the shooter’s distance from the smartphone.

“When we began, we didn’t think you could detect the crack with a smartphone because it’s really short,” Hauptmann said. “But it turns out today’s cell phone microphones are pretty good.”

By using video from three or more smartphones, the direction from which the shots were fired — and the shooter’s location — can be calculated based on the differences in how long it takes the muzzle blast to reach each camera.

With the proliferation of mass protests occurring in places such as Hong Kong, Egypt, and Iraq, identifying where a shot originated can be critical to determining whether protesters, police or other groups might be responsible when a shooting takes place, Aronson said.

But VERA is not limited to detecting gunshots. It is an event analysis system that can be used to locate a variety of other sounds relevant to human rights and war crimes investigations, he said. He and Hauptmann hope that other groups will add functionalities to the open-source software.

“Once it’s open source, the journalism and human rights communities can build on it in ways we don’t have the imagination for or time to do,” Aronson added.

The National Institute of Standards and Technology provided partial support for this work. The MacArthur Foundation and the Oak Foundation also have supported this work.

Reference: “Shooter Localization Using Social Media Videos” by Junwei Liang, Jay D. Aronson and Alexander Hauptmann, 15 October 2019, MM ’19: Proceedings of the 27th ACM International Conference on Multimedia.
DOI: 10.1145/3343031.3350536

Be the first to comment on "New System Uses Smartphone Video to Accurately Locate Shooters"

Leave a comment

Email address is optional. If provided, your email will not be published or shared.