Who needs a peep hole when a wifi network will do? Researchers from MIT have developed technology that uses wireless signals to see your silhouette through a wall—and it can even tell you apart from other people, too.
The team from MIT’s Computer Science and Artificial Intelligence Lab are no strangers to using wireless signals to see what’s happening on the other side of a wall. In 2013, they showed off software that could use variations in wifi signal to detect the presence of human motion from the other side of a wall. But in the last two years they’ve been busy developing the technique, and now they’ve unveiled the obvious — if slightly alarming — natural progression: they can use the wireless reflections bouncing off a human body to see the silhouette of a person standing behind a wall.
Not only that, the team’s technique, known is RF-Capture, is accurate enough to track the hand of a human and, with some repeated measurements, the system can even be trained to recognise different people based just on their wifi silhouette. The research, which is to be presented at SIGGRAPH Asia next month, was published this morning on the research group’s website.
So how does it work? It’s actually relatively straightforward: a device transmits wireless signals on one side of the wall, which propagate through it and are then reflected by bodies on the other side. The device then captures the reflections, which are passed to software to be cleaned up. As you might expect, this part requires some pretty serious processing, as different body parts, humans and objects introduce all kinds of interference.
First, the team captures a series of frames of data before it does anything, to reduce the effects of random noise. “At a high level, we suppress noise by combining information across time and fitting the data into a model,” explained Fadel Adi, one of the researchers, to me via email. “For example, if you look at the video, you see that we capture consecutive time snapshots, before we can construct the human silhouette.”
Then, the team take the data they’ve managed to gather and feed it through algorithms that are trained to detect body-like features. “The algorithms that we developed fit all of these snapshots into a coarse human model with major body parts — such as head, chest, arms, and feet,” continues Adi. “That is, we combine these snapshots in a manner that maximizes the ability of the reconstructed silhouette in representing the human body.”
Given the small amount of information that’s received in the reflections, the system keeps constant track of what it can identify: sometimes it might be an arm and a head, other times a torso and a leg. It’s capable of stitching these glimpses together, in turn forming a full human silhouette. And given the world is made up of humans with a wide variety of body types, that data can be scrutinized a little further.
In fact, the team has been using distinctive measures from these images, such as height, shoulder width and other body shape metrics, to identify different humans hiding behind a wall.
Using machine learning techniques, the researchers can train algorithms to spot the subtle differences in different people’s body shapes. “[W]e use the captured human silhouettes from our reconstruction algorithm [to] train a classifier on these silhouettes which allows us to distinguish between people,” explains Adi. “The classifier captures features like height and body builds, which allows us to distinguish between people using RF-Capture.”
In a series of tests, they’ve shown that the recognition capability can distinguish between 15 different people through a wall with nearly 90 percent accuracy. And in another series of experiments, where the team simply tracked the patterns they were capturing, they were able to trace a person’s hand as they wrote in the air. “The accuracy of tracking the moving hand is about an inch,” explains Adi.
The team is, understandably, excited about the applications that this kind of technology could provide. They’re already working, for instance, to develop a device that could sit in the home of your elderly grandparents, constantly scanning the house for their presence; if they were observed to have fallen over, the system could phone 911 on their behalf.
The team suggests that RF-Capture will get more accurate over time, too. “We are very excited about future research along two fronts: first, getting finer resolution to recover the silhouette with higher accuracy, and second, we would like to gain deeper understanding on the health front,” explains Professor Dina Katabi, one of the researchers involved in the project. “For example, can we track human fingers from behind a wall? We can [already] use RF-Capture to extract a person’s breathing and heart rate...can we detect heart problems using this wireless technology? We believe the answer is yes.”
If they can achieve that kind of detection, they suggest the technology could allow your smart home to sense your motion to, say, control video-games, control appliances depending on the way you move, or even perform on-the-fly health checks.
There is, of course, an elephant in the room here that you wouldn’t even need RF-Capture to identify — and that’s privacy. First off, the team insists that any device using this kind of technology, such as a router that can tell when your relative has fallen, uses encryption from the get-go. “We [also] want to ensure that people do not use it for malicious purposes,” adds Katabi. “To that end, we are working along two fronts: first, we are designing blockers that can prevent someone from being tracked except by their own device. And, second, we need to have regulations that dictate how and when these devices can be used. Privacy is always a chief concern.”
All images and video by MIT Computer Science & Artificial Intellegence Laboratory