System Recognizes Objects Touched by User, Enabling Context-Aware Smartwatch Apps-CMU News - Carnegie Mellon University

Monday, November 9, 2015

System Recognizes Objects Touched by User, Enabling Context-Aware Smartwatch Apps

Carnegie Mellon, Disney Method Exploits Conductivity of Human Body

By Byron Spice / Canegie Mellon / 412-268-9068 / bspice@cs.cmu.edu
and Jennifer Liu / Walt Disney Imagineering / 818-544-6130 / Jennifer.C.Liu@disney.com

Context Aware App Door

A new technology developed by Carnegie Mellon University and Disney Research could enable smartwatches to automatically recognize what objects users are touching, for instance, whether the wearer is using a laptop, operating a saw, or riding a motorcycle, creating new opportunities for context-aware apps.

The technique, called EM-Sense, takes advantage of the body’s natural electrical conductivity to detect whether a person is touching an electrical or electromechanical device and, based on the distinctive electromagnetic noise emitted by such devices, automatically identify the object.

“This could be a great feature for smartwatches, making them much smarter than before,” said Gierad Laput, a Ph.D. student in Carnegie Mellon’s Human-Computer Interaction Institute (HCII). A smartwatch equipped with EM-Sense would have a much more detailed understanding of what the user is doing than is possible with common mobile sensors, such as accelerometers or pulse monitors, he added.

“This could be a great feature for smartwatches, making them much smarter than before.” — Gierad Laput

The smartwatch therefore could automatically start a timer when the wearer begins using an electric toothbrush, unlock a keyboard without a password when users touch their laptops, or play the latest news when breakfast is being prepared. Linking a smartwatch with a smartphone or other mobile device would expand the possibilities even further.

Context Aware App Stove

“We are now able to gain a greater contextual understanding of user activities by recognizing what objects they are interacting with,” said Alanson Sample, a research scientist at Disney Research. EM-Sense is able to differentiate between scores of objects based on the ambient electromagnetic noise they emit, so objects need not be modified or have RFID tags for the technique to work.

The researchers will discuss EM-Sense at UIST 2015, the ACM Symposium on User Interface Software and Technology, Nov. 8-11, in Charlotte, N.C.

“By extending the Internet to physical objects — what’s being called the Internet of Things (IoT) — we are creating new ways for people to interact with the world around them,” said Jessica Hodgins, vice president of Disney Research. “EM-Sense can make the IoT experience even richer by enabling people to get information or additional functionality simply by touching everyday objects.”

The human body serves as an antenna for EM-Sense. From any body part an object touches, its distinctive electromagnetic emissions propagate through the body to an electrode worn at the wrist. Kitchen appliances, power tools, electronic scales and door handles with electrically triggered locks are among the items that can be detected and identified. It’s even possible to differentiate between different models of cellphones.

Laput said the proof-of-concept sensor was built with off-the-shelf components – principally a small, software-defined radio. A smartwatch would provide the perfect platform, he added, because the watch is in contact with the skin and also houses a highly capable computer for deciphering signals and running apps.

The researchers are able to filter out background noise, so the technique is able to recognize powered devices at any location. The technique also can work with large, non-powered conductive objects, such as doors, ladders and other structural elements.

In addition to Laput and Sample, the research team included Chouchang Yang, a post-doctoral researcher at Disney; Chris Harrison, assistant professor of Human-Computer Interaction; and Robert Xiao, a Ph.D. student in HCII.

Find out more about the project. 

Watch the video.