search

UMD     This Site





Who’s being deceptive? In (a) smaller hand movements and in (d) the velocity of hand and foot joints provide deception cues.

Who’s being deceptive? In (a) smaller hand movements and in (d) the velocity of hand and foot joints provide deception cues.

 

The Geometric Algorithms for Modeling, Motion and Animation (GAMMA) research team at the University of Maryland and the University of North Carolina has developed a data-driven deep neural algorithm for detecting deceptive walking behavior using nonverbal cues like gaits and gestures.

“The Liar’s Walk: Detecting Deception with Gait and Gesture” is part of GAMMA’s project in social perception of pedestrians and virtual agents using movement features, which it calls GAIT. The researchers include Kurt Gray, Tanmay Randhavane, Kyra Kapsaskis of the University of North Carolina, and Uttaran Bhattacharya, Aniket Bera, and ISR-affiliated Professor Dinesh Manocha (CS/ECE/UMIACS) of the University of Maryland.

In recent years, AI and vision communities have focused on learning human behaviors, and human-centric video analysis has rapidly developed. While conventional analysis of video pays attention to the actions recorded, human-centric video analysis focuses on the behaviors, emotions, personalities, actions, anomalies, dispositions and intentions of the humans in the video.

Surprisingly, detecting deliberate deception has not been a major research focus. Detection is challenging because deception is subtle, and because deceivers are, of course, attempting to conceal their actual cues and expressions. However, much is known about the verbal and nonverbal deception cues of deception. Implicit cues include facial and body expressions, eye contact, and hand movements. Deceivers try to alter or control the cues they believe get the most attention, like facial expressions. Body movements such as walking gaits are less likely to be controlled, making gait an excellent way to detect deceptive behaviors.

The authors developed a data-driven approach for detecting deceptive walking, based on gaits and gestures extracted from videos. Gaits are extracted from videos as a series of 3D poses using state-of-the-art human pose estimation. Various gestures were then annotated. Using this data, the researchers computed psychology-based gait features, gesture features, and deep features (deceptive features) learned using an LSTM-based neural network. The features were fed into fully connected layers of the neural network to classify normal and deceptive behaviors. The researchers trained this neural network “Deception Classifier” to learn the deep features as well as classify the data into behavior labels on a novel public-domain dataset called “DeceptiveWalk,” which contains annotated gaits and gestures with both deceptive and natural walks.

The Deception Classifier can achieve an accuracy of 93.4% when classifying deceptive walking, an improvement of 5.9% and 10.1% over classifiers based on the state-of-the-art emotion and action classification algorithms, respectively.

Additionally, the DeceptiveWalk dataset provided interesting observations about participants’ behavior in deceptive and natural conditions. For example, deceivers put their hands in their pockets and looked around more than participants in the natural condition. Deceivers also display the opposite of an expected behavior or make a controlled movement.

The researchers caution that their work is preliminary and limited in nature, with videos produced in a controlled laboratory setting, and thus may not be directly applicable for detecting many kinds of deceptive walking behaviors in the real world. In addition, while they present an algorithm that detects deception using walking, they do not claim that performing certain gestures and walking in a certain way conveys deception in all cases.



Related Articles:
Forecasting traffic for autonomous vehicles
RoadTrack algorithm could help autonomous vehicles navigate dense traffic scenarios
A learning algorithm for training robots' deep neural networks to grasp novel objects
AlphaGo family of AI programs grew from AMS simulation-based algorithms developed at UMD
IFIG framework helps robots follow instructions
ISR alum honors former advisor Benjamin Kedem with quantile frequency analysis paper
ISR, ECE, CS, UMIACS faculty present 12 talks at Northrop Grumman University Research Symposium
NSF funds Shamma, Espy-Wilson for neuromorphic and data-driven speech segregation research
New affiliate faculty Mark Fuge is expert in machine learning and artificial intelligence
Maryland research contributes to Google’s AlphaGo AI system

January 9, 2020


«Previous Story  

 

 

Current Headlines

RoadTrack algorithm could help autonomous vehicles navigate dense traffic scenarios

Who should Google Scholar update more often?

Alum Evandro Valente testifies on two economic development bills

Warren Savage delivers cybersecurity keynote at DesignCon 2020

Fast estimation algorithm aids investigations of the cortical origins of underlying neural processes

Nau and colleagues develop APE, an integrated acting-and-planning system

IFIG framework helps robots follow instructions

A learning algorithm for training robots' deep neural networks to grasp novel objects

Alumnus Donald Martin honored by Network of Minorities in Mathematical Sciences

Dean Pines Named University of Maryland's 34th President

 
 
Back to top  
Home Clark School Home UMD Home