Facial recognition tech taken to the next level in virtual reality

[This ScienceDaily version of information from the University of South Australia describes a recently published study that tested the use of EEG to measure facial expressions that allow users to control their movement in virtual reality, as a replacement for or supplement to hand controllers; the study explicitly measured participants’ sense of presence. The link to the published study is at the end and see new coverage in Science for Students for more pictures and details. For a related story see an October 2020 ISPR Presence News post. –Matthew]

[Image: Participants used three facial expressions — a smile, a frown and clenched teeth — to interact with their virtual environment. Credit: Arindam Dey/University of Queensland]

Forget handheld virtual reality controllers: a smile, frown or clench will suffice

Facial recognition tech taken to the next level in virtual reality

February 18, 2022

Our face can unlock a smartphone, provide access to a secure building and speed up passport control at airports, verifying our identity for numerous purposes.

An international team of researchers from Australia, New Zealand and India has now taken facial recognition technology to the next level, using a person’s expression to manipulate objects in a virtual reality setting without the use of a handheld controller or touchpad.

In a world first study led by University of Queensland researcher Dr Arindam Dey, human computer interaction experts used neural processing techniques to capture a person’s smile, frown and clenched jaw and used each expression to trigger specific actions in virtual reality environments. “The primary motivation of this work was to make the metaverse more accessible and inclusive,” says Dr Dey. “At the same time facial expressions can also be used to enable interactions like kissing and blowing air within the virtual environments in a more realistic way than now.”

One of the researchers involved in the experiment, University of South Australia’s Professor Mark Billinghurst, says the system has been designed to recognise different facial expressions via an EEG headset.

“A smile was used to trigger the ‘move’ command; a frown for the ‘stop’ command and a clench for the ‘action’ command, in place of a handheld controller performing these actions,” says Prof Billinghurst.

“Essentially we are capturing common facial expressions such as anger, happiness and surprise and implementing them in a virtual reality environment.”

The researchers designed three virtual environments — happy, neutral and scary — and measured each person’s cognitive and physiological state while they were immersed in each scenario.

By reproducing three universal facial expressions — a smile, frown and a clench — they explored whether changes in the environment triggered one of the three expressions, based on emotional and physiological responses.

For example, in the happy environment, users were tasked with moving through a park to catch butterflies with a net. The user moved when they smiled and stopped when they frowned.

In the neutral environment, participants were tasked with navigating a workshop to pick up items strewn throughout. The clenched jaw triggered an action — in this case picking up each object — while the start and stop movement commands were initiated with a smile and frown.

The same facial expressions were employed in the scary environment, where participants navigated an underground base to shoot zombies.

“Overall, we expected the handheld controllers to perform better as they are a more intuitive method than facial expressions, however people reported feeling more immersed in the VR experiences controlled by facial expressions.”

Prof Billinghurst says relying on facial expressions in a VR setting is hard work for the brain but gives users a more realistic experience.

“Hopefully with some more research we can make it more user friendly,” he says.

In addition to providing a novel way to use VR, the technique will also allow people with disabilities — including amputees and those with motor neurone disease — to interact hands free in VR, no longer needing to use controllers designed for fully abled people.

Researchers say the technology may also be used to complement handheld controllers where facial expressions are a more natural form of interaction.

CITATION:

Dey et al. Effects of interacting with facial expressions and controllers in different virtual environments on presence, usability, affect, and neurophysiological signals. International Journal of Human – Computer Studies. Vol. 160, April 2022, p. 102762. doi: 10.1016/j.ijhcs.2021.102762.

This entry was posted in Presence in the News. Bookmark the permalink. Trackbacks are closed, but you can post a comment.

Post a Comment

Your email is never published nor shared. Required fields are marked *

You may use these HTML tags and attributes: <a href="" title=""> <abbr title=""> <acronym title=""> <b> <blockquote cite=""> <cite> <code> <del datetime=""> <em> <i> <q cite=""> <s> <strike> <strong>

*
*

  • Find Researchers

    Use the links below to find researchers listed alphabetically by the first letter of their last name.

    A | B | C | D | E | F| G | H | I | J | K | L | M | N | O | P | Q | R | S | T | U | V | W | X | Y | Z