Robohub.org
 

Signal processing technology to extract required information from image, sound, and biometric signals

by
26 May 2014



share this:

At Keio University, the Mitsukura Laboratory, in the Department of System Design Engineering, researches how to extract required information from biometric, image, audio signal data. To achieve this, the researchers use technologies such as signal processing, machine learning, pattern recognition, artificial intelligence, and statistical processing.

“Our research has three research projects : image processing, audio signal processing, and biometric signal processing. In image processing, we’re researching how to embed and extract motion. In audio signal processing, we construct the systems which divide into the scores automatically. And in biometric signal processing, we determine the meaning of brain waves(EEG) and convert thoughts to text.”

In image signal processing research, the Mitsukura Lab works on position matching between real and virtual space, which is the basis for AR technology. These researchers are developing a method for estimating the orientation of a face at high speed with high precision, to link the motion of an actual person with an animation. Another research topic at the Mitsukura Lab is expression recognition, using information about changes in the eyebrows and corners of the mouth. The Lab is also researching how to overlay virtual data used in head-mount displays in the real world.

In audio signal processing research, we studied the ability to automatically determine which direction a sound is coming from, by using multiple mic signals. The researchers are developing a system to automatically direct the mic toward the person talking in a videoconference, and a robot to track a talking person. The Lab is also working to develop a system for inputting information by whistling.

In research on biometric signal processing, the Lab is developing a system to operate computers and devices by reading brain waves(EEG) and EOGs.

In brain-wave(EEG) signal processing, the Lab is doing a wide range of R&D. One research project is a system to determine the user’s likes / dislikes and convert them to text form. Another research project is to evaluate sound quality from brain waves(EEG).

Signal processing technology using EOGs, obtained around the eyes, is being utilized in R&D on steering a wheelchair by winking.

Signal processing technology is expected to have a diverse range of applications, including healthcare, entertainment, and wearable devices.

In the Mitsukura Lab, the researchers are using the many algorithms they’ve developed so far to advance such R&D further, in ways that contribute to society.

“At the present stage, for example, it’s said that Google Glass will become very popular, but with that, you have to wear glasses. Wearing glasses itself is unnatural for people. The same is true for brain-wave(EEG) systems; if something’s unnatural, people won’t do it for long. The first thing we have to do is create a system that can read brain waves(EEG) in a natural manner. Next, in conjunction with that, we need a system that can communicate emotions between people, in a way that resembles using the telephone to communicate thoughts. I think we might be able to communicate emotions even without a telephone.”



tags: , , ,


DigInfo TV is a Tokyo-based online video news platform dedicated to producing original coverage of cutting edge technology, research and products from Japan.
DigInfo TV is a Tokyo-based online video news platform dedicated to producing original coverage of cutting edge technology, research and products from Japan.





Related posts :



Open Robotics Launches the Open Source Robotics Alliance

The Open Source Robotics Foundation (OSRF) is pleased to announce the creation of the Open Source Robotics Alliance (OSRA), a new initiative to strengthen the governance of our open-source robotics so...

Robot Talk Episode 77 – Patricia Shaw

In the latest episode of the Robot Talk podcast, Claire chatted to Patricia Shaw from Aberystwyth University all about home assistance robots, and robot learning and development.
18 March 2024, by

Robot Talk Episode 64 – Rav Chunilal

In the latest episode of the Robot Talk podcast, Claire chatted to Rav Chunilal from Sellafield all about robotics and AI for nuclear decommissioning.
31 December 2023, by

AI holidays 2023

Thanks to those that sent and suggested AI and robotics-themed holiday videos, images, and stories. Here’s a sample to get you into the spirit this season....
31 December 2023, by and

Faced with dwindling bee colonies, scientists are arming queens with robots and smart hives

By Farshad Arvin, Martin Stefanec, and Tomas Krajnik Be it the news or the dwindling number of creatures hitting your windscreens, it will not have evaded you that the insect world in bad shape. ...
31 December 2023, by

Robot Talk Episode 63 – Ayse Kucukyilmaz

In the latest episode of the Robot Talk podcast, Claire chatted to Ayse Kucukyilmaz from the University of Nottingham about collaboration, conflict and failure in human-robot interactions.
31 December 2023, by





Robohub is supported by:




Would you like to learn how to tell impactful stories about your robot or AI system?


scicomm
training the next generation of science communicators in robotics & AI


©2024 - Association for the Understanding of Artificial Intelligence


 












©2021 - ROBOTS Association