Why
-
If you can’t speak in a classroom, venue, etc., you want to interact without speaking out
-
Tell the story without pronunciation
Background
-
Researchers at Oxford University and Google DeepMind have developed artificial intelligence (AI) that reads the movement of a person’s lips through training with thousands of hours of content broadcast by the BBC.
-
Little research on facial muscle activity
-
Difficult to discriminate micro electromyographic activity
-
Obtained action potential is a continuous waveform that fluctuates in time and is difficult to analyze usually.
How
Different mouth shape when speaking different vowels
By collecting masticatory EMG data when speaking different vowels(chin–plus,cheek–minus,forehead–reference)
Result
-
I analyzed the EMG data by analyzing envelope of different signals, different signals can be easily distinguished
-
By using a simple neural network from MATLAB, different EMG signals of different vowels can be correctly distinguished.