- Main
Application of machine learning to signal entrainment identifies predictive processing in sign language.
Abstract
We present the first analysis of multi-frequency neural entrainment to dynamic visual features which drives sign language comprehension. Using the measure of EEG coherence to optical flow in video stimuli, we are able to classify fluent signers’ brain states as denoting online language comprehension, or non-comprehension during watching of non-linguistic videos that are equivalent in low-level spatiotemporal features and high-level scene parameters. The data also indicates that lower frequencies, such as 1 Hz and 4 Hz, contribute substantially to brain state classification, indicating relevance of neural coherence to the signal at these frequencies to language comprehension. These findings suggest that fluent signers rely on predictive processing during online comprehension.
Main Content
Enter the password to open this PDF file:
-
-
-
-
-
-
-
-
-
-
-
-
-
-