Investigating neural speech processing with functional near infrared spectroscopy: considerations for temporal response functions

This study demonstrates that temporal response functions (TRFs), a method traditionally used for EEG and MEG, can be successfully applied to fNIRS data to yield statistically significant and meaningful neural responses during continuous speech perception, outperforming conventional GLM approaches.

Original authors: Wilroth, J., Sotero Silva, N., Tafakkor, A., de Avo Mesquita, B., Ip, E. Y. J., Lau, B. K., Hannah, J., Di Liberto, G. M.

Published 2026-03-23
📖 3 min read☕ Coffee break read
⚕️

This is an AI-generated explanation of a preprint that has not been peer-reviewed. It is not medical advice. Do not make health decisions based on this content. Read full disclaimer

Imagine your brain is a bustling city, and when you listen to someone speak, different neighborhoods (brain areas) light up to process the sounds, words, and meaning. Scientists have long used tools like EEG (which measures electrical sparks) and fNIRS (which measures blood flow) to watch this city come alive.

The Problem: The "Slow Motion" Camera
For a long time, researchers studying how we understand speech used a method called Temporal Response Functions (TRFs). Think of TRFs as a high-speed camera that can track the brain's reaction to every single syllable of a sentence in real-time. This works great with EEG, which is like a camera with a super-fast shutter speed.

However, fNIRS is a bit different. It measures blood flow, which is the brain's way of delivering oxygen to active areas. But blood flow is slow—it's like watching a movie in slow motion. Because the signal is sluggish, scientists weren't sure if they could use the "high-speed" TRF method on fNIRS data. It felt like trying to catch a hummingbird's wingbeat with a camera designed for a sloth.

The Experiment: A Hyperscanning Party
In this study, the researchers decided to test if this "slow-motion" camera could actually keep up with the "fast-paced" speech. They set up a hyperscanning session (imagine a group of friends sitting together, all wearing special headgear at the same time). Eight people listened to continuous speech while their brains were scanned with fNIRS.

Instead of just looking at the brain's reaction to isolated words (like a standard test), they used the TRF method to see if they could map how the brain tracked the entire flow of the conversation, just like the fast cameras do.

The Results: A Surprising Success
The results were fantastic! Here is what they found, using some simple comparisons:

  1. It Works: The TRF method worked perfectly on the slow blood-flow data. It was like discovering that even though your sloth is slow, it can still run a perfect marathon if you give it the right map.
  2. High Quality: The connection between the speech and the brain's reaction was very strong. In fact, the clarity of the signal was just as good as what you get from the super-fast EEG cameras and comparable to the even more expensive MEG machines.
  3. Better Than the Old Way: When they compared this new TRF method to the traditional way of analyzing fNIRS data (called GLM), the TRF method was like a sharp, high-definition lens compared to the old method's blurry, standard-definition lens. It explained more of the brain's activity and gave clearer answers.

The Bottom Line
This paper is a green light for researchers. It proves that you don't need a super-fast electrical camera to study how we understand speech in real-time. Even the "slow-motion" blood-flow camera (fNIRS) can capture the brain's conversation with the world, provided you use the right mathematical tools (TRFs) to interpret the data. This opens the door for studying speech in more natural, moving, and real-world settings where other tools might fail.

Drowning in papers in your field?

Get daily digests of the most novel papers matching your research keywords — with technical summaries, in your language.

Try Digest →