AUDITORY EEG SIGNAL PROCESSING (AESOP) SYMPOSIUM - SEPTEMBER 16-18, 2019 LEUVEN, BELGIUM

Page created by Greg Hale
 
CONTINUE READING
AUDITORY EEG SIGNAL PROCESSING (AESOP) SYMPOSIUM - SEPTEMBER 16-18, 2019 LEUVEN, BELGIUM
Auditory
EEG Signal Processing (AESoP)
         Symposium

         September 16-18, 2019
            Leuven, Belgium
AUDITORY EEG SIGNAL PROCESSING (AESOP) SYMPOSIUM - SEPTEMBER 16-18, 2019 LEUVEN, BELGIUM
AUDITORY EEG SIGNAL PROCESSING (AESOP) SYMPOSIUM - SEPTEMBER 16-18, 2019 LEUVEN, BELGIUM
Local Organisation
Tom Francart
Jan Wouters

Steering Committee
Alain de Cheveigné (CNRS, France)
Andrew Dimitrijevic (Sunnybrook Research Institute, Toronto, Canada)
Mounya Elhilali (Johns Hopkins, Maryland, USA)
Tom Francart (KU Leuven, Belgium)
Ed Lalor (University of Rochester, New York, USA)
Tobias Reichenbach (Imperial College London, United Kingdom)
Jonathan Simon (University of Maryland, USA)
Malcolm Slaney (Google Research, USA)
Jan Wouters (KU Leuven, Belgium)

Invited Speakers
Samira Anderson (University of Maryland, USA)
Gopala Krishna Anumanchipalli (University of California, San Francisco, USA)
Behtash Babadi (University of Maryland, USA)
Mathieu Bourguignon (BCBL, Spain)
Joachim Gross (University of Münster, Germany)
Karen Livescu (TTI-Chicago, USA)
Ross Maddox (University of Rochester, USA)
Myles Mc Laughlin (KU Leuven, Belgium)
Nima Mesgarani (Columbia University, USA)
Sarah Verhulst (Universiteit Gent, Belgium)

The symposium will be held under the auspices of

               AESoP symposium, Leuven, 16–18 September 2019                   i
Sponsors

                                           EEG
                                 REIMAGINED

ii    AESoP symposium, Leuven, 16–18 September 2019
Contents
 Organizers . . . . . . . . . . . . . . . . . . .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .     i
 Sponsors . . . . . . . . . . . . . . . . . . . .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .    ii
 Introduction . . . . . . . . . . . . . . . . . .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .    1
 Practical Information . . . . . . . . . . . . .    .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .    2
 Program . . . . . . . . . . . . . . . . . . . .    .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .    3
 Speaker Abstracts . . . . . . . . . . . . . .      .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .    7
      Signal Processing 1 . . . . . . . . . . .     .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .    7
      Deep Learning . . . . . . . . . . . . .       .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .    9
      Attention Decoding 1 . . . . . . . . . .      .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   13
      Attention Decoding 2 . . . . . . . . . .      .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   17
      Special Populations and Applications 1        .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   20
      Brainstem Responses and FFR . . . .           .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   23
      Signal Processing 2 . . . . . . . . . . .     .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   26
      Special Populations and Applications 2        .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   30
 Poster Abstracts . . . . . . . . . . . . . . .     .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   33

              AESoP symposium, Leuven, 16–18 September 2019                                                             iii
Introduction
Dear colleagues and friends,

It is our great pleasure to welcome you to Leuven. We hope you will not only enjoy
excellent science, but also great company, some history of our university and city,
and culinary delights.
Two years ago, when we discussed the idea or organising the first edition of the
AESoP symposium, we never thought it would be so successful as to organise a
next edition only 18 months later. The idea for this symposium was born from a
desire to bring together researchers from two fields that already interact, but could
do so even more: signal processing and neuroscience. With the wide availability
of multi-channel EEG/MEG systems and ever more powerful computers, it has be-
come possible to analyse brain signals more thoroughly than with the traditional
visual assessment of peaks in the time domain, and to do so with a single presen-
tation of natural running speech. A number of powerful quantitative methods for
EEG/MEG signal analysis are now available and under intensive further develop-
ment. The goals of the AESoP symposium are to further encourage development
of such methods, but also to make sure they are widely applied to tackle important
neuroscience questions. We therefore want to bring together engineers and neuro-
scientists (or any combination within the same person), for three days of intense
scientific exchange.
To stimulate discussion, we’ve allowed time for questions after each presentation,
plenty of breaks, and long poster sessions. The posters are an integral part of the
program, so we highly encourage viewing them all.
I would like to express my sincere thanks to everyone who contributed to the organ-
isation of this symposium: the steering committee, the enthusiastic Leuven team,
the invited speakers, and of course all participants for making this a success.

Yours sincerely,
Tom Francart,
supported by the steering committee:
Jan Wouters, Alain de Cheveigné, Andrew Dimitrijevic, Mounya Elhilali, Ed Lalor,
Tobias Reichenbach, Malcolm Slaney, Jonathan Simon

                AESoP symposium, Leuven, 16–18 September 2019                      1
Practical Information
Posters
To guarantee the smooth running of the poster sessions, we ask you to put up your
poster at the designated number as soon as you arrive. This number can be found
in the program book. Poster panels are 2 m high and 1 m wide. Tuesday after the
poster session, we kindly ask you to remove your poster.

    • The authors of odd-numbered posters should be present at their poster during
      the Monday poster session.

    • The authors of even-numbered posters should be present at their poster during
      the Tuesday poster session.

Presentations
As a speaker, you are requested to upload your presentation no later than during
the break preceding your session (see program). Speakers are kindly requested to
respect their presentation time to allow time for questions and keep on schedule.

Airport
To reach the airport from the symposium location you can take bus 2 (direction
Kessel-Lo) from bus stop “Leuven Sint-Michielskerk”. Alternatively you can take
one of the buses: 3, 4, 5, 6, 8, 9, 284, 285, 315, 316, 333, 352, 358, 370, 371, 373,
380, 395 from bus stop “Leuven Rector De Somerplein zone A”. In all cases you
need to get off at stop “Leuven Station” from where you can take train to Brussels
Airport—Zaventem (cost 9,30 EUR including Diabolo fee).
A bus ticket, valid for 60 minutes, costs 3 EUR and can be bought on the bus (or
cheaper from a machine or shop).
Alternatively, you can take a 20 min walk to from the symposium venue to the train
station.

2               AESoP symposium, Leuven, 16–18 September 2019
Program

                  Monday, September 16

11:00           Registration
12:00 - 13:00 Lunch
13:00 - 13:05 Welcome, Tom Francart
13:05 - 14:05 Signal Processing 1
              Moderator: Jonathan Simon
         13:05 Alain de Cheveigné, Tutorial: making the most of the data
               with linear methods
         13:35 Behtash Babadi, Neuro-current response functions: a uni-
               fied approach to MEG source analysis under the continuous
               stimuli paradigm
14:05 - 14:35 Break
14:35 - 16:25 Deep Learning
              Moderator: Malcolm Slaney
         14:35 Malcolm Slaney, The problems with attention decoding
         15:05 Karen Livescu, Deep multi-view representation learning: old
               and new approaches inspired by canonical correlation analysis
         15:35 Greg Ciccarelli, A neural network training approach for high
               accuracy, low latency auditory attention decoding
         15:55 Gopala Krishna Anumanchipalli, Physiology grounded deep
               networks for decoding speech and language from the brain
16:25 - 18:25 Poster Session

           AESoP symposium, Leuven, 16–18 September 2019                       3
Tuesday, September 17

     9:00 - 10:30 Attention Decoding 1
                  Moderator: Alain de Cheveigné
              9:00 Nima Mesgarani, Hierarchical encoding of attended auditory
                   objects in multi-talker speech perception
              9:30 Ali Aroudi, Binary-masking-based auditory attention decod-
                   ing without access to clean speech signals
              9:50 Emina Alickovic, An overview of auditory attention decipher-
                   ing methods
             10:10 Simon Geirnaert, A comparative study of auditory attention
                   decoding algorithms
    10:30 - 11:00 Break
    11:00 - 12:00 Attention Decoding 2
                  Moderator: Alain de Cheveigné
             11:00 Cantisani Giorgia, Decoding auditory attention in polyphonic
                   music based on EEG: a new dataset and a preliminary study
             11:20 Emily Graber, The focus of attention in music modulates
                   neural responses
             11:40 Lars Hausfeld, Cortical processing of distractor voices in nat-
                   ural auditory scenes depends on perceptual load
    12:00 - 13:00 Lunch
    13:00 - 14:25 Special Populations and Applications 1
                  Moderator: Jan Wouters
             13:00 Jan Wouters, Introduction
             13:05 Mathhieu Bourguignon, Speech tracking in noise: develop-
                   ment from childhood to adulthood
             13:35 Samira Anderson, Age-related deficits in neural processing
                   revealed for specific temporal components of speech
             14:05 Ben Somers, Temporal response functions in cochlear implant
                   users
    14:25 - 14:55 Break
                     continued on the next page

4              AESoP symposium, Leuven, 16–18 September 2019
14:55 - 16:35 Brainstem Responses and FFR
              Moderator: Tobias Reichenbach
         14:55 Tobias Reichenbach, Introduction
         15:05 Sarah Verhulst, Using computational models to improve the
               diagnostic sensitivity of auditory brainstem EEG
         15:35 Ross Maddox, Measuring the frequency-specific auditory
               brainstem response to naturalistic speech
         16:05 Jonathan Simon, High frequency cortical processing of con-
               tinuous speech in younger and older listeners
16:35 - 18:35 Poster Session
19:00          Dinner, symposium venue

          AESoP symposium, Leuven, 16–18 September 2019                     5
Wednesday, September 18

     9:00 - 10:40 Signal Processing 2
                  Moderator: Tom Francart
              9:00 Joachim Gross, Studying processing of continuous audiovisual
                   speech with MEG
              9:30 Christoph Daube, Slow and fast components in MEG re-
                   sponses to naturalistic speech
             10:00 Maryam Faramarzi Yazd, EEG correlates of illusory auditory
                   perception
             10:20 Ehsan Darestani Farahani, Source modeling of auditory
                   steady-state responses using minimum-norm imaging
    10:40 - 11:10 Break
    11:10 - 12:30 Special Populations and Applications 2
                  Moderator: Andrew Dimitrijevic
             11:10 Octave Etard, Neural speech tracking in the delta and theta
                   frequency bands differentially encodes comprehension and in-
                   telligibility of speech in noise
             11:40 Myles McLaughlin, Transcranial alternating current stimu-
                   lation: mechanisms and future directions
             12:10 Melissa Polonenko, Optimal parameters for obtaining robust
                   parallel auditory brainstem responses at different intensities
    12:30 - 12:40 Final Notes & Best Poster Award, Tom Francart
    12:40 - 14:00 Lunch
            14:00 End Symposium

6              AESoP symposium, Leuven, 16–18 September 2019
Speaker Abstracts

Monday, September 16: Signal Processing 1
13:05–14:05 Moderator: Jonathan Simon

Tutorial: making the most of the data with linear methods
Alain de Cheveigné (1,2)
(1) CNRS, École normale supérieure (University Paris Sciences Lettres); (2) UCL

EEG and MEG both record signals on multiple channels, electric or magnetic,
each of which mirrors the activity of billions of neural sources. The mixing pro-
cess is thought to be linear: each channel is the weighted sum of neural (and noise)
sources. It is thus no surprise that EEG and MEG data are most often analyzed
using linear methods involving spatial or temporal filters. Given the ubiquity of
linear operations, it is useful to reason in terms of the vector spaces spanned by the
data. The N time series recorded by the device together span a space of dimension
N, containing all their linear combinations, itself a subspace of the billion-dimension
space of neural activity. This tutorial discusses analysis methods (such as PCA,
ICA, CCA, etc.) in terms of their ability to find useful projections within these
spaces and subspaces. An important concept is dimensionality. Dimensionality de-
termines both the ability of a linear transform to separate signal from noise, and the
tendency to overfit of the data-driven methods that are used discover such trans-
forms. Greater dimensionality (more channels, time shifts, etc.) increases resolution
power. Lesser dimensionality (channel selection, PCA followed by truncation, etc.)
reduces overfitting. Progress in data analysis depends on finding an optimal path
between these conflicting requirements. I will review a number of ideas that may
help us move forward, to make the most of the data that we can record from the
brain.

                  AESoP symposium, Leuven, 16–18 September 2019                      7
Neuro-current response functions: a unified approach to MEG
source analysis under the continuous stimuli paradigm
Proloy Das (1,2), Christian Brodbeck (2), Jonathan Z. Simon (1,2,3),
Behtash Babadi (1,2)
(1) Department of Electrical & Computer Engineering, University of Maryland; (2) Institute for
Systems Research, University of Maryland; (3) Department of Biology, University of Maryland

Existing work suggests that certain features of speech, such as the acoustic en-
velope, can be used as reliable linear predictors of the neural response manifested
in MEG/EEG. The corresponding linear filters are referred to as temporal response
functions (TRFs). While the functional roles of specific components of the TRF are
well-studied, the cortical origins of the underlying neural processes are not as well
understood. Existing methods for obtaining cortical representations of the TRF
work in a two-stage fashion: either the TRFs are first estimated at the sensor level,
and then mapped to the cortex via source localization, or the data are first mapped
to the cortex followed by estimating TRFs for each of the resulting cortical sources.
Given that each stage is biased towards specific requirements, the end result typ-
ically suffers from destructive propagation of biases. In this work we address this
issue by introducing Neuro-Current Response Functions (NCRFs), which are three-
dimensionally oriented linear filters distributed throughout the cortex, that predict
the cortical current dipoles giving rise to the observed MEG (or EEG) data in re-
sponse to speech. We present a fast estimation algorithm and demonstrate its utility
through application to simulated and real MEG data under auditory experiments.
Our results show significant improvements over existing work, in terms of both spa-
tial resolution and reliance on post-processing steps such as clustering and denoising.

Acknowledgements: This work was supported by the National Science Foundation
(Award No. 1552946 and 1734892), the Defense Advanced Research Projects Agency
(Award No. N6600118240224) and the National Institutes of Health (Award No.
R01-DC014085).

8                AESoP symposium, Leuven, 16–18 September 2019
Monday, September 16: Deep Learning
14:35–16:25 Moderator: Malcolm Slaney

 The problems with attention decoding
Malcolm Slaney (1)
(1) Google Machine Hearing Research

I would like to review the status of attention decoding, and open a discussion about
the interesting scientific and engineering problems that remain. Attention decoding
is popular enough to be considered for real-life applications and here the engineer-
ing and machine-learning problems rule. But there are also a lot of questions about
how the brain processes sound and how to model attention. I would like to start a
discussion about the next steps, some of which I hope will be answered in the talks
to follow.

                 AESoP symposium, Leuven, 16–18 September 2019                    9
Deep multi-view representation learning: old and new ap-
proaches inspired by canonical correlation analysis
Karen Livescu (1)
(1) TTI-Chicago

Many scientific and engineering applications involve settings where we have access
to multiple types of sensor data, or multiple “views”, corresponding to the same
phenomenon. For example, we may have access to audio, video, and physiological
measurements of a person speaking. In such settings, it is often possible to learn
better representations (features) of each view by taking advantage of the statistical
relationships between the multiple views. One classic approach for learning linear
projections of two views is canonical correlation analysis (CCA). However, for many
types of data it is preferable to learn non-linear features. This talk will review several
approaches, including deep and variational CCA, for learning non-linear representa-
tions in a multi-view setting. It will also include new extensions tailored for sequence
data. Most of the experiments will be on speech recognition, but the techniques are
broadly applicable to any multi-view learning setting.

10                AESoP symposium, Leuven, 16–18 September 2019
A neural network training approach for high accuracy, low
latency auditory attention decoding
G. Ciccarelli (1), S. Haro (2), C. Calamia (1), M. Brandstein (1), T.
Quatieri (1,2), C. Smalt (1)
(1) Bioengineering Systems and Technologies Group, MIT Lincoln Laboratory, Lexington; (2) Speech
and Hearing Bioscience and Technology, Harvard Medical School, Boston

Auditory attention decoding (AAD) is the process of determining to which acoustic
source a listener desires to attend. High accuracy and low latency decoding is a key
requirement of a brain computer interface based hearing aid or hearing enhance-
ment device. To date, research into noninvasive, wearable electroencephalography
(EEG) AAD has concentrated primarily on improving decoding accuracy and under-
standing how decoding accuracy declines with smaller segments of evaluation data.
However, developing techniques to improve response time to attention switching
in an acoustic environment with multiple simultaneous speakers (the cocktail party
problem) is just as important as improving decoding accuracy. In this work, we show
how artificially introducing attention switches during training cannot only increase
response times but also improve accuracy relative to standard training procedures.
Our highest decoding accuracy is 81 % with a 3.75 s attentional switching latency
compared to a baseline system of 79 % with a 5.03 s latency. Our most responsive
system has an accuracy of 68 % and latency of 1.19 s compared to a baseline sys-
tem accuracy of 63 % and latency of 1.24 s. Future work will evaluate this training
paradigm with directed and at-will switches of auditory attention.

Acknowledgements: This material is based upon work supported by the Under
Secretary of Defense for Research and Engineering under Air Force Contract No.
FA8702-15-D-0001.

                  AESoP symposium, Leuven, 16–18 September 2019                              11
Physiology grounded deep networks for decoding speech and
language from the brain
Gopala K. Anumanchipalli (1), Josh Chartier (1), Pengfei Sun (1), Ed-
ward F. Chang (1)
(1) University of California, San Francisco

Spoken communication is basic to who we are. Neurological conditions that re-
sult in loss of speech can be devastating for affected patients. This work will sum-
marize recent efforts in decoding neural activity directly from the surface of the
speech cortex during fluent speech production, monitored using intracranial electro-
corticography (ECoG). Decoding speech from neural activity is challenging because
speaking requires very precise and rapid multi-dimensional control of vocal tract ar-
ticulators. We first describe the articulatory encoding characteristics in the speech
motor cortex and compare them against other representations like the phonemes.
We then describe deep learning approaches to convert neural activity into these
articulatory physiological signals that can then be transformed into audible speech
acoustics or decoded to text. In closed vocabulary tests, listeners could readily
identify and transcribe speech synthesized from cortical activity. We show that the
described biomimetic strategies make optimal use of available data; generalize well
across subjects, and also demonstrate silent speech decoding. These results set a
new benchmark in the development of high performance Brain-Computer Interfaces
for assistive communication in paralyzed individuals with intact cortical function.

12                AESoP symposium, Leuven, 16–18 September 2019
Tuesday, September 17: Attention Decoding 1
9:00–10:30 Moderator: Alain de Cheveigné

 Hierarchical encoding of attended auditory objects in multi-
talker speech perception
Nima Mesgarani (1,2), James O’Sullivan (1,2)
(1) Zuckerman Mind Brain Behavior Institute, Columbia University; (2) Electrical Engineering De-
partment, Columbia University

Humans can easily focus on one speaker in multi-talker acoustic environments. How
different areas in the human auditory cortex (AC) represent the acoustic components
of mixed speech is unknown. We recorded invasively from primary and nonprimary
AC in neurosurgical patients as they listened to multi-talker speech. We found
that neural sites in primary AC responded selectively to the individual speakers
in the mixture and were relatively unchanged with attention. In contrast, neural
sites in nonprimary AC were less discerning of individual speakers but selectively
represented the attended speaker. Moreover, the encoding of the attended speaker
in nonprimary AC was insensitive to the varying degree of acoustic overlap with
the unattended speaker. Finally, this emergent representation of attended speech
in nonprimary AC was linearly predictable from the responses in primary AC. Our
results reveal the neural computations underlying the hierarchical formation of au-
ditory objects in human AC during multi-talker speech perception.

                  AESoP symposium, Leuven, 16–18 September 2019                              13
Binary-masking-based auditory attention decoding without
access to clean speech signals
Ali Aroudi (1), Hendrik Kayser (1), Simon Doclo (1)
(1) Department of Medical Physics and Acoustics and Cluster of Excellence “Hearing4all”, University
of Oldenburg

Objectives: During the last decade several EEG-based auditory attention decoding
(AAD) methods have been proposed, which often require the clean speech signals of
the speakers as reference signals for decoding. However, in many applications these
clean speech signals are obviously not available in practice. In this contribution we
aim at generating appropriate reference signals for decoding from noisy and rever-
berant microphone signals in an acoustic scene with two competing speakers.

Methods: We propose a reference signal generation approach based on binary mask-
ing, which disregards the intervals mainly containing interfering speech and back-
ground noise. The binary masks are determined from the directional speech presence
probability of both speakers, which is estimated from the microphone signals. The
reference signals are generated by either masking the microphone signals or the out-
put signals of an MVDR beamformer. In addition, the estimated binary masks are
directly considered as reference signals for decoding.

Conclusions: The simulation results show that the proposed binary-masking-based
approach significantly improves the decoding performance (especially in the rever-
berant condition) compared to using the (non-masked) microphone and MVDR
output signals. In addition, the results show that directly using the estimated bi-
nary masks as reference signals yields a decoding performance that is comparable
to using the masked MVDR output signals.

Acknowledgements: This work was supported by the Deutsche Forschungsgemein-
schaft (DFG, German Research Foundation)—Project ID 390895286 EXC 2177/1.

14                AESoP symposium, Leuven, 16–18 September 2019
An overview of auditory attention deciphering methods
Emina Alickovic (1,2), Thomas Lunner (1,2,3,4), Fredrik Gustafsson (1),
Lennart Ljung (1)
(1) Linkoping University, Department of Electrical Engineering; (2) Oticon A/S, Eriksholm Research
Centre; (3) Technical University of Denmark, Department of Health Technology, Hearing Systems;
(4) Linkoping University, Swedish Institute for Disability Research, Linnaeus Centre HEAD

Speech communication in everyday listening environments often requires selective
attending to a particular talker amid a crowd of competing talkers and other un-
wanted sounds. Auditory attention deciphering (AAD) methods attempt to deci-
pher which of the talkers received the listener’s attention in such complex envi-
ronments by analyzing electrophysiological data—measured brain activity elicited
during selective listening. The main focus in AAD methods is on the linear re-
lations between electrophysiological and speech data. We present an overview of
the recent trends in multivariate correlation-based and model-based learning frame-
works that have been developed in recent decades as AAD methods to expose the
talker of a listener’s interest. The two frameworks compute these linear relations in
different ways. Whereas canonical correlation analysis (CCA)—a representative of
correlation-based framework—finds two different linear subsets, one that best cor-
relates electrophysiological to speech data and one that best correlates speech to
electrophysiological data, encoding and decoding approaches—two representatives
of model-based framework—focus on finding only one of these two subsets. Similar-
ities and differences between these frameworks together with different approaches
to estimate these subsets, for example, sparse vs. dense estimation, are discussed
so as to provide better understanding and practical use of state-of-the-art methods
available in the literature.

                  AESoP symposium, Leuven, 16–18 September 2019                                15
A comparative study of auditory attention decoding algo-
rithms
Simon Geirnaert (1,2), Servaas Vandecappelle (2,1), Tom Francart (2),
Alexander Bertrand (1)
(1) ESAT-STADIUS, Department of Electrical Engineering, KU Leuven; (2) Research Group Ex-
pORL, Department of Neurosciences, KU Leuven

Hearing aids and cochlear implants have major difficulties when operating in a so-
called cocktail party scenario. It is not only essential to separate a recorded mixture
of the speech or audio signals in its original contributions, but also to determine to
which speech signal the hearing aid user intends to listen. Recently, there has been
a growing interest in this second problem. In particular, it has been shown that the
brain synchronizes with the envelope of the attended speech signal. This creates
the possibility to decode the attention directly from the brain activity, recorded by,
e.g., electroencephalography sensors, which has lead to the development of different
auditory attention decoding (AAD) algorithms. In this study, we evaluate different
AAD algorithms on two independent datasets and compare and interpret the AAD
algorithms using, among others, a recently designed new metric (MESD). We take
their various characteristics into account, for example, by comparing linear (CCA,
regression-based, …) with nonlinear ((deep) neural networks) algorithms, backward
with hybrid models, different regularization techniques (ridge, sparse, …), etc.

Acknowledgements: This research is funded by an Aspirant Grant from the Research
Foundation Flanders (FWO) (for S. Geirnaert), the KU Leuven Special Research
Fund C14/16/057, FWO project nr. G0A4918N, the European Research Council
(ERC) under the H2020 programme (No. 802895 and No. 637424).

16              AESoP symposium, Leuven, 16–18 September 2019
Tuesday, September 17: Attention Decoding 2
11:00–12:00 Moderator: Alain de Cheveigné

 Decoding auditory attention in polyphonic music based on
EEG: a new dataset and a preliminary study
Giorgia Cantisani (1), Slim Essid (1), Gaël Richard (1)
(1) LTCI, Télécom Paris, Institut Polytechnique de Paris

In this work, we address the problem of EEG-based decoding of auditory attention
to a target instrument in realistic polyphonic music, i.e. recorded music, featuring
two or three instruments played concurrently. To this end, we exploit a stimulus
reconstruction model able to reconstruct an audio representation of the attended
instrument from single-trial multi-channel EEG recordings. This model was proven
to decode successfully the attention to speech in multi-speaker environments but, to
our knowledge, was never applied to musical stimuli. For this purpose, we acquired a
new dataset, named MAD-EEG, which consists of 20-channel EEG signals recorded
from 8 subjects listening to solo, duo and trio music excerpts and attending to one
pre-specified instrument in the mixture. It is worth noting that these stimuli were
not specifically designed to elicit ERPs, and were played to the subjects using loud-
speakers instead of headphones, rendering different spatial configurations. We were
thus able to investigate how the decoding performance is influenced by properties of
the musical stimuli, such as number and type of instruments, spatial rendering, mu-
sic genre, and the melody/rhythmical pattern that is played. The obtained results
are comparable to those obtained on speech data in previous works, and confirm
that it is thus possible to correlate the human brain’s activity with musically rele-
vant features of the attended source.

Acknowledgements: This project has received funding from the European Union’s
Horizon 2020 research and innovation program under the Marie Skłodowska-Curie
grant agreement No. 765068.

                  AESoP symposium, Leuven, 16–18 September 2019                   17
The focus of attention in music modulates neural responses
Emily Graber (1), Natalie Nguyen (1), Andrew Dimitrijevic (1)
(1) Sunnybrook Research Institute, Toronto

Objectives: Music is a complex stimulus often comprised of simultaneous voices
with multiple features per voice. Here, we used EEG to investigate how a listener’s
focus of attention to one or another musical voice in a duet changes the brain’s
representation of the voice/features under attention.

Methods: 18 NH volunteers, self-reportedly non-musicians, listened to 14 two-part
Bach Inventions passively, while watching a movie, and actively, while attending to
a specific musical voice. Multivariate temporal response function (TRF) encoders
were computed using the time-aligned features of pitch, meter, amplitude, and on-
sets for both musical voices during passive and active listening. Encoders were
compared between passive and active listening, and under active attention to the
upper and lower musical voices. Prediction accuracy will also be evaluated with
attended/unattended features separately under the hypothesis that attended ones
best model the EEG data.

Conclusions: Preliminary analyses suggest that active listening to any musical voice
increases the TRF amplitudes compared to passive listening. Currently, in our NH
participants, effects of attention are most evident in TRFs to the amplitude en-
velopes. Whether a similar relationship holds for participants with hearing loss can
be explored in the future. Overall, these data demonstrate the feasibility of using
musical stimuli in an attention paradigm.

Acknowledgements: Support provided by Oticon Medical.

18                AESoP symposium, Leuven, 16–18 September 2019
Cortical processing of distractor voices in natural auditory
scenes depends on perceptual load
Lars Hausfeld (1,2), Martha Shiell (1,2), Elia Formisano (1,2,3), Lars
Riecke (1,2)
(1) Department of Cognitive Neuroscience, Maastricht University; (2) Maastricht Brain Imaging
Centre (M-BIC), The Netherlands; (3) Maastricht Centre for Systems Biology, Maastricht Univer-
sity

Objectives: Selective attention is essential for processing multi-speaker auditory
scenes. It has been proposed that for visual selective attention the depth of process-
ing distractors is determined by the perceptual load of the stimulus. Whether this
“load dependency” exists in the auditory system is unclear.

Methods: We hypothesized that higher perceptual load would reduce the processing
resources available for the segregation of distractor sounds.
Participants (N = 20) were presented with auditory scenes including 3 speakers and
asked to selectively attended to one speaker. We varied perceptual load by spa-
tial cues (ITD = 0 and ITD 6= 0). EEG data was predicted by temporal response
functions including envelopes of the target speaker and either averaged or individ-
ual distractor speakers (AD or ID, respectively). We reasoned that hypothesized
reduced segregation for ITD = 0, would be reflected by higher differences in model
performance between AD and ID vs. ITD 6= 0.

Results: Conform with our hypothesis, results show that performance differences
between AD and ID models were significantly higher during ITD 6= 0 vs. ITD = 0
conditions. This effect was strongest in the theta band and at early delays (0–
200ms).

Conclusions: Our results indicate that during lower perceptual load, the cortex
represents individual distractor speech signals as more segregated. This suggests
that, in addition to purely acoustical properties, the processing of multiple distrac-
tor speakers depends on perceptual load.

Acknowledgements: This work was supported by the Netherlands Organisation for
Scientific Research (NWO; VENI grant 451-17-033 to L.H.).

                 AESoP symposium, Leuven, 16–18 September 2019                             19
Tuesday, September 17: Special Populations and
Applications 1
13:00–14:25 Moderator: Jan Wouters

 Speech tracking in noise: development from childhood to
adulthood
Mathieu Bourguignon (1,2,3)
(1) Laboratoire de Cartographie fonctionnelle du Cerveau, UNI-ULB Neuroscience Institute, Univer-
sité libre de Bruxelles; (2) Laboratoire Cognition, Langage et Développement, UNI-ULB Neuroscience
Institute, Université libre de Bruxelles; (3) BCBL, Basque Center on Cognition, Brain and Language,
Spain

During connected speech listening, oscillatory activity within auditory cortices tracks
speech rhythmicity at syllable (4–7 Hz) and phrasal (below 1 Hz) rates. In adults,
such speech brain tracking (SBT) is also observed in speech-in-noise (SiN) condi-
tions, and in that context, oscillatory brain activity aligns more to the attended
speech than to the heard sound. This talk will present recent studies relying on
SBT to better understand 1) the neuronal basis of the well-described difficulty of
children to perceive SiN, 2) the impact of noise properties and audiovisual integra-
tion on SiN perception, and 3) the relation between SiN perception and reading
abilities.

This research shows that tracking of SiN at phrasal and syllable rates parallels well
documented behavioral effects: SBT and SiN perception (i) are more compromised
by informational than non-informational masking, (ii) are enhanced when seeing lip
movements, and (iii) increase from age 6 to 12. Our results also revealed substantial
differences between SBT at phrasal and syllable rates: both maturation and seeing
lip movements boosted syllable tracking globally and phrasal tracking selectively in
informational noise conditions. Finally, these effects are behaviorally relevant since
the ability to maintain SBT at phrasal rate in SiN conditions and to leverage visual
speech information to boost it directly relate to reading abilities.

Acknowledgements: Mathieu Bourguignon is supported by the program Attract
of Innoviris (grant 2015-BB2B-10), by the Marie Skłodowska-Curie Action of the
European Commission (grant 743562) and by the Spanish Ministry of Economy and
Competitiveness (grant PSI2016-77175-P).

20                AESoP symposium, Leuven, 16–18 September 2019
Age-related deficits in neural processing revealed for specific
temporal components of speech
Samira Anderson (1)
(1) University of Maryland, Department of Hearing and Speech Sciences

Older adults often report that they have trouble with clarity rather than with loud-
ness when conversing with others. This lack of clarity may stem from decreased
neural processing of the temporal cues that help to distinguish one word from an-
other. Identification of neural processing deficits for specific temporal components
of speech may assist in developing management strategies for older adults who strug-
gle to hear. To that end, we recorded the frequency-following response (FFR) to
words that are distinguished by temporal duration cues and analyzed responses in
the time-frequency domain. We recruited normal-hearing middle-aged (55–70 yrs)
and young adult participants (18–30 yrs) and recorded FFRs to word pairs differing
in vowel duration (WHEAT vs. WEED) and consonant-transition duration (BEAT
vs. WHEAT). We also recorded auditory brainstem responses (ABRs) in quiet
and noise to obtain measures of auditory nerve and early brainstem function. We
found that age-related deficits were more pronounced for the word containing the
longer vowel duration (WEED) and for the word containing the shorter consonant
transition duration (BEAT). ABR measures of early neural function contributed to
variance in the neural representation of these temporal components, but only in
the middle-aged participants. These results have implications for management, and
efforts are underway to determine if training on the perception of these temporal
components leads to improved speech understanding.

Acknowledgements: This research was supported by the National Institute on Deaf-
ness and Other Communication Disorders of the National Institutes of Health under
Award Number R21DC015843.

                  AESoP symposium, Leuven, 16–18 September 2019                  21
Temporal response functions in cochlear implant users
Ben Somers (1), Eline Verschueren (1), Tom Francart (1)
(1) Research Group ExpORL, Department of Neurosciences, KU Leuven

Objective: EEG recorded while listening to natural running speech can be used
to estimate the Temporal Response Function (TRF), which models the encoding
of the speech envelope in the listener’s brain signals. The computation of TRFs
in cochlear implant (CI) users is complicated because of the electrical stimulation
artifacts, which are highly correlated with the speech envelope and distort TRFs.
The objective of this study is to compute TRFs in CI users, and assess the effect of
electrical stimulation artifacts and their removal on the TRFs.

Methods: EEG was measured from CI users while they listened to natural run-
ning speech. A previously-validated CI artifact removal technique was applied, in
which short gaps were inserted in the speech stimulus to create artifact-free EEG
samples. TRFs were computed using the resulting EEG and speech envelope, and
were compared with conditions where no CI artifact removal was applied.

Conclusions: Typical TRF morphologies were found for both artifact and artifact-
free conditions. Without CI artifact removal, TRF coefficients at short latencies
were dominated by artifacts, rendering these TRFs useless as a model to predict
EEG. This study demonstrates for the first time the possibility to compute TRFs in
CI users, and stresses the need for adequate CI artifact removal. TRFs and changes
in their morphologies can be used as an objective measure in future studies to eval-
uate encoding of speech in challenging listening conditions.

Acknowledgements: ERC H2020 grant No. 637424, and FWO PhD grants 1S46117N
and 1S86118N.

22              AESoP symposium, Leuven, 16–18 September 2019
Tuesday, 17 September: Brainstem Responses and
FFR
14:55–16:35 Moderator: Tobias Reichenbach

 Using computational models to improve the diagnostic sen-
sitivity of auditory brainstem EEG
Sarah Verhulst (1)
(1) Ghent University, Department of Information Technology

Auditory brainstem EEG has a long history in the diagnosis of sensorineural or
central hearing deficits, but because its dominant sources are generated in periph-
eral/brainstem structures, they offer only an indirect quantification of the underly-
ing hearing deficit. Head size, background noise, cochlear mechanics, and various
hearing deficits can all impact response amplitude and hence require (i) differential
paradigms in which confounding factors are minimized, and (ii), a thorough under-
standing of the transformation between the stimulus and the measured response.
Here, I will discuss how computational models of the auditory periphery can be
used to develop robust and sensitive EEG paradigms to isolate different aspects
of sensorineural hearing loss. Both stimulus characteristics and recording analysis
techniques are introduced and discussed on the basis of recordings from normal-
hearing and hearing-impaired listeners.

Acknowledgements: Work supported by the European Research Council (RobSpear;
678120).

                 AESoP symposium, Leuven, 16–18 September 2019                    23
Measuring the frequency-specific auditory brainstem response
to ongoing naturalistic speech
R. K. Maddox (1-3), M. J. Polonenko (2,3)
(1) University of Rochester, Department of Biomedical Engineering; (2) University of Rochester,
Department of Neuroscience; (3) University of Rochester, Del Monte Insitute for Neuroscience

Objectives: 1) To measure the auditory brainstem reponse to natural speech in
a way that yields canonical waveforms, 2) To use speech to separately assess hearing
across frequency bands.

Methods: Natural speech was resynthesized so that during periodic vowel segments
the phase of the harmonics were aligned so that they all crossed zero once per period,
leading to highly peaky waveforms. This was accomplished by extracting the fun-
damental frequency of the speech and synthesizing the harmonics with the desired
phase. The time-varying amplitude of each harmonic was taken from the spectro-
gram of the original speech. The result was a stimulus whose spectrogram was nearly
identical to the original speech, but was comprised of a series of pulses with known
times. These pulse times were then used to create an impulse train regressor and
derive the response waveform through deconvolution with the recorded EEG. This
process was extended to measuring separate responses across frequency bands by
resynthesizing speech several times at slightly different fundamental frequencies and
then filtering. Single-band responses were computed by deconvolution with each of
the bands’ respective pulse trains.

Conclusions: Natural speech can be processed and resyenthesized to elicit canonical
ABR waveforms with all standard component waves, with application in under-
standing the role of distinct subcortical nuclei in cognitive processes. This can be
extended for multi-band audiological purposes.

Acknowledgements: National Institute for Deafness and other Communication Dis-
orders, R00DC014288.

24                AESoP symposium, Leuven, 16–18 September 2019
High frequency cortical processing of continuous speech in
younger and older listeners
Joshua P. Kulasingham (1), Christian Brodbeck (2), Alessandro Pre-
sacco (2), Stefanie E. Kuchinsky (5), Samira Anderson (3), Jonathan Z.
Simon (1,2,4)
(1) Department of Electrical and Computer Engineering; (2) Institute for Systems Research; (3)
Department of Hearing and Speech Sciences; (4) Department of Biology, University of Maryland; (5)
Audiology and Speech Pathology Center, Walter Reed National Military Medical Center

Neural processing along the ascending auditory pathway shows a progressive re-
duction in frequencies. The frequency-following response (FFR) of midbrain, from
electroencephalography (EEG), is dominated by time-locked responses from ∼100
to several hundred Hz. In contrast, cortical speech responses, from EEG or mag-
netoencephalography (MEG), are dominated by time-locked responses of a few to
tens of Hz. However, fast (∼100 Hz) MEG responses time-locked to fast envelope
changes in speech have been reported. We investigate such MEG responses to con-
tinuous speech using neural source-localized temporal response functions (TRF)
analysis. Continuous speech stimuli were presented to 40 subjects (17 younger) and
fast MEG responses were analyzed. The spatiotemporal response profile indicates a
predominantly cortical origin with ∼35 ms latency and right hemisphere bias. TRF
analysis was performed using the: a) 70–300 Hz band of the waveform/carrier, and
b) 70–300 Hz envelope of the high frequency (300–4000 Hz) band; the latter was seen
to dominate. Age-related differences were analyzed to investigate a previously seen
reversal, whereby older listeners have weaker midbrain FFR responses, but, para-
doxically, stronger low frequency cortical responses. In contrast, this study found
no such age-related differences in high frequency cortical responses. In conclusion,
FFR-like cortical responses share properties with both midbrain at the same fre-
quencies and cortex at much lower frequencies.

Acknowledgements: Funding for this study was provided by the National Insti-
tute on Deafness and Other Communication Disorders (R01-DC014085), the Na-
tional Institute of Aging (P01-AG055365), and the National Science Foundation
(SMA1734892).

                  AESoP symposium, Leuven, 16–18 September 2019                               25
Wednesday, September 18: Signal Processing 2
9:00–10:40 Moderator: Tom Francart

 Using MEG to study the dynamics of information process-
ing in the human brain
Joachim Gross (1,2)
(1) University of Muenster, Institute for Biomagnetism and Biosignalanalysis; (2) University of
Glasgow, Centre for Cognitive Neuroimaging

There is a growing number of studies demonstrating a temporal reorganization of
human brain oscillations in response to complex quasi-rhythmic stimuli such as
speech. The reorganization is characterized by a temporal alignment of frequency-
specific brain activity to stimulus features. However, the differential contributions
of bottom-up and top-down processes to this alignment have remained largely un-
known. Furthermore, we are just beginning to understand what physical stimulus
features and what linguistic structures are entraining brain activity. Recent studies
suggest that this entrainment reflects cognitive processes of temporal coding, seg-
mentation and prediction that are orchestrated by hierarchically organized brain
oscillations. In my presentation I will give an overview of our recent studies in this
field and present new developments. I will discuss how rhythmic brain activity could
support the processing of complex, naturalistic stimuli and, ultimately, facilitates
human communication.

26                AESoP symposium, Leuven, 16–18 September 2019
Slow and fast components in MEG responses to naturalistic
speech
Christoph Daube (1), Joachim Gross (1,2), Robin A. A. Ince (1)
(1) University of Glasgow, Institute of Neuroscience and Psychology; (2) University of Muenster,
Institute of Biomagnetism and Biosignalanalysis

Over the recent years, countless studies have demonstrated that non-invasive record-
ings of brain responses to naturalistic speech constitute a rich experimental setup.
Often, such investigations focus on the low-frequency portion of the recorded signal
and explore how different features of the stimulus explain these responses. Con-
sistent effects for other parts of the neuronal responses are however more rarely
reported. Using an information theoretic approach building on Transfer Entropy
(TE), we here report evidence of a coupling signature at faster frequencies.
In TE, the idea is to describe the delayed relationship between a source and a target
variable while accounting for auto-correlations in the target. We find that in simplis-
tic scenarios with spectrally located and strongly auto-correlated signals, classic TE
formulations fail in recovering simulated effect sizes and delays. The combination
of a time-delayed non-uniform embedding and local conditional entropy however
mitigates these problems. Applying it to a passive story-listening MEG dataset,
we first find the known low-frequency (
EEG correlates of illusory auditory perception
Maryam Faramarzi Yazd (1,2), André Aleman (2), Branislava Ćurčić-
Blake (2), Christoph S. Herrmann (1)
(1) Experimental Psychology Laboratory, University of Oldenburg; (2) University Medical Center
Groningen

Objectives: We employed an auditory signal detection task to investigate the neural
markers in electroencephalogram (EEG) recordings corresponding to illusory voice
perception and compare it to neural correlates of detecting near-threshold voices in
noise. More specifically, we contrasted hits to trials classified as false alarms which
allowed us to investigate the neural mechanisms that are at least in part indepen-
dent of the acoustic properties of the external auditory stimuli.

Methods: In order to address this question, we recorded EEG in 21 normal-hearing
participants while they were performing an adapted speech recognition in noise
(SRN) task for 30 minutes, composed of 180 randomly presented 1s speech snippets
at near threshold intensity masked by a continuous steady state speech-shaped noise.

Conclusions: Neuro-oscillatory responses corresponding to illusory perceptions can-
not be fully captured by ERPs due to the absence of external stimuli and lack of
information in the exact onset of elicited perceptual processes. Event-related spec-
tral perturbation (ERSP) suffers less from this ambiguity. The performed analysis
revealed an early fronto-central lower theta subband and a later midparietal delta
power enhancement in hits. A similar pattern, although less prominently, was also
found in false alarms suggesting that the observed neural activity leading to the
conscious perception can be elicited in both absence and presence of the external
auditory stimuli.

28               AESoP symposium, Leuven, 16–18 September 2019
Source modeling of auditory steady-state responses using
minimum-norm imaging
Ehsan Darestani Farahani (1), Jan Wouters (1), Astrid van Wieringen (1)
(1) Research Group ExpORL, Department of Neurosciences, KU Leuven

Objectives: As auditory steady-state responses (ASSRs) have a wide range of clini-
cal and research applications, there is a great interest to investigate the underlying
neural generators. The main sources of ASSRs are located bilaterally in the audi-
tory cortex (AC), although some studies avoiding prior assumptions regarding the
number and location of the sources have also reported activity of sources outside
the AC. However, little is known about the number and location of these sources.
We present a novel extension to minimum-norm imaging (MNI) which facilitates
ASSR source reconstruction and provides a comprehensive and consistent picture of
sources.

Methods: MNI was applied to the preprocessed EEG data and a source distri-
bution map was obtained for each time point. Subsequently, the source map was
transformed into frequency domain and the ASSRs were calculated for each dipole
in order to develop the ASSR map.

Conclusions: Results demonstrate that the proposed MNI approach is successful
in reconstructing sources located both within (primary) and outside (non-primary)
of the AC. The non-primary sources are consistent with, and extends the literature.
The primary sources are detected in every experimental conditions indicating the
robustness of the approach. Furthermore, we demonstrate that the MNI approach is
capable of reconstructing the subcortical activities of ASSRs. Lastly, the results in-
dicate that the MNI approach outperform the previously used method of group-ICA.

Acknowledgements: Our special thanks go to Dr. Tine Goossens for sharing the
data used in this work. This work was supported by the Research Council, KU Leu-
ven through project OT/12/98 and by the Research Foundation Flanders through
FWO-project ZKC9024 and FWO-project ZKC5655.

                AESoP symposium, Leuven, 16–18 September 2019                      29
Wednesday, September 18: Special Populations and
Applications 2
11:10–12:30 Moderator: Andrev Dimitrijevic

  Neural speech tracking in the delta and theta frequency
bands differentially encodes comprehension and intelligibil-
ity of speech in noise
Octave Etard (1), Tobias Reichenbach (1)
(1) Imperial College London, Department of Bioengineering

Objectives: Speech processing may be aided by cortical activity in the delta and
theta frequency bands that tracks the speech envelope. Change of neural responses
with the clarity and comprehension of speech have previously been established, but
it remains unclear which aspects of neural speech tracking represent the processing
of acoustic features, related to speech clarity, and which aspects reflect higher-level
linguistic processing related to speech comprehension.

Methods: We employed EEG to record neural responses of native English speakers
listening to continuous speech in varying conditions of noise. We obtained EEG
recordings in response to English stimuli as well as in response to a foreign unknown
language (Dutch). When listening to English the subjects’ comprehension was mod-
ulated by the noise level, but remained nil in the matched Dutch conditions. This
allowed us to separate neural correlates of changes in the acoustic properties of the
stimuli (clarity) and of speech comprehension. We used regularised linear spatio-
temporal models to relate clarity and comprehension to the neural responses.

Conclusion: We were able to predict speech comprehension and clarity in the dif-
ferent acoustic conditions based on envelope tracking. We investigated the relative
importance of the delta, theta and alpha frequency bands. We found that cortical
tracking in the theta frequency band is mainly correlated to clarity, while the delta
band contributed most to speech comprehension.

30                AESoP symposium, Leuven, 16–18 September 2019
Transcranial alternating current stimulation: mechanisms
and future directions
Myles Mc Laughlin (1), Boateng Asamoah (2), Ahmad Khatoun (3)
(1) Research Group ExpORL, Department of Neurosciences, KU Leuven

Transcranial alternating current stimulation (tACS) is a noninvasive neuromodu-
lation method which has been shown to modulate hearing, motor, cognitive and
memory function. However, the mechanisms underpinning these findings are con-
troversial, as studies show that the current reaching the cortex may not be strong
enough to entrain neural activity.

In this talk we present data that highlights a previously overlooked mechanism
through which tACS may work, and based on this propose a new minimally invasive
tACS approach.

Firstly, we suggest that some tACS effects are actually caused by transcutaneous
stimulation of peripheral nerves in the skin and not transcranial stimulation of cor-
tical neurons. Rhythmic activity from peripheral nerves then entrains cortical neu-
rons. We present data from a series of experiments in rats and humans that isolated
the transcranial and transcutaneous mechanisms and showed that the reported ef-
fects of tACS on the motor system can be caused by transcutaneous stimulation of
peripheral nerves.

Secondly, we propose that some of these issues can be overcome with a novel mini-
mally invasive approach where an electrode is implanted under the skin, directly on
the skull. We present data from animal and computational modeling studies that
support the feasibility of this novel approach.

                AESoP symposium, Leuven, 16–18 September 2019                     31
Optimal parameters for obtaining robust parallel auditory
brainstem responses at different intensities
Melissa J. Polonenko (1), Ross K. Maddox (1,2)
(1) University of Rochester Medical Center, Department of Neuroscience; (2) University of Rochester,
Department of Biomedical Engineering

Objectives: To evaluate 1) how presentation rate and level interact to affect auditory
brainstem responses collected in parallel (pABR); and 2) the optimal parameters to
generate pABR responses with maximal signal-to-noise ratios (SNR) in minimal
recording times.

Methods: Two-channel pABRs for 0.5, 1, 2, 4 and 8 kHz tonebursts were mea-
sured for both ears of 20 adults with normal hearing. Twelve interleaved conditions
were recorded in 2 hours and comprised 6 rates from 20–120 stimuli/s at 75 and
45 dBpeSPL. Responses were analyzed by cross-correlating the rectified impulse
trains with the raw EEG and then averaged by weighting each epoch by the inverse
variance of its pre-stimulus baseline.

Results: Robust canonical responses were present and at both intensities and showed
adaptation at higher rates. Latencies and amplitudes of wave V characteristically
lengthened and decreased with a decrease in level respectively. A 40 Hz rate was
optimal for yielding 0 dB SNR responses within
Poster Abstracts
 P1. On miniaturization effects and optimal channel selection
for auditory attention decoding with EEG sensor networks
Abhijith Mundanad Narayanan (1), Panagiotis Patrinos (1), Alexander Bertrand (1)
(1) ESAT-STADIUS, Department of Electrical Engineering, KU Leuven

Chronic EEG monitoring in daily life will require highly miniaturized EEG sensors, which by de-
sign can only record within a small scalp area to avoid wires running over large distances across
the head. To increase the spatial coverage, we envision a modular platform where a multitude of
such wireless mini-EEG sensors are deployed, forming a so-called “wireless EEG sensor network” or
WESN. We explore methods to identify the optimal scalp locations to position these sensors and
analyze the performance loss due to miniaturization effects in an auditory attention decoding (AAD)
task. To this end, candidate WESN nodes were created by pairing each electrode of a standard
64-channel EEG cap with its neighbors at
P3. Wave V eABR amplitude and latency in response to
multi-pulse train stimulation
Ludwig Englert (1), Ali Saeedi (1), Werner Hemmert (1)
(1) Technical University of Munich, Department of Electrical Engineering and Computer Engineer-
ing, Bio-Inspired Information Processing
The electrically-evoked auditory brainstem responses (eABR) are signals which can be excited in
cochlear impant (CI) users and are considered as helpful in diagnostic measures. In contrast to the
typically used single pulses for eABR measurements in this study multi-pulse trains are used due to
its similarity to the stimulus pattern used in clinics, whereby its still possible to evaluate the eABR
signal.

In this work thresholds (THRs) and mostcomfortable Levels (MCLs) were obtained for each sub-
ject and multi-pulse condition. Datapoints were defined on DR, which is the difference of MCL and
THR for each multi-pulse condition. For each datapoint the eABR signal was measured and evalu-
ated. The dataset for each multi-pulse condition was used to analyse the latency-intensity function
and the amplitude-growth function.

The latency-intensity function shows a decreasing pattern for higher stimulation amplitudes across
all multi-pulse conditions. The slope of the amplitude-growth function decreases for higher number
of pulses. The multi-pulses as used in this work are delivered in a time span in order of the refractori-
ness period of a single neuron. Therefore these results provide more information about the electrical
excitation of auditory nerve fibers and the progression of the signal through the auditory brainstem.
Furthermore the data is used in a subsequent study to estimate clinical THRs.

P4. Estimation of clinical thresholds by means of eABR
thresholds in response to multi-pulse trains in cochlear im-
plant users
Ali Saeedi (1), Ludwig Englert (1), Werner Hemmert (1)
(1) Technical University of Munich, Department of Electrical Engineering and Computer Engineer-
ing, Bio-Inspired Information Processing

It is known that behavioural thresholds (THR) depend not only on the stimulation amplitude but
also on stimulation rate. When stimulating with single pulses, eABR THRs show strong correlation
with behavioural THRs. However, for high-rate stimulation, such as used for clinical purposes, the
correlation is weaker. In this study, we employed multi-pulse trains as stimuli to see if they can
estimate clinical THRs.

Multi-pulse trains of 1-, 2-, 4-, 8-, and 16-pulses with a repetition rate of 37 Hz were delivered
to a medial electrode of six CI users with MEDEL implants. Clinical THRs and MCLs were also
measure in response to stimuli of rate of 1000 pps. EABR amplitude growth functions (AGF) were
then measured starting from a level of 95 % of the DR down to a point where eABR waves amplitudes
were just above the noise floor. EABR THR was defined where the extrapolated fitted curve was zero.
Over all subjects, the median correlation between estimated eABR THR and psychophysical THR
was 0.79 and 0.78 for waves III and V, respectively. Generally, the estimated eABR THR decreased
with increasing number of pulses. The estimated eABR THRs for 1- and 2-pulse conditions were
still well-above clinical THRs (median difference = 147 and 101 µA, respectively for wave V). For
the 4-, 8-, and 16-pulse conditions, the estimated eABR THRs were much closer to the clinical THR
(median difference = −22, 4, and 12 µA, respectively for wave V).

34                 AESoP symposium, Leuven, 16–18 September 2019
You can also read