IEEE Systems, Man and Cybernetics Magazine - April 2023 - 34

The feature functions fl : a R1 128
#
A # "
take as input the CA features a and v, respectively,
and location l and output a feature of size 1 128#
R1 128#
. The feature
outputs are combined at each location using the outer
matrix product; i.e., the bilinear combination of fA
and fV
at location l is given by
(, ,, ,) (, )( ,)
bilinear la vf ff la fl v
AV A
=
where a is ACA R1 128
#
"
BL (, )avF
"
T
V
and v is VCA R1 128
#
.
aims to aggregate the bilinear combination
of two cross-modal attention features at each pixel location
l along a temporal direction t [45]:
BLFbilinear la vf f
lL
(, )(,, ,, )
(, )( ,).
av
=
=
|
|
!
fl af lv
T
A
lL
!
V
The resulting feature BLF R1 256
#
"
(9)
is generated to capture
the multiplicative interaction at the corresponding
spatial location. The BLF is added after the audiovisual CA
layers fuse the audio and video attention features to generate
the fused feature ().Eav
SA
SA, also known as intra-attention, computes the response
at a position in a sequence by attending to all positions
within the same sequence [44]. We adopt SA to find a relevant
feature or long-term cue. It inputs the fused features
()Eav
(),Kav
and value (),Vav
AV
Experiments
Implementation Details, Model Complexity,
and Inference Time
The whole network was trained in an end-to-end manner.
The PyTorch library has been used to develop ADNet and
adopted the stochastic gradient descent optimizer. The initial
learning rate is set to 10 4(8)
and
:fl v# "
V
In this work, we considered AD as a frame-level classification
task. The cross-entropy loss is adopted to
compare the predicted sequence with the ground truth
sequence. The adopted loss function is illustrated in
(10), where pi
and y ,i
jN1!6
respectively, are the predicted
and the ground truth AD labels of the jth video frame,
,,
L=- |
j =1
@ and N refers to the number of video frames:
lo () ().glog
N yp yp
1
N
jj ij
+11(10)
and
decreases by 5% for
every epoch. One transformer layer with eight attention
heads is set for the CA and SA networks. We undertake
random flipping, rotation, and crop operations to augment
the images.
The model inference time was tested on an NVIDIA
GeForce GPU with 12-GB memory. It predicts a single face
within 1.45 ms. The complexity analysis [44], [46] of the
proposed frameworks is shown in Table 4.
Ablation Experiments
from the BLF to model the audiovisual utterance-level
temporal information. Apart from the query (),Qav
which come from the joint audiovisual
feature, the SA architecture is identical to the CA
network, shown in Figure 7(b). This module aims to distinguish
between addressing the robot and addressing other
subject frames.
Loss Function
Following the SA layer, a fully connected layer and
softmax operation are attached to project the output
features of the SA network to the AD label sequence.
Table 4. Per-layer complexity.
Layer Type
Fully connected layer
SA
Convolution
BLF
key
Complexity per Layer
On2
()
On d2
(. )
Ok nd2
()
(. .)
On2
n is the sequence length, d is the representation dimension, and k is the
kernel size of convolution.
BLF Versus Concatenation
We investigate the performance of the proposed model,
comparing the BLF and deterministic feature fusion
approaches to choose a better fusion approach. The
experiments were conducted in the same settings on the
E-MuMMER test for a fair comparison. The only difference
is that we replace the BLF with concatenation on
the baseline framework (ADNet). As shown in Table 5,
though the concatenation-based approach showed better
prediction results in terms of recall, we found out that
BLF showed a 1% accuracy improvement over the concatenation
approach and better precision. Unlike the concatenation
module, which blindly fuses heterogeneous
features, BLF considers the interaction of local features
between ACA and VCA features, capturing the pairwise
feature relationships and leading to a better classification
accuracy than the concatenation module.
Does the Attention Module Help?
We undertake the ablation experiment to observe the contribution
of the attention module, removing two attention
modules, one at a time. The results are shown in Table 6
and discussed as follows:
◆ Without CA: When the CA module is removed, the
model does not show a performance gain in terms of
Bal.Accu. We found that SA better distinguishes speech
34 IEEE SYSTEMS, MAN, & CYBERNETICS MAGAZINE April 2023

IEEE Systems, Man and Cybernetics Magazine - April 2023

Table of Contents for the Digital Edition of IEEE Systems, Man and Cybernetics Magazine - April 2023

IEEE Systems, Man and Cybernetics Magazine - April 2023 - Cover1
IEEE Systems, Man and Cybernetics Magazine - April 2023 - Cover2
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 1
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 2
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 3
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 4
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 5
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 6
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 7
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 8
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 9
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 10
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 11
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 12
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 13
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 14
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 15
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 16
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 17
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 18
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 19
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 20
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 21
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 22
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 23
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 24
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 25
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 26
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 27
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 28
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 29
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 30
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 31
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 32
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 33
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 34
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 35
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 36
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 37
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 38
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 39
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 40
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 41
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 42
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 43
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 44
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 45
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 46
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 47
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 48
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 49
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 50
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 51
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 52
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 53
IEEE Systems, Man and Cybernetics Magazine - April 2023 - 54
IEEE Systems, Man and Cybernetics Magazine - April 2023 - Cover3
IEEE Systems, Man and Cybernetics Magazine - April 2023 - Cover4
https://www.nxtbook.com/nxtbooks/ieee/smc_202310
https://www.nxtbook.com/nxtbooks/ieee/smc_202307
https://www.nxtbook.com/nxtbooks/ieee/smc_202304
https://www.nxtbook.com/nxtbooks/ieee/smc_202301
https://www.nxtbook.com/nxtbooks/ieee/smc_202210
https://www.nxtbook.com/nxtbooks/ieee/smc_202207
https://www.nxtbook.com/nxtbooks/ieee/smc_202204
https://www.nxtbook.com/nxtbooks/ieee/smc_202201
https://www.nxtbook.com/nxtbooks/ieee/smc_202110
https://www.nxtbook.com/nxtbooks/ieee/smc_202107
https://www.nxtbook.com/nxtbooks/ieee/smc_202104
https://www.nxtbook.com/nxtbooks/ieee/smc_202101
https://www.nxtbook.com/nxtbooks/ieee/smc_202010
https://www.nxtbook.com/nxtbooks/ieee/smc_202007
https://www.nxtbook.com/nxtbooks/ieee/smc_202004
https://www.nxtbook.com/nxtbooks/ieee/smc_202001
https://www.nxtbook.com/nxtbooks/ieee/smc_201910
https://www.nxtbook.com/nxtbooks/ieee/smc_201907
https://www.nxtbook.com/nxtbooks/ieee/smc_201904
https://www.nxtbook.com/nxtbooks/ieee/smc_201901
https://www.nxtbook.com/nxtbooks/ieee/smc_201810
https://www.nxtbook.com/nxtbooks/ieee/smc_201807
https://www.nxtbook.com/nxtbooks/ieee/smc_201804
https://www.nxtbook.com/nxtbooks/ieee/smc_201801
https://www.nxtbook.com/nxtbooks/ieee/systems_man_cybernetics_1017
https://www.nxtbook.com/nxtbooks/ieee/systems_man_cybernetics_0717
https://www.nxtbook.com/nxtbooks/ieee/systems_man_cybernetics_0417
https://www.nxtbook.com/nxtbooks/ieee/systems_man_cybernetics_0117
https://www.nxtbook.com/nxtbooks/ieee/systems_man_cybernetics_1016
https://www.nxtbook.com/nxtbooks/ieee/systems_man_cybernetics_0716
https://www.nxtbook.com/nxtbooks/ieee/systems_man_cybernetics_0416
https://www.nxtbook.com/nxtbooks/ieee/systems_man_cybernetics_0116
https://www.nxtbook.com/nxtbooks/ieee/systems_man_cybernetics_1015
https://www.nxtbook.com/nxtbooks/ieee/systems_man_cybernetics_0715
https://www.nxtbook.com/nxtbooks/ieee/systems_man_cybernetics_0415
https://www.nxtbook.com/nxtbooks/ieee/systems_man_cybernetics_0115
https://www.nxtbookmedia.com