Next Article in Journal
Mapping Pharmacologically Evoked Neurovascular Activation and Its Suppression in a Rat Model of Tremor Using Functional Ultrasound: A Feasibility Study
Previous Article in Journal
Research on the Improvement of Semi-Global Matching Algorithm for Binocular Vision Based on Lunar Surface Environment
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Zero-Shot Neural Decoding with Semi-Supervised Multi-View Embedding

1
Graduate School of Information Science and Technology, Hokkaido University, N-14, W-9, Kita-ku, Sapporo 060-0814, Hokkaido, Japan
2
Faculty of Information Science and Technology, Hokkaido University, N-14, W-9, Kita-ku, Sapporo 060-0814, Hokkaido, Japan
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(15), 6903; https://doi.org/10.3390/s23156903
Submission received: 15 April 2023 / Revised: 20 July 2023 / Accepted: 31 July 2023 / Published: 3 August 2023
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Zero-shot neural decoding aims to decode image categories, which were not previously trained, from functional magnetic resonance imaging (fMRI) activity evoked when a person views images. However, having insufficient training data due to the difficulty in collecting fMRI data causes poor generalization capability. Thus, models suffer from the projection domain shift problem when novel target categories are decoded. In this paper, we propose a zero-shot neural decoding approach with semi-supervised multi-view embedding. We introduce the semi-supervised approach that utilizes additional images related to the target categories without fMRI activity patterns. Furthermore, we project fMRI activity patterns into a multi-view embedding space, i.e., visual and semantic feature spaces of viewed images to effectively exploit the complementary information. We define several source and target groups whose image categories are very different and verify the zero-shot neural decoding performance. The experimental results demonstrate that the proposed approach rectifies the projection domain shift problem and outperforms existing methods.

1. Introduction

Neural decoding has enabled the interpretation of a person’s cognitive state from evoked brain activity. This attempt makes a significant contribution to the development of brain–computer interfaces (BCIs) [1] that would establish communication between computer systems and human brain activity. Several machine learning methods [2,3,4] have revealed what a person viewed from evoked brain activity. In these methods, functional magnetic resonance imaging (fMRI) activity patterns that were measured when a subject viewed several images (e.g., fish, chair, and face) were classified into a valid image category. Since these methods focused on the relationship between fMRI activity patterns and image categories used in the training phase, the predicted categories were restricted to trained categories only. It is not feasible to acquire fMRI activity patterns for all possible categories; therefore, the results of these methods are significantly limited.
To overcome this limitation, zero-shot neural decoding approaches attempt to decode viewed images [5,6,7,8,9,10] and meanings of presented words [11,12,13] that were not trained previously from corresponding brain activity. For example, the studies [6,7,8,14] estimated mappings between fMRI activity patterns and visual features that were extracted from the viewed images via convolutional neural networks (CNN) [15]. Then, visual features were predicted from fMRI activity measured when a subject viewed a novel category. Finally, the decoding was made feasible by comparing the predicted features with visual features from various categories. On the other hand, zero-shot learning, which aims to recognize novel classes without labeled training samples, suffers from the projection domain shift problem [16,17,18]. Zero-shot learning can be considered as transfer learning that transfers the knowledge from the training/source domain to the test/target domain. When two domains are potentially unrelated, the mappings learned from one source domain may not correctly capture the relationship of the target domain, and this is known as the projection domain shift problem (see Figure 1a). Furthermore, collecting brain activity patterns is a very laborious task since the measurement requires a heavy burden on subjects. Thus, the number of source categories is small, and source categories may be limited to certain categories only. Therefore, the projection domain shift problem is remarkable, resulting in poor decoding performance.
In this paper, we propose a zero-shot neural decoding approach with semi-supervised multi-view embedding. To rectify the projection domain shift problem, we introduce a semi-supervised framework that utilizes images related to the target categories without fMRI activity patterns. Our framework can be seen as domain adaptation relying on images from the target domain, which can be collected at a reasonable cost. Furthermore, we project fMRI activity patterns into a multi-view embedding space: the visual feature space and the semantic feature space. Specifically, visual features are extracted from images via CNN, and semantic features are extracted from image categories via a distributed word representation model. Our semi-supervised multi-view approach can consider different feature spaces that contain complementary information while introducing additional visual and semantic features extracted from images related to the target categories (see Figure 1b). Most importantly, fMRI activity patterns corresponding to additional visual and semantic features are not necessary for the proposed method.
Towards the realization of our approach, we construct a semi-supervised multi-view generative model. Our proposed model assumes that fMRI activity, visual features, and semantic features are generated from a shared latent variable under a Bayesian framework [19,20]. Furthermore, additional visual and semantic features are incorporated into the model to solve the projection domain shift problem. We consider unobserved fMRI activity patterns corresponding to additional features as missing values and estimate these missing values while optimizing model parameters. Figure 2 demonstrates the embedding space in the previous method [6] and the proposed model by plotting visual and semantic features predicted from fMRI activity (we used fMRI activity collected from subject 3 in the publicly available fMRI dataset [6]) onto two dimensions using t-SNE [21]. In the visual feature space of the method [6] illustrated in Figure 2a, the target category embeddings of artifact are biased towards the source categories of living thing and are separated from their prototypes. On the other hand, this bias is alleviated in the visual feature space of our method, and target category embeddings of artifact are projected properly in the semantic feature space illustrated in Figure 2b. We solve the projection domain shift problem by the semi-supervised approach and consider a better feature space by the multi-view embedding approach.
The main contributions of this paper are as follows:
  • To the best of our knowledge, this is the first study to address the projection domain shift problem in zero-shot neural decoding. For this problem, we introduce the semi-supervised framework that employs images related to the target categories without fMRI activity patterns.
  • We propose multi-view embeddings that associate fMRI activity patterns with visual features from images and semantic features from image categories.
  • We address the difficulty in collecting fMRI data and estimate unobserved fMRI activity patterns in a fully probabilistic manner [22]. Furthermore, the Bayesian framework of our model automatically selects a small set of appropriate components from high dimensional fMRI voxels.

2. Related Work

2.1. Multi-View Learning

Multi-view learning [23,24] is machine learning which considers learning with multiple views to improve the generalization performance and provide complementary information. Canonical correlation analysis (CCA) [25] is a classical but still powerful tool for analyzing multi-view paired samples. CCA finds projection matrices such that the correlation between projected paired samples in the shared latent space is maximized. Semi-supervised CCA (SemiCCA) [26] utilizes additional unpaired samples to prevent the performance degradation of CCA when the number of paired samples is limited. The CCA framework is also applied to neural decoding [7,14], where pairs of fMRI data and visual features of images that a subject viewed are associated. The method in [14] predicts visual features from fMRI data utilizing CCA to estimate the viewed image categories. The method in [7] introduces semi-supervised fuzzy discriminative CCA (Semi-FDCCA), where additional images not used in measuring fMRI data are utilized by SemiCCA and the similarity information of image categories is incorporated. However, one of the central problems in CCA is the model selection or how to select the dimensions to be retained [27]. In particular, since the number of dimensions of fMRI data is high, dimension reduction must be applied before input to the CCA, resulting in performance degradation.
The proposed model is constructed under the same framework as Bayesian canonical correlation analysis (BCCA) [19,27] and group factor analysis (GFA) [20]. BCCA is a generative model that links two observed variables via a shared latent variable, and GFA also treats more than two observed variables. BCCA and GFA introduce automatic relevance determination (ARD) [28] via the Bayesian approach, which automatically selects the appropriate dimensions from the data. Therefore, we can automatically select a small set of appropriate components from high dimensional fMRI voxels. The proposed model handles three observed variables (i.e., fMRI activity, visual features, and semantic features) and can be easily extended to semi-supervised learning. We also refer to the proposed model without the semi-supervised learning scenario as the multi-view generative model (MGM) [29].

2.2. Zero-Shot Learning

Zero-shot learning (ZSL) is the recognition of novel visual categories without labeled training samples for visual recognition. Inspired by humans’ ability to recognize a new object category without ever seeing a visual instance, ZSL aims to recognize a visual instance of a new category that has never been seen before [17]. Common ZSL methods learn a projection function from a visual feature space of images to a semantic embedding space (e.g., a semantic attribute space or a semantic word vector space) using the labeled training data consisting of seen classes only. At test time for recognizing unseen objects, this projection function is then used to project the visual representation of an unseen class image into the semantic embedding space. However, ZSL models mostly suffer from the projection domain shift problem [16]. That is, if the projection is learned only from the seen classes, the projections of unseen class images are likely to be misplaced (shifted) due to the bias of the seen classes [18]. Zero-shot neural decoding also suffers from the domain shift problem when projecting fMRI activity to the visual feature space and semantic feature space.

2.3. Neural Decoding

Neural decoding (or brain decoding) is the task of estimating images or words that a person sees or imagines from his/her brain activity (especially fMRI activity). In the traditional neural decoding methods [2,3,4], fMRI activity that was measured when a person viewed an image was classified into an image category, which enables the estimation of the viewed object. Subsequent neural decoding methods [5,6,7,8,9,10,11,12,13] estimate images or words that were not trained before, i.e., zero-shot neural decoding, by projecting fMRI activity into an intermediate-level feature space (e.g., visual feature space [6] and semantic feature space [12,13]). Recent neural decoding approaches [30,31,32,33] employ semi-supervised learning to utilize additional images without corresponding fMRI activity patterns. Beliy et al. [30] aimed to reconstruct high-quality images from evoked fMRI activity while viewing images. Their method utilizes unlabeled data (i.e., images without fMRI recording, and fMRI recording without images) to train fMRI-to-image reconstruction networks. Akamatsu et al. [31] and Du et al. [32] aimed to accurately estimate viewed image categories from evoked fMRI activity. They associate fMRI activity with visual features and semantic [31] or textual features [32] extracted from image categories. They also introduce semi-supervised learning by incorporating additional visual and semantic or textual features without fMRI activity. Liu et al. [33] proposed BrainCLIP, which unifies the visual stimulus classification and image reconstruction from fMRI activity. They leverage the semantic space of CLIP [34] learned from a large-scale vision–language corpus to perform neural decoding tasks. The method is pre-trained with large-scale unlabeled images without fMRI activity.
This paper is an extended version of the previous work [31]. This work aims to address the projection domain shift problem in zero-shot neural decoding, especially when the source domain and the target domain are potentially unrelated. Specifically, we define several source and target groups (e.g., mammal, device, and structure) and validate their decoding performances in zero-shot learning settings. In the previous works [31,32,33], image categories related to the target group were included in the source group (e.g., source: dolphin, airship, and beer mug; target: killer whale, airliner, and coffee mug, respectively). On the other hand, the target and source groups are very different in our setting; therefore, we are tackling a more challenging problem than in previous works.

3. Proposed Method

Our proposed method consists of the following three phases: the construction of semi-supervised multi-view generative model, the optimization of the model parameters, and the decoding of viewed image categories from evoked fMRI activity.

3.1. Semi-Supervised Multi-View Generative Model

Suppose that X ( f ) = [ x 1 ( f ) , , x N ( f ) ] R D f × N , X ( v ) = [ x 1 ( v ) , , x N ( v ) ] R D v × N , and X ( s ) = [ x 1 ( s ) , , x N ( s ) ] R D s × N represent fMRI activity patterns, visual features, and semantic features, respectively. Here, D f , D v , and D s denote the dimensions of each sample of X ( f ) , X ( v ) , and X ( s ) , respectively. N denotes the size of training samples. Visual features X ( v ) are extracted from viewed images via VGG19 [35], which is a CNN model that was used in a previous neural decoding study [36]. Semantic features X ( s ) are extracted from viewed image categories based on a word2vec model [37]. Furthermore, suppose that X add ( v ) = [ x N + 1 ( v ) , , x N + M ( v ) ] R D v × M and X add ( s ) = [ x N + 1 ( s ) , , x N + M ( s ) ] R D s × M represent additional visual and semantic features from images related to the target categories, respectively. Here, a set of additional visual and semantic features is obtained from each category, i.e., M denotes the number of additional image categories. Our approach considers unobserved fMRI activity patterns corresponding to additional visual and semantic features as missing values. These missing values X miss ( f ) = [ x N + 1 ( f ) , , x N + M ( f ) ] R D f × M are estimated by Bayesian inference [22] while we are optimizing the model parameters. Hereafter, the observed and missing variables are redefined as X ^ ( f ) = [ X ( f ) , X miss ( f ) ] R D f × ( N + M ) , X ^ ( v ) = [ X ( v ) , X add ( v ) ] R D v × ( N + M ) , and X ^ ( s ) = [ X ( s ) , X add ( s ) ] R D s × ( N + M ) . Note that [ · , · ] represents the horizontal concatenation. We also introduce latent variables Z = [ z 1 , , z N + M ] R D z × ( N + M ) that generate fMRI activity X ^ ( f ) , visual features X ^ ( v ) , and semantic features X ^ ( s ) , where D z is the dimension of each sample of Z . In our model, D z was set to the smallest dimension of X ^ ( f ) , X ^ ( v ) , and X ^ ( s ) . The prior distributions of the missing values X miss ( f ) and latent variables Z are initialized by random variables following a multivariate normal distribution as
p 0 ( X miss ( f ) ) = n = N + 1 N + M N ( x n ( f ) | 0 , I D f ) ,
p 0 ( Z ) = n = 1 N + M N ( z n | 0 , I D z ) ,
where N ( x | μ , Σ ) represents a multivariate Gaussian distribution with a mean μ and a covariance matrix Σ . 0 denotes the zero vector and I D denotes the identity matrix with D × D size. Our proposed model assumes that fMRI activity X ^ ( f ) , visual features X ^ ( v ) , and semantic features X ^ ( s ) are generated from shared latent variables Z by the following likelihood function:
p ( X ^ ( k ) | W ( k ) , Z ) = n = 1 N + M N ( x n ( k ) | W ( k ) z n , β k 1 I D k ) ,
where k { f , v , s } . In Equation (3), W ( k ) R D k × D z is a projection matrix that connects Z with X ^ ( k ) , and β k 1 is the scalar variable representing noise variance of X ^ ( k ) . The prior distribution of the projection matrix W ( k ) and its hyperprior distribution α ( k ) are assumed as
p 0 ( W ( k ) | α ( k ) ) = i = 1 D k j = 1 D z N ( W i , j ( k ) | 0 , α i , j ( k ) 1 ) ,
p 0 ( α ( k ) ) = i = 1 D k j = 1 D z G ( α i , j ( k ) | α ¯ 0 ( i , j ) ( k ) , γ 0 ( i , j ) ( k ) ) ,
where α i , j ( k ) 1 is a variance of the elements in W ( k ) . In Equation (5), we introduce hyperprior distribution for the inverse variances α ( k ) , and G ( · | α ¯ , γ ) represents the Gamma distribution with a mean α ¯ and a shape parameter γ . In our model, all of the means α ¯ 0 ( i , j ) ( k ) and the shape parameters γ 0 ( i , j ) ( k ) were set to 1 and 0, respectively, as in the previous study [19]. This form of prior and hyperprior distributions is motivated by the idea of automatic relevance determination (ARD). Since ARD automatically selects important components in the model, we can extract a small number of appropriate features from observed variables. Especially, this sparse representation is effective for fMRI data analysis since fMRI data have thousands of voxels and it is reported that only a small number of voxels is relevant to a visual stimulus [19,38]. This ARD form avoids overfitting due to the high dimensionality and improves generalization performance. Finally, the prior distribution of inverse variance β k is assumed to be an uninformative prior described as p 0 ( β k ) = β k 1 . The graphical model of the proposed model is illustrated in Figure 3. The meaning and role of the observed variables and model parameters are summarized in Table 1.

3.2. Optimization of Model Parameters via Variational Inference

In this subsection, we estimate the missing values X miss ( f ) and optimize model parameters Z , W ( k ) , α ( k ) , and β k ( k { f , v , s } ) . Hereafter, the observed variables (except for the missing values), the projection matrices, and the inverse variances are summarized as X = { X ( f ) , X ^ ( v ) , X ^ ( s ) } , W = { W ( f ) , W ( v ) , W ( s ) } , and Θ = { α ( f ) , α ( v ) , α ( s ) , β f , β v , β s } , respectively. We optimize X miss ( f ) , Z , W , and Θ by calculating a posterior distribution p ( X miss ( f ) , Z , W , Θ | X ) . Since this posterior distribution cannot be calculated analytically, we approximate it by using variational inference [39]. In variational inference, we consider an approximate distribution q ( X miss ( f ) , Z , W , Θ ) as a factorized form described as q ( X miss ( f ) , Z , W , Θ ) q ( X miss ( f ) ) q ( Z ) q ( W ) q ( Θ ) . Then the approximate distribution is optimized so that the Kullback–Leibler (KL) divergence D K L ( q , p ) between the approximate distribution and the true posterior distribution can be minimized. This optimization is equivalent to the maximization of a lower bound L ( q ) as follows:
L ( q ) = log p ( X ) D K L ( q , p ) = q ( X miss ( f ) ) q ( Z ) q ( W ) q ( Θ ) log p ( X , X miss ( f ) , Z , W , Θ ) q ( X miss ( f ) ) q ( Z ) q ( W ) q ( Θ ) d X miss ( f ) d Z d W d Θ .
The distributions of q ( X miss ( f ) ) , q ( Z ) , q ( W ( k ) ) , q ( α ( k ) ) , and q ( β k ) ( k { f , v , s } ) are updated iteratively to maximize the lower bound L ( q ) .
First, we update the distribution of the projection matrix q ( W ( k ) ) according to variational inference as
q ( W ( k ) ) = i = 1 D k j = 1 D z N ( W i , j ( k ) | W ¯ i , j ( k ) , σ i , j ( k ) 1 ) , k { f , v , s }
where
σ i , j ( k ) = β ¯ k n = 1 N + M z ¯ n ( j ) 2 + ( N + M ) σ z ( j , j ) 1 + α ¯ i , j ( k ) ,
W ¯ i , j ( k ) = β ¯ k σ i , j ( k ) 1 n = 1 N + M x n ( i ) ( k ) z ¯ n ( j ) .
In Equations (8) and (9), z ¯ n ( j ) represents the mean of the jth variable of z n . Next, the distribution of the latent variables q ( Z ) is updated as
q ( Z ) = n = 1 N + M N ( z n | z ¯ n , σ z 1 ) ,
where
σ z = β ¯ f W ¯ ( f ) T W ¯ ( f ) + σ W ( f ) 1 + β ¯ v W ¯ ( v ) T W ¯ ( v ) + σ W ( v ) 1 + β ¯ s W ¯ ( s ) T W ¯ ( s ) + σ W ( s ) 1 + I D z ,
z ¯ n = σ z 1 β ¯ f W ¯ ( f ) T x n ( f ) + β ¯ v W ¯ ( v ) T x n ( v ) + β ¯ s W ¯ ( s ) T x n ( s ) .
In Equation (11), σ W ( k ) ( k { f , v , s } ) is defined as
σ W ( k ) = diag i = 1 D k σ i , 1 ( k ) , , i = 1 D k σ i , D z ( k ) .
Furthermore, the distribution of the missing values q ( X miss ( f ) ) is updated as
q ( X miss ( f ) ) = n = N + 1 N + M N ( x n | W ¯ ( f ) z ¯ n , β f 1 I D f ) .
In Equation (14), we estimate the missing values, i.e., unobserved fMRI activity patterns, by using the current model parameters W ( f ) , Z , and β f . Finally, we update the distribution of the variances q ( α ( k ) ) and q ( β k ) ( k { f , v , s } ) . The distribution q ( α ( k ) ) is updated as
q ( α ( k ) ) = i = 1 D k j = 1 D z G ( α i , j ( k ) | α ¯ ( i , j ) ( k ) , γ ( i , j ) ( k ) ) , k { f , v , s }
where
γ ( i , j ) ( k ) = 1 2 + γ 0 ( i , j ) ( k ) ,
α ¯ i , j ( k ) 1 = γ ( i , j ) ( k ) 1 1 2 W ¯ i , j ( k ) 2 + 1 2 σ i , j ( k ) 1 + γ 0 ( i , j ) ( k ) σ 0 ( i , j ) ( k ) 1 .
In our model, σ 0 ( i , j ) ( k ) 1 was set to 0 as in [19]. The distribution q ( β k ) is updated as
q ( β k ) = G ( β k | β ¯ k , γ β k ) , k { f , v , s }
where
γ β k = 1 2 D k ( N + M ) ,
β ¯ k 1 = 1 2 γ β k 1 { n = 1 N + M | | x n ( k ) W ¯ ( k ) z ¯ n | | 2 + Tr [ σ W ( k ) 1 n = 1 N + M z n z n T + ( N + M ) σ z 1 + ( N + M ) σ z 1 W ¯ ( k ) T W ¯ ( k ) ] } .
These update procedures of the model parameters are summarized in Algorithm 1.
Algorithm 1 Update procedures of model parameters
Input: Observed variables X = { X ( f ) , X ^ ( v ) , X ^ ( s ) }
 Initialize X miss ( f ) , Z , W ( k ) , α ( k ) , and β k ( k { f , v , s } ) by prior distributions in Equations (1), (2), (4), (5)
for number of training iterations do
   Update the projection matrix W ( k ) in Equation (7)
   Update the shared latent variables Z in Equation (10)
   Update the missing values X miss ( f ) in Equation (14)
   Update the inverse variances α ( k ) in Equation (15)
   Update the inverse variance β k in Equation (18)
end for
Output: Updated distributions of X miss ( f ) , Z , W ( k ) , α ( k ) , and β k ( k { f , v , s } )

3.3. Decoding of Viewed Image Categories from fMRI Activity

After the optimization of model parameters, image categories are decoded from fMRI activity evoked by test visual stimuli, i.e., unobserved images. In zero-shot neural decoding settings, the categories of test visual stimuli are not included in the training categories. Suppose that x test ( f ) R D f is the fMRI activity measured when a subject viewed a test image. We decode the viewed image category by predicting visual features x test ( v ) R D v and semantic features x test ( s ) R D s from test fMRI activity x test ( f ) . For the prediction of visual and semantic features, we calculate the following predictive distribution constructed from the likelihood function p ( x test ( i ) | W ( i ) , z test ) , the distribution of projection matrix q ( W ( i ) ) , and a posterior distribution p ( z test | x test ( f ) ) as follows:
p ( x test ( i ) | x test ( f ) ) = p ( x test ( i ) | W ( i ) , z test ) q ( W ( i ) ) p ( z test | x test ( f ) ) d W ( i ) d z test , i { v , s } .
In Equation (21), z test is a latent variable shared by x test ( f ) , x test ( v ) , and x test ( s ) . Since the posterior distribution p ( z test | x test ( f ) ) is an unknown distribution, we approximate it based on the distribution of the latent variables q ( Z ) (Equations (10)–(12)). We derive an approximate distribution q ˜ ( z test ) by omitting the terms with respect to visual features and semantic features in Equations (10)–(12) as follows:
q ˜ ( z test ) = N ( z test | z ¯ test , σ z test 1 ) ,
where
σ z test = β ¯ f W ¯ ( f ) T W ¯ ( f ) + σ W ( f ) 1 + I D z ,
z ¯ test = β ¯ f σ z test 1 W ¯ ( f ) T x test ( f ) .
Since the multiple integral of Equation (21) cannot be calculated analytically, we replace q ( W ( i ) ) with W ¯ ( i ) obtained from Equation (9). Moreover, p ( x test ( i ) | W ( i ) , z test ) is obtained from the likelihood function in Equation (3). From the above approximations, the predictive distribution in Equation (21) becomes
p ( x test ( i ) | x test ( f ) ) N ( x test ( i ) | x ¯ test ( i ) , σ x test ( i ) 1 ) , i { v , s } ,
where
σ x test ( i ) = W ¯ ( i ) σ z test 1 W ¯ ( i ) T + β ¯ i 1 I D i ,
x ¯ test ( i ) = β ¯ f W ¯ ( i ) σ z test 1 W ¯ ( f ) T x test ( f ) .
Finally, these predicted visual and semantic features x test ( i ) ( i { v , s } ) are obtained by averaging samples x ¯ test ( i ) through many sets of model training and prediction,
x test ( i ) = 1 T t = 1 T x ¯ test ( i ) , i { v , s } .
In our analysis, the number of sets of model training and prediction was set to T = 100 .
These predicted features are compared with visual and semantic features obtained from various candidate categories. Specifically, we calculate the Pearson correlation coefficient r ( v ) between x test ( v ) and visual features obtained from a candidate category as in previous studies [6,8]. We also calculate the Pearson correlation coefficient r ( s ) between x test ( s ) and semantic features obtained from a candidate category. r ( v ) and r ( s ) are calculated separately from visual and semantic features, respectively. The values of r ( v ) and r ( s ) range from −1 to 1, and higher values indicate that these features are similar. Then these correlation coefficients are integrated as r ( v + s ) = η · r ( v ) + ( 1 η ) · r ( s ) , where η ( 0 η 1 ) is the trade-off parameter of balancing visual and semantic features. The decoding of the viewed image category is feasible by identifying a candidate category that scored the max correlation coefficient r ( v + s ) .

4. Experimental Results

4.1. Dataset

We utilized a publicly available fMRI dataset [6] in this study. This dataset includes fMRI activity patterns measured when five subjects viewed images collected from ImageNet [40]. In this dataset, fMRI activity patterns that were measured when each subject viewed 1200 images from 150 categories (eight images per category) were collected for the training data. Furthermore, fMRI activity patterns that were measured when each subject viewed 50 images from 50 categories (one image per category), which differ from training categories, were collected for the test data. Since test fMRI activity was measured 35 times for each test image, we used the average fMRI activity patterns across all trials, as in [6]. We used fMRI activity with approximately 4500-dimensional (the number of voxels) vectors from the visual cortex (V1-V4, the lateral occipital complex, fusiform face area, and parahippocampal place area) defined in the study [6]. Visual features with 512-dimensional vectors were extracted from images via the last convolution layer with an average pooling of VGG19 that was pre-trained on ImageNet. Semantic features with 300-dimensional vectors were obtained from image categories via a word2vec that was pre-trained on a Google News dataset. When the categories were not in the word2vec corpus, we utilized hypernyms on ImageNet whose semantic features were available. For example, since the category ‘beer mug’ was not in the corpus, we utilized ‘mug’, which is a hypernym of ‘beer mug’.

4.2. Conditions

To validate the effectiveness of the semi-supervised multi-view generative model, we compared the proposed model with the following seven methods: existing neural decoding methods [6,7,8,14], BCCA [19], and the multi-view generative model (MGM). The methods in [6,8] predicted visual features of viewed images from fMRI activity via sparse logistic regression (SLR) [38] and multilayer perceptrons (MLP) [41], respectively. The method in [14] projected fMRI activity and visual features into the same latent space by CCA [25]. The method in [7] constructed semi-supervised fuzzy discriminative CCA (Semi-FDCCA) that incorporated similarities of image categories and visual features from various images into the framework of CCA. BCCA and MGM are in the same framework as our proposed model. BCCA assumes that the number of observed variables is two. Here, we constructed the following two BCCA methods: BCCA-V and BCCA-S. The observed variables of BCCA-V are fMRI activity and visual features, and those of BCCA-S are fMRI activity and semantic features. MGM associated fMRI activity with visual features and semantic features, i.e., the proposed model without a semi-supervised learning scenario. We verified the effectiveness of the multi-view approach by comparing ours with BCCA-V and BCCA-S, and verified the effectiveness of the semi-supervised approach by comparing ours with MGM. In all methods, we standardized fMRI activity, visual features, and semantic features before the training and testing phases. In BCCA-V, BCCA-S, MGM, and the proposed model, the number of training iterations was set to 10. In MGM and the proposed model, the trade-off parameter η was set to the best value from η { 0 , 0.1 , , 1 } for each model.
In this experiment, we used identification accuracy to evaluate the decoding performance as in previous studies [6,8]. This evaluation metric represents the accuracy for identifying its ground truth image category between the two candidate categories: one is the ground truth category and another is a randomly selected category (chance level being 50%). This identification was performed based on the correlation coefficient r ( v + s ) between predicted features and each candidate category feature. We prepared 10,000 candidate categories (including 50 test categories) randomly selected from ImageNet. Therefore, we performed this identification between all combinations of a ground truth category and one of the other 9,999 candidate categories. Furthermore, we also evaluated by using rank-n accuracy to confirm where the correct category is ranked in all candidates. Rank-n accuracy refers to the ratio that shows that a correct category is ranked at top-n ranks across 10,000 candidates based on the correlation coefficient r ( v + s ) .

4.3. Results and Discussion

We evaluate the decoding performance in several zero-shot neural decoding settings where the source domain and the target domain are potentially unrelated. In this experiment, we classified categories used in the dataset [6] into eight target groups: mammal, bird, invertebrate, device, container, equipment, structure, and commodity. These groups are defined as subgroups in ImageNet. For zero-shot analysis, several categories that belong to each target group were excluded from the 150 training categories of the dataset [6], and the remaining categories were used for training. Moreover, several categories that belong to each target group were collected from 50 test categories of the dataset [6], and fMRI activity patterns corresponding to these categories were used for test fMRI activity. Furthermore, the proposed model incorporated additional image categories that belong to each target group, which were collected for as many categories as possible from ImageNet. As additional visual features, we used the average visual features extracted from multiple images of an additional category. The numbers of training, test, and additional categories for each target group are described in Table 2. Note that we have eight training samples per training category.
Decoding performance Table 3 shows the identification accuracy for each target group. The accuracy was averaged across five subjects. As shown in this table, the proposed model outperforms the seven other methods. By comparing our model with MGM, we confirmed the effectiveness of the semi-supervised approach that employs additional image categories belonging to target groups. We also confirmed that the identification accuracy of SLR and BCCA-V for mammal are remarkably low, whereas the accuracy of BCCA-S is comparatively high. These results indicate that mammal can be decoded more successfully from the semantic feature space than the visual feature space. On the other hand, the accuracy for other target groups such as container and structure from the visual feature space is higher than that of the semantic feature space. Therefore, considering a multi-view embedding space is effective since the better feature space depends on each target group. Table 4 presents the rank-n ( n = 100 , 1000 , 5000 ) accuracy in 10,000 random candidate categories. From the table, we also confirmed that the rank accuracy of the proposed model is higher than that of comparative methods and outperforms the chance level.
In Table 3, we used 10,000 candidate categories when identification accuracy was calculated. This evaluation metric indicates how accurately the correct category can be identified from a large number of random categories. Here, we also validate how accurately the correct category can be identified from similar categories (i.e., target groups). Table 5 shows the identification accuracy of comparative methods and the proposed method when the candidate categories are in each target group. For example, when test categories of mammal were decoded, we performed identification between all combinations of the ground truth category and one of the candidate categories from mammal. As seen in Table 5, the identification accuracy of almost all target groups are degraded in comparison with Table 3, since it is more difficult to identify a ground truth category from similar candidate categories. Nevertheless, the proposed model outperforms comparison methods for all target groups. By comparing with MGM, we confirmed that the utilization of additional visual and semantic features from each target group contributes to the identification from similar candidate categories.
Projection domain shift problem To examine the projection domain shift problem, we visualized the L2 norm between target group embedding and target group prototype in the visual and semantic feature spaces, as seen in Figure 4. Target group embedding is the average vector of embedding features of categories from each target group, and target group prototype is the average vector of ground truth values of categories from each target group. The L2 norm was averaged across five subjects. CCA [14] and Semi-FDCCA [7] were not considered in Figure 4 since these methods calculate the distance between features in the canonical space. From Figure 4, the L2 norm of the proposed model is shorter than that of comparative methods in both the visual and semantic feature spaces. This result shows that the semi-supervised approach reduces the embedding biases towards source categories and realizes better target embeddings towards their prototypes. This alleviation of the projection domain shift problem is effective not only in identifying from a large number of random categories (see Table 3), but also in identifying from similar categories (see Table 5). (Identification accuracy tends to improve when the L2 norm is short. However, the L2 norm is not perfectly consistent with identification accuracy. This is because the L2 norm is calculated purely on the distance of the embedding features, while the identification accuracy is calculated by comparing the embedding features with other candidate features. Since the distribution of candidate features is not uniform but depends on the target group, the relationship of the L2 norm between target groups is not consistent with identification accuracy).
Normal setting vs. Zero-shot setting To confirm the performance limit in zero-shot neural decoding settings, we evaluate the performance of the proposed model in normal settings, where target groups are included in the training set (i.e., the original setting of dataset [6]). Table 6 shows the decoding performance of our model when each target group is included in the training set (normal) and not included in the training set (zero-shot). The results are averaged across five subjects. We can see that the performance in the normal setting is higher than in the zero-shot setting for all target groups. Although there is still a gap in performance between the normal and zero-shot settings, the proposed method achieves comparable performance for several target groups. (e.g., bird, container, and equipment). It is a natural result that the normal setting outperforms the zero-shot setting, and we found that the performance of the zero-shot setting is comparable to that of the normal setting for several target groups.
Balance between visual and semantic features We investigate the balance between visual and semantic features in decoding each target group. Table 7 and Table 8 show the best trade-off parameters η representing the importance of visual and semantic features in the normal and zero-shot settings, respectively. In the tables, visual features are important when the value is large, while semantic features are important when the value is small. From Table 7 in the normal setting, we see that the visual and semantic features are equally important in all target groups (i.e., η is around 0.5). On the other hand, Table 8 in the zero-shot setting shows that η is changed from the normal setting in some target groups (e.g., mammal, invertebrate, and device). In groups of living thing (i.e., mammal, bird, and invertebrate), η tends to become smaller, i.e., the importance of semantic features increases. In groups of artifact (i.e., device, container, equipment, structure, and commodity), η tends to become larger, i.e., the importance of visual features increases. The results suggest that visual features of living thing and semantic features of artifact suffer from the projection domain problem in the zero-shot setting. Therefore, we can see that another view (i.e., semantic features of living thing and visual features of artifact), which is less affected by the problem, contributes to the zero-shot neural decoding.
Different additional group than the target group In the above results, we utilized additional image categories belonging to each target group. Here, we discuss the decoding performance when incorporating a different additional group than the target group. Figure 5 shows the identification accuracy improvement of the proposed model for all combinations of additional and target groups in comparison with the accuracy of MGM. Accuracy improvement was averaged across five subjects. From the result, we confirmed that the method using the same additional group as the target group is the best model, while the accuracy is improved when the additional group is similar to the target group (e.g., device and structure). On the other hand, the utilization of other additional groups degrades the decoding performance when the target group is equipment. Thus, as expected, it is important to add categories that are relevant to the target group.
Effect of decoding performance on the source group We investigate the decoding performance of the source group when each additional group is incorporated. Figure 6 shows the decoding performance of the source group for each target and additional group (the additional group is the same as the target group) in MGM and the proposed model. Note that the source group means the other seven groups except for each target and additional group. The result indicates that the utilization of additional image categories slightly improves the decoding performance of the source group. Therefore, these results give us confidence that the semi-supervised approach contributes to the decoding of the target group and does not negatively affect the decoding of the source group.

5. Conclusions

This paper has proposed a semi-supervised multi-view embedding approach for zero-shot neural decoding. In zero-shot neural decoding, the training data are scarce because of the difficulty in collecting brain activity patterns; therefore, the projection domain shift problem between the source domain and the target domain becomes remarkable. We address this problem by introducing the semi-supervised approach that employs additional image categories related to the target domain. Furthermore, to exploit the complementary information, we assume that fMRI activity patterns are projected into the visual feature space and semantic feature space. The experimental results show the advantages of the proposed model over existing methods.
Given the difficulty in collecting fMRI activity, we incorporate additional visual and semantic features from abundant images while estimating corresponding fMRI activity patterns. This framework can be applied to not only neural decoding tasks but also other difficult learning tasks where one type of modal data are insufficient whereas other modal data are available abundantly.

Author Contributions

Conceptualization, Y.A., T.O. and M.H.; methodology, Y.A. and T.O.; software, Y.A.; validation, Y.A., K.M., T.O. and M.H.; data curation, Y.A.; writing—original draft preparation, Y.A.; writing—review and editing, Y.A., K.M., T.O. and M.H.; visualization, Y.A.; funding acquisition, T.O. and M.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partly supported by JSPS KAKENHI Grant Numbers JP21H03456 and JP23K11211.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The public dataset used in our experiment is available at https://github.com/KamitaniLab/GenericObjectDecoding (accessed on 15 April 2023).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wolpaw, J.R.; Birbaumer, N.; McFarland, D.J.; Pfurtscheller, G.; Vaughan, T.M. Brain–computer interfaces for communication and control. Clin. Neurophysiol. 2002, 113, 767–791. [Google Scholar] [CrossRef] [PubMed]
  2. Haxby, J.V.; Gobbini, M.I.; Furey, M.L.; Ishai, A.; Schouten, J.L.; Pietrini, P. Distributed and overlapping representations of faces and objects in ventral temporal cortex. Science 2001, 293, 2425–2430. [Google Scholar] [CrossRef] [Green Version]
  3. Cox, D.D.; Savoy, R.L. Functional magnetic resonance imaging (fMRI) “brain reading”: Detecting and classifying distributed patterns of fMRI activity in human visual cortex. NeuroImage 2003, 19, 261–270. [Google Scholar] [CrossRef] [PubMed]
  4. Reddy, L.; Tsuchiya, N.; Serre, T. Reading the mind’s eye: Decoding category information during mental imagery. NeuroImage 2010, 50, 818–825. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Kay, K.N.; Naselaris, T.; Prenger, R.J.; Gallant, J.L. Identifying natural images from human brain activity. Nature 2008, 452, 352–355. [Google Scholar] [CrossRef] [Green Version]
  6. Horikawa, T.; Kamitani, Y. Generic decoding of seen and imagined objects using hierarchical visual features. Nat. Commun. 2017, 8, 15037. [Google Scholar] [CrossRef]
  7. Akamatsu, Y.; Harakawa, R.; Ogawa, T.; Haseyama, M. Estimating viewed image categories from human brain activity via semi-supervised fuzzy discriminative canonical correlation analysis. In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 1105–1109. [Google Scholar]
  8. Papadimitriou, A.; Passalis, N.; Tefas, A. Visual representation decoding from human brain activity using machine learning: A baseline study. Pattern Recognit. Lett. 2019, 128, 38–44. [Google Scholar] [CrossRef]
  9. O’Connell, T.P.; Chun, M.M.; Kreiman, G. Zero-shot neural decoding of visual categories without prior exemplars. bioRxiv 2019, 700344. [Google Scholar]
  10. McCartney, B.; Martinez-del Rincon, J.; Devereux, B.; Murphy, B. A zero-shot learning approach to the development of brain-computer interfaces for image retrieval. PLoS ONE 2019, 14, e0214342. [Google Scholar] [CrossRef] [Green Version]
  11. Mitchell, T.M.; Shinkareva, S.V.; Carlson, A.; Chang, K.; Malave, V.L.; Mason, R.A.; Just, M.A. Predicting human brain activity associated with the meanings of nouns. Science 2008, 320, 1191–1195. [Google Scholar] [CrossRef] [Green Version]
  12. Palatucci, M.; Pomerleau, D.; Hinton, G.E.; Mitchell, T.M. Zero-shot learning with semantic output codes. In Proceedings of the Advances in Neural Information Processing Systems 22 (NIPS 2009), Vancouver, BC, Canada, 7–10 December 2009; pp. 1410–1418. [Google Scholar]
  13. Pereira, F.; Lou, B.; Pritchett, B.; Ritter, S.; Gershman, S.J.; Kanwisher, N.; Botvinick, M.; Fedorenko, E. Toward a universal decoder of linguistic meaning from brain activation. Nat. Commun. 2018, 9, 963. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Akamatsu, Y.; Harakawa, R.; Ogawa, T.; Haseyama, M. Estimation of viewed image categories via CCA using human brain activity. In Proceedings of the 2018 IEEE 7th Global Conference on Consumer Electronics (GCCE), Nara, Japan, 9–12 October 2018; pp. 171–172. [Google Scholar]
  15. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. In Proceedings of the Advances in Neural Information Processing Systems 25 (NIPS 2012), Lake Tahoe, NV, USA, 3–6 December 2012; pp. 1097–1105. [Google Scholar]
  16. Fu, Y.; Hospedales, T.M.; Xiang, T.; Gong, S. Transductive multi-view zero-shot learning. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 2332–2345. [Google Scholar] [CrossRef] [PubMed]
  17. Kodirov, E.; Xiang, T.; Fu, Z.; Gong, S. Unsupervised domain adaptation for zero-shot learning. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 2452–2460. [Google Scholar]
  18. Kodirov, E.; Xiang, T.; Gong, S. Semantic autoencoder for zero-shot learning. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 3174–3183. [Google Scholar]
  19. Fujiwara, Y.; Miyawaki, Y.; Kamitani, Y. Modular encoding and decoding models derived from Bayesian canonical correlation analysis. Neural Comput. 2013, 25, 979–1005. [Google Scholar] [CrossRef]
  20. Klami, A.; Virtanen, S.; Leppäaho, E.; Kaski, S. Group factor analysis. IEEE Trans. Neural Netw. Learn. Syst. 2015, 26, 2136–2147. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Maaten, L.v.d.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  22. Oba, S.; Sato, M.; Takemasa, I.; Monden, M.; Matsubara, K.I.; Ishii, S. A Bayesian missing value estimation method for gene expression profile data. Bioinformatics 2003, 19, 2088–2096. [Google Scholar] [CrossRef] [Green Version]
  23. Xu, C.; Tao, D.; Xu, C. A survey on multi-view learning. arXiv 2013, arXiv:1304.5634. [Google Scholar]
  24. Zhao, J.; Xie, X.; Xu, X.; Sun, S. Multi-view learning overview: Recent progress and new challenges. Inf. Fusion 2017, 38, 43–54. [Google Scholar] [CrossRef]
  25. Hotelling, H. Relations between two sets of variates. Biometrika 1936, 28, 321–377. [Google Scholar] [CrossRef]
  26. Kimura, A.; Sugiyama, M.; Nakano, T.; Kameoka, H.; Sakano, H.; Maeda, E.; Ishiguro, K. SemiCCA: Efficient semi-supervised learning of canonical correlations. Inf. Media Technol. 2013, 8, 311–318. [Google Scholar] [CrossRef] [Green Version]
  27. Wang, C. Variational Bayesian approach to canonical correlation analysis. IEEE Trans. Neural Netw. 2007, 18, 905–910. [Google Scholar] [CrossRef] [PubMed]
  28. Neal, R.M. Bayesian Learning for Neural Networks; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2012; Volume 118. [Google Scholar]
  29. Akamatsu, Y.; Harakawa, R.; Ogawa, T.; Haseyama, M. Estimating viewed image categories from fMRI activity via multi-view Bayesian generative model. In Proceedings of the 2019 IEEE 8th Global Conference on Consumer Electronics (GCCE), Osaka, Japan, 15–18 October 2019; pp. 127–128. [Google Scholar]
  30. Beliy, R.; Gaziv, G.; Hoogi, A.; Strappini, F.; Golan, T.; Irani, M. From voxels to pixels and back: Self-supervision in natural-image reconstruction from fMRI. In Proceedings of the Advances in Neural Information Processing Systems 32 (NeurIPS 2019), Vancouver, BC, Canada, 8–14 December 2019; pp. 6517–6527. [Google Scholar]
  31. Akamatsu, Y.; Harakawa, R.; Ogawa, T.; Haseyama, M. Brain decoding of viewed image categories via semi-supervised multi-view Bayesian generative model. IEEE Trans. Signal Process. 2020, 68, 5769–5781. [Google Scholar] [CrossRef]
  32. Du, C.; Fu, K.; Li, J.; He, H. Decoding Visual Neural Representations by Multimodal Learning of Brain-Visual-Linguistic Features. IEEE Trans. Pattern Anal. Mach. Intell. 2023; early access. [Google Scholar]
  33. Liu, Y.; Ma, Y.; Zhou, W.; Zhu, G.; Zheng, N. BrainCLIP: Bridging Brain and Visual-Linguistic Representation via CLIP for Generic Natural Visual Stimulus Decoding from fMRI. arXiv 2023, arXiv:2302.12971. [Google Scholar]
  34. Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; et al. Learning transferable visual models from natural language supervision. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 18–24 July 2021; pp. 8748–8763. [Google Scholar]
  35. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  36. Horikawa, T.; Aoki, S.C.; Tsukamoto, M.; Kamitani, Y. Characterization of deep neural network features by decodability from human brain activity. Sci. Data 2019, 6, 190012. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Mikolov, T.; Sutskever, I.; Chen, K.; Corrado, G.; Dean, J. Distributed representations of words and phrases and their compositionality. In Proceedings of the Advances in Neural Information Processing Systems 26 (NIPS 2013), Lake Tahoe, NV, USA, 5–10 December 2013; pp. 3111–3119. [Google Scholar]
  38. Yamashita, O.; Sato, M.; Yoshioka, T.; Tong, F.; Kamitani, Y. Sparse estimation automatically selects voxels relevant for the decoding of fMRI activity patterns. NeuroImage 2008, 42, 1414–1429. [Google Scholar] [CrossRef] [Green Version]
  39. Attias, H. Inferring parameters and structure of latent variable models by variational Bayes. In Proceedings of the Fifteenth Conference on Uncertainty in Artificial Intelligence (UAI1999), Stockholm, Sweden, 30 July–1 August 1999; pp. 21–30. [Google Scholar]
  40. Deng, J.; Dong, W.; Socher, R.; Li, L.; Li, L.J.; Li, K.; Li, F. ImageNet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  41. Haykin, S.S. Neural Networks and Learning Machines; Prentice Hall: New York, NY, USA, 2009. [Google Scholar]
Figure 1. (a) The projection domain shift problem in zero-shot neural decoding. When the source categories (i.e., ostrich, kangaroo, and dolphin) and the target categories (i.e., mailbox and iPod) are potentially different, the projections that connect fMRI activity with the visual feature space are biased towards the source categories. This bias could keep target fMRI activity embeddings away from the actual target category embeddings. (b) Our proposed semi-supervised multi-view embedding. We embedded additional visual and semantic features extracted from images related to the target categories without fMRI activity patterns. By introducing additional visual and semantic features, the biased projections towards the source categories are compensated, and the projection domain shift problem is alleviated.
Figure 1. (a) The projection domain shift problem in zero-shot neural decoding. When the source categories (i.e., ostrich, kangaroo, and dolphin) and the target categories (i.e., mailbox and iPod) are potentially different, the projections that connect fMRI activity with the visual feature space are biased towards the source categories. This bias could keep target fMRI activity embeddings away from the actual target category embeddings. (b) Our proposed semi-supervised multi-view embedding. We embedded additional visual and semantic features extracted from images related to the target categories without fMRI activity patterns. By introducing additional visual and semantic features, the biased projections towards the source categories are compensated, and the projection domain shift problem is alleviated.
Sensors 23 06903 g001
Figure 2. (a) The visual feature space by the previous method [6] and (b) the visual and semantic feature spaces by the proposed model. As seen in Figure 1, source categories belong to only living thing, and target categories of living thing (red circles) and artifact (blue circles) are projected into the embedding space. Prototypes of living thing (red stars) and artifact (blue stars) represent the ground truth embeddings.
Figure 2. (a) The visual feature space by the previous method [6] and (b) the visual and semantic feature spaces by the proposed model. As seen in Figure 1, source categories belong to only living thing, and target categories of living thing (red circles) and artifact (blue circles) are projected into the embedding space. Prototypes of living thing (red stars) and artifact (blue stars) represent the ground truth embeddings.
Sensors 23 06903 g002
Figure 3. Graphical model of the proposed model. The gray nodes represent observed variables and the white nodes represent model parameters. The mixture node of gray and white includes both observed variables and model parameters (i.e., missing values).
Figure 3. Graphical model of the proposed model. The gray nodes represent observed variables and the white nodes represent model parameters. The mixture node of gray and white includes both observed variables and model parameters (i.e., missing values).
Sensors 23 06903 g003
Figure 4. Distance between target group embeddings and their prototypes in the visual and semantic feature spaces. The L2 norm was averaged across five subjects and the error bars represent standard errors.
Figure 4. Distance between target group embeddings and their prototypes in the visual and semantic feature spaces. The L2 norm was averaged across five subjects and the error bars represent standard errors.
Sensors 23 06903 g004
Figure 5. Identification accuracy improvement compared with MGM for all combinations of additional and target groups.
Figure 5. Identification accuracy improvement compared with MGM for all combinations of additional and target groups.
Sensors 23 06903 g005
Figure 6. Decoding performance of the source group for each additional group in MGM and our proposed model. For example, when the target and additional groups are mammal, the source group is the other seven groups except for mammal. The identification accuracy was averaged across five subjects and the error bars represent standard errors.
Figure 6. Decoding performance of the source group for each additional group in MGM and our proposed model. For example, when the target and additional groups are mammal, the source group is the other seven groups except for mammal. The identification accuracy was averaged across five subjects and the error bars represent standard errors.
Sensors 23 06903 g006
Table 1. Meaning and role of observed variables and model parameters.
Table 1. Meaning and role of observed variables and model parameters.
SymbolMeaningRole
Observed variables ( k { f , v , s } ):
X ( f ) fMRI activityInput for neural decoding
X ( v ) Visual featuresVisual information of viewed image
X ( s ) Semantic featuresSemantic information of viewed image
X add ( v ) Additional visual featuresAlleviate the projection domain shift problem
X add ( s ) Additional semantic featuresAlleviate the projection domain shift problem
X ^ ( k ) Concatenated observed variablesConcatenate original and additional features
Model parameters ( k { f , v , s } ):
Z Shared latent variablesLink X ^ ( f ) with X ^ ( v ) and X ^ ( s )
X miss ( f ) Unobserved fMRI activityfMRI activity corresponding to additional features
W ( k ) Projection matrixProject Z into X ^ ( f ) , X ^ ( v ) , and X ^ ( s )
α ( k ) Hyperprior distribution for W ( k ) Control the variance of W ( k )
β k Inverse variance for X ^ ( k ) Control the noise of X ^ ( k )
Table 2. Numbers of training, test, and additional categories for each target group.
Table 2. Numbers of training, test, and additional categories for each target group.
MammalBirdInvertebrateDeviceContainerEquipmentStructureCommodity
# Training134144138106135135145144
# Test63487346
# Additional107781462610386394171100551
Table 3. Decoding performance in 10,000 random candidate categories. The first line for each method represents the identification accuracy (%) and the second line represents the standard error of five subjects. Existing neural decoding methods (SLR, CCA, Semi-FDCCA, and MLP), a part of the proposed model (BCCA-V, BCCA-S, and MGM), and the proposed model are separated by blocks.
Table 3. Decoding performance in 10,000 random candidate categories. The first line for each method represents the identification accuracy (%) and the second line represents the standard error of five subjects. Existing neural decoding methods (SLR, CCA, Semi-FDCCA, and MLP), a part of the proposed model (BCCA-V, BCCA-S, and MGM), and the proposed model are separated by blocks.
MethodTarget Group
MammalBirdInvertebrateDeviceContainerEquipmentStructureCommodity
SLR [6]49.29 ± 4.1783.90 ± 2.2785.50 ± 2.2687.11 ± 1.8396.67 ± 0.3394.25 ± 1.3190.72 ± 1.0887.39 ± 3.02
CCA [14]76.90 ± 4.3873.12 ± 2.4374.85 ± 6.9878.23 ± 2.1187.20 ± 1.0177.73 ± 1.8972.44 ± 7.2373.61 ± 5.86
Semi-FDCCA [7]85.38 ± 0.9186.60 ± 2.4289.73 ± 0.9883.76 ± 1.9489.43 ± 2.4291.95 ± 3.8677.67 ± 6.4580.10 ± 3.57
MLP [8]75.14 ± 2.3195.73 ± 0.9891.25 ± 1.9189.98 ± 1.6597.63 ± 0.3292.74 ± 1.3490.90 ± 0.9488.44 ± 1.98
BCCA-V [19]45.65 ± 3.6084.41 ± 3.6882.90 ± 3.4688.34 ± 1.7896.56 ± 0.5193.81 ± 0.7989.13 ± 1.0485.99 ± 2.49
BCCA-S [19]83.51 ± 0.3894.22 ± 0.4884.64 ± 0.4173.54 ± 1.7165.39 ± 1.5482.48 ± 1.3277.14 ± 0.9581.03 ± 0.46
MGM83.97 ± 0.2895.15 ± 0.5990.46 ± 1.0989.18 ± 1.5097.19 ± 0.3895.30 ± 0.6191.47 ± 0.3789.14 ± 1.62
Ours86.91± 0.3397.69± 0.4993.10± 1.2091.61± 1.4998.19± 0.2395.36± 0.6393.11± 0.2291.63± 1.55
Table 4. Rank-n accuracy (%) in 10,000 random candidate categories. The results are averaged across all target groups and five subjects.
Table 4. Rank-n accuracy (%) in 10,000 random candidate categories. The results are averaged across all target groups and five subjects.
Metric (Chance Level)SLRCCASemi-FDCCAMLPBCCA-VBCCA-SMGMOurs
Rank-100 accuracy (1.0%)16.039.09211.5820.7614.320.416721.3230.22
Rank-1000 accuracy (10%)58.3541.4156.6771.1057.1131.2470.1677.66
Rank-5000 accuracy (50%)91.7782.7294.6497.8190.6394.0698.7599.06
Table 5. Decoding performance in similar candidate categories. The first line for each method represents identification accuracy (%) and the second line represents the standard deviation of five subjects.
Table 5. Decoding performance in similar candidate categories. The first line for each method represents identification accuracy (%) and the second line represents the standard deviation of five subjects.
MethodTarget Group
MammalBirdInvertebrateDeviceContainerEquipmentStructureCommodity
SLR [6]65.18 ± 3.0386.42 ± 4.8675.17 ± 5.8181.31 ± 3.0092.75 ± 2.0776.96 ± 9.1179.96 ± 3.3592.85 ± 2.86
CCA [14]60.29 ± 1.7773.62 ± 16.3865.88 ± 17.4575.17 ± 6.5082.93 ± 3.2971.35 ± 4.4859.99 ± 16.5869.38 ± 13.15
Semi-FDCCA [7]73.34 ± 3.3080.50 ± 9.9379.18 ± 5.4579.44 ± 6.6582.87 ± 7.1780.90 ± 12.1260.40 ± 11.2583.01 ± 7.30
MLP [8]73.98 ± 3.6291.26 ± 2.3480.42 ± 6.7686.57 ± 3.2894.23 ± 1.9375.40 ± 7.2876.61 ± 4.4492.59 ± 2.30
BCCA-V [19]65.38 ± 1.6183.90 ± 2.7774.06 ± 6.0684.41 ± 5.2092.80 ± 2.5373.99 ± 5.8977.80 ± 2.4193.42 ± 1.49
BCCA-S [19]61.83 ± 1.5178.79 ± 2.7058.96 ± 2.9074.80 ± 1.9450.63 ± 4.1186.23 ± 0.9687.14 ± 1.0087.04 ± 0.27
MGM74.68 ± 2.1092.62 ± 2.8078.71 ± 5.0887.11 ± 3.3193.48 ± 1.9490.14 ± 1.5688.59 ± 1.1395.87 ± 0.68
Ours77.38± 2.7294.45± 1.8682.54± 6.1490.46± 2.5995.07± 1.3390.37± 1.5888.62± 1.9896.03± 0.77
Table 6. Identification accuracy (%) of the proposed model for normal and zero-shot settings.
Table 6. Identification accuracy (%) of the proposed model for normal and zero-shot settings.
MammalBirdInvertebrateDeviceContainerEquipmentStructureCommodity
Normal96.4998.8497.2895.3398.8897.5897.0494.21
Zero-shot86.9197.6993.1091.6198.1995.3693.1191.63
Table 7. The trade-off parameter η of the proposed model in the normal setting. Visual features are important when the value is large, while semantic features are important when the value is small.
Table 7. The trade-off parameter η of the proposed model in the normal setting. Visual features are important when the value is large, while semantic features are important when the value is small.
MammalBirdInvertebrateDeviceContainerEquipmentStructureCommodity
Subject 10.30.50.40.70.80.40.50.3
Subject 20.70.80.70.60.70.50.50.5
Subject 30.70.70.80.60.80.40.50.6
Subject 40.50.70.70.50.80.50.60.7
Subject 50.60.70.70.60.80.50.60.5
Mean0.560.680.660.600.780.460.540.52
Table 8. The trade-off parameter η of the proposed model in the zero-shot setting. Visual features are important when the value is large, while semantic features are important when the value is small.
Table 8. The trade-off parameter η of the proposed model in the zero-shot setting. Visual features are important when the value is large, while semantic features are important when the value is small.
MammalBirdInvertebrateDeviceContainerEquipmentStructureCommodity
Subject 100.40.30.70.80.40.50.3
Subject 200.70.61.00.70.70.60.5
Subject 30.30.80.60.70.70.40.50.7
Subject 40.20.70.50.60.90.50.70.7
Subject 500.30.40.70.80.50.60.6
Mean0.100.580.480.740.780.50.580.56
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Akamatsu, Y.; Maeda, K.; Ogawa, T.; Haseyama, M. Zero-Shot Neural Decoding with Semi-Supervised Multi-View Embedding. Sensors 2023, 23, 6903. https://doi.org/10.3390/s23156903

AMA Style

Akamatsu Y, Maeda K, Ogawa T, Haseyama M. Zero-Shot Neural Decoding with Semi-Supervised Multi-View Embedding. Sensors. 2023; 23(15):6903. https://doi.org/10.3390/s23156903

Chicago/Turabian Style

Akamatsu, Yusuke, Keisuke Maeda, Takahiro Ogawa, and Miki Haseyama. 2023. "Zero-Shot Neural Decoding with Semi-Supervised Multi-View Embedding" Sensors 23, no. 15: 6903. https://doi.org/10.3390/s23156903

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop