Abstract
Brain networks characterize the temporal and/or spectral connections between brain regions and are inherently represented by multi-way arrays (tensors). In order to discover the underlying factors driving such connections, we need to derive compact representations from brain network data. Such representations should be discriminative so as to facilitate the identification of subjects performing different cognitive tasks or with different neurological disorders. In this paper, we propose semiBAT, a novel semi-supervised Brain network Analysis approach based on constrained Tensor factorization. semiBAT (1) leverages unlabeled resting-state brain networks for task recognition, (2) explores the temporal dimension to capture the progress, (3) incorporates classifier learning procedure to introduce supervision from labeled data, and (4) selects discriminative latent factors for different tasks. The Alternating Direction Method of Multipliers (ADMM) framework is utilized to solve the optimization objective. Experimental results on EEG brain networks illustrate the superior performance of the proposed semiBAT model on graph classification with a significant improvement \(31.60\,\%\) over plain vanilla tensor factorization. Moreover, the data-driven factors can be readily visualized which should be informative for investigating cognitive mechanisms. The software related to this paper is available at https://www.cs.uic.edu/~bcao1/code/semibat.zip.
You have full access to this open access chapter, Download conference paper PDF
1 Introduction
Brain networks (a.k.a, connectome [25]) obtained from neuroimaging data have been commonly employed to study neuropsychiatric disorders [3, 15, 27, 30]. Connectivity patterns are usually embedded within the graph structures by a set of vertices and edges where vertices correspond to regions of interest in the brain and edges represent the connectivity strength or correlation between brain regions. Considering the temporal and spectral domain, original brain networks are typically represented in multi-way arrays (i.e., tensors) which make the conventional vector-based classification algorithms inapplicable. Moreover, directly reshaping tensors into vectors would result in the curse of dimensionality, and the number of brain network samples is usually small, thereby making it challenging to train an effective classifier in a high dimensional feature space with a limited number of samples.
In order to apply conventional machine learning algorithms and train pattern classifiers, it is preferable to first derive vector representations from the brain network data. In general, researchers have proposed to extract two types of features: (1) graph-theoretical measures [13, 27] and (2) subgraph patterns [6, 17]. However, the expressiveness of these features is limited to the predefined formulations. To explore a larger space of potentially informative features to represent brain networks, it motivates us to learn latent representations from the brain network data. It is desirable to let the latent representations be discriminative so that brain networks with different labels can easily be separated. Learning such representations is a non-trivial task due to the following problems:
-
(P1) Although labeled brain network data for specific tasks or diseases are usually costly to obtain, brain networks under resting-state from healthy subjects are recorded in many neuroimaging experiments. How can we leverage the unlabeled data, i.e., resting-state brain networks, to facilitate classification?
-
(P2) Existing studies usually compute time-averaged brain networks before further analysis [28] which may result in formidable information loss. How can we directly fully utilize the temporal information in our model?
-
(P3) In order to obtain discriminative representations, we should incorporate the classifier training procedure into the representation learning process for leveraging the supervision information. How can we effectively fuse these two procedures together?
-
(P4) Different classes (or tasks) are usually associated with different subsets of the latent factors. How can we achieve feature selection in the latent space?
In this paper, we propose semiBAT, a semi-supervised Brain network Analysis approach based on constrained Tensor factorization. The proposed framework is illustrated in Fig. 1. The contributions of this work are fourfold:
-
We leverage unlabeled resting-state brain networks together with labeled data to collectively learn a latent space, which alleviates the problem that labeled brain network data for specific tasks or diseases are usually very limited.
-
We model brain networks through partially symmetric tensor factorization which is suitable for inherently undirected graphs, e.g., EEG brain networks. The temporal dimension is modeled as one of modes in the fourth-order tensor.
-
We blend representation learning and classifier training into a unified optimization problem, which allows classifier parameters to interact with discriminative latent factors by leveraging the supervision information.
-
We incorporate the \(\ell _{2,1}\) norm to conduct feature selection in the latent space, thereby identifying discriminative latent factors for different tasks.
2 Preliminaries
Table 1 lists some basic symbols that will be used throughout the paper. We introduce the concept of tensors which are higher order arrays that generalize the notions of vectors (the first-order tensors) and matrices (the second-order tensors), whose elements are indexed by more than two indices. Each index expresses a mode of the data and corresponds to a coordinate direction. The number of variables in each mode indicates the dimensionality of a mode. The order of a tensor is determined by the number of its modes. An mth-order tensor can be represented as \(\mathcal {X}=(x_{i_1,\cdots ,i_m})\in \mathbb {R}^{I_{1}\times \cdots \times I_{m}}\), where \(I_i\) is the dimension of \(\mathcal {X}\) along mode i. An overview of tensor notation and operators is given as follows which will be used to formulate the problem.
Definition 1
(Tensor Product). The tensor product \(\mathcal {X}\circ \mathcal {Y}\) of a tensor \(\mathcal {X} \in \mathbb {R}^{I_1\times \cdots \times I_m}\) and another tensor \(\mathcal {Y} \in \mathbb {R}^{I'_1\times \cdots \times I'_{m'}}\) is defined by \((\mathcal {X}\circ \mathcal {Y})_{i_1,...,i_m,i'_1,...,i'_{m'}} = x_{i_1,...,i_m}y_{i'_1,...,i'_{m'}}\).
Tensor product is also referred to as outer product in some literature [9, 29]. An mth-order tensor is a rank-one tensor if it can be defined as the tensor product of m vectors.
Definition 2
(Mode- k Product). The mode-k product \(\mathcal {X}\times _k\mathbf {A}\) of a tensor \(\mathcal {X} \in \mathbb {R}^{I_1\times \cdots \times I_m}\) and a matrix \(\mathbf {A} \in \mathbb {R}^{J\times I_k}\) is of size \(I_1\times \cdots \times I_{k-1}\times J\times I_{k+1}\times \cdots \times I_m\) and is defined by \((\mathcal {X}\times _k\mathbf {A})_{i_1,...,i_{k-1},j,i_{k+1},...,i_m}=\sum _{i_k=1}^{I_k}x_{i_1,...,i_m}a_{j,i_k}\).
Definition 3
(Kronecker Product). The Kronecker product of two matrices \(\mathbf {A} \in \mathbb {R}^{I \times J}, \mathbf {B} \in \mathbb {R}^{K \times L}\) is of size \(IK \times JL\) and is defined by
Definition 4
(Khatri-Rao Product). The Khatri-Rao product of two matrices \(\mathbf {A} \in \mathbb {R}^{I \times K}, \mathbf {B} \in \mathbb {R}^{J \times K}\) is of size \(IJ \times K\) and is defined by \(\mathbf {A}\odot \mathbf {B}=(a_1\otimes b_1,\cdots ,a_K\otimes b_K)\) where \(a_1,\cdots ,a_K,b_1,\cdots ,b_K\) are the columns of matrices \(\mathbf {A}\) and \(\mathbf {B}\), respectively.
Definition 5
(Partially Symmetric Tensor). A rank-one mth-order tensor \(\mathcal {X}\in \mathbb {R}^{I_{1}\times \cdots \times I_{m}}\) is partially symmetric if it is symmetric on modes \(i_1,...,i_j\in \{1,...,m\}\) and can be written as the tensor product of m vectors: \(\mathcal {X}=\mathbf {x}^{(1)}\circ \cdots \circ \mathbf {x}^{(m)}\) where \(\mathbf {x}^{(i_1)}=\cdots =\mathbf {x}^{(i_j)}\).
Definition 6
(Mode- k Matricization). The mode-k matricization of a tensor \(\mathcal {X}\in \mathbb {R}^{I_{1}\times \cdots \times I_{m}}\) is denoted by \(\mathbf {X}_{(k)}\) and arranges the mode-k fibers to be the columns of the resulting matrix. The dimension of \(\mathbf {X}_{(k)}\) is \(\mathbb {R}^{I_k\times J}\), where \(J=I_1\cdots I_{k-1}I_{k+1}\cdots I_m\). Each tensor element \((i_1,\cdots ,i_m)\) maps to the matrix element \((i_k,j)\): \(j=1+\sum _{p=1,p\ne k}^m(i_p-1)J_p~~\text {with}~~J_p=\prod _{q=1,q\ne k}^{p-1}I_q\).
3 SEMIBAT Framework
3.1 Problem Formulation
Let \(\mathcal {D}=\{G_1,\cdots ,G_n\}\) denote a dynamic graph dataset of brain networks where \(|\mathcal {D}|=n\) is the number of graph objects. All graphs in the dataset share a given set of vertices V which corresponds to a brain parcellation scheme. Suppose the brain is parcellated via an atlas into \(|V|=m\) regions, and the temporal dimensionality is t. A brain network \(G_i\) can be represented by a partially symmetric tensor \(\mathcal {Z}_i \in \mathbb {R}^{m \times m \times t}\). We assume that the first l graphs within \(\mathcal {D}\) are labeled and \(\mathbf {Y} \in \mathbb {R}^{l \times c}\) is the class label matrix where c is the number of class labels. \(\mathbf {Y}(i, j) = 1\) if \(G_i\) belongs to the j-th class, otherwise \(\mathbf {Y}(i, j) = 0\). For convenience, we also denote the labeled graph dataset by \(\mathcal {D}_l = \{G_1, \cdots , G_l\}\), and the unlabeled graph dataset as \(\mathcal {D}_u = \{G_{l+1}, \cdots , G_n\}\), \(\mathcal {D} = \mathcal {D}_l \cup \mathcal {D}_u\). In our experiments, brain networks under emotion regulation tasks compose labeled graphs, while those under resting-state compose unlabeled graphs.
3.2 Tensor Modeling
We first address the problem (P1) discussed in Sect. 1 by stacking the brain network dataset \(\mathcal {D}\) of n graphs, i.e., \(\{\mathcal {Z}_i\}_{i=1}^n\), as a tensor \(\mathcal {X} \in \mathbb {R}^{m \times m \times t \times n}\). Through joint tensor factorization, unlabeled graphs in \(\mathcal {D}_u\) could facilitate the representation learning of labeled graphs in \(\mathcal {D}_l\) by affecting latent factors.
Note that \(\mathcal {X}\) is a fourth-order partially symmetric tensor (symmetric on the first two modes) and it naturally models the temporal dimension discussed as the problem (P2). We assume that \(\mathcal {X}\) can be decomposed into k factors in the following manner
where \(\mathbf {B} \in \mathbb {R}^{m \times k}\) is the factor matrix for vertices, \(\mathbf {T} \in \mathbb {R}^{t \times k}\) is the factor matrix for time points, \(\mathbf {A} \in \mathbb {R}^{n \times k}\) is the factor matrix for graphs, and \(\mathcal {C} \in \mathbb {R}^{k \times \cdots \times k}\) is a fourth-order identity tensor, i.e., \(\mathcal {C}(i_1,\cdots ,i_4)=\delta (i_1=\cdots =i_4)\). Basically, Eq. (1) is a CANDECOMP/PARAFAC (CP) factorization [16] as shown in Fig. 2. It is desirable to discover distinct latent factors to obtain more concise and interpretable results, and thus we include orthogonality constraints \(\mathbf {A}^\mathrm {T}\mathbf {A}=\mathbf {I}\).Footnote 1
One of the targets is task recognition based on the brain network data. We assume that there is a matrix of regression coefficients \(\mathbf {W} \in \mathbb {R}^{k \times c}\) which assigns graphs with labels based on the graph factor matrix \(\mathbf {A}\), i.e., \(\mathbf {Y}=\mathbf {D}\mathbf {A}\mathbf {W}\) where \(\mathbf {D}=[\mathbf {I}^{l\times l},~\mathbf {0}^{l\times (n-l)}] \in \mathbb {R}^{l \times n}\).
An intuitive idea is to first learn latent representations of brain networks and then train a classifier on them in a serial two-step manner, which however would make these two procedures independent with each other and fail to introduce the supervision information to the representation learning process. Moreover, the advantage is established in [5] of directly searching for classification-relevant structure in the original data, rather than solving the supervised and unsupervised problems independently. To address the problem (P3) discussed in Sect. 1, we propose to incorporate the classifier learning process (i.e., \(\mathbf {W}\)) into the framework of learning latent feature representations of graphs (i.e., \(\mathbf {A}\)). In this manner, the weight matrix \(\mathbf {W}\) and the feature matrix \(\mathbf {A}\) can interact with each other in the same learning framework. Note that it is similar to coupled matrix and tensor factorization [2], however \(\mathcal {X}\) and \(\mathbf {Y}\) are coupled only in part of the graph mode, as shown in Fig. 3.
In summary, the proposed brain network analysis framework can be mathematically formulated as solving the following optimization problem
where \(\Vert \mathbf {W}^\mathrm {T}\Vert _{2,1}\) is the sparsity-promoting regularization term that controls the complexity of \(\mathbf {W}\) and has the effects of feature selection thereby addressing the problem (P4), and \(\alpha \), \(\lambda \) are positive parameters which control contributions of classification loss and regularization, respectively.
3.3 Optimization Framework
The model parameters that have to be estimated include \(\mathbf {B}\in \mathbb {R}^{m \times k}\), \(\mathbf {T}\in \mathbb {R}^{t \times k}\), \(\mathbf {A}\in \mathbb {R}^{n \times k}\) and \(\mathbf {W}\in \mathbb {R}^{k \times c}\). The optimization problem in Eq. (2) is non-convex with respect to \(\mathbf {B}\), \(\mathbf {T}\), \(\mathbf {A}\) and \(\mathbf {W}\) together. There is no closed-form solution for the problem. We now introduce an alternating scheme to solve the optimization problem. The key idea is to optimize the objective with respect to one variable, while fixing others, and decouple constraints using an Alternating Direction Method of Multipliers (ADMM) scheme [4]. The algorithm will keep updating the variables until convergence. First, we define the following notations
where \(\mathbf {P}\in \mathbb {R}^{m \times k}\) is the auxiliary variable.
Update the vertex factor matrix \(\mathbf {B}\) while fixing \(\mathbf {T}\), \(\mathbf {A}\) and \(\mathbf {W}\). Note that \(\mathcal {X}\) is a partially symmetric tensor and the objective function in Eq. (2) involves a fourth-order term w.r.t. \(\mathbf {B}\) which is difficult to optimize directly. To obviate this problem, we use a variable substitution technique and minimize the following objective function
The augmented Lagrangian function for problem in Eq. (3) is
where \(\varUpsilon \in \mathbb {R}^{m \times k}\) are Lagrange multipliers, \(\upsilon \) is the penalty parameter which can be adjusted efficiently according to [19].
By setting the derivative of Eq. (4) w.r.t. \(\mathbf {B}\) to zero, we obtain the closed-form solution
To efficiently compute \(\mathbf {E}^\mathrm {T}\mathbf {E}\), we consider the following property of the Khatri-Rao product of two matrices [16]
Similarly, the auxiliary matrix \(\mathbf {P}\) can be optimized successively
The Lagrange multipliers \(\varUpsilon \) can be updated using gradient ascent
Update the temporal factor matrix \(\mathbf {T}\) while fixing \(\mathbf {B}\), \(\mathbf {A}\) and \(\mathbf {W}\). Since there is no constraint on \(\mathbf {T}\), we directly obtain the closed-form solution
Update the graph factor matrix \(\mathbf {A}\) while fixing \(\mathbf {B}\), \(\mathbf {T}\) and \(\mathbf {W}\). By variable substitution, we need to minimize the following objective function
The augmented Lagrangian function for problem in Eq. (10) is
where \(\varPhi \in \mathbb {R}^{n \times k}\) and \(\varPsi \in \mathbb {R}^{k \times k}\) are Lagrange multipliers, \(\phi \) and \(\psi \) are penalty parameters. By setting the derivative of Eq. (11) w.r.t. \(\mathbf {A}\) to zero, we obtain the Sylvester equation
which can be solved by several numerical approaches, e.g., the lyap function in MATLAB.
The closed-form update for \(\mathbf {Q}\) is
The Lagrange multipliers \(\varPhi \) and \(\varPsi \) can be updated by
Update the weight matrix W while fixing B, T and A. According to the analysis of the \(\ell _{2,1}\) norm in [22], we need to minimize the following objective function
where \(\varOmega \in \mathbb {R}^{c \times c}\) is an auxiliary diagonal matrix of the \(\ell _{2,1}\) norm. The diagonal elements of \(\varOmega \) are computed as \(\varOmega (i,i) = \frac{1}{2\sqrt{||\mathbf {W}(:,i)||^2_2+\epsilon }}\) where \(\epsilon \) is a smoothing term which is usually set to a small constant.
By setting the derivative of Eq. (15) w.r.t. \(\mathbf {W}\) to zero, we obtain the Sylvester equation
Based on the above analysis, we develop the optimization framework for brain network analysis based on tensor factorization, as described in Algorithm 1. The code has been made available at the author’s homepageFootnote 2.
3.4 Time Complexity
Each ADMM iteration consists of simple matrix operations. Therefore, rough estimates of its computational complexity can be easily derived [18].
-
The estimate for the update of \(\mathbf {B}\) according to Eq. (5) is: (1) \(O(m^2ntk)\) for the computation of the term \(2\mathbf {X}_{(1)}\mathbf {E}+\upsilon \mathbf {P}+\varUpsilon \), (2) \(O((m+n+t)k^2)\) for the computation of the term \(2\mathbf {E}^\mathrm {T}\mathbf {E}+\upsilon \mathbf {I}\) due to Eq. (6), \(O(k^3)\) for its Cholesky decomposition, and (3) \(O(mk^2)\) for the computation of the system solution that gives the updated value of \(\mathbf {B}\). An analogous estimate can be derived for the update of \(\mathbf {P}\) and \(\mathbf {T}\) which cost \(O(k^3+(m+n+t)k^2+m^2ntk)\).
-
Considering \(l<n\) and c is usually a small constant, the estimate for the update of \(\mathbf {A}\) according to Eq. (12) is: (1) \(O(n^2k)\) for the computation of the term X, (2) \(O((m+n+t)k^2)\) for the computation of the term Y, (3) \(O(nk^2+m^2ntk+n^2)\) for the computation of the term Z, and (4) \(O(nk^2+n^2k)\) for the computation of the Sylvester equation [14].
-
The estimate for the update of \(\mathbf {Q}\) according to Eq. (13) is \(O(nk^2+n^2k+n^3)\).
-
The estimate for the update of \(\mathbf {W}\) according to Eq. (16) is \(O(nk^2+n^2k)\).
Overall, the updates of all model parameters require \(O(k^3+(m+n+t)k^2+(m^2nt+n^2)k+n^3)\) arithmetic operations in total.
4 Experiments
4.1 Data Collections
Data were collected from 22 healthy participants at the University of Illinois at Chicago (UIC) and from 11 healthy participants at the University of Michigan (UMich), respectively. Each participant underwent an emotion regulation task, while UIC participants further underwent an eight-minute resting-state recording session which served as unlabeled data. During the ERT session, participants were instructed to look at pictures displayed on the screen. Emotionally neutral pictures (e.g., landscape, everyday objects) and negative pictures (e.g., car crash, natural disasters) would appear on the screen for seven seconds in random orders. One second after the picture on display, a corresponding auditory guide would instruct the participant to look: viewing the neutral pictures; to maintain: viewing the negative pictures as they normally would; or to reappraise: viewing the negative pictures while attempting to reduce their emotion response by re-interpreting the meaning of pictures. All EEG data were recorded using the Biosemi system equipped with an elastic cap with 34 scalp channels. A detailed description about data acquisition and preprocessing is available in [28].
Overall, the dataset contains \(n=121\) EEG brain network samples that are based upon \(m=34\) vertices and \(t=130\) time points. The target is to train a classifier on the UIC source (66 training samples and 22 unlabeled samples) to predict which task (Neutral, Maintain, or Reappraise) a subject in the UMich source (33 test samples) is performing. The average brain networks are shown in Fig. 4 where the x and y axes represent the vertex id, and the color of the cell represents the strength of the connection between vertex x and y. Although the group difference appears to be significant, it is non-trivial to identify the tasks for each individual. It will be validated in the experiments that simply using edge values as features to train a classifier could not lead to a good classification performance.
4.2 Compared Methods
The compared methods are summarized as follows:
-
semiBAT: the proposed semi-supervised brain network analysis approach based on constrained tensor factorization.
-
BAT-ridge: replacing the \(\ell _{2,1}\) norm in semiBAT with a regular ridge term.
-
BAT-supv: a fully supervised variant of semiBAT without leveraging the unlabeled data.
-
BAT-unsupv: an unsupervised variant of semiBAT that first learns latent representations of brain networks and then trains a classifier on them in a serial two-step manner.
-
BAT-3d: applying semiBAT on time-averaged brain networks.
-
ALS: plain vanilla tensor factorization using alternating least squares without any constraint [8].
-
Subgraph: a discriminative subgraph selection method for uncertain graph classification [7, 17].
-
CC: extracting local clustering coefficients as features, one of the most popular graph-theoretical measures that quantify the cliquishness of the vertices [24].
-
Edge: using edge values as features by flatting adjacency matrices of brain networks into vectors.
For a fair comparison, we used a regularized regression in semiBAT as the base classifier for all the compared methods. The parameters \(\alpha \) and \(\lambda \) were tuned in the range of \(2^{-10},...,2^{10}\), the rank k was tuned in the range of 1, ..., 20. The accuracy with the best parameter configuration was reported, as well as the corresponding precision, recall and F1 score.
4.3 Classification Performance
Experimental results in Table 2 show the classification performance of compared methods on distinguishing the three tasks. Edge serves as the basis for comparison that treats a brain network as a collection of edges, thereby blinding the connectivity structures of brain networks, which surprisingly outperforms CC. Although clustering coefficients have been widely used to identify Alzheimer’s disease [13, 27], they appear to be less useful for distinguishing the emotion regulation tasks. Subgraph achieves a better performance by extracting connectivity patterns within brain networks.
Factorization models demonstrate themselves with significantly better accuracy. According to the low-rank assumption, a low-dimensional latent factor of each graph is obtained by first stacking all the brain network data and then factorizing the constructed tensor. ALS is a direct application of the alternating least squares technique to the standard tensor factorization problem without incorporating any constraint or supervision. A significant improvement of \(31.60\,\%\) by semiBAT over ALS can be observed, mainly due to the fact that the unsupervised ALS approach fails to interact with the classifier training procedure which shows comparable performance with BAT-unsupv. It indicates the importance of addressing the problem (P3) discussed in Sect. 1. Moreover, semiBAT outperforms BAT-ridge thereby demonstrating that it is critical to apply feature selection in the tensor factorization framework (i.e., the problem (P4)). The advantages of semiBAT over BAT-supv and BAT-3d are attributed to leveraging unlabeled resting-state brain network data (i.e., the problem (P1)) and modeling the temporal dimension (i.e., the problem (P2)), respectively.
4.4 Parameter Sensitivity
In all experiments, the regularization parameter \(\lambda \) was tuned for all the baselines, the rank k was tuned for all the factorization models, and \(\alpha \) was tuned for semiBAT and its variants. We first investigate the influence of \(\alpha \) and \(\lambda \) in semiBAT and present the results in Fig. 5. It illustrates that neither a small nor a large \(\alpha \) or \(\lambda \) would be preferred, and in general, a good choice of \(\alpha \) and \(\lambda \) can be found in the range of \(2^{5},...,2^{7}\) and \(2^{0},...,2^{2}\), respectively. Moreover, experimental results of factorization models with different k are shown in Fig. 6. In general, a small k would rarely be a wise choice, and the best performance can usually be achieved around \(k=17\).
4.5 Factor Analysis
We first investigate the factor matrices derived from semiBAT in a row-wise manner. Note that initially with the best parameter configuration as reported in the last section where \(k=17\), we obtain a 17-dimensional feature vector for each brain network (i.e., \(\mathbf {A}(i,:)\)) and each time point (i.e., \(\mathbf {T}(i,:)\)). For visualization we use t-SNE [21] to reduce them into a 2-dimensional space. In Fig. 7, we show the distribution of brain networks, where there are 99 points representing 33 samples from each of the three tasks (22 resting-state samples are omitted). A relatively clear separation between Neutral and Reappraise can be observed, while Maintain usually mix with the other two conditions which make the classification problem challenging. Figure 8 illustrates the distribution of time points, where there are 130 points and each of them represents an exact time point indicated by the color. Basically, adjacent time points are colored similarly. From this figure, we can see that continuous time points form distinct clusters and brain activities change over time, so it is important to capture the temporal dimension explicitly.
Next, we visualize and interpret the factor matrices in a column-wise manner. A k-factor semiBAT model extracts the factors \(\mathbf {B}(:,i)\), \(\mathbf {T}(:,i)\) and \(\mathbf {A}(:,i)\), for \(i=1,...,k\), where these factors indicate the signatures of sources in vertex, time and graph domain, respectively. We show the two largest factors in terms of magnitude in Fig. 9. In the left panel, points indicate the spatial layout of electrodes (i.e., vertices) on the scalp, and factor values of electrodes are demonstrated on a colormap using EEGLAB [10]. The middle panel shows the temporal changes of the factor. The right panel shows the strength of 66 brain networks in the training set performing different tasks where red, green and blue stand for Neutral, Maintain and Reappraise, respectively. A domain expert can identify the brain activity pattern in the left panel, the corresponding coefficients of time points in the middle panel, and the graph difference on such pattern in the right panel. We can see that different latent factors capture activity of different brain regions. The first factor appears to highlight a quantitative anterior-posterior gradient (maximum values of the first factor appear in the occipital lobe) that is shared across all three conditions, thus may be related to visual processing, while the second factor, which primarily differentiates neutral from maintain and reappraisal, predominantly involves electrodes around the frontal-parietal junction and thus may be related to the late positive potential [11, 12].
5 Related Work
Tensor factorization has become an effective technique in many healthcare applications. For example, Acar et al. identify spatial, spectral and temporal signatures of an epileptic seizure as well as an artifact through the application of tensor models [1]. Davidson et al. propose a constrained alternating least squares framework for network discovery of fMRI data [9]. Papalexakis et al. present a scalable solution for the coupled matrix-tensor factorization problem, and find latent variables that jointly explain both the brain activity and the behavioral responses [23]. Wang et al. introduce knowledge guided tensor factorization for computational phenotyping [26]. Ma et al. propose a spatio-temporal tensor kernel approach for whole-brain fMRI image analysis [20]. However, these frameworks are not directly applicable to partially symmetric tensor factorization or further task recognition.
For graph classification on brain networks, literatures have been focused on first deriving vector presentations from the brain network data which are then fed into conventional pattern classifiers. In general, two types of features are usually extracted: (1) graph-theoretical measures and (2) subgraph patterns. Wee et al. extract weighted local clustering coefficients of each brain region in relation to other regions in brain networks to quantify the prevalence of clustered connectivity around brain regions for diagnosis on Alzheimer’s disease [27]. In addition to the local network property, Jie et al. use a topology-based graph kernel to measure the topological similarity between paired fMRI brain networks [13]. Kong et al. propose a discriminative subgraph feature selection method based on dynamic programming to compute the probability distribution of discrimination scores for each subgraph pattern within a set of weighted graphs [17]. In contrast to focusing on the graph view alone, Cao et al. introduce a subgraph mining algorithm using side information guidance to find an optimal set of subgraph features for graph classification [6]. However, the expressiveness of these features is limited to the predefined formulations. It is critical to explore a larger space of potentially informative features to represent brain networks through data-driven approaches.
6 Conclusion
This paper presents semiBAT, a novel semi-supervised brain network analysis approach based on constrained tensor factorization. It leverages unlabeled resting-state brain networks for task recognition, explores the temporal dimension to capture the progress, incorporates classifier learning procedure to introduce supervision from labeled data, and selects discriminative latent factors for different tasks. ADMM is used to solve the optimization problem. In the experiments on EEG datasets, we demonstrate the superior performance of semiBAT on graph classification tasks over the state-of-art methods.
Notes
- 1.
We considered adding non-negativity constraints to enhance interpretability, but our preliminary results showed that it would degrade performance.
- 2.
References
Acar, E., Aykut-Bingol, C., Bingol, H., Bro, R., Yener, B.: Multiway analysis of epilepsy tensors. Bioinformatics 23(13), i10–i18 (2007)
Acar, E., Kolda, T.G., Dunlavy, D.M.: All-at-once optimization for coupled matrix and tensor factorizations (2011). arXiv:1105.3422
Ajilore, O., Zhan, L., GadElkarim, J., Zhang, A., Feusner, J.D., Yang, S., Thompson, P.M., Kumar, A., Leow, A.: Constructing the resting state structural connectome. Front. Neuroinformatics 7, 1427–1437 (2013)
Boyd, S., Parikh, N., Chu, E., Peleato, B., Eckstein, J.: Distributedoptimization and statistical learning via the alternating direction method ofmultipliers. Found. Trends. Mach. Learn. 3(1), 1–122 (2011)
Bzdok, D., Eickenberg, M., Grisel, O., Thirion, B., Varoquaux, G.: Semi-supervised factored logistic regression for high-dimensional neuroimaging data. In: NIPS, pp. 3330–3338 (2015)
Cao, B., Kong, X., Zhang, J., Yu, P.S., Ragin, A.B.: Mining brain networks using multiple side views for neurological disorder identification. In: ICDM, pp. 709–714. IEEE (2015)
Cao, B., Zhan, L., Kong, X., Yu, P.S., Vizueta, N., Altshuler, L.L., Leow, A.D.: Identification of discriminative subgraph patterns in fMRI brain networks in bipolar affective disorder. In: Brain Informatics and Health, pp. 105–114. Springer, Heidelberg (2015)
Comon, P., Luciani, X., De Almeida, A.L.: Tensor decompositions, alternating least squares and other tales. J. Chemometr. 23(7–8), 393–405 (2009)
Davidson, I., Gilpin, S., Carmichael, O., Walker, P.: Network discovery via constrained tensor analysis of fMRI data. In: KDD, pp. 194–202. ACM (2013)
Delorme, A., Makeig, S.: EEGLAB: an open source toolbox for analysis of single-trial EEG dynamics including independent component analysis. J. Neurosci. Methods 134(1), 9–21 (2004)
Dennis, T.A., Hajcak, G.: The late positive potential: a neurophysiological marker for emotion regulation in children. J. Child Psychol. Psychiatry 50(11), 1373–1383 (2009)
Hajcak, G., MacNamara, A., Olvet, D.M.: Event-related potentials, emotion, and emotion regulation: an integrative review. Dev. Neuropsychol. 35(2), 129–155 (2010)
Jie, B., Zhang, D., Gao, W., Wang, Q., Wee, C., Shen, D.: Integration of network topological and connectivity properties for neuroimaging classification. Biomed. Eng. 61(2), 576 (2014)
Jonsson, I., Kågström, B.: Recursive blocked algorithms for solving triangular systems part i: One-sided and coupled sylvester-type matrix equations. ACM Trans. Math. Softw. 28(4), 392–415 (2002)
Kim, J., Calhoun, V.D., Shim, E., Lee, J.H.: Deep neural network with weight sparsity control and pre-training extracts hierarchical features and enhances classification performance: Evidence from whole-brain resting-state functional connectivity patterns of schizophrenia. NeuroImage 124, 127–146 (2016)
Kolda, T.G., Bader, B.W.: Tensor decompositions and applications. SIAM Rev. 51(3), 455–500 (2009)
Kong, X., Yu, P.S., Wang, X., Ragin, A.B.: Discriminative feature selection for uncertain graph classification. In: SDM (2013)
Liavas, A.P., Sidiropoulos, N.D.: Parallel algorithms for constrained tensor factorization via the alternating direction method of multipliers (2014). arXiv:1409.2383
Lin, Z., Liu, R., Su, Z.: Linearized alternating direction method with adaptive penalty for low-rank representation. In: NIPS, pp. 612–620 (2011)
Ma, G., He, L., Lu, C.T., Yu, P.S., Shen, L., Ragin, A.B.: Spatio-temporal tensor analysis for whole-brain fMRI classification. In: SDM. SIAM (2016)
Van der Maaten, L., Hinton, G.: Visualizing data using t-sne. J. Mach. Learn. Res. 9(2579–2605), 85 (2008)
Nie, F., Huang, H., Cai, X., Ding, C.H.: Efficient and robust feature selection via joint 2, 1-norms minimization. In: NIPS, pp. 1813–1821 (2010)
Papalexakis, E.E., Mitchell, T.M., Sidiropoulos, N.D., Faloutsos, C., Talukdar, P.P., Murphy, B.: Turbo-smt: Accelerating coupled sparse matrix-tensor factorizations by 200x. In: SDM. SIAM (2014)
Rubinov, M., Sporns, O.: Complex network measures of brain connectivity: uses and interpretations. Neuroimage 52(3), 1059–1069 (2010)
Sporns, O., Tononi, G., Kötter, R.: The human connectome: a structural description of the human brain. PLoS Comput. Biol. 1(4), e42 (2005)
Wang, Y., Chen, R., Ghosh, J., Denny, J.C., Kho, A., Chen, Y., Malin, B.A., Sun, J.: Rubik: Knowledge guided tensor factorization and completion for health data analytics. In: KDD, pp. 1265–1274. ACM (2015)
Wee, C.Y., Yap, P.T., Zhang, D., Denny, K., Browndyke, J.N., Potter, G.G., Welsh-Bohmer, K.A., Wang, L., Shen, D.: Identification of MCI individuals using structural and functional connectivity networks. Neuroimage 59(3), 2045–2056 (2012)
Xing, M., Tadayonnejad, R., MacNamara, A., Ajilore, O., Phan, K.L., Klumpp, H., Leow, A.: EEG based functional connectivity reflects cognitive load during emotion regulation. In: ISBI. IEEE (2016)
Ye, J., Chen, K., Wu, T., Li, J., Zhao, Z., Patel, R., Bae, M., Janardan, R., Liu, H., Alexander, G., Reiman, E.: Heterogeneous data fusion for Alzheimer’s disease study. In: KDD. pp. 1025–1033. ACM (2008)
Zhang, J., Cao, B., Xie, S., Lu, C.T., Yu, P.S., Ragin, A.B.: Identifying connectivity patterns for brain diseases via multi-side-view guided deep architectures. In: SDM. SIAM (2016)
Acknowledgement
This work is supported in part by NSF through grants III-1526499.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2016 Springer International Publishing AG
About this paper
Cite this paper
Cao, B., Lu, CT., Wei, X., Yu, P.S., Leow, A.D. (2016). Semi-supervised Tensor Factorization for Brain Network Analysis. In: Frasconi, P., Landwehr, N., Manco, G., Vreeken, J. (eds) Machine Learning and Knowledge Discovery in Databases. ECML PKDD 2016. Lecture Notes in Computer Science(), vol 9851. Springer, Cham. https://doi.org/10.1007/978-3-319-46128-1_2
Download citation
DOI: https://doi.org/10.1007/978-3-319-46128-1_2
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-46127-4
Online ISBN: 978-3-319-46128-1
eBook Packages: Computer ScienceComputer Science (R0)