Keywords

1 Introduction

Brain has been known to be a network of cortical regions, yet until the relatively recent advances in magnetic resonance imaging (MRI), it was not possible to build network models of in-vivo brain. Current functional MRI (fMRI) and diffusion MRI (dMRI) technologies allow us to delineate functional (fNET) and structural network (sNET) models, collectively called the brain connectome, at a cortical parcellation scale. The analysis of these network models has the potential of shedding light on how the brain works, as well as the cause and progress of neurodegenerative diseases.

The majority of the analysis approaches has been focused on the changes in the global network features, such as the clustering coefficient, the average pathlength, the small-worldness index, etc. Despite their high sensitivity to abnormalities, these features are poor in classification and/or staging due to their global nature [1]. A more promising approach is to assess the changes in sub-networks of the connectome, towards which purely statistical techniques, such as Network Based Statistics (NBS) [2], have gained popularity. Another concern is the lack of standardized techniques for building connectomes, which introduces an unavoidable uncertainty on the derived conclusions.

We propose to use the powerful tensor factorization techniques to simultaneously address the aforementioned problems, namely sub-network based and multi-modal analysis of the connectome. We introduce the Brain Tensor (\(\mathbf{B}\)-tensor) as a multi-dimensional multi-modal connectome representation and factorize it in terms of apriori known canonical subnetworks. In a preliminary study with Alzheimer’s Disease patients, we demonstrate that the factorization coefficients not only have high discrimination power in a binary classification task, but also provide an insight into the most affected/discriminative canonical subnetworks. We conclude with a discussion on the potential extensions of \(\mathbf{B}\)-tensor factorization.

2 Background

Initial efforts on network analysis have focused on global, and local characteristics in order to identify components of networks, and to assess similarities or differences between networks. Utilizing global features such as clustering coefficients, average path length, small-worldness is useful when a given network is compared with a reference network, or to examine differences of neural networks from different species. Briefly, global features look at the network as a whole and fail to identify local difference. On the other hand, local features such as local clustering coefficient, shortest path etc. have shown their significance when properties of individual components are examined [3].

NBS has been proposed to overcome the limitation of global assessment. It identifies statistically significantly different edges between two classes of networks. The identified edges are used to detect discriminative subnetworks. The distribution of sizes of these subnetworks is used to assign a p-value to the subnetwork identified as discriminative between the positive classes. NBS is purely statistical, oversees the a priori known structure of brain.

Karahan et al. utilized coupled tensor factorization to fuse EEG, FMRI, and DTI data represented in \(3^{rd}\), \(2^{nd}\), and \(3^{rd}\) order tensors, respectively. Coupling is enforced over temporal and spatial domains for EEG/FMRI and over subjects for EEG/DTI. Two different representation is used for EEG where first one is time-varying and the other one is subject-varying EEG [4]. However, in this research, a priori is not used and multimodal structural and functional representation of brain networks have not been considered. Utilized tensors do not represent network sets row signals/spectra with a spatial and/or subject dimension. Williams et al. generate a \(3^{rd}\) order tensor model by using a trial-structured neural data with dimensions represent neurons, time and trials. With the help of TCA (Tensor Component Analysis), they have managed to decompose this tensor into three interpretable factors (neuron factors, temporal factors, trial factors). In addition, TCA has been utilized for dimension reduction [5].

We propose a 4th order \(\mathbf {B}\)-tensor that represents multi-modal networks (structural and/or functional) over a population. A modality is defined as a network construction method independent from data source (i.e. structural or functional data). Decomposition of \(\mathbf {B}\)-tensor over a priori known subnetwork is studied.

3 Method

3.1 B-Tensor Construction

For a multi-modal (R-modal) connectome defined over \(I \times J\) nodes (cortical parcels) for a population of K cases, the \(4^{th}\) order B-tensor (\(\mathbf{B} \in \mathbb {R}^{K \times I \times J \times R}\)) is defined. In this work, we used 6 variants of sNET definitions, hence \(R=6\), over a population of 20 cases, hence \(K=20\). We used the 148-parcel Destrieux atlas [6], hence \(I=J=148\).

Following the co-registration of T1-weighted MRI and dMRI volumes, the T1-weighted MRI volume is parcellated using FreeSurferFootnote 1, into 148 parcels which are used to define the 148 nodes (\(\{V_i\}\)) of the sNETs. Diffusion tensor (DT) volume is computed from diffusion weighted MRI (DWI) using an in-house software built upon the MITK platformFootnote 2. Fiber tracts are constructed using the \(4^{th}\)-order Runge-Kutta (RK4) deterministic tractography algorithm [7] with minimum fractional anisotropy (FA) set to 0.15, stepsize set to 0.7 mm (\(\approx \) half the voxel size), minimum fiber length set to 14 mm and the maximum curvature set to \(35^\circ \). RK4 was initiated from 30 randomly selected seeds per voxels with \(FA>0.15\).

In order to construct the sNETs, each fiber (\(\{f_k\}\)) is associated with the nodes in the vicinity of its end-points using a symmetric 3D Gaussian kernel with a standard deviation (\(\sigma \)) of 0.155 mm, centered at the fiber endpoints. \(\sigma \) is optimized by minimizing the integrated square error (ISE), as described in [8]. The numeric volumetric integral of the Gaussian kernel positioned at one of the end points of \(f_k\), within the node \(V_i\) (and up to a radial distance of \(2\sigma \)) is used as the fiber-parcel/node association and is denoted by \(W_{ik}\).

Six different sNETs are constructed using 6 different structural connectivity (network edge weight) definitions, \(C_{ij}\), between pairs of nodes, (\(V_i,V_j\)), as follows:

$$\begin{aligned} C_{ij} = \sum _k W_{ik} W_{jk}: \text{ Weighted } \text{ Connectivity } \end{aligned}$$
(1)
$$\begin{aligned} C_{ij}^{N}= & {} \frac{2\times \frac{C_{ij}}{\mathcal {V}}}{V_i + V_j} : \text{ Normalized } \text{ Connectivity } \end{aligned}$$
(2)
$$\begin{aligned} \nonumber C_{ij}^{stat}= & {} \frac{1}{C_{ij}}\sum _k W_{ik} W_{jk} \times \Psi (FA(f_k(t))) \text{, } \forall C_{ij} \ne 0 \\&\qquad \qquad \qquad \,\,\text {FA-based Connectivities} \end{aligned}$$
(3)

where \(\mathcal {V}\) is the voxel volume in \(mm^3\), \(V_i\) and \(V_j\) are the volumes of corresponding parcels, and \(\Psi \) represents the statistics operator (\(\in \){min, max, mean, median}) operating over the fiber parametrized by t.

3.2 B-Tensor Factorization

The CP factorization of \(\mathbf{B}\) is given as [9, 10],

$$\begin{aligned} \mathbf{B}_{k,i,j,r}\approx \sum _{q=1}^Q \mathbf{A}_{k,q}{} \mathbf{C}_{i,q}{} \mathbf{D}_{j,q}{} \mathbf{E}_{r,q} \end{aligned}$$
(4)

where the decomposition is performed over Q (free parameter) factors. Each one of the components represents the factorization of a single dimension of \(\mathbf{B}\) over the Q factors. While A and E are associated with the individual cases and the sNET definitions, C and D are solely associated with the network topology. Hence, we combined C and D into a single component that represents network topologies across Q factors, while A and E are merged to represent per case per connectivity (per-modality) represented in terms of those network topologies. Namely,

$$\begin{aligned} \mathbf{B}_{k,i,j,r} \approx \sum _{q=1}^Q \mathbf{M}_{k,r,q} \mathbf{G}_{i,j,q} = \sum _{q=1}^Q \Big (\sum _s \mathbf{A}_{k,s}{} \mathbf{E}_{r,s}{} \mathbf{I}_{q,s}\Big )\Big (\sum _s \mathbf{C}_{i,s}{} \mathbf{D}_{j,s}{} \mathbf{I}_{q,s}\Big ) \end{aligned}$$
(5)

where \(\mathbf{I}_{q,s}=\delta _{qs}\), i.e. the identity matrix and \(s\in \{1,2,\cdots ,Q\}\). This allows us to decouple the network topology from the cases and the modalities (connectivity definitions). Thus, we can work with case and modality independent network topologies, namely the canonical subnetworks, \(\mathbf{G}\). With a further simplification, we constrained \(\mathbf{G}\) to be a binary valued tensor representing the apriori known (fixed) canonical subnetworks. Following Yeo et al., we defined 7 canonical subnetworks, namely the visual, the somatomotor and auditory, the dorsal attention, the salience, the limbic, the frontoparietal and the default mode subnetworks [11]. They are expressed in terms of the node definitions of B and numbered from 1 to 7, respectively. \(\mathbf{G}\) is assumed known and fixed for the rest. \(\mathbf{M}\), on the other hand, represents the factorization coefficients over the canonical dimensions (subnetworks). This gives us 7D representations per case and per modality as \(\mathbf{F}_{k,r}\in \mathbb {R}^7\).

Following [12], Eq. 5 can be matricised as

$$\begin{aligned} \mathbf{B}_{(1,4;2,3)}=\mathbf{M}_{(3;1,2)}^T\mathbf{G}_{(3;1,2)} \end{aligned}$$
(6)

where \(\mathbf{M}_{(3;1,2)}=\mathbf{M}_{(3)}=(\mathbf{E}\odot \mathbf{A})^T\) and \(\mathbf{G}_{(3;1,2)}=(\mathbf{G}_{(3)}=\mathbf{D}\odot \mathbf{C})^T\).Footnote 3 Fixing \(\mathbf{G}\) as described above, we can solve for M using any matrix inversion technique, such as QR decomposition, to get \(\mathbf{M}_{(3)}=\mathbf{G}_{(3)}^{-T}{} \mathbf{B}_{(1,4;2,3)}^T\). \(\mathbf{G}_{(3)}^{-T}\) is computed once and used throughout the analysis. Finally, the computed \(\mathbf{M}_{(3)}\) is tensorized back to its original form to get the estimated \(\mathbf{M}\) as a \(3^{rd}\) order tensor representing per case, per connectivity unconstrained real-valued factorization coefficients.

4 Experiments

4.1 Data

T1-weighted MRI and dMRI images were acquired by using a Philips Achieva 3.0T X scanner with a 32-channel head coil from 7 AD patients and 13 controls with written consent. The AD patients were diagnosed by means of standard clinical evaluation tests by a team of expert neurologists. We used 3D FFE (Fast Field Echo) pulse sequence with multi-shot TFE (Turbo Field Echo) imaging mode for T1-weighted MRI. The acquisition parameters were TE/TR = 3.8ms/8.3ms, flip-angle = \(8^\circ \), SENSE reduction 2 (Foot-Head), FOV = \(220(RL) \times 240(AP)\) mm\(^2\), voxel size = \(1.0 \times 1.0 \times 1.0\) mm\(^3\) and number of slices = 180. dMRI were acquired with a maximum gradient strength of 40 mT/m, and 200 mT/m/ms slew rate, using a single-shot, pulse-gradient spin echo (PGSE), echo planar imaging (EPI) sequence. The acquisition parameters were FOV= \(200 \times 236\) mm\(^2\), 2.27 mm isotropic voxel size, \(112\times 112\) reconstruction matrix, 71 slices and \(TE/TR = 92\,\text {ms}{/}9032\,\text {ms}\). 120 diffusion weighting gradient directions were used at various b-values between \(3000-0\) s/mm\(^2\).

4.2 Analysis and Results

The separation of the AD patients and the controls in the 7D canonical space of subnetworks was assessed by Linear Discriminant Analysis(LDA) [13]. Binary classifiers are trained for each one of the 6 connectivity definitions in the corresponding 7D space (\(\mathbf {F}_r\)) and the training accuracies are measured. We also computed the unit normal of separating hyperplanes, i.e. the canonical axis, along which the separation of the two groups is maximized. \(Acc_{B}\) represents the binary classification accuracy. We also trained a separate binary classifier using LDA on the standard global network parameters (global clustering coefficients [14], average shortest path length [15], small-worldness index [16]). \(Acc_{glb}\) represents the classification accuracy using these global parameters. The results are given in Table 1 together with the corresponding canonical axes for the B-tensor analysis. All accuracies were above \(90\%\), with the weighted and normalized connectivity definitions being the best performing ones for \(\mathbf {B}\)-tensor. The corresponding classification accuracies using global connectome features in LDA are between 75%–90%.

Table 1. Classification accuracies of LDA classifier for different connectivity definitions (modalities) and the associated canonical axes.

The components of the canonical axes unit vectors provide an insight with regard to the relevance of the corresponding canonical subnetwork in discriminating the AD patients from the controls. The higher the absolute value of a component of a canonical axes, the more important that canonical dimension (subnetwork) is in discriminating the two groups. Figure 1 shows the mean and standard deviation of the magnitude of each component computed over all modalities.

Fig. 1.
figure 1

Mean and standard deviations of the canonical axes’ components computed over all modalities. The \(4^{th}\) subnetwork (Salience subnetwork) is consistently observed to be the most important one.

In order to compare the connectivity definitions with regard to their discriminating power, we ordered the canonical subnetworks based on the mean values given in Fig. 1 and computed the classification accuracy of LDA using the top K (\(\in [1,7]\)) canonical subnetworks, separately for each connectivity definition. Figure 2 shows the results. In general, increasing the dimension of the canonical space improves the accuracy, except for the 2D case (i.e. using the dorsal attention and the salience networks only). However, the weighted connectivity definition performed the best almost unanimously where as the FA-statistics based connectivities performed relatively poorly in general.

Fig. 2.
figure 2

Accuracy of LDA for all modalities using top K (\(\in [1,7]\)) canonical dimensions (subnetworks). The weighted connectivity definition performed the best almost unanimously where as the FA-statistics based connectivities performed relatively poorly in general.

5 Discussion

The B-tensor factorization allows us to represent the multi-modal (multi-connectivity) brain connectome in a canonical space of subnetworks with an intuitive interpretation. The AD patients are clearly separated from the controls in this space. The salience network is consistently observed to be the most important subnetwork among the 7 subnetworks used. The dorsal attention and the limbic subnetworks seem to be the second most important networks whereas the fronto-parietal network is the least important one. Although this result seems to be counter intuitive as the memory loss (primary function of limbic subnetwork) is the major symptom of AD, there is also evidence supporting our findings [17]. Furthermore, the AD cases in our dataset are described as early stage AD by our collaborating neurologists, which may also explain the observed importance of the salience subnetwork. These preliminary results are limited by the fixed definition of the canonical subnetworks. However, the B-tensor factorization framework can be utilized for searching for discriminative subnetworks. This would provide a further insight into the causes and progression mechanism of neurodegenerative diseases.

The assessment of different connectivity definitions within the aforementioned canonical space reveals that the weighted and normalized connectivity definitions’ discriminative power outperforms those of FA-statistics based connectivities. Although it has been discussed that the FA is an indirect measure of the quality of communication between cortical regions, these preliminary results suggest the opposite. This is potentially due to well-known deficiencies of diffusion tensor model that underlies the FA measurements. A similar assessment using the microstructural integrity of the fiber tracts by means of compartment models, such as NODDI [18], can potentially show a higher discriminative power.

The continuous representation of brain connectome in the canonical space can also be used for staging the disease progression. A regression analysis between this low-dimensional representation and clinical test results should be carried out, which is left for future work. Current study is limited to different sNET definitions, yet the framework is suitable to include fNETs in the B-tensor simultaneously with sNETs. Such a multi-modal analysis may uncover non-trivial relations between the structural and functional changes during the course of the disease, by means of the correlations of different modalities as represented in the canonical space. A fundamental limitation of the present study is the small dataset size which might have caused an overfitting of the LDA classifier. Further experiments on a larger dataset are due to arrive at stronger conclusions.

6 Conclusion

We have presented a novel tensor based multi-modal representation of brain connectome and described how it can be factorized to get a continuous, low-dimensional representation in a canonical space offering an intuitive understanding of neurodegenerative diseases’ causes and progression. Preliminary results with a small cohort of AD patients and controls revealed high classification accuracy and identified the salience subnetwork as the most discriminative network component among the 7 known.

Future work will include the experiments with a larger cohort, the extension of the model to joint analysis of structural and functional networks, the assessment of the canonical representation as a disease staging/monitoring biomarker and developing a canonical subnetwork search strategy optimized for classification/regression accuracy.