Next Article in Journal
A Compensatory Fuzzy Logic Model in Technical Trading
Next Article in Special Issue
Non-Standard Discrete RothC Models for Soil Carbon Dynamics
Previous Article in Journal
Outsourcing Reverse Logistics for E-Commerce Retailers: A Two-Stage Fuzzy Optimization Approach
Previous Article in Special Issue
A Cellular Potts Model for Analyzing Cell Migration across Constraining Pillar Arrays
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Role of Spectral Complexity in Connectivity Estimation

by
Elisabetta Vallarino
1,
Alberto Sorrentino
1,2,*,
Michele Piana
1,2 and
Sara Sommariva
1
1
Dipartimento di Matematica, Università di Genova, via Dodecaneso 35, 16146 Genova, Italy
2
CNR–SPIN, Corso Ferdinando Maria Perrone 24, 16152 Genova, Italy
*
Author to whom correspondence should be addressed.
Submission received: 30 November 2020 / Revised: 10 March 2021 / Accepted: 13 March 2021 / Published: 16 March 2021
(This article belongs to the Special Issue Differential Models, Numerical Simulations and Applications)

Abstract

:
The study of functional connectivity from magnetoecenphalographic (MEG) data consists of quantifying the statistical dependencies among time series describing the activity of different neural sources from the magnetic field recorded outside the scalp. This problem can be addressed by utilizing connectivity measures whose computation in the frequency domain often relies on the evaluation of the cross-power spectrum of the neural time series estimated by solving the MEG inverse problem. Recent studies have focused on the optimal determination of the cross-power spectrum in the framework of regularization theory for ill-posed inverse problems, providing indications that, rather surprisingly, the regularization process that leads to the optimal estimate of the neural activity does not lead to the optimal estimate of the corresponding functional connectivity. Along these lines, the present paper utilizes synthetic time series simulating the neural activity recorded by an MEG device to show that the regularization of the cross-power spectrum is significantly correlated with the signal-to-noise ratio of the measurements and that, as a consequence, this regularization correspondingly depends on the spectral complexity of the neural activity.

1. Introduction

Magnetoencephalography (MEG) provides high temporal resolution measurements of the magnetic field associated to neural currents. The MEG device relies on superconducting sensors, named SQUIDs, organized in a helmet array close and around the scalp. MEG experimental time series can be used essentially to address two neuroscientific problems, whose solution requires both an accurate mathematical modelization based on Maxwell’s equations, and the numerical reduction of such formal models [1].
The first problem is concerned with the dynamical ill-posed inverse problem of estimating parameters associated with the neural sources inducing the magnetic field signal [2,3,4,5,6,7,8]. The second problem is concerned with the quantification of the interactions among neural sources located in different cortical areas and intertwined by means of either anatomical or functional connectivity [9,10,11,12,13,14].
In particular, the connectivity problem can be addressed by either computing proper connectivity metrics directly from the experimental time series provided by the MEG sensors or searching for connections in the source space, i.e., among the neural time series estimated as solutions of the inversion process. This second approach has the advantages of reducing the impact of volume conduction and providing results that can be more easily interpreted in the framework of neuroscientific models [15,16,17]. Several approaches for identifying connectivity paths rely on physiological models assuming that the functional communication between different brain areas is regulated by the synchronization of their activity at specific temporal frequencies [18,19]. This implies that, for these models, the frequency domain represents the natural computational framework where to perform the connectivity analysis. This is the reason why, in the present paper, we focus on the analysis of the cross-power spectrum, which is the mathematical quantity of reference for the computation of most frequency-domain connectivity measures [11,20,21]. From an operational viewpoint, the computation of the cross-power spectrum in the source space typically relies on a two-step procedure: first the neural activity is estimated by applying a regularized inversion method on the recorded time series and then the cross-power spectrum is computed from the Fourier transform of the estimated neural time series [22].
This paper investigates how to optimally select the regularization parameter in the inversion procedure in order to obtain the best possible estimate of the neural cross-power spectrum. In fact, we consider the Tikhonov method (better known as Minimum Norm Estimation (MNE) in the MEG world [2]) as it is one of the most commonly used inverse methods in connectivity studies [23,24,25]; we study the interplay between the regularization parameter providing the reconstructed neural time series minimizing the relative error in 2 -norm, and the one that allows the optimal estimate of the cross-power spectrum according to the normalized Frobenius norm. The conceptual motivation of this problem is illustrated in Figure 1, which tentatively sketches the result of recent investigations in MEG-based connectivity research, i.e., that the regularization parameter leading to the optimal estimate of the neural activity may not lead to the optimal estimate of the cross-power spectrum and vice versa. In fact, in [26] the authors used numerical simulations to compare the parameter that provides the best estimate of the power spectrum with the one that provides the best estimate of coherence and showed that the latter is in general two orders of magnitude smaller than the former. More recently, Vallarino et al. [27] addressed an analogous problem via analytical computations, considering a simplified model. Specifically, under the assumption that the neural time series are realizations of white Gaussian processes, the authors proved that the parameter providing the best neural activity estimate is more than twice as large as the one providing the best estimate of the cross-power spectrum.
The present paper focuses on an analysis of the impact of spectral complexity of the actual neural signal on the value of the two regularization parameters. Specifically, we simulate synthetic MEG signals and discuss how the optimal parameter for the reconstruction of the cross-power spectrum depends on its signal-to-noise ratio and how this latter quantity is related to the spectral richness of the neural sources. To this aim, we considered a simulation setting in which the signal is modeled as a multivariate autoregressive process.
This paper is structured as follows. Section 2 introduces the problem in a formal way. Section 3 describes how the synthetic data are simulated and analyzed. Section 4 presents the results of the analysis. Our conclusions are offered in Section 5.

2. Definition of the Problem

2.1. Forward Model

Let X ( t ) = ( X 1 ( t ) , , X N ( t ) ) R N be a multivariate stationary stochastic process whose realizations x ( t ) can not be observed and let Y ( t ) = ( Y 1 ( t ) , , Y M ( t ) ) R M be the process whose realizations y ( t ) are used to infer information on x ( t ) . Let Y ( t ) and X ( t ) be related by the following equation
Y ( t ) = G X ( t ) + N ( t ) ,
where G R M × N is the forward matrix and N ( t ) = ( N 1 ( t ) , , N M ( t ) ) R M is the measurement noise, that is here assumed to be a white Gaussian process with zero mean and covariance matrix α 2 I , i.e.,  N ( t ) N ( 0 , α 2 I ) , independent from X ( t ) .

2.2. Cross-Power Spectrum

We are interested in reconstructing the cross-power spectrum of X ( t ) , which describes the statistical dependencies between each pair of time series ( X j ( t ) , X k ( t ) ) j , k { 1 , , N } . The cross-power spectrum is a one parameter family of N × N matrices S X ( f ) , whose ( j , k ) - t h element is defined as
S j , k X ( f ) = lim T 1 T E [ X ^ j ( f , T ) X ^ k ( f , T ) H ] ,
where X ^ j ( f , T ) is the Fourier transform of X j ( t ) over the interval [ 0 , T ] , defined as
X ^ j ( f , T ) = 0 T X j ( t ) e 2 π i f t d t
and X H is the Hermitian transpose of X [28].
Given a realization x ( t ) of the process X ( t ) , the cross-power spectrum S X ( f ) can be estimated via the Welch method [29], which consists in partitioning the data in P overlapping segments multiplied by a window function, { w ( t ) x p ( t ) } p = 1 P , computing their discrete Fourier transform x ^ p ( f ) = 1 L t = 0 L 1 x p ( t ) w ( t ) e 2 π i t f L and averaging:
S x ( f ) = L P W p = 1 P x ^ p ( f ) x ^ p ( f ) H , f = 0 , , L 1 ,
where L is the length of each segment and W = 1 L t = 0 L 1 w ( t ) 2 .
It is often the case that the data reach a high dimension, and visual inspection of the cross-power spectrum is not doable. In such cases a metric that describes the spectral properties of the signals would be useful. Here we use the spectral complexity coefficient, defined as follows.
Definition 1.
Given a realization x ( t ) of the process X ( t ) , and the corresponding cross-power spectrum S x ( f ) , we define the spectral complexity coefficient as the average of the elements of the upper triangular part of the matrix obtained by computing the squared 2 norm over the frequencies of S j , k x ( f ) , j , k = 1 , , N , that is
c = 2 N ( N + 1 ) j = 1 N k = j N f S j , k x ( f ) 2 .
The spectral complexity coefficient assumes small values if the elements of the cross-power spectrum are flat, that is when time series do not present any periodic trend and no dependencies among the pairs of time series are present. On the contrary, it assumes large values if the elements of the cross-power spectrum are peaked, that is when time series present periodic trends and complex relations among them. Finally, we observe that in Definition 1 only the elements on the upper triangular part of S x ( f ) are considered because S x ( f ) is Hermitian.

2.3. Two-Step Approach for Cross-Power Spectrum Estimation

Let us now consider a realization of Equation (1). Further than an estimate of the hidden data x ( t ) , an estimate of the cross-power spectrum can be obtained from y ( t ) . Such estimate can be achieved through a two-step process [22]:
  • First, a regularized estimate x λ ( t ) of x ( t ) is obtained by solving the inverse problem associated to Equation (1). Here we consider the Tikhonov regularized solution [30] of the problem which is defined as
    x λ ( t ) = arg min x ( t ) Gx ( t ) y ( t ) 2 2 + λ x ( t ) 2 2 ;
    where λ is a proper regularization parameter and · 2 is the 2 -norm.
  • Then, the corresponding estimate of the cross-power spectrum S x λ ( f ) is computed from the reconstructed time series using the Welch method, as described in the previous section.
Remark 1.
In many applied fields, Tikhonov regularization with an 2 penalty term has been outdated by more modern techniques that use sparsity-inducing penalization terms such as 1 or p with 0 < p < 1 . Indeed, also in the M/EEG literature there has been considerable effort in developing 1 solutions [31,32], and mixed norm solutions [33]; both these approaches have proved to provide superior performances in terms of localization of neural activity. However, these newer methods are seldom used in connectivity studies, for good reasons: 1 solutions computed independently at each time point produce extremely jittering reconstructions, resulting in highly sparse time courses that are not suitable for computing connectivity metrics. Mixed norms, that have been developed precisely to overcome this jittering problem, are computationally very expensive, and this actually prevents their use with the large datasets typically involved in connectivity studies.
When applying the described two-step process, the regularization parameter λ in Equation (6) has to be set for the computation of x λ ( t ) . Thus, the problem naturally arises of the choice of such parameter, which can be set in order to optimally reconstruct either x λ ( t ) or S x λ ( f ) . We define optimality through the minimization of the normalized norm of the discrepancy between the true and the reconstructed time series and cross-power spectra as follows.
Definition 2.
Given the regularized solution (6) and the cross-power spectrum (4), we define the optimal regularization parameter for the reconstruction of x ( t ) as
λ x * = arg min λ ε x ( λ ) with ε x ( λ ) = t x λ ( t ) x ( t ) 2 2 t x λ ( t ) 2 2 + t x ( t ) 2 2 ;
and the optimal parameter for the reconstruction of S x ( f ) as
λ S * = arg min λ ε S ( λ ) w i t h ε S ( λ ) = f S x λ ( f ) S x ( f ) F 2 f S x λ ( f ) F 2 + f S x ( f ) F 2 ;
where · F is the Frobenius norm; ε x ( λ ) and ε S ( λ ) will be called reconstruction errors.
The reconstruction errors range from 0 to 1 and penalize both a too small and a too large value of λ . In fact, they assume their maximum value when either λ is very high and thus x λ ( t ) is negligible with respect to x ( t ) , or when λ is too small and thus, vice versa, x ( t ) is negligible with respect to x λ ( t ) . This definition may appear overly complex compared to, e.g., a mere 2 -norm of the difference; however, in the presence of sparse data where only a few time series are non-zero, the simple 2 -norm would prefer a very high regularization parameter in order to minimize the error on the null time series, at the expense of the error on the non-zero ones; our definition aims to cope with this limitation of the 2 -norm. A similar definition has been introduced in [34].
In experimental contexts, where x ( t ) is not known, the choice of the optimal regularization parameter is crucial. This matter is widely discussed in literature [35,36,37,38], and many criteria have been proposed. Such criteria apply to Equation (1) and can be used to set the regularization parameter λ x . A possibility is to set the regularization parameter as a function of the signal-to-noise ratio (SNR), which describes the level of the desired signal with respect to that of the measurement noise; for Equation (1) the SNR is defined as follows.
Definition 3.
Consider the linear model (1). We define the signal-to-noise ratio of X ( t ) related to such model as
SNR X = 10 log 10 t G X ( t ) 2 2 t N ( t ) 2 2 .
To the best of our knowledge, the choice of the optimal regularization parameter for the reconstruction of the cross-power spectrum has never been related to the signal-to-noise ratio. This relation is presented in Section 4; however we first need to relate the cross-power spectrum of the unknown S X ( f ) with that of the data S Y ( f ) .
By computing the cross-power spectrum of both sides of Equation (1) and from the linearity of the Fourier transform it follows that
S Y ( f ) = G S X ( f ) G + S N ( f ) ,
where the mixed terms S XN ( f ) and S NX ( f ) are negligible thanks to the independence between X ( t ) and N ( t ) . Just like for Equation (1), we can define the signal-to-noise ratio for Equation (10) as follows.
Definition 4.
Consider the linear model (10). We define the signal-to-noise ratio of S X ( f ) related to such model as
SNR S = 10 log 10 f G S X ( f ) G F 2 f S N ( f ) F 2 .
This definition is in line with the definition of SNR X for the signal, the main difference being in the use of the Frobenius norm rather than the 2 -norm, motivated by the fact that we are working with matrices rather than vectors.

2.4. Multivariate Autoregressive Models

To model the statistical relationships between the different components of the stochastic process X ( t ) , in this work the latter is assumed to follow a stable multivariate autoregressive model of order P [39].
Definition 5.
A zero-mean stochastic process X ( t ) R N is said to follow a multivariate autoregressive (MVAR) model of order P if
X ( t ) = k = 1 P A ( k ) X ( t k ) + ε ( t ) t ,
where A ( k ) R N × N are fixed coefficient matrices, and  ε ( t ) R N is a white Gaussian noise process.
Moreover, the MVAR model described by Equation (12) is said to be stable if
det ( I k = 1 P A ( k ) z k ) 0 z C s . t . | z | 1 ,
where I is the identity matrix of size N.
Remark 2.
From Equation (12) it can be easily seen that the process X ( t ) is uniquely determined by the process ε ( t ) and by the first P time points, X ( 0 ) , ⋯, X ( P 1 ) . Indeed, consider for example an MVAR model of order 1 (a similar proof holds for the general case P > 1 and can be found in [39]); then, for each time point t
X ( t ) = A ( 1 ) X ( t 1 ) + ε ( t ) = A ( 1 ) 2 X ( t 2 ) + A ( 1 ) ε ( t 1 ) + ε ( t ) = A ( 1 ) t X ( 0 ) + k = 0 t 1 A ( 1 ) k ε ( t k ) .
Such a model satisfies the stability condition defined in Equation (13) if all the eigenvalues of the coefficient matrix A ( 1 ) have a modulus of less then one, the condition that guarantees the sequence of exponential matrices A ( 1 ) k k is absolutely summable.
According to Equation (12), if the process X ( t ) follows an MVAR model, then at each time point the value of X ( t ) can be derived as a weighted sum of the values of the process at the previous P time points, X ( t 1 ) , ..., X ( t P ) , plus a random perturbation ε ( t ) . In particular, the  ( i , j ) -th elements of the coefficient matrices, a i j ( 1 ) , ..., a i j ( P ) , describe how the value of the i-th component of the process depends on the past of the j-th component. Different connectivity patterns, with various levels of complexity, can thus be obtained by tuning the off-diagonal values of the coefficient matrices. Due to their flexibility and simplicity, MVAR models have been used by various authors in the framework of MEG functional connectivity estimation as a benchmark for testing and comparing different connectivity metrics [21,23,34,40,41,42,43]. Other models have been proposed to simulate different connectivity patterns, such as coherent sinusoidal time series [26], neural mass models [44,45], or Kuramoto models [46,47]. However a comprehensive comparison of all possible generative models is beyond the scope of this work.

3. Generation and Analysis Pipeline of the MEG Simulated Data

In this section we describe the numerical simulation that led to the main results of our study. First we introduce the continuous MEG forward problem and its discretized version, then we describe how we generated the data and, finally, we describe the inverse model and how we numerically computed the optimal regularization parameters.

3.1. MEG Forward Model

The MEG forward problem aims at computing the magnetic field produced outside the head by an electric current that flows inside the brain. The quasi static approximation of Maxwell’s equations provides the local relationship between the recorded magnetic field and the neural currents [1,48,49]. The two equations that are of interest here read as
× E ( r , t ) = 0
× B ( r , t ) = μ 0 J ( r , t ) ;
where E ( r , t ) and B ( r , t ) are the electric and magnetic fields at location r and time t, μ 0 is the magnetic permeability in vacuum and J ( r , t ) is the total electric current that flows inside the brain. The latter is the sum of two contributions
J ( r , t ) = J p ( r , t ) + J v ( r , t ) ,
J p ( r , t ) being the primary current directly related to the brain activity, while J v ( r , t ) = σ ( r ) V ( r , t ) is the induced volume current due the non-null conductivity σ ( r ) of the brain, V ( r , t ) being the electric scalar potential.
The manipulation of Maxwell’s Equations leads to the Biot–Savart Equation
B ( r , t ) = B 0 ( r , t ) μ 0 4 π Ω σ ( r ) V ( r , t ) × r r | r r | 3 d v ,
where Ω is the volume occupied by the brain, the first term B 0 ( r , t ) = μ 0 4 π Ω J ( r , t ) r r | r r | 3 d v is the magnetic field induced by the primary current, whereas the second term is related to the volume current.
Solving the forward problem requires the computation of these two contributions knowing the primary current. Although for the first one straightforward numerical integration is feasible, for the second one it is common to model the head as the union of nested homogeneous volumes { Ω j } j = 1 , , J and to replace volume integration with surface integration. In this way, the Biot–Savart equation becomes
B ( r , t ) = B 0 ( r , t ) + μ 0 4 π i , j ( σ i σ j ) Ω i , j V ( r , t ) r r | r r | 3 × n i , j ( r ) d s ,
where Ω i , j is the contact surface between regions Ω i and Ω j and n i , j ( r ) is the unit vector normal to the surface Ω i , j at r from region i to region j.
The forward problem can now be solved by computing the second term at the right hand side of Equation (18) after having computed V ( r , t ) by solving the equation
· J p ( r , t ) · σ ( r ) V ( r , t ) = 0 ,
which follows from Equation (15) by applying the divergence.
For further details on the MEG forward problem, we refer the reader to [1].

3.2. The Leadfield Matrix

Experimental contexts require the discretization of the forward problem. This involves a discretization of both the volume occupied by the brain and the volume outside the head.
When using a distributed model for the primary current J p , the brain volume is uniformly divided in N small parcels. If N is sufficiently big and thus each parcel has a sufficient small area, the activity in each brain parcel is approximated by a point-like source, henceforth denoted as dipole. From a mathematical point of view, each dipole is a vector whose strength and direction represent the intensity and orientation of the primary current in the corresponding brain area [1].
As for the volume outside the brain, it is natural to discretize it in correspondence of the MEG sensors. Let us denote the measured magnetic field as y ( t ) = ( y 1 ( t ) , , y M ( t ) ) . Now, observing that the magnetic field B depends linearly on the primary current J p , the magnetic field in correspondence of the sensors of the instrument is
y ( t ) = k = 1 N G ( r k ) q k ( t ) + n ( t ) ,
where r k , k = 1 , , N , is the location of the k-th brain parcel, G ( r k ) R M × 3 is the corresponding leadfield matrix, and  { q k ( t ) } k = 1 , , N are the electric current intensities along the three orthogonal direction of the N dipoles within the brain at time t and n ( t ) is the measurement noise. The l-th column of G ( r k ) contains the measurement at a sensor level when a unit current dipole is placed at location r k and oriented along the l-th orthogonal direction.
In this work, we assume dipoles to be located only on the brain cortical mantle and their orientation to be normal to the local cortical surface [50]. In this case, the electric current intensities are scalar quantities (we refer to them as { q k } k = 1 , , N ) and the leadfield matrices are column vectors (we refer to them as { G k } k = 1 , , N ).
Let us define
x ( t ) : = ( q 1 ( t ) , , q N ( t ) )
and
G : = [ G 1 , , G N ] R M × N ;
reassembling Equations (21) and (22) in to Equation (20), we get
y ( t ) = G x ( t ) + n ( t ) ,
which can be interpreted as a realization of Equation (1). From now on we refer to G as to the leadfield matrix.
For the simulation presented in this work, we used the leadfield matrix available in the sample dataset of MNE Python [51]. We selected magnetometers and set a fixed orientation. For computational reasons, the available source space, containing 1884 sources, was uniformly down-sampled to obtain 274 sources. Thus, our model has M = 102 sensors and N = 274 dipole sources.

3.3. Data Generation

We simulated N m o d = 10 pairs of active sources, ( z 1 ( t ) , z 2 ( t ) ) , with unidirectional coupling from the first to the second; their time series follow a multivariate autoregressive (MVAR) model of order P = 5 [39,43]
z 1 ( t ) z 2 ( t ) = k = 1 P a 1 , 1 ( k ) 0 a 2 , 1 ( k ) a 2 , 2 ( k ) z 1 ( t k ) z 2 ( t k ) + ε 1 ( t ) ε 2 ( t ) , t = P , , T .
The non-zero elements a i , j ( k ) of the coefficient matrices were drawn from a normal distribution of zero mean and standard deviation γ , and T = 10,000. We retained only coefficient matrices providing (i) a stable MVAR model [39] and (ii) pairs of signals ( z 1 ( t ) , z 2 ( t ) ) such that the 2 -norm of the strongest one was less than three times the 2 -norm of the weakest one. In order to obtain time series with different spectral complexity coefficients we set γ to N m o d different values randomly drawn in the interval [ 0.1 , 1 ] . The values of the spectral complexity coefficient of the N m o d simulated time series are reported in Table 1. Finally, the resulting time series ( z 1 ( t ) , z 2 ( t ) ) were normalized by the mean of their standard deviations over time, so that pairs of time series drawn from different models had similar magnitude. Figure 2 shows a sample of the the cross-power spectra among the simulated pairs of time series. The figure shows that for increasing values of the spectral complexity coefficient the cross-power spectrum of the corresponding time series becomes more peaked. Each pair of simulated time series was then assigned to N l o c = 20 pairs of point like sources randomly chosen in the source space, so that the ratio of the norms of the corresponding columns of the leadfield matrix was close to one, i.e., they had similar intensity at a sensor level, and their distance was grater than 7 cm . The remaining N 2 sources were set to have null activity.
Source space activity was then projected to sensor level by multiplying the simulated source activity by the leadfield matrix and white Gaussian noise was added to obtain N s n r = 6 levels of SNR X evenly spaced in the interval [ 20 dB , 5 dB ] .
Summarizing, we generated N m o d · N l o c · N s n r = 1200 different sensor level configurations. The green box in Figure 3 shows a visual representation of the simulation pipeline.

3.4. Inverse Model

Source space time series were reconstructed using the Tikhonov method, also known as the minimum norm estimate (MNE) [2] within the MEG community. For each combination of source time series, source locations, and SNR X level, we computed the optimal regularization parameters λ x * and λ S * by minimizing the reconstruction errors ε x ( λ ) and ε S ( λ ) , defined in Definition 2. The minimization procedure was achieved by using the Matlab built in function fminsearch that implements an iterative procedure based on the simplex method developed by Lagarias and colleagues [52]. In more detail, λ x * and λ S * have been obtained by applying such procedure to ε x ( λ ) and ε S ( λ ) , respectively; in both cases the starting point of the simplex method was set equal to 10 SNR X 10 , which corresponds to the optimal value of λ x in the case of white Gaussian signals [27]. The blue box in Figure 3 describes the inverse procedure to obtain an estimate of the cross-power spectrum and stresses the role of the regularization parameter in the two-step process.

4. Results

In this section we illustrate the results of our analysis. We begin with the description of the analytical dependence between SNR X and SNR S , then we highlight how the optimal parameter for the reconstruction of the cross-power spectrum depends on SNR S and how this implies that the spectral complexity of the signal is behind such dependence. Finally we show how the reconstruction error ε S ( λ ) behaves for different values of the regularization parameters. As a byproduct, this analysis also confirms the results of Vallarino et al. [27] in the case of a more complex setting.

4.1. Analytical Relation between SNR X and SNR S

From Equations (9) and (11) and reminding that N ( t ) N ( 0 , α 2 I ) it follows that
SNR X = 10 log 10 t G X ( t ) 2 2 M T α 2 ;
and
SNR S = 10 log 10 f G S X ( f ) G F 2 M N f α 2 ,
where T is the number of time points and N f is the number of frequencies used to compute the cross-power spectrum. Observe that to derive Equation (26) we used the fact that the cross-power spectrum of a white noise Gaussian process of zero mean and covariance matrix α 2 I is S N ( f ) = α 2 I .
By isolating α 2 from Equation (25) and substituting in Equation (26) we obtain
SNR S = 10 log 10 T 2 M f G S X ( f ) G F 2 N f t G X ( t ) 2 4 + 2 SNR X .
Equation (27) relates the signal-to-noise ratio of X ( t ) with that of S X ( f ) . It shows that, for same levels of SNR X , SNR S changes with the spectral complexity coefficient of the signals. In fact, the higher the spectral complexity coefficient, the higher the quantity G S X ( f ) G F 2 . Intuitively, this happens because when the signal has a higher spectral complexity coefficient its cross-power spectrum is more peaked and thus it is stronger over the cross-power spectrum of the noise with respect to a signal with a lower spectral complexity coefficient.

4.2. Dependence of λ S * on SNR S

As described in Section 3 we simulated several sensor level configurations, based on different combinations of spectral complexity coefficients, source locations, and SNR X levels. For each configuration, we collected the two optimal parameters λ x * and λ S * and we investigated their dependence on the signal-to-noise-ratio. In accordance with classical results from inverse theory [36], we found that λ x * depends on the signal-to-noise ratio. What is novel here is the relation between λ S * and both SNR X and SNR S . Indeed, for increasing SNR X , less regularization is needed, but such dependence varies with the MVAR models. On the other side, the dependence of λ S * on SNR S is neater and does not depend on the models. Figure 4 shows this result; on the left the regularization parameters for the cross-power spectrum reconstruction versus SNR X are shown, while on the right the same parameters are shown with respect to SNR S . For the ease of presentation the figure shows the parameters related to one source location; while on the left lines corresponding to different MVAR models have different heights, on the right they overlap.

4.3. λ S * < λ x * and Dependency from the Spectral Complexity

We also investigated the relation between the two optimal regularization parameters. Figure 5 shows the ratio λ S * λ X * versus SNR X for the simulated MVAR models. The ratio between the two parameters is always smaller than 1 2 , meaning that λ S * < 1 2 λ x * , as it was analytically proved in a simplified case in [27]. Further to this, the figure shows that for increasing spectral complexity coefficients this ratio gets smaller. This latter result is directly related to Equation (27). In fact, for same levels of SNR X , signals with higher spectral complexity have higher SNR S and, thus, need less regularization.

4.4. The Reconstruction Errors

To show the benefit of using a value of the regularization parameter different from λ x * when estimating the cross-power spectrum, in Figure 6 we plotted the reconstruction errors ε S ( λ ) as a function of the regularization parameter (normalized by λ x * ) obtained when considering two illustrative realizations of the simulated sensor data. Specifically, we fixed the locations and time courses of the pair of interacting sources and we considered the corresponding simulated MEG data for two levels of SNR X , namely SNR X = 20 dB and SNR X = 5 dB. Similar results where obtained when considered the other source configurations.
As shown by Figure 6, for both the values of SNR X the value of the reconstruction error significantly decreases when λ S * is used instead of λ x * . Specifically in this simulation, ε S ( λ ) drops from 0.99 to 0.96 when SNR X = 20 dB, and from 0.92 to 0.77 when SNR X = 5 dB.
Notably, one may observe that the relative reconstruction errors shown in Figure 6 are rather large, being above 90 % in the low-SNR case and remaining above 75 % even in the high-SNR scenario. We point out that this fact is mainly due to the combined effect of two factors: first, Tikhonov regularization tends to produce reconstructions that are small but non-zero almost everywhere, as it reduces but does not cancel entirely backprojection of noise; second, in our simulations the true activity is zero everywhere but in two points. These two facts inevitably lead to large relative errors that, however, pleasantly decrease for increasing values of SNR X .

5. Discussion and Conclusions

In the present work, we investigated the role of the spectral complexity of a time series, x ( t ) , in the design of an optimal inverse technique for estimating its cross-power spectrum, S x ( f ) , from indirect measurements of the time series itself. Motivated by an analysis pipeline widely used for estimating brain functional connectivity from MEG data, we reconstructed the cross-power spectrum in two steps: first, we estimated the unknown time series by using the Tikhonov method, then, we computed the cross-power spectrum of the reconstructed time series. In the present work, we used numerical simulations to study how the spectral complexity of x ( t ) impacts the value of the regularization parameter that provides the best reconstruction of the cross-power spectrum.
As a first analytical result, we related SNR X to SNR S , i.e., the signal-to-noise ratio of the time series and the signal-to-noise ratio of the corresponding cross-power spectra. The obtained formula suggests that, for a fixed level of SNR X , SNR S depends on the spectral complexity of x ( t ) : the higher the spectral complexity coefficient the higher SNR S . Intuitively this happens because a higher value of the spectral complexity coefficient corresponds to a more peaked cross-power spectrum that will emerge over the cross-power spectrum of the noise.
To test the effect of this result on the choice of the Tikohonov regularization parameter in a practical scenario, we simulated a large set of MEG data and applied the described two-step approach for estimating the cross-power spectrum of the underlying neural sources. In details, we simulated 1200 synthetic MEG data with varying SNR X generated by pairs of coupled point-like sources at varying locations and with different spectral complexities. For each simulated data, we computed the two parameters providing the best estimates of the time series ( λ x * ) and of the cross-power spectrum ( λ S * ), defined as the ones minimizing the relative 2 norm of the difference between the true and the reconstructed time series/cross-power spectrum according to Definition 2. As shown by Figure 4, the results of our simulations highlighted a high correlation between the values of λ S * and of SNR S .
Eventually, we focused on the relationship between the two parameters λ x * and λ S * , whose ratio is shown in Figure 5. The figure points out that this ratio depends on the spectral complexity of the simulated time series. This fact may be understood in lights of the previous results, as λ S * depends on SNR S that in turns depends on the spectral complexity coefficient. Additionally, we found that, for all the simulated data, λ S * λ x * < 1 2 , in line with the results shown in [27] for a simplified model where the neural time series were assumed to be white Gaussian processes. Moreover, when the spectral complexity coefficient increases ( c > 5 in our simulations) the ratio between the two parameters approaches 0.01. This agrees with the results shown in [26] where, by simulating sinusoidal signals, the authors suggested to use for connectivity estimation a parameter of two orders of magnitude lower. In fact, our numerical results indicate that the use of λ S * results in a substantially lower reconstruction error on the cross-power spectrum, particularly when the data have a high SNR.
The present work focuses on the cross-power spectrum as a connectivity metric. Even though the cross-power spectrum is the starting point for the computation of many connectivity metrics it would be interesting to directly investigate the behavior of the Thikonov regularization parameters when using such metrics. Future works will be devoted to this. It is also worth noticing that the definition of optimality when defining the regularization parameters is not univocal, since many metrics can be used. A common example is the area under the curve (AUC), which is the metric that was used in [26]. The use of different metrics would firstly strengthen our results and would also allow a more straightforward comparison with the results of [26]. Finally, the dependence of λ S * on SNR S suggests that an analysis of such dependence could be considered for the definition of a rule for choosing λ S * in practical scenarios.

Author Contributions

Conceptualization: M.P., S.S., and A.S.; methodology: S.S. and E.V.; software: E.V.; validation and writing: all authors. All authors have read and agreed to the published version of the manuscript.

Funding

E.V., A.S., S.S. and M.P. have been partially supported by Gruppo Nazionale per il Calcolo Scientifico.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hämäläinen, M.; Hari, R.; Ilmoniemi, R.J.; Knuutila, J.; Lounasmaa, O.V. Magnetoencephalography—theory, instrumentation, and applications to noninvasive studies of the working human brain. Rev. Mod. Phys. 1993, 65, 413. [Google Scholar] [CrossRef] [Green Version]
  2. Hämäläinen, M.S.; Ilmoniemi, R.J. Interpreting magnetic fields of the brain: Minimum norm estimates. Med. Biol. Eng. Comput. 1994, 32, 35–42. [Google Scholar] [CrossRef]
  3. Calvetti, D.; Pascarella, A.; Pitolli, F.; Somersalo, E.; Vantaggi, B. A hierarchical Krylov–Bayes iterative inverse solver for MEG with physiological preconditioning. Inverse Probl. 2015, 31, 125005. [Google Scholar] [CrossRef]
  4. Costa, F.; Batatia, H.; Oberlin, T.; D’Giano, C.; Tourneret, J.Y. Bayesian EEG source localization using a structured sparsity prior. NeuroImage 2017, 144, 142–152. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Sorrentino, A.; Piana, M. Inverse Modeling for MEG/EEG data. In Mathematical and Theoretical Neuroscience; Springer: Cham, Switzerland, 2017; pp. 239–253. [Google Scholar]
  6. Bekhti, Y.; Lucka, F.; Salmon, J.; Gramfort, A. A hierarchical Bayesian perspective on majorization-minimization for non-convex sparse regression: Application to M/EEG source imaging. Inverse Probl. 2018, 34, 085010. [Google Scholar] [CrossRef]
  7. Luria, G.; Duran, D.; Visani, E.; Sommariva, S.; Rotondi, F.; Sebastiano, D.R.; Panzica, F.; Piana, M.; Sorrentino, A. Bayesian multi-dipole modelling in the frequency domain. J. Neurosci. Methods 2019, 312, 27–36. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Ilmoniemi, R.J.; Sarvas, J. Brain Signals: Physics and Mathematics of MEG and EEG; The MIT Press: Cambridge, MA, USA, 2019. [Google Scholar]
  9. Geweke, J. Measurement of linear dependence and feedback between multiple time series. J. Am. Stat. Assoc. 1982, 77, 304–313. [Google Scholar] [CrossRef]
  10. Baccalá, L.A.; Sameshima, K. Partial directed coherence: A new concept in neural structure determination. Biol. Cybern. 2001, 84, 463–474. [Google Scholar] [CrossRef]
  11. Nolte, G.; Bai, O.; Wheaton, L.; Mari, Z.; Vorbach, S.; Hallett, M. Identifying true brain interaction from EEG data using the imaginary part of coherency. Clin. Neurophysiol. 2004, 115, 2292–2307. [Google Scholar] [CrossRef]
  12. Pereda, E.; Quiroga, R.Q.; Bhattacharya, J. Nonlinear multivariate analysis of neurophysiological signals. Prog. Neurobiol. 2005, 77, 1–37. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Sakkalis, V. Review of advanced techniques for the estimation of brain connectivity measured with EEG/MEG. Comput. Biol. Med. 2011, 41, 1110–1117. [Google Scholar] [CrossRef] [PubMed]
  14. Chella, F.; Marzetti, L.; Pizzella, V.; Zappasodi, F.; Nolte, G. Third order spectral analysis robust to mixing artifacts for mapping cross-frequency interactions in EEG/MEG. NeuroImage 2014, 91, 146–161. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Schoffelen, J.M.; Gross, J. Source connectivity analysis with MEG and EEG. Hum. Brain Mapp. 2009, 30, 1857–1865. [Google Scholar] [CrossRef] [PubMed]
  16. Barzegaran, E.; Knyazeva, M.G. Functional connectivity analysis in EEG source space: The choice of method. PLoS ONE 2017, 12, e0181105. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Van de Steen, F.; Faes, L.; Karahan, E.; Songsiri, J.; Valdes-Sosa, P.A.; Marinazzo, D. Critical comments on EEG sensor space dynamical connectivity analysis. Brain Topogr. 2019, 32, 643–654. [Google Scholar] [CrossRef] [Green Version]
  18. Fries, P. A mechanism for cognitive dynamics: Neuronal communication through neuronal coherence. Trends Cogn. Sci. 2005, 9, 474–480. [Google Scholar] [CrossRef] [PubMed]
  19. Fries, P. Rhythms for cognition: Communication through coherence. Neuron 2015, 88, 220–235. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Nunez, P.L.; Silberstein, R.B.; Shi, Z.; Carpenter, M.R.; Srinivasan, R.; Tucker, D.M.; Doran, S.M.; Cadusch, P.J.; Wijesinghe, R.S. EEG coherency II: Experimental comparisons of multiple measures. Clin. Neurophysiol. 1999, 110, 469–486. [Google Scholar] [CrossRef]
  21. Nolte, G.; Ziehe, A.; Nikulin, V.V.; Schlögl, A.; Krämer, N.; Brismar, T.; Müller, K.R. Robustly estimating the flow direction of information in complex physical systems. Phys. Rev. Lett. 2008, 100, 234101. [Google Scholar] [CrossRef] [Green Version]
  22. Schoffelen, J.M.; Gross, J. Studying dynamic neural interactions with MEG. In Magnetoencephalography: From Signals to Dynamic Cortical Networks; Springer: Cham, Switzerland, 2019; pp. 1–23. [Google Scholar]
  23. Anzolin, A.; Presti, P.; Van De Steen, F.; Astolfi, L.; Haufe, S.; Marinazzo, D. Quantifying the effect of demixing approaches on directed connectivity estimated between reconstructed EEG sources. Brain Topogr. 2019, 32, 655–674. [Google Scholar] [CrossRef]
  24. Mahjoory, K.; Nikulin, V.V.; Botrel, L.; Linkenkaer-Hansen, K.; Fato, M.M.; Haufe, S. Consistency of EEG source localization and connectivity estimates. NeuroImage 2017, 152, 590–601. [Google Scholar] [CrossRef]
  25. Hincapié, A.S.; Kujala, J.; Mattout, J.; Pascarella, A.; Daligault, S.; Delpuech, C.; Mery, D.; Cosmelli, D.; Jerbi, K. The impact of MEG source reconstruction method on source-space connectivity estimation: A comparison between minimum-norm solution and beamforming. NeuroImage 2017, 156, 29–42. [Google Scholar] [CrossRef]
  26. Hincapié, A.S.; Kujala, J.; Mattout, J.; Daligault, S.; Delpuech, C.; Mery, D.; Cosmelli, D.; Jerbi, K. MEG Connectivity and Power Detections with Minimum Norm Estimates Require Different Regularization Parameters. Comput. Intell. Neurosci. 2016, 2016. [Google Scholar] [CrossRef] [Green Version]
  27. Vallarino, E.; Sommariva, S.; Piana, M.; Sorrentino, A. On the two-step estimation of the cross-power spectrum for dynamical linear inverse problems. Inverse Probl. 2020, 36, 045010. [Google Scholar] [CrossRef]
  28. Bendat, J.S.; Piersol, A.G. Random Data: Analysis and Measurement Procedures; John Wiley & Sons: Hoboken, NJ, USA, 2011; Volume 729. [Google Scholar]
  29. Welch, P. The use of fast Fourier transform for the estimation of power spectra: A method based on time averaging over short, modified periodograms. IEEE Trans. Audio Electroacoust. 1967, 15, 70–73. [Google Scholar] [CrossRef] [Green Version]
  30. Tikhonov, A.N.; Goncharsky, A.; Stepanov, V.; Yagola, A.G. Numerical Methods for the Solution of Ill-Posed Problems; Springer Science & Business Media: Dordrecht, The Netherlands, 2013; Volume 328. [Google Scholar]
  31. Matsuura, K.; Okabe, Y. Selective minimum-norm solution of the biomagnetic inverse problem. IEEE Trans. Biomed. Eng. 1995, 42, 608–615. [Google Scholar] [CrossRef] [PubMed]
  32. Uutela, K.; Hämäläinen, M.; Somersalo, E. Visualization of magnetoencephalographic data using minimum current estimates. NeuroImage 1999, 10, 173–180. [Google Scholar] [CrossRef] [Green Version]
  33. Gramfort, A.; Kowalski, M.; Hämäläinen, M. Mixed-norm estimates for the M/EEG inverse problem using accelerated gradient methods. Phys. Med. Biol. 2012, 57, 1937–1961. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. Chella, F.; Marzetti, L.; Stenroos, M.; Parkkonen, L.; Ilmoniemi, R.J.; Romani, G.L.; Pizzella, V. The impact of improved MEG–MRI co-registration on MEG connectivity analysis. NeuroImage 2019, 197, 354–367. [Google Scholar] [CrossRef] [PubMed]
  35. Thompson, A.M.; Brown, J.C.; Kay, J.W.; Titterington, D.M. A study of methods of choosing the smoothing parameter in image restoration by regularization. IEEE Trans. Pattern Anal. Mach. Intell. 1991, 13, 326–339. [Google Scholar] [CrossRef]
  36. Hanke, M.; Hansen, P.C. Regularization methods for large-scale problems. Surv. Math. Ind. 1993, 3, 253–315. [Google Scholar]
  37. Hansen, P.C. Rank-Deficient and Discrete Ill-Posed Problems: Numerical Aspects of Linear Inversion; SIAM: Philadelphia, PA, USA, 2005. [Google Scholar]
  38. Vogel, C.R. Computational Methods for Inverse Problems; SIAM: Philadelphia, PA, USA, 2002. [Google Scholar]
  39. Lütkepohl, H. New Introduction to Multiple Time Series Analysis; Springer-Verlag: Berlin/Heidelberg, Germany, 2005. [Google Scholar]
  40. Haufe, S.; Nikulin, V.V.; Müller, K.R.; Nolte, G. A critical assessment of connectivity measures for EEG data: A simulation study. NeuroImage 2013, 64, 120–133. [Google Scholar] [CrossRef]
  41. Haufe, S.; Ewald, A. A simulation framework for benchmarking EEG-based brain connectivity estimation methodologies. Brain Topogr. 2019, 32, 625–642. [Google Scholar] [CrossRef]
  42. Liuzzi, L.; Quinn, A.J.; O’Neill, G.C.; Woolrich, M.W.; Brookes, M.J.; Hillebrand, A.; Tewarie, P. How sensitive are conventional MEG functional connectivity metrics with sliding windows to detect genuine fluctuations in dynamic functional connectivity? Front. Neurosci. 2019, 13, 797. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  43. Sommariva, S.; Sorrentino, A.; Piana, M.; Pizzella, V.; Marzetti, L. A comparative study of the robustness of frequency-domain connectivity measures to finite data length. Brain Topogr. 2019, 32, 675–695. [Google Scholar] [CrossRef] [Green Version]
  44. Wendling, F.; Bartolomei, F.; Bellanger, J.; Chauvel, P. Epileptic fast activity can be explained by a model of impaired GABAergic dendritic inhibition. Eur. J. Neurosci. 2002, 15, 1499–1508. [Google Scholar] [CrossRef]
  45. Astolfi, L.; Cincotti, F.; Mattia, D.; Marciani, M.G.; Baccala, L.A.; de Vico Fallani, F.; Salinari, S.; Ursino, M.; Zavaglia, M.; Ding, L.; et al. Comparison of different cortical connectivity estimators for high-resolution EEG recordings. Hum. Brain Mapp. 2007, 28, 143–157. [Google Scholar] [CrossRef] [Green Version]
  46. Acebrón, J.A.; Bonilla, L.L.; Vicente, C.J.P.; Ritort, F.; Spigler, R. The Kuramoto model: A simple paradigm for synchronization phenomena. Rev. Mod. Phys. 2005, 77, 137. [Google Scholar] [CrossRef] [Green Version]
  47. Cabral, J.; Luckhoo, H.; Woolrich, M.; Joensson, M.; Mohseni, H.; Baker, A.; Kringelbach, M.L.; Deco, G. Exploring mechanisms of spontaneous functional connectivity in MEG: How delayed network interactions lead to structured amplitude envelopes of band-pass filtered oscillations. NeuroImage 2014, 90, 423–435. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  48. Baillet, S.; Mosher, J.C.; Leahy, R.M. Electromagnetic brain mapping. IEEE Signal Process. Mag. 2001, 18, 14–30. [Google Scholar] [CrossRef]
  49. Sorrentino, A. Particle filters for magnetoencephalography. Arch. Comput. Methods Eng. 2010, 17, 213–251. [Google Scholar] [CrossRef]
  50. Dale, A.M.; Sereno, M.I. Improved localization of cortical activity by combining EEG and MEG with MRI cortical surface reconstruction: A linear approach. J. Cogn. Neurosci. 1993, 5, 162–176. [Google Scholar] [CrossRef] [PubMed]
  51. Gramfort, A.; Luessi, M.; Larson, E.; Engemann, D.A.; Strohmeier, D.; Brodbeck, C.; Parkkonen, L.; Hämäläinen, M.S. MNE software for processing MEG and EEG data. NeuroImage 2014, 86, 446–460. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  52. Lagarias, J.C.; Reeds, J.A.; Wright, M.H.; Wright, P.E. Convergence properties of the Nelder–Mead simplex method in low dimensions. SIAM J. Optim. 1998, 9, 112–147. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Schematic representation of the differences between the regularization parameter providing the best time series estimate ( λ x ) and the one providing the best cross-power spectrum estimate ( λ S ). The first one provides an optimal reconstruction of the neural activity, but it may not lead to an optimal estimate of the cross-power spectrum; vice versa, λ S provides an optimal reconstruction of the cross-power spectrum at the expense of a sub-optimal estimate of the time series.
Figure 1. Schematic representation of the differences between the regularization parameter providing the best time series estimate ( λ x ) and the one providing the best cross-power spectrum estimate ( λ S ). The first one provides an optimal reconstruction of the neural activity, but it may not lead to an optimal estimate of the cross-power spectrum; vice versa, λ S provides an optimal reconstruction of the cross-power spectrum at the expense of a sub-optimal estimate of the time series.
Axioms 10 00035 g001
Figure 2. Real and imaginary part of the cross–power spectra of three simulated time series. Higher values of spectral complexity correspond to more peaked spectra.
Figure 2. Real and imaginary part of the cross–power spectra of three simulated time series. Higher values of spectral complexity correspond to more peaked spectra.
Axioms 10 00035 g002
Figure 3. Pipeline of the simulation of the data (green box) and of the estimation of the cross–power spectrum (blue box).
Figure 3. Pipeline of the simulation of the data (green box) and of the estimation of the cross–power spectrum (blue box).
Axioms 10 00035 g003
Figure 4. Optimal regularization parameters for the reconstruction of the cross–power spectrum ( λ S * ) as a function of SNR X (left) and SNR S (right). Different colors correspond to different MVAR models. On the left, the lines have different heights, while on the right they overlap, meaning that the dependence of λ S * on SNR S is neater with respect to SNR X .
Figure 4. Optimal regularization parameters for the reconstruction of the cross–power spectrum ( λ S * ) as a function of SNR X (left) and SNR S (right). Different colors correspond to different MVAR models. On the left, the lines have different heights, while on the right they overlap, meaning that the dependence of λ S * on SNR S is neater with respect to SNR X .
Axioms 10 00035 g004
Figure 5. Ratio between the optimal parameters ( λ S * λ x * ) as a function of SNR X . Different colors correspond to MVAR model with different spectral complexities. Dashed lines are the mean of the ratio over the different sources location; solid colors correspond to the standard deviation of the mean.
Figure 5. Ratio between the optimal parameters ( λ S * λ x * ) as a function of SNR X . Different colors correspond to MVAR model with different spectral complexities. Dashed lines are the mean of the ratio over the different sources location; solid colors correspond to the standard deviation of the mean.
Axioms 10 00035 g005
Figure 6. Reconstruction error ε S ( λ ) for two simulated data mimicking MEG signals with SNR X = 20 dB (lowest considered signal–to–noise ratio (SNR), left panel) and SNR X = 5 dB (highest considered SNR, right panel). In each panel, black and red vertical lines highlight the values of ε S ( λ ) in correspondence of λ S * and λ x * , respectively.
Figure 6. Reconstruction error ε S ( λ ) for two simulated data mimicking MEG signals with SNR X = 20 dB (lowest considered signal–to–noise ratio (SNR), left panel) and SNR X = 5 dB (highest considered SNR, right panel). In each panel, black and red vertical lines highlight the values of ε S ( λ ) in correspondence of λ S * and λ x * , respectively.
Axioms 10 00035 g006
Table 1. The table reports the values of the spectral complexity coefficients, c j , associated to each simulated multivariate autoregressive (MVAR) model, m j , j = 1 , , N m o d .
Table 1. The table reports the values of the spectral complexity coefficients, c j , associated to each simulated multivariate autoregressive (MVAR) model, m j , j = 1 , , N m o d .
Model12345678910
Spectral complexity coefficient1.411.962.143.103.444.174.645.676.698.67
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Vallarino, E.; Sorrentino, A.; Piana, M.; Sommariva, S. The Role of Spectral Complexity in Connectivity Estimation. Axioms 2021, 10, 35. https://0-doi-org.brum.beds.ac.uk/10.3390/axioms10010035

AMA Style

Vallarino E, Sorrentino A, Piana M, Sommariva S. The Role of Spectral Complexity in Connectivity Estimation. Axioms. 2021; 10(1):35. https://0-doi-org.brum.beds.ac.uk/10.3390/axioms10010035

Chicago/Turabian Style

Vallarino, Elisabetta, Alberto Sorrentino, Michele Piana, and Sara Sommariva. 2021. "The Role of Spectral Complexity in Connectivity Estimation" Axioms 10, no. 1: 35. https://0-doi-org.brum.beds.ac.uk/10.3390/axioms10010035

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop