Table of Contents Author Guidelines Submit a Manuscript
Complexity
Volume 2018, Article ID 1918753, 19 pages
https://doi.org/10.1155/2018/1918753
Research Article

A Complex Network Framework to Model Cognition: Unveiling Correlation Structures from Connectivity

1Departament de Física de la Matèria Condensada, Universitat de Barcelona, Barcelona, Spain
2Universitat de Barcelona Institute of Complex Systems (UBICS), Universitat de Barcelona, Barcelona, Spain

Correspondence should be addressed to Gemma Rosell-Tarragó; moc.liamg@ogarratllesorammeg

Received 24 January 2018; Revised 20 April 2018; Accepted 3 May 2018; Published 12 July 2018

Academic Editor: Hiroki Sayama

Copyright © 2018 Gemma Rosell-Tarragó et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Abstract

Several approaches to cognition and intelligence research rely on statistics-based model testing, namely, factor analysis. In the present work, we exploit the emerging dynamical system perspective putting the focus on the role of the network topology underlying the relationships between cognitive processes. We go through a couple of models of distinct cognitive phenomena and yet find the conditions for them to be mathematically equivalent. We find a nontrivial attractor of the system that corresponds to the exact definition of a well-known network centrality and hence stresses the interplay between the dynamics and the underlying network connectivity, showing that both of the two are relevant. Correlation matrices evince there must be a meaningful structure underlying real data. Nevertheless, the true architecture regarding the connectivity between cognitive processes is still a burning issue of research. Regardless of the network considered, it is always possible to recover a positive manifold of correlations. Furthermore, we show that different network topologies lead to different plausible statistical models concerning the correlation structure, ranging from one to multiple factor models and richer correlation structures.

1. Introduction

Individuals differ from one another in their ability to learn from experience, to adapt to new situations and overcome challenges, to understand simple to complex ideas, to solve real-world and abstract problems, and to engage in different forms of reasoning and thinking. Such differences in performance occur even in the same person, in different domains, across time, and using distinct yardsticks [14].

These complex cognitive processes are intended to be clarified and put together by the concept of intelligence. Although many advances have been made, there are still open questions regarding its building blocks and nature yet to be solved [57].

There is a fair amount of research carried out and still going on about the theory of intelligence, and few statements have been unequivocally established. Nowadays, there are several unsolved questions which lead to prevailing discussions and active research. However, incorrect generalizations or misleading results may bring forth social and educational moves [6, 8, 9]. For this reason, there is an urgent need to understand the most important root causes, validate existing theories, and shed light to people who are responsible of educational and even social and health decision-making.

Nowadays, there are a significant number of approaches to intelligence. Developmental psychologists are often more concerned about intelligence as a subset of evolving processes throughout life, rather than about individual differences [10]. Several theorists stress the role of culture in the very conceptualization of intelligence and its influence in individuals [11], while others point to the existence of different intelligences, either measurable or not [12]. There is also an increasing interest in contributions coming from biology and neuroscience [1317]. Yet, the most influential approach so far is based on psychometric testing [1824].

Psychometrics has enabled successful and systematic measures of a wide range of cognitive abilities like verbal, visual-spatial, fluid reasoning, working memory, and processing speed through standardized tests [25, 26]. Even if distinct, these assessed abilities turn out to be intercorrelated rather than autonomous prowesses. That is, people who perform well in a given test tend to obtain higher scores on the others as well. This well-documented evidence concerning positive correlations between tests, regardless of its nature, is called the positive manifold. And precisely because of the existence of such complex relations, one of the main aims of this approach is to unveil the structure which best describes the relationships between a number of distinguishable factors or aptitudes that may exist. On this basis, many studies use exploratory and confirmatory factor analysis techniques, starting off from between-test correlation matrices.

Furthermore, there exists a complex correlation structure between abilities which may unveil the underlying connection between cognitive processes. Factor analysis might help clearing up such patterns and yet bring about discussion on the meaning of the outcome.

A brief historical overview since the early days of intelligence research and its development may help us understand the spectrum of existing models. Some theorists relied on the shared variance among abilities, which Charles Spearman, pioneer of factor analysis, called the factor or general intelligence [27], that is, one common factor which explains most of the variances within a population and source of improvement or decline of all other abilities, and it is still a cause for controversial.

Alternatively, hierarchical models of intelligence where each layer accounts for the variations in the correlations within the previous one were also well accepted [18, 28, 29].

Nevertheless, a fair number of scholars argued against theories of cognitive abilities or intelligence drawn upon the concept, measure, and meaning of general intelligence. Namely, Howard Gardner, stated that an individual has a number of relatively autonomous intellectual capacities, with a degree of correlation empirically yet to be determined, called multiple intelligences, among which noncognitive abilities are included [12].

Two different approaches with reference to the relationship between observable variables and attributes or constructs prevail in present research and theorizing not only in psychology but also in clinical psychology, sociology, and business research amongst others: formative and reflective models [30]. In the first of this conceptualization, observed scores define the attribute, whereas in the latter, the attribute is considered as the common cause of all observables. As an example, the classic definition of general intelligence could fall into a reflective model. But also, in clinical psychology, a mental disorder may be thought to be a reflective construct that brings about its observable symptoms [31]. Possible correlation between observables might be therefore due to its underlying common cause. Conversely, the aggregate outcome of education, job, neighbourhood, and salary leads to socioeconomic status (SES), a standard example of formative model.

A more recent approach aims to combine distinct possible factor models by only using the information about the factorial structure found by each study [32].

Both formative and reflective models, along with similar alternatives, may elicit discussion regarding two different issues: one first source of debate is rooted in the meaning and interpretation of such models, while a second cause stems from disregarding the role of time, that is, the dynamics of the system is not explicitly considered.

The abovementioned problems can potentially be overcome if we consider that variables, that is, observables, scores, or indicators, are the characteristics of nodes in a network. These latter are directly connected through edges, which reflect the coupling between variables. Dynamical systems theory is therefore the proper framework to formalize and study the behaviour of such systems [33]. Starting from an initial state, the system evolves in time according to a system of coupled differential equations and eventually reaches an attractor state of the system.

Examples of previous works on a dynamic systems approach to social and developmental psychology are the modeling of language acquisition and growth, the dynamics of scaffolding, dyadic interaction in children, teaching and learning processes, and coconstruction of scientific understanding among many others [3443].

Noteworthy, a substantive piece is prevalently missing: the topology of the network on where the process is taking place, which may be a determinant fact that enables nodes to communicate between each other and brings about correlations not explicitly enforced in the model. Therefore, the objective and contribution of the present work is exploring the significant role of the network topology or connectivity structure between the variables deemed meaningful to the case of cognitive abilities or intelligence models.

In this work, we evince the tight connection between a centrality measure of the network and the stable solution of the studied models. Moreover, we show that distinct network topologies may explain different correlation structures.

The paper is organized as follows: Section 2 introduces basic notions of networks and explored topologies. Section 3 describes and formalizes the two studied models of cognition. Section 4 and Section 5 go through the main results, concerning dynamics and correlations, respectively. The final discussion and the conclusions are presented in last section. Further mathematical methods can be found in the appendix.

2. Network Topology

A network, , is a collection of vertices or nodes, , linked by edges, , which are given meaning and attributes. Networks can describe complex interconnected systems such as social relationships, transportation maps, and economic, biological, and ecological systems. We consider networks that have neither self-edges nor multiedges, called simple networks [44].

The adjacency matrix of , written , is the -by- matrix which entries equal 1 if node is linked to node and 0 otherwise. Networks can be directed or undirected, although we stay on the latter case.

The topology of a network characterizes its shape or structure and the distribution of connections between nodes. Besides the attributes of nodes and edges, the topology of a network determines its main properties and makes it distinguishable from others. One main property is the degree of a node , , which is the number of edges connected to it. Although networks may describe particular real systems, regardless of its nature, they can be classified to one of the most well-known families of networks. Right after, we briefly describe the four network models explored in the present work. (a)Complete network (Figure 1): within the family of deterministic networks, a complete network is characterized by its nodes being fully connected, that is, each node is connected to the others, such that all off-diagonal elements of the adjacency matrix are equal to 1, .(b)Erdös-Rényi network (Figure 2): one of the most renowned random networks is generated by the Erdös-Rényi (ER) model [45]. Given the number of nodes, , and the probability of an edge, , this model, , chooses each of the possible edges with probability . However, generally, real networks are better described by heterogeneous rather than ER networks. Therefore, ER networks are often used as null hypothesis to reject or accept models concerning more complex situations.(c)Heterogeneous network (Figure 3): there is a wide range of networks coming from real systems (either found in nature or human driven) which topology is far from being homogeneous, but it rather entails degree distributions which are characterized by a power law, also called scale-free when the networks are large enough [46]. The Internet network, protein regulatory network, research collaboration, online social network, airline system, cellular metabolism, company, and industry interlinks are few examples of them [47, 48].(d)Newman modular network (Figure 4): in addition to the degree distribution, another important feature is the presence of communities or modules within a network, mainly in social but also in metabolic or economic networks [49, 50]. A module or community can be defined as a subset of nodes which is more densely linked within it than with other subsets of nodes.One particular method to generate such modules within a network is the Newman model, which distributes the nodes in a number, , of modules not necessarily isolated from the others [5153]. Similarly as ER networks, with a probability , an edge between pairs of nodes belonging to the same community is created, whereas pairs belonging to different communities are linked with probability . In the model, the number of nodes, ; the total average degree, ; and , which stands for the average degree within a community, are fixed. Hence, and are given by where and .As grows, the network modularity increases [54], that is, the communities become easier to identify.

Figure 1: Complete network of size 16 nodes.
Figure 2: Erdös-Rényi network of size 16 nodes.
Figure 3: Heterogeneous network of size 16 nodes.
Figure 4: Newman modular network of size 16 nodes.

3. Models of Cognition

Within the framework of dynamical systems and network theory, there exists a one-to-one map between variables and nodes, such that variable is represented by node . The value of variable , , is set as an attribute of its corresponding node. In this new space, the adjacency matrix, which maps the interactions between variables on a network, can lump exogenous effects together in a very compact way:

Therefore, (2) is the most general expression which integration determines the temporal evolution of each variable, . accounts for endogenous effects, that is, a function that depends only on variable . takes into account exogenous effects on , that is, a function that describes the influence of its neighbouring variables, . The intensity of such individual interactions is included in in the form of weights. The adjacency matrix, , determines whether variables are coupled between them: if variables and are directly linked, then the corresponding element . Otherwise, . In addition, , , and can, in general, depend explicitly on time and could be potentially inferred from data: firstly, static or time-based correlations between observables may provide a reliable approximation of the adjacency matrix. Secondly, endogenous parameters could be inferred from single-variable evolution or biological and genetic properties. Finally, exogenous parameters account for the temporal deviation of self-evolution due to effect of other variables. Although real data can be fitted to individual trajectories, we are mostly interested in the qualitative and interpretable outcome.

Two models are addressed in this work: a networked dynamical model to explain the development of excellent human performance [43] and a dynamical model of general intelligence [36], both sharing great resemblance (Section 4).

The equations describing both models share similar function structures: logistic growth and bounded multiplicative effect of connected variables. Developmental curves are mostly characterized by a strong initial increase followed by some kind of asymptote [55]. Furthermore, they make the assumption that cognitive variables or processes are mutually beneficial.

3.1. Model A: A Networked Dynamical Model to Explain the Development of Excellent Human Performance

Den Hartigh et al. [43] were interested in the excellent level of performance of some individuals across different domains. They argued that the key to excellence does not reside in specific underlying components but rather in the ongoing interactions among them and hence leading to the emergence of excellence out of the network integrated by genetic endowment, motivation, practice, and coaching inter alia.

They attempted to render well-known characteristics of abilities leading to excellence: the absence of early indicators of ultimate exceptional abilities, the fact that a similar ability level may be shifted in time between individuals, the change of abilities during a person’s life span, and the existence of unique pathways leading to excellence, that is, individuals may have diverse ways to achieve it.

They considered a networked dynamical model which can be mathematically defined as a set of coupled logistic growth equations, each of which represents the growth of a single variable. One of such variables represents the domain-specific ability. The growth of the variable depends on the already attained level, available resources that remain relatively constant during development (), resources that vary on the time scale of ability development, the degree in which a variable profits from the constant resources (), and a general limiting factor (): the ultimate carrying capacity, which captures the physical limits of growth. Moreover, accounts for the effect of variable on .

Using (2), model A can be written as

Equation (3) is better understood as a modified logistic growth:

Figure 5 shows 3 different possible temporal evolutions of the system, determined by both the topology of the connections between variables and the parameters of the dynamical model.

Figure 5: Temporal evolution of all variables in the case of model A, described in (3), for increasing values of the links’ weight: (a), (b), and (c). The system is attracted to one of the three possible global stable states, depending on both the parameters of the dynamic model and the topology: metric stable state, (a), mixed stable state (b), and optimal stable state (c). All plots are represented for an ER network with 50 nodes and edge probability . Parameters are equally set to , , for all nodes.
3.2. Model B: Dynamical Model of General Intelligence

Van Der Maas et al. [36] were concerned with the conceptualization and models of intelligence or cognitive ability system by means of a general latent factor, as widely stated. They proposed an alternative explanation to the positive manifold based on a dynamical model built upon mutualistic interactions between cognitive processes, such as perception, memory, decision, and reasoning, which are captured by psychometric test scores to some extent. Such connections between items bring about another plausible explanation to the existence of one common factor. Hence, the latter being used as a valid model does not imply that there is a real latent process or cause, such as speed of processing or brain size, or one general intelligence module in the brain.

Inspired by Lotka-Volterra models commonly used in population dynamics [56, 57], they proposed to model the cognitive system as a developing ecosystem with primarily cooperative relations between cognitive processes. Variables () represent the distinct cognitive processes, in which growth function is parametrized by the steepness of the growth () and the limited resources for each process (). Matrix contains the relation between pairs of processes, which they assume being positive, that is, involved cognitive processes have mutual beneficial interactions.

Starting from uncorrelated initial conditions and parameters, that is, following uncorrelated random distributions, the dynamical connections between variables gradually lead the system to specific correlation patterns.

Using (2), model B can be written as

Equation (5) is better understood as a modified logistic growth:

Section 4 puts stress on the resemblance between (4) and (6).

A recently published paper [58] makes further comments on the model and its framework.

4. Interplay between Dynamics and Network Topology

A dynamical model which captures the network structure of the connection between variables by using an expression similar to (2) enables further analysis of the process as it considers the effect of topology, embodied in the adjacency matrix, .

Neither of the two described models is geared toward a particular cognitive architecture or brain model with regard to connectivity structure. Rather, much effort is devoted to understanding the effect of nonzero correlations between the parameters of the models or a heterogeneous landscape of parameters [36]. The former approach, however, requires certain constrains or assumptions which, in general, may not be easy to proof. Alternatively, in the present work we consider the dynamical model to be parametrized by a homogeneous configuration, that is, all nodes with equally fixed parameters, and explore the role of different connectivity structures between variables, which can be mapped on a network.

Although a dynamical model describes the temporal evolution of several variables, it is usually the final state that provides more useful and interpretable information. For instance, children may achieve a given level of achievement regardless of its learning rate, as long as they have had enough time to develop.

4.1. Mapping between Models through Weight Rescaling

The space of parameters is large and hence so is the number of possible stable states. However, we focus our interest on solutions given by one unique analytical expression. Therefore, depending on the stability conditions (Section 4.3), we can distinguish two of such stable states. For model A, an optimal stable solution, , is achieved when all variables reach the maximum allowed value: where we assume .

Otherwise, final state, , is determined by matrix expression (8):

matrix in expression (8) captures the entanglement between network topology, , and the parameters of the dynamical model. The influence of variable on variable is thus rescaled by its carrying capacity, , and growing rate, , as follows:

All intermediate states, which lay in the transition between metric and optimal stable states, are called mixed stable states.

Analogously, for model B, there is one unique stable state, :

Expressions (8) and (10), referred to stable states, for models A and B, respectively, are equivalent under rescaling (9). Both models behave differently with regard to the temporal evolution as well as possible states. However, when their final state is given by the metric stable state, a mapping between them exists. Moreover, we highlight the absence of initial conditions in the attractor state.

The case when parameters are constant throughout variables, that is, when , , and , may enable an explicit average solution to (10). An individual can be characterized by the average of the achieved values of all variables. We follow the notation

For a complete network,

For an Erdös-Rényi network (Appendix A), where is the degree of node variance.

Average solutions to (8) for a complete network and an Erdös-Rényi network are equivalent to (12) and (13), respectively, with .

4.2. Katz-Bonacich Centrality as Stable State

Centrality measures seek the most important or central nodes in a network [44]. Among many possible centralities, the generalized Katz-Bonacich centrality [59] is given by the stable solution of the recursive equation:

Solving (14), the vector of centralities is given by

Unlike eigenvector centrality, Katz-Bonacich centrality solves the issue of zero centrality values for acyclic or not strongly connected networks by introducing a constant term for each node. Therefore, Katz-Bonacich centrality gives each node a score proportional to the sum of the scores of its neighbours plus a constant value. parameter rules the balance between the first term in (14), which is the normal eigenvector centrality [54], and the second. The longest walks become more significant as increases, and hence the global topology of the network is considered, resembling eigenvector centrality. On the contrary, small values of make Katz-Bonacich centrality a local measure which approaches degree centrality. When , , and as increases, so do the centralities until they diverge when where is the maximum eigenvalue of matrix. Hence, Katz-Bonacich centrality is defined as long as [44].

Equivalently, for weighted networks, generalized Katz-Bonacich centrality is defined as and .

Equation (17) can also be expanded to

Element in (18) stands for the number of walks of length from node to node taking the strength of connections into account. This value is attenuated by a factor , and hence accounts for the strength of all walks from node to node , with greater weakening as gets larger.

Comparing (17) with (8) or (10), we conclude that generalized Katz-Bonacich centrality vector is the exact solution of the stable state of model B with and . Furthermore, when rescaling (9) is considered, so it is of model A or any other model in which dynamics can be included in the weighted adjacency matrix in a similar way.

In general, the stable state of a dynamic system may not have an analytical solution but rather needs to be described and characterized by means of numerical simulations. The considered systems, however, can achieve a particular state which can be mathematically obtained from the parameters of the network and the model, without the full integration of the coupled set of differential equations in time. This finding has several implications: (i)Firstly, variables which obtain the best score in Katz-Bonacich centrality achieve the highest values of performance on the long run. For this reason, we call the stable state (8) as “metric” stable state. The most central nodes according to (18) are those which reach the largest number of nodes enabling all possible path lengths, with a penalization to greater longitudes. This, intuitively, is a characterization of the node’s ability to spread information along the network. In case we consider the transposed of the weighted adjacency matrix, the most central nodes are those which are reached by the largest number of nodes and hence can be interpreted as the best receiver nodes.(ii)Secondly, we note that Katz-Bonacich centrality is the stable solution of the process . Equation (14) corresponds to the equation of a nonconservative process, that is, a process in which the sum of the values of the variables is not constant over time. We could provide further details, but these can be easily found in the literature [60, 61]. The metric stable state could hence be interpreted as the distribution of a given quantity along the network if all nodes are explored by a random walker with a bias given by . We stress the fact that these processes are not restricted to the presented models of cognition but can be found in many other systems in biology and society, which may be well described by similar equations. For this reason, the generalization of the models, as well as its interpretation in terms of centrality measures, may help understanding cognitive and other types of phenomena.

4.3. Stability Conditions

For model A, stability analysis is rather complex since many different stable states may exist, depending on a sizeable number of parameters which characterize both the topology and the dynamics. Nevertheless, we focus our interest on the most extreme situations: the optimal stable state, , given by (7), and the metric stable state, , given by (8). All other configurations are described by a mixed pattern which falls between optimal and metric stable states.

solution is stable when (see Appendix C) or using rescaled weighted matrix defined in (9), when

Provided that a given node does not meet condition (20), the stable state is no longer , and hence, starting from node , nodes will start getting values lower than the threshold.

On the other hand, solution is stable when (Appendix C) where is the maximum eigenvalue of matrix , which is defined as follows:

Eigenvalues of depend explicitly on and therefore can only be computed numerically (see equation C.51). However, Perron-Frobenius theorem [62] allows us to obtain an upper threshold for analytically:

Equations (21) and (23) imply the existence of an upper bound for the stability condition of metric stable state (Appendix C):

For model B, stability analysis concerns only the metric stable state (10), and stability conditions are given by (21) with

From Figure 6, we conclude that network topology is enough to obtain information about the stability of the system. Due to the fact that we are in a situation of homogeneous configuration, is tightly connected to , although it is not the same as seen in (22) and (25). Concerning Erdös-Rényi network (Figure 6(a)), there is little variability in , and therefore the critical value of the parameter for which stability changes, , is confined within a narrow range. Using (24) and assuming a Poisson degree distribution, we can obtain an approximate value for in case of a homogeneous configuration: where .

Figure 6: Set of boxplots of , defined in (22), for different values of the characteristic parameter of the network: edge probability for an Erdös-Rényi network of size (a), degree distribution exponent for a heterogeneous network of size (b), and intramodule degree for a Newman modular network of size and (c). Metric state is given by (8) and it is stable only if (dashed blue line). For each value of ,we make 1000 realization of the corresponding network class, each one representing one single individual, to capture the distribution of the maximum eigenvalue of matrix . Edge probability, , of an Erdös-Rényi network makes for a narrower threshold of stability than exponent of a heterogeneous network. Conversely, parameter of the Newman modular network does not lead to any change in stability. Parameters of the model are set for all nodes to and for all networks, and for Erdös-Rényi network, and for heterogeneous network, and and for Newman modular network.

The critical value of obtained from (26) when the parameters are the same as for the Erdös-Rényi network of Figure 6 gives , which is in agreement with the numerical solution.

Conversely, the stability with regard to heterogeneous network is rather diffuse (Figure 6(b)). The broad spectrum of [63] is somehow captured by the variability on . Consequently, there exist outlier networks which eventually break the stability condition. Nevertheless, they become less frequent as exponent increases. The effect of in a Newman modular network is completely different (Figure 6(c)): in spite of being a parameter which rules the modularity of the network, the average total degree, , is still fixed and hence both matrix and solutions (8) and (10) remain essentially the same, except fluctuations.

5. Unveiling Correlation Structures from Network Topology

Provided that condition (21) is met, the stable state is given by (8) and (10), for models A and B, respectively. This latter result provides one subject with as many values as the existing variables. However, studies based on large batteries of psychometric tests rely on a sample from a population, made up of tens to thousands of individuals, from where intervariable correlations are inferred. Section 5.1 describes the generation of distinct individuals from the models and the handling of the correlation matrix out of them.

5.1. From Dynamics to Correlation Matrix

These models have been studied by assuming that parameters corresponding to different variables are correlated. If, in addition, all variables are considered to be equally interconnected in a mutualistic scenario, that is, positive interactions and a complete network topology is considered, it is possible to recover well-known correlation structures [36]. Alternatively, our hypothesis is that the network topology, which determines the connectivity between variables, is sufficient to recover equivalent results concerning correlations and also enables us to avoid stronger constraints on parameters.

Since we are interested in studying the effect of topology, each individual is considered to be one random instance of the same network class. In other words, a new individual is obtained out of the possible random , and the corresponding correlation matrix is characterized by fixed values of and , where is the number of variables or nodes and is the set of characteristic attributes of a given network model. For instance, a network generated by Erdös-Rényi model has only one parameter corresponding to (Section 2).

Despite each variable holding its own set of parameters, we make all variables equivalent in what we call homogeneous configuration, such that individual differences result only from topological properties:

Independently of others, each individual reaches a stable state given by (8) or (10), for models A or B, respectively, as long as condition (21) is true. If this is not the case (see Section 4.3), model A requires full temporal evolution being simulated and model B comes to divergence, and hence the considered configuration is not acceptable.

Once stable states of the entire population are obtained, the computation of pairwise correlations between variables is straightforward. In line with most studies conveyed by psychometrics, Pearson standard correlation coefficient is used to finally get the correlation matrix [64], although other methods may excel this procedure [65, 66].

In conclusion, the variability within one same family of networks brings about individual differences concerning variables which entail particular patterns or structures of the correlation matrix (28), with no explicit constrains on parameters, but considering them as homogeneous both along variables and individuals (see Figure 7).

Figure 7: Flowchart to generate correlation matrix of variables out of a population of individuals.

From Figure 8, we conclude that the positive manifold, that is, positive correlations, , can come out regardless of the topology, as long as of the network is properly set. Thus, connectivity by itself allows positive interactions, regardless of the structure and independently of further constrains on dynamic parameters. However, the structure of the correlation matrix indeed relies on the topology: Erdös-Rényi network’s correlation histogram displays a narrow symmetric peak which captures the homogeneity of the topology, that is, all nodes being equivalent. The distribution which comes out when considering this topology and identical parameters for all nodes is equivalent to the one obtained by [36] in the case of a complete network and heterogeneity in parameters. On the contrary, heterogeneous network’s correlation histogram follows a much wider asymmetric distribution, with a longer tail for larger values. This behaviour captures the presence of few hubs, which leads to nontrivial correlation structures, as we will see in Section 5.2. Finally, the Newman modular network presents two characteristic peaks corresponding to intracluster, that is, within the same cluster, and intercluster, that is, between different clusters, correlations. As increases, both peaks become undistinguishable and the pattern more closely resembles an Erdös-Rényi network. For large , intercluster correlations tend to value.

Figure 8: Histogram of the correlation between variables, , for 3 distinct network topologies: Erdös-Rényi of size and edge probability (a), heterogeneous network of size and degree distribution exponent (b), and Newman modular of size , and (c). The histogram captures the effect of on . Correlation matrix is computed from analytic metric stable state (8) of 5000 generated individuals. The distribution of correlations between variables connected as an ER network is peaked at a certain value, whereas assuming a heterogeneous network leads to a much broader spectrum of correlation values. The Newman modular network brings about two clear peaks, corresponding to intracluster and intercluster correlations. Parameters of the model are set for all nodes to and for all networks, and for Erdös-Rényi network, and for heterogeneous network, and and for Newman modular network.

So far, in addition to considering a homogeneous configuration of parameters, we have constrained ourselves to solutions within the regime where metric stable solution is stable. Nevertheless, we expect richer phenomena when neither of the two restrictions exists. If we allow parameter to be a random variable distributed among variables following a noncorrelated normal distribution, , increasing values of lead to lower values of , the average correlation between variables.

Figure 9 captures not only the positive manifold but also the effect of network topology on the correlations. is indeed modulated by the characteristic parameter of the network, . For Erdös-Rényi and heterogeneous networks, peaks around the transition between metric and optimal stable states, with a profile shaped by the network topology. The average correlation increases as connectivity does so, until nodes are recurrently being absorbed by the optimal state and, consequently, is scaled down. Namely, Erdös-Rényi network peaks around , that is, when stability changes landscape. As increases, more and more variables reach the optimal value and become independent from others. The impact of more variability in parameter across variables is the decrease in the correlations, and the peak shifting towards lower values of , since stability condition (21) is more likely to be broken. The peak generated by a heterogeneous network is more diffuse, owing to a wider distribution of the spectra. Moreover, the effect of increasing the variability in has a stronger attenuation effect. On the contrary, the absence of a peak coming out from the Newman modular network is explained in Section 4.3. Nevertheless, a clear pattern emerges if we split intracluster from intercluster mean correlations. While intracluster correlations increase as does so, intercluster correlations decrease. However, the total mean correlation remains unchanged, as long as condition (21) is true.

Figure 9: For model A, defined in (3), the average correlation between variables, , is plotted as a function of the characteristic network parameter, , for 3 network topologies: Erdös-Rényi of size (a), heterogeneous network of size (b), and Newman modular of size and , where intracluster (dimmer upper) and intercluster (solid bottom) correlations are held separately (c). The distribution of each value is captured by a boxplot computed from 50 independent realizations. Six different values of account for the variability of parameter. As increases, is scaled down, being the effect much larger in heterogeneous networks. In the case of ER network, grows with increasing until metric state eventually becomes unstable, and nodes are gradually absorbed by optimal state. Larger dispersion on shifts the peak towards lower values of . Conversely, for a heterogenous network, an increase on exponent leads to more stability of metric state, although stable conditions are more fragile. Larger dispersion on shifts the peak towards much lower values of . Finally, in the case of Newman modular network, stability conditions of metric state, (21), are always true for these parameters and hence the absence of the peak. For a better understanding of change of stability landscape, see Figure 6. The parameters of the model are set as Figure 8.

For model B, continuously increases as connectivity does so, until divergence conditions are met, and solutions do not longer exist.

5.2. From Correlation Matrix to Statistical Models

When it comes to describing variability among observed, correlated variables, a bunch of statistical models comes along, aiming to approximate and understand reality. Factor analysis is a statistical method developed and widely used in psychometrics [64, 6769], inter alia. Observed variables are described as linear combinations of unobserved latent variables or factors, plus individual error terms, such that covariance or correlation matrix may be explained by fewer latent variables.

Historically, the most widely held theories of cognition and intelligence are built upon factor analysis raising from large batteries of conducted psychometric tests. There is still no agreement on the proper model and the underlying process which bring about the observed outcome. We list the models stood for the most outstanding theories: one factor models, multifactorial models, hierarchical models, and other more complex structural models [70].

Section 5.1 evinces that connectivity structure between variables, mapped on a particular network topology, gives rise to different correlation matrices, even no explicit constrains on correlations between parameters are imposed. We show that certain factor models can be explained by a mutualistic dynamical model running on a particular network of variables.

In order to avoid subjective criteria when deciding the number of factors, several methods have been developed: Horn’s parallel analysis, Velicer’s map test, Kaiser criteria, Cattell scree plot, or variance explained criteria [7174]. We proceed to compute the scree plot so as to obtain the number of main principal components and hence potential latent factors. Thereupon, the statistical significance of the factor model is assessed by means of the value, which enables us to explore the effect of changing the parameter which best characterizes the network, .

In Figure 10, we have selected one specific network for each topology, given by the value of the characteristic parameter. The choice is such that solution is given by the metric stable state (8).

Figure 10: Scree plot, in logarithmic vertical scale, calculated from the correlation matrix, , for 3 network topologies: Erdös-Rényi (triangular orange markers), heterogeneous (circular blue markers), and Newman modular (squared green markers) networks, with parameters as in Figure 9. A scree plot shows the value of eigenvalues, , in descending order, as components (principal orthogonal directions) are gradually being included (up to the number of variables). is labelled in red, as a possible selection criterion for the number of components or factors. Other used criteria: much larger than the others and describing rather straight angles with successive values. 1-factor model and 4-factor model may be suitable for ER and Newman modular networks, respectively, whereas a more complex model is needed for heterogeneous network.

The number of retained components, or factors, can be obtained according to different criteria. We look at values which are much larger than 1 and display rather straight angles with the successive values.

With these criteria and from Figure 10, we hypothesize that the correlation matrix of variables which are connected following an Erdös-Rényi network is well described by a 1-factor model. In the case where their connectivity structure is better captured by a Newman modular network with clusters, then an -factor model cannot be rejected. Conversely, considering a heterogeneous network, a factorial model is no longer proposed, as eigenvalues are not clearly separated according to former criteria, but rather they follow a smooth decreasing curve. Alternatively, a statistical model which accounts for hierarchy between variables or more complex structural modelling and path analysis may be more realistic. However, this latter analysis is out of the scope of this paper.

For the cases of Erdös-Rényi and Newman modular networks with 4 clusters, we show that a 1-factor model and a 4-factor model, respectively, can be accurate models to obtain results. To do so, we compute the value of such models for different numbers of retained factors. value in this case is testing the hypothesis that the model fits the results perfectly, and hence we seek values > 0.05.

In the case of an Erdös-Rényi network, value ≈ 1 ≫ 0.05 for , and hence a 1-factor model is highly likely. Similarly, in the case of a Newman modular network with , value > 0.05 only when , as suggested (see Figure 11).

Figure 11: -values obtained for a factorial model with a number of factors ranging 1 to 5, for 2 network topologies: Erdös-Rényi of size and (a) and Newman modular of size , , and (b) networks, with parameters as in Figure 9. value = 0.05 is labelled in orange, as threshold for not rejecting the model.

Once we have figured out the number of factors with respect to each network, we explore the effect of connectivity on the reinforcement of the statistical models, by looking at the explained variance of the most important components:

We make use of (29), which gives the proportion of explained variance of correlation matrix for each principal component of PCA. Although PCA and FCA are not equivalent statistical models [75, 76], after having sustained the validity of a factor model, the former approach is acceptable in order to justify the suitability of the number of factors or components considered.

Figure 12 confirms the assumptions underlying the models for both networks: in the case of an Erdös-Rényi network, the first eigenvalue increases as connectivity, characterized by , does so, and hence a 1-factor model is being reinforced. Similarly, in the case of a Newman modular network, besides the first eigenvalue, which is always large, second to fourth eigenvalues increase as modularity, characterized by , does so, in contrast to eigenvalues on further positions, which remain unchanged or smaller.

Figure 12: Percentage of the total explained variance of correlation matrix , (29), for each of the first five components as a function of the edge probability, , for an Erdös-Rényi of size (a), and intracluster degree, , for a Newman modular network of size and (b). Both networks enable a factor model as a good descriptor of the outcome. Selected significant number of components are highlighted in orange and green colors, for ER and Newman modular networks, respectively. As nodes become more connected and communities more delimited, factor model moves from mirroring a noisy identity matrix to be a clear indicator of the correlation structure. First component (circular orange marker) increases as does so, strengthening the validity of a 1-factor model (a), while second to fourth components (circular, starry, and triangular green markers) increase as does so. Hence, an increase in modularity reinforces the validity of a 4-factor model (b). Parameters are set as in Figure 6, constraining the values to lay within the metric stable state regime.

We are mostly interested in the qualitative results. Different network sizes and dynamic parameters lead to different values for the eigenvalues and variances, while showing the same patterns [77].

We have provided an alternative explanation of data being well described by different factor models. However, alternative statistical models ought to be considered in the case of Newman modular network to obtain more precise insights, namely, bifactor and second-order factor models, which are the most popular factor models in the intelligence literature.

Covariance matrix can be analytically computed in the case of a complete network and a heterogeneous parameter landscape, (Appendix B).

6. Conclusions

Despite mainstream approaches to cognition and intelligence research are built on static and statistics-based models, we explore the emerging dynamical systems perspective putting a greater emphasis on the role of the network topology underlying the relationships between cognitive processes.

We go through a couple of models of distinct cognitive phenomena and yet find the conditions for them to be mathematically equivalent. Both models meet the requirements set out by empirical observations and established theories regarding the corresponding cognitive phenomena to which they aim to provide an explanation, namely, the positive manifold and the suitability of several statistical models. Furthermore, the applied mathematical formulation may well enlight models of many real mutualistic systems, other than cognitive.

The topology of the network defined by the dynamical influence between processes indeed underlays further analysis of the results. We find the principal attractor of the system to be the exact definition of Katz-Bonacich centrality, a measure of a node importance which can also be understood as a nonconservative biased random walk along a network. We propose that heterogeneities in the dynamical parameters can be absorbed by a rescaling of the adjacency matrix weights and hence leading to the same result.

Individuals may differ not only in the genetic-environmental markers captured by the parameters of the model, but also in the connectivity structure between brain regions, either structural or functional. Two individuals might achieve the same performance through different neuronal routes or architectures and cognitive strategies when solving cognitive tasks. Certain common brain structures and functional pathways may be more likely to be involved in intelligence than others. However, despite of the heterogeneity between subjects, cognitive outcomes may result similar [14]. For instance, as Newman modular network increases its modularity, the corresponding correlation matrix becomes more and more likely to be well described by a factorial model with as many factors as communities it has. However, whereas the inner structure gradually changes, the average result of its stable state remains unchanged.

The connectivity structure between cognitive processes is not known but yet it is not any. We show that network topology by its own leads to different plausible statistical models. Regardless of the network considered, it is always possible to set a parameter configuration such that the positive manifold results from the dynamical model. However, the correlation structure is determined by the network topology. Complete and Erdös-Rényi networks are constrained to bring about a one-factor model, more clearly defined as connectivity increases. The Newman modular network enables higher-order factor models, depending on the number of defined communities. Latent factors turn to be more distinguishable as modules grow to be more isolated. Conversely, heterogeneous networks lead to richer correlation structures, which may be better described by more complex statistical models.

In the present article, we exploit the interplay between the dynamics and the underlying network topology to model cognitive abilities, and we conclude that both of the two are relevant. Although scholars are not yet sure of the relationship between cognitive processes and of the nature of intelligence, we can shed a bit of light by proposing an alternative framework which captures the real meaning of “process” and “relationship”: a dynamic complex network framework to model cognition.

This work is an open door to further research: we show that different network topologies lead to different correlation structures. Still, richer topologies can be considered and may bring about other interesting and eventually more realistic structures. We have restricted ourselves to static networks, though the more general definition of a network is not time constrained. What if the network which captures the connection between cognitive processes or brain modules was time dependent? What if cognitive processes could be modelled as a multilayer network, from generalists to specialists layers? Moreover, when looking only at the attractors of the system, we are missing the temporal evolution of such processes and its real causality. Therefore, are these models able to explain evolving properties of the considered variables? We finally highlight that there exist several limitations in models based on ecologic systems, as exhaustively studied in population dynamics research, namely, inconsistent results coming from unbounded models and discrepancy with the behaviour of some real systems [7881]. Hence, alternative mathematical models which overcome some of these problems shall be investigated.

Appendix

A. Average and Variance of Stable Solution for a Complete and an Erdös-Rényi Network

We define homogeneous configuration as follows:

In this case, metric stable solution (10) for a complete network is straighforward:

In the case of a complete network, as solutions are exactly the same for all nodes, the variance of stable state is null.

For an Erdös-Rényi network, we compute the average of metric stable state:

Using (18) and considering is a symmetric matrix, though the expression is equivalent,

Taking the average on (A.4),

We proceed to calculate each term of the average: where is the mean degree of a node.

Second and following terms account for the average number of next-nearest neighbours, denoted with . The general expression for any network is given by [82] where .

In the case of an Erdös-Rényi network following a Poisson distribution, we get

Using (A.6), (A.7), and (A.9),

We point out the fact that a Poisson distribution is not always a good model for an Erdös-Rényi network, and hence (A.10) is considered an approximation for .

The variance of , , is a rather difficult computation, and therefore we explore the behaviour at the second order of , : where corresponds to the average of the ensemble.

The second term, , is already known (A.10), as it is given by in the thermodynamic limit. Using (10), the first term can be expanded as

Using (A.8) and splitting the cases and ,

Therefore, in order to obtain further information of the structure of the covariance and correlation matrices, we ought to compute higher orders on .

Nevertheless, we have proved that the covariance matrix is made out of two different values: the diagonal and off-diagonal elements. The covariance matrix, , can be written as follows:

A requirement for a covariance matrix, , to be explained by a one-factor model is that the spectral gap is large enough, that is, :

Hence, , which increases as the number of variables or do so and enables a one-factor model to fit the results.

B. Covariance Matrix for a Complete Network and Heterogeneous Parameter Configuration

Although the solution when connectivity structure is captured by a complete network is trivial, we can go a step further when taking into account variability in the dynamic parameters, as it was similarly done in the appendix of [36]. Let us consider parameter taken from a normal uncorrelated distribution, :

If we separate according to the contribution regarding parameter, where and .

Using (A.11), covariance matrix can finally be written as follows: where and

From (B.3), it turns out that a 1-factor model is indeed valid for the description of the covariance matrix. In cases where , then adequacy decreases. The particular case is achieved when . However, such condition is never reached:

As seen from (12), stable state diverges when .

C. Stability Conditions

To obtain the stability conditions of optimal fixed point (7), we expand (3) around this fixed point :

If we linearize (C.1) keeping only terms , we obtain

On the other hand, to obtain the stability conditions of metric fixed point (8), we expand (3) around this fixed point :

Equation (C.3) can be written in matrix forms as (22).

We can derive a threshold for the stability condition using the Perron-Frobenius theorem, (23), applied to (22) which enables us to write

Looking into the extreme conditions of (C.4), we conclude

Data Availability

The data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare that there is no conflict of interests regarding the publication of this paper.

Acknowledgments

The authors acknowledge the financial support from MINECO under Project no. FIS2012-38266-C02-02 and no. FIS2015-71582-C2-2-P (MINECO/FEDER) and Generalitat de Catalunya under the Grant no. 2014SGR608. Gemma Rosell-Tarragó acknowledges MECD under Grant no. FPU15/03053. The authors wish to express their gratitude to Joan Guàrdia for his revision and comments on the work.

References

  1. E. M. Tucker-Drob, “Differentiation of cognitive abilities across the life span,” Developmental Psychology, vol. 45, no. 4, pp. 1097–1118, 2009. View at Publisher · View at Google Scholar · View at Scopus
  2. R. Kanfer, “Motivation and individual differences in learning: an integration of developmental, differential and cognitive perspectives,” Learning and Individual Differences, vol. 2, no. 2, pp. 221–239, 1990. View at Publisher · View at Google Scholar · View at Scopus
  3. D. H. Jonassen and B. L. Grabowski, Handbook of Individual Differences, Learning, and Instruction, L. Erlbaum Associates, 1993.
  4. R. J. Sternberg and L.-F. Zhang, Perspectives on Thinking, Learning, and Cognitive Styles, Lawrence Erlbaum Associates, 2001.
  5. U. Neisser, G. Boodoo, T. J. Bouchard Jr. et al., “Intelligence: knowns and unknowns,” American Psychologist, vol. 51, no. 2, pp. 77–101, 1996. View at Publisher · View at Google Scholar
  6. E. E. Smith and S. M. Kosslyn, Cognitive Psychology: Mind and Brain, Prentice Hall, 2009.
  7. H. J. Eysenck, “The concept of “intelligence”: useful or useless?” Intelligence, vol. 12, no. 1, pp. 1–16, 1988. View at Publisher · View at Google Scholar · View at Scopus
  8. R. J. Herrnstein and C. A. Murray, The Bell Curve: Intelligence and Class Structure in American Life, Simon & Schuster, 1996.
  9. H. Rindermann, “The g-factor of international cognitive ability comparisons: the homogeneity of results in PISA, TIMSS, PIRLS and IQ-tests across nations,” European Journal of Personality, vol. 21, no. 5, pp. 667–706, 2007. View at Publisher · View at Google Scholar · View at Scopus
  10. Z. Stein and K. Heikkinen, “Models, metrics, and measurement in developmental psychology,” Integral Review, vol. 5, no. 1, pp. 4–24, 2009. View at Google Scholar
  11. G. S. Lesser, G. Fifer, and D. H. Clark, “Mental abilities of children from different social-class and cultural groups,” Monographs of the Society for Research in Child Development, vol. 30, no. 4, pp. 1–115, 1965. View at Publisher · View at Google Scholar
  12. H. Gardner, Frames of Mind, Basic Books, New York, NY, USA, 1983.
  13. L. Erlenmeyer-Kimling and L. F. Jarvik, “Genetics and intelligence: a review,” Science, vol. 142, no. 3598, pp. 1477–1479, 1963. View at Publisher · View at Google Scholar
  14. I. J. Deary, L. Penke, and W. Johnson, “The neuroscience of human intelligence differences,” Nature Reviews Neuroscience, vol. 11, no. 3, pp. 201–211, 2010. View at Publisher · View at Google Scholar · View at Scopus
  15. J. R. Gray and P. M. Thompson, “Neurobiology of intelligence: science and ethics,” Nature Reviews Neuroscience, vol. 5, no. 6, pp. 471–482, 2004. View at Publisher · View at Google Scholar
  16. M. G. Mattar and D. S. Bassett, “Brain network architecture: implications for human learning,” 2016, https://arxiv.org/abs/1609.01790.
  17. J. D. Bransford, A. L. Brown, and R. R. Cocking, How People Learn, National Academy Press, Washington, DC, USA, 2000.
  18. J. B. Carroll, Human Cognitive Abilities. A Survey of Factor-Analytic Studies, Cambridge University Press, Cambridge, UK, 1993. View at Publisher · View at Google Scholar
  19. P. L. Ackerman, “Individual differences in skill learning: an integration of psychometric and information processing perspectives,” Psychological Bulletin, vol. 102, no. 1, pp. 3–27, 1987. View at Publisher · View at Google Scholar · View at Scopus
  20. S. B. Kaufman, M. R. Reynolds, X. Liu, A. S. Kaufman, and K. S. McGrew, “Are cognitive g and academic achievement g one and the same g? An exploration on the Woodcock–Johnson and Kaufman tests,” Intelligence, vol. 40, no. 2, pp. 123–138, 2012. View at Publisher · View at Google Scholar · View at Scopus
  21. J. L. Castejon, A. M. Perez, and R. Gilar, “Confirmatory factor analysis of Project Spectrum activities. A second-order g factor or multiple intelligences?” Intelligence, vol. 38, no. 5, pp. 481–496, 2010. View at Publisher · View at Google Scholar · View at Scopus
  22. D. Molenaar, C. V. Dolan, J. M. Wicherts, and H. L. J. van der Maas, “Modeling differentiation of cognitive abilities within the higher-order factor model using moderated factor analysis,” Intelligence, vol. 38, no. 6, pp. 611–624, 2010. View at Publisher · View at Google Scholar · View at Scopus
  23. W. Johnson, J. t. Nijenhuis, and T. J. Bouchard Jr., “Still just 1 g: consistent results from five test batteries,” Intelligence, vol. 36, no. 1, pp. 81–95, 2008. View at Publisher · View at Google Scholar · View at Scopus
  24. P. L. Ackerman, M. E. Beier, and M. O. Boyle, “Working memory and intelligence: the same or different constructs?” Psychological Bulletin, vol. 131, no. 1, pp. 30–60, 2005. View at Publisher · View at Google Scholar · View at Scopus
  25. D. Wechsler, The Measurement of Adult Intelligence, Williams & Wilkins Co, Baltimore, MD, USA, 1939. View at Publisher · View at Google Scholar
  26. L. M. Terman and M. A. Merrill, Stanford-Binet Intelligence Scale: Manual for the Third Revision, Form L-M, Houghton Mifflin, Oxford, UK, 1960.
  27. C. Spearman, ““General intelligence”: objectively determined and measured,” The American Journal of Psychology, vol. 15, no. 2, p. 201, 1904. View at Publisher · View at Google Scholar
  28. R. B. Cattell, “Theory of fluid and crystallized intelligence: a critical experiment,” Journal of Educational Psychology, vol. 54, no. 1, pp. 1–22, 1963. View at Publisher · View at Google Scholar · View at Scopus
  29. J. L. Horn and R. B. Cattell, “Refinement and test of the theory of fluid and crystallized general intelligences,” Journal of Educational Psychology, vol. 57, no. 5, pp. 253–270, 1966. View at Publisher · View at Google Scholar · View at Scopus
  30. V. D. Schmittmann, A. O. J. Cramer, L. J. Waldorp, S. Epskamp, R. A. Kievit, and D. Borsboom, “Deconstructing the construct: a network perspective on psychological phenomena,” New Ideas in Psychology, vol. 31, no. 1, pp. 43–53, 2013. View at Publisher · View at Google Scholar · View at Scopus
  31. C. V. Dolan, F. J. Oort, R. D. Stoel, and J. M. Wicherts, “Testing measurement invariance in the target rotated multigroup exploratory factor model,” Structural Equation Modeling, vol. 16, no. 2, pp. 295–314, 2009. View at Publisher · View at Google Scholar · View at Scopus
  32. D. Massidda, M. Giovanni Carta, and G. Altoè, “Integrating different factorial solutions of a psychometric tool via social network analysis,” Methodology, vol. 12, no. 3, pp. 97–106, 2016. View at Publisher · View at Google Scholar · View at Scopus
  33. A. Juarrero, “Complex Dynamical Systems Theory,” Tech. Rep., Cognitive Edge, 2010. View at Google Scholar
  34. P. Van Geert, “A dynamic systems model of cognitive and language growth,” Psychological Review, vol. 98, no. 1, pp. 3–53, 1991. View at Publisher · View at Google Scholar · View at Scopus
  35. P. van Geert and H. Steenbeek, “The dynamics of scaffolding,” New Ideas in Psychology, vol. 23, no. 3, pp. 115–128, 2005. View at Publisher · View at Google Scholar · View at Scopus
  36. H. L. J. Van Der Maas, C. V. Dolan, R. P. P. P. Grasman, J. M. Wicherts, H. M. Huizenga, and M. E. J. Raijmakers, “A dynamical model of general intelligence: the positive manifold of intelligence by mutualism,” Psychological Review, vol. 113, no. 4, pp. 842–861, 2006. View at Publisher · View at Google Scholar · View at Scopus
  37. H. W. Steenbeek and P. L. C. van Geert, “A theory and dynamic model of dyadic interaction: concerns, appraisals, and contagiousness in a developmental context,” Developmental Review, vol. 27, no. 1, pp. 1–40, 2007. View at Publisher · View at Google Scholar · View at Scopus
  38. S. van Vondel, H. Steenbeek, M. van Dijk, and P. van Geert, “Ask, don’t tell; a complex dynamic systems approach to improving science education by focusing on the co-construction of scientific understanding,” Teaching and Teacher Education, vol. 63, pp. 243–253, 2017. View at Publisher · View at Google Scholar · View at Scopus
  39. C. H. Geveke, H. W. Steenbeek, J. M. Doornenbal, and P. L. C. van Geert, “Attractor states in teaching and learning processes: a study of out-of-school science education,” Frontiers in Psychology, vol. 8, 299 pages, 2017. View at Publisher · View at Google Scholar · View at Scopus
  40. R. R. Vallacher and A. Nowak, Eds., Dynamical Systems in Social Psychology, Academic Press, 1994.
  41. E. M. A. Blijd-Hoogewys and P. L. C. van Geert, “Non-linearities in theory-of-mind development,” Frontiers in Psychology, vol. 7, p. 1970, 2017. View at Publisher · View at Google Scholar · View at Scopus
  42. R. R. Vallacher, P. van Geert, and A. Nowak, “The intrinsic dynamics of psychological process,” Current Directions in Psychological Science, vol. 24, no. 1, pp. 58–64, 2015. View at Publisher · View at Google Scholar · View at Scopus
  43. R. J. R. Den Hartigh, M. W. G. van Dijk, H. W. Steenbeek, and P. L. C. van Geert, “A dynamic network model to explain the development of excellent human performance,” Frontiers in Psychology, vol. 7, p. 532, 2016. View at Publisher · View at Google Scholar · View at Scopus
  44. M. Newman, Networks: An Introduction, Oxford University Press, 2010. View at Publisher · View at Google Scholar · View at Scopus
  45. P. Erdös and A. Rényi, “On random graphs,” Publicationes Mathematicae Debrecen, vol. 6, pp. 290–297, 1959. View at Google Scholar
  46. A.-L. Barabasi, “Scale-free networks: a decade and beyond,” Science, vol. 325, no. 5939, pp. 412-413, 2009. View at Publisher · View at Google Scholar · View at Scopus
  47. G. Caldarelli, Scale-Free Networks. Complex Webs in Nature and Technology. Reprint of the 2007 Hardback, Oxford University Press, 2013.
  48. R. Albert and A.-L. Barabási, “Statistical mechanics of complex networks,” Reviews of Modern Physics, vol. 74, no. 1, pp. 47–97, 2002. View at Publisher · View at Google Scholar · View at Scopus
  49. E. Ravasz, A. L. Somera, D. A. Mongru, Z. N. Oltvai, and A. L. Barabási, “Hierarchical organization of modularity in metabolic networks,” Science, vol. 297, no. 5586, pp. 1551–1555, 2002. View at Publisher · View at Google Scholar · View at Scopus
  50. M. Girvan and M. E. J. Newman, “Community structure in social and biological networks,” Proceedings of the National Academy of Sciences, vol. 99, no. 12, pp. 7821–7826, 2002. View at Publisher · View at Google Scholar · View at Scopus
  51. L. Danon, A. Díaz-Guilera, J. Duch, and A. Arenas, “Comparing community structure identification,” Journal of Statistical Mechanics: Theory and Experiment, vol. 2005, no. 9, 2005. View at Publisher · View at Google Scholar · View at Scopus
  52. M. E. J. Newman and M. Girvan, “Finding and evaluating community structure in networks,” Physical Review E, vol. 69, no. 2, article 026113, 2004. View at Publisher · View at Google Scholar · View at Scopus
  53. M. E. J. Newman, “Fast algorithm for detecting community structure in networks,” Physical Review E, vol. 69, no. 6, article 066133, 2004. View at Publisher · View at Google Scholar · View at Scopus
  54. M. E. J. Newman, “Modularity and community structure in networks,” Proceedings of the National Academy of Sciences, vol. 103, no. 23, pp. 8577–8582, 2006. View at Publisher · View at Google Scholar · View at Scopus
  55. K. M. Newell, Y. T. Liu, and G. Mayer-Kress, “Time scales in motor learning and development,” Psychological Review, vol. 108, no. 1, pp. 57–82, 2001. View at Publisher · View at Google Scholar · View at Scopus
  56. M. T. Hannan and J. Freeman, “The population ecology of organizations,” American Journal of Sociology, vol. 82, no. 5, pp. 929–964, 1977. View at Publisher · View at Google Scholar
  57. A. Tsoularis and J. Wallace, “Analysis of logistic growth models,” Mathematical Biosciences, vol. 179, no. 1, pp. 21–55, 2002. View at Publisher · View at Google Scholar · View at Scopus
  58. H. van der Maas, K. J. Kan, M. Marsman, and C. E. Stevenson, “Network models for cognitive development and intelligence,” Journal of Intelligence, vol. 5, no. 2, p. 16, 2017. View at Publisher · View at Google Scholar
  59. L. Katz, “A new status index derived from sociometric analysis,” Psychometrika, vol. 18, no. 1, pp. 39–43, 1953. View at Publisher · View at Google Scholar · View at Scopus
  60. R. Ghosh, S.-h. Teng, K. Lerman, and X. Yan, “The interplay between dynamics and networks,” in Proceedings of the 20th ACM SIGKDD international conference on Knowledge discovery and data mining - KDD '14, pp. 1406–1415, New York, NY, USA, 2014. View at Publisher · View at Google Scholar · View at Scopus
  61. S. P. Borgatti, “Centrality and network flow,” Social Networks, vol. 27, no. 1, pp. 55–71, 2005. View at Publisher · View at Google Scholar · View at Scopus
  62. S. U. Pillai, T. Suel, and Seunghun Cha, “The Perron-Frobenius theorem. Some of its applications,” IEEE Signal Processing Magazine, vol. 22, no. 2, pp. 62–75, 2005. View at Publisher · View at Google Scholar · View at Scopus
  63. K. I. Goh, B. Kahng, and D. Kim, “Spectra and eigenvectors of scale-free networks,” Physical Review E, vol. 64, no. 5, article 051903, 2001. View at Publisher · View at Google Scholar · View at Scopus
  64. I. Cohen, Y. Huang, J. Chen, and J. Benesty, Noise Reduction in Speech Processing, Vol. 2, Chapter 5, Springer, 2009. View at Publisher · View at Google Scholar
  65. C. Spearman, “The proof and measurement of association between two things,” The American Journal of Psychology, vol. 15, no. 1, pp. 72–101, 1904. View at Publisher · View at Google Scholar
  66. J. C. F. de Winter, S. D. Gosling, and J. Potter, “Comparing the Pearson and Spearman correlation coefficients across distributions and sample sizes: a tutorial using simulations and empirical data,” Psychological Methods, vol. 21, no. 3, pp. 273–290, 2016. View at Publisher · View at Google Scholar
  67. R. B. Cattell, The Scientific Use of Factor Analysis in Behavioral and Life Sciences, Springer US, 1978. View at Publisher · View at Google Scholar
  68. E. McAuley, T. Duncan, and V. V. Tammen, “Psychometric properties of the intrinsic motivation inventory in a competitive sport setting: a confirmatory factor analysis,” Research Quarterly for Exercise and Sport, vol. 60, no. 1, pp. 48–58, 1989. View at Publisher · View at Google Scholar · View at Scopus
  69. E. S. Barratt, “Factor analysis of some psychometric measures of impulsiveness and anxiety,” Psychological Reports, vol. 16, no. 2, pp. 547–554, 1965. View at Publisher · View at Google Scholar
  70. P. Kline, The New Psychometrics: Science, Psychology, and Measurement, Routledge, 1998.
  71. R. B. Cattell, “The scree test for the number of factors,” Multivariate Behavioral Research, vol. 1, no. 2, pp. 245–276, 1966. View at Publisher · View at Google Scholar · View at Scopus
  72. J. L. Horn, “A rationale and test for the number of factors in factor analysis,” Psychometrika, vol. 30, no. 2, pp. 179–185, 1965. View at Publisher · View at Google Scholar · View at Scopus
  73. W. R. Zwick and W. F. Velicer, “Factors influencing four rules for determining the number of components to retain,” Multivariate Behavioral Research, vol. 17, no. 2, pp. 253–269, 1982. View at Publisher · View at Google Scholar · View at Scopus
  74. A. B. Costello and J. W. Osborne, “Best practices in exploratory factor analysis: four recommendations for getting the most from your analysis,” Practical Assessment, Research & Education, vol. 10, no. 7, pp. 1–9, 2005. View at Google Scholar
  75. L. R. Fabrigar, D. T. Wegener, R. C. Maccallum, and E. J. Strahan, “Evaluating the use of exploratory factor analysis in psychological research,” Psychological Methods, vol. 4, no. 3, pp. 272–299, 1999. View at Publisher · View at Google Scholar · View at Scopus
  76. N. L. Ritter, “A comparison of distribution free and non-distribution free factor analysis methods,” in Annual Meeting of the Southwest Educational Research Association, p. 15, New Orleans, LA, USA, February 2012.
  77. R. R. Nadakuditi and M. E. J. Newman, “Graph spectra and the detectability of community structure in networks,” Physical Review Letters, vol. 108, no. 18, 2012. View at Publisher · View at Google Scholar · View at Scopus
  78. J. M. Montoya, S. L. Pimm, and R. V. Solé, “Ecological networks and their fragility,” Nature, vol. 442, no. 7100, pp. 259–264, 2006. View at Publisher · View at Google Scholar · View at Scopus
  79. B. S. Goh, “Global stability in many-species systems,” The American Naturalist, vol. 111, no. 977, pp. 135–143, 1977. View at Publisher · View at Google Scholar
  80. W. Feng and R. M. Bailey, “Stability and critical transitions in mutualistic ecological systems,” bioRxiv, 2017. View at Publisher · View at Google Scholar
  81. J. M. Pastor, J. García-Algarra, J. M. Iriondo, J. J. Ramasco, and J. Galeano, “Dragging in mutualistic networks,” Networks and Heterogeneous Media, vol. 10, no. 1, pp. 37–52, 2015. View at Publisher · View at Google Scholar · View at Scopus
  82. M. E. J. Newman, D. J. Watts, and S. H. Strogatz, “Random graph models of social networks,” Proceedings of the National Academy of Sciences, vol. 99, Supplement 1, pp. 2566–2572, 2002. View at Publisher · View at Google Scholar · View at Scopus