更改

跳到导航 跳到搜索
添加5,882字节 、 2021年8月12日 (四) 16:18
第86行: 第86行:     
A simple way of quantifying complexity on a structural basis would be to count the number of components and/or interactions within a system.  An examination of the number of structural parts (McShea, 1996) and functional behaviors of organisms across evolution demonstrates that these measures increase over time, a finding that contributes to the ongoing debate over whether complexity grows as a result of natural selection.  However, numerosity alone may only be an indicator of complicatedness, but not necessarily of complexity.  Large and highly coupled systems may not be more complex than those that are smaller and less coupled.  For example, a very large system that is fully connected can be described in a compact manner and may tend to generate uniform behavior, while the description of a smaller but more heterogeneous system may be less compressible and its behavior may be more differentiated.
 
A simple way of quantifying complexity on a structural basis would be to count the number of components and/or interactions within a system.  An examination of the number of structural parts (McShea, 1996) and functional behaviors of organisms across evolution demonstrates that these measures increase over time, a finding that contributes to the ongoing debate over whether complexity grows as a result of natural selection.  However, numerosity alone may only be an indicator of complicatedness, but not necessarily of complexity.  Large and highly coupled systems may not be more complex than those that are smaller and less coupled.  For example, a very large system that is fully connected can be described in a compact manner and may tend to generate uniform behavior, while the description of a smaller but more heterogeneous system may be less compressible and its behavior may be more differentiated.
 +
 +
在结构基础上量化复杂性的一个简单方法是计算一个系统中的'''组件和/或交互的数量'''。对生物进化过程中结构部分的数量(McShea,1996)和功能行为的研究表明,这些指标会随着时间的推移而增加,这一发现有助于持续辩论复杂性是否是自然选择的结果。然而,数量本身可能只是复杂性的'''一个'''指标,而不完全是复杂性的指标。大型和高度耦合的系统可能并不比那些较小和耦合较少的系统复杂。例如,一个完全连接的非常大的系统可以用一种紧凑的方式来描述,并且可能会产生统一的行为,而一个较小但是更多异质的系统的描述可能不那么可压缩,其行为可能更加不同。相同节点数量下,[[无标度网络]]往往比全连接网络更复杂。
 +
    
Effective complexity (Gell-Mann, 1995) measures the minimal description length of a system’s regularities.  As such, this measure is related to AIC, but it attempts to distinguish regular features from random or incidental ones and therefore belongs within the family of complexity measures that aim at capturing how much structure a system contains. The separation of regular features from random ones may be difficult for any given empirical system, and it may crucially depend on criteria supplied by an external observer.
 
Effective complexity (Gell-Mann, 1995) measures the minimal description length of a system’s regularities.  As such, this measure is related to AIC, but it attempts to distinguish regular features from random or incidental ones and therefore belongs within the family of complexity measures that aim at capturing how much structure a system contains. The separation of regular features from random ones may be difficult for any given empirical system, and it may crucially depend on criteria supplied by an external observer.
 +
 +
'''有效复杂性 Effective complexity''' (Gell-Mann,1995)度量了系统规律性的最小描述长度。这种测量方法与 AIC 有关,但它试图区分常规特征和随机或偶然特征,因此属于旨在获取系统包含多少结构的复杂性测量方法的范畴。对于任何给定的经验系统来说,区分开规则特征和随机特征都是困难的,而且它可能取决于外部观察者提供的标准。
 +
    
Physical complexity (Adami and Cerf, 2000) is related to effective complexity and is designed to estimate the complexity of any sequence of symbols that is about a physical world or environment.  As such the measure is particularly useful when applied to biological systems.  It is defined as the Kolmogorov complexity (AIC) that is shared between a sequence of symbols (such as a genome) and some description of the environment in which that sequence has meaning (such as an ecological niche).  Since the Kolmogorov complexity is not computable, neither is the physical complexity.  However, the average physical complexity of an ensemble of sequences (e.g. the set of genomes of an entire population of organisms) can be approximated by the mutual information between the ensembles of sequences (genomes) and their environment (ecology).  Experiments conducted in a digital ecology (Adami, 2002) demonstrated that the mutual information between self-replicating genomes and their environment increased along evolutionary time.  Physical complexity has also been used to estimate the complexity of biomolecules.  The structural and functional complexity of a set of RNA molecules were shown to positively correlate with physical complexity (Carothers et al., 2004), indicating a possible link between functional capacities of evolved molecular structures and the amount of information they encode.
 
Physical complexity (Adami and Cerf, 2000) is related to effective complexity and is designed to estimate the complexity of any sequence of symbols that is about a physical world or environment.  As such the measure is particularly useful when applied to biological systems.  It is defined as the Kolmogorov complexity (AIC) that is shared between a sequence of symbols (such as a genome) and some description of the environment in which that sequence has meaning (such as an ecological niche).  Since the Kolmogorov complexity is not computable, neither is the physical complexity.  However, the average physical complexity of an ensemble of sequences (e.g. the set of genomes of an entire population of organisms) can be approximated by the mutual information between the ensembles of sequences (genomes) and their environment (ecology).  Experiments conducted in a digital ecology (Adami, 2002) demonstrated that the mutual information between self-replicating genomes and their environment increased along evolutionary time.  Physical complexity has also been used to estimate the complexity of biomolecules.  The structural and functional complexity of a set of RNA molecules were shown to positively correlate with physical complexity (Carothers et al., 2004), indicating a possible link between functional capacities of evolved molecular structures and the amount of information they encode.
 +
 +
'''物理复杂度 Physical complexity''' (Adami 和 Cerf,2000)与有效复杂性有关,其目的是估计任何符号序列的复杂性,这些符号是关于物理世界或环境的。因此,这一措施在应用于生物系统时特别有用。给定一个符号序列(比如基因组),和一些使得该序列具有意义的环境(比如生态位)的描述,它们之间共有的算法复杂度(AIC),就是物理复杂性。由于算法复杂度是不可计算的,物理复杂度也不能。然而,一组序列(例如整个生物体的基因组集合)的平均物理复杂性可以用序列(基因组)集合与其环境(生态学)之间的相互信息来近似。在数字生态学(Adami,2002)中进行的实验表明,自我复制的基因组与其环境之间的互信息随着进化时间的增加而增加。物理复杂度也被用来估计生物分子的复杂性。一组 RNA 分子的结构和功能复杂性与物理复杂性呈正相关(Carothers 等人,2004) ,这表明进化的分子结构的功能与它们编码的信息量之间可能存在联系。
 +
    
Statistical complexity (Crutchfield and Young, 1989) is a component of a broader theoretic framework known as computational mechanics, and can be calculated directly from empirical data.  To calculate the statistical complexity, each point in the time series is mapped to a corresponding symbol according to some partitioning scheme, so that the raw data is now a stream of consecutive symbols.  The symbol sequences are then clustered into causal states according to the following rule: two symbol sequences (i.e. histories of the dynamics) are contained in the same causal state if the conditional probability of any future symbol is identical for these two histories.  In other words, two symbol sequences are considered to be the same if, on average, they predict the same distribution of future dynamics. Once these causal states have been identified, the transition probabilities between causal states can be extracted from the data, and the long-run probability distribution over all causal states can be calculated. The statistical complexity is then defined as the Shannon entropy of this distribution over causal states.  Statistical complexity can be calculated analytically for abstract systems such as the logistic map, cellular automata and many basic Markov processes, and computational methods for constructing appropriate causal states in real systems, while taxing, exist and have been applied in a variety of contexts.
 
Statistical complexity (Crutchfield and Young, 1989) is a component of a broader theoretic framework known as computational mechanics, and can be calculated directly from empirical data.  To calculate the statistical complexity, each point in the time series is mapped to a corresponding symbol according to some partitioning scheme, so that the raw data is now a stream of consecutive symbols.  The symbol sequences are then clustered into causal states according to the following rule: two symbol sequences (i.e. histories of the dynamics) are contained in the same causal state if the conditional probability of any future symbol is identical for these two histories.  In other words, two symbol sequences are considered to be the same if, on average, they predict the same distribution of future dynamics. Once these causal states have been identified, the transition probabilities between causal states can be extracted from the data, and the long-run probability distribution over all causal states can be calculated. The statistical complexity is then defined as the Shannon entropy of this distribution over causal states.  Statistical complexity can be calculated analytically for abstract systems such as the logistic map, cellular automata and many basic Markov processes, and computational methods for constructing appropriate causal states in real systems, while taxing, exist and have been applied in a variety of contexts.
 +
 +
'''统计复杂度 Statistical Complexity'''(Crutchfield and Young,1989)是一个更广泛的、被称为'''计算力学 Computational Mechanics'''的理论框架的组成部分,并可以直接从经验数据估算。为了计算统计复杂度,时间序列中的每个点根据某种划分方案映射到对应的符号,使原始数据成为一个连续的符号流。然后,符号序列按照以下规则聚类成因果状态: 如果任何未来符号的条件概率对于这两个历史是相同的,那么两个符号序列(即动力学的历史)就属于相同的因果状态。换句话说,如果两个符号序列预测了同一个未来的动力学分布,那么可以认为这两个符号序列是相同的。一旦确定了这些因果状态,就可以从数据中提取因果状态之间的转换概率,并计算出所有因果状态的长期概率分布。统计复杂度被定义为因果状态上这种分布的香农熵。对于逻辑斯谛映射、元胞自动机和许多基本的马尔可夫过程等抽象系统,统计复杂性可以通过解析方法计算出来。
 +
    
Predictive information (Bialek et al., 2001), while not in itself a complexity measure, can be used to separate systems into different complexity categories based on the principle of the extensivity of entropy.  Extensivity manifests itself, for example, in systems composed of increasing numbers of homogeneous independent random variables.  The Shannon entropy of such systems will grow linearly with their size.  This linear growth of entropy with system size is known as extensivity.  However, the constituent elements of a complex system are typically inhomogeneous and interdependent, so that as the number of random variables grows the entropy does not always grow linearly.  The manner in which a given system departs from extensivity can be used to characterize its complexity.
 
Predictive information (Bialek et al., 2001), while not in itself a complexity measure, can be used to separate systems into different complexity categories based on the principle of the extensivity of entropy.  Extensivity manifests itself, for example, in systems composed of increasing numbers of homogeneous independent random variables.  The Shannon entropy of such systems will grow linearly with their size.  This linear growth of entropy with system size is known as extensivity.  However, the constituent elements of a complex system are typically inhomogeneous and interdependent, so that as the number of random variables grows the entropy does not always grow linearly.  The manner in which a given system departs from extensivity can be used to characterize its complexity.
 +
 +
'''预测信息 Predictive Information''' (Bialek 等,2001) ,虽然本身不是一个复杂性度量,但可以用来根据熵的可扩展性原理将系统分成不同的复杂性类别。例如,可扩展性表现在由越来越多的同质独立随机变量组成的系统中。这类系统的香农熵将随其大小线性增长。熵随系统大小的线性增长称为扩展性。然而,复杂系统的组成元素通常是异质的和相互依赖的。因此当随机变量数量的增长,熵并不总是线性增长的。一个给定系统偏离扩展性的方式也可以用来描述其复杂性。
 +
    
[[Image:complexity_figure2.jpg|thumb|400px|right|F2|Movie frames from a demonstration model of neural dynamics, consisting of 1600 spontaneously active Wison-Cowan neural mass units arranged on a sphere and coupled by excitatory connections.  Three cases are shown: sparse coupling (local connections only), uniform coupling (global connections only), and a mixture of local and global connections (forming a small-world network).  Neural complexity (Tononi et al., 1994; Sporns et al., 2000) is high only for the last case.]]
 
[[Image:complexity_figure2.jpg|thumb|400px|right|F2|Movie frames from a demonstration model of neural dynamics, consisting of 1600 spontaneously active Wison-Cowan neural mass units arranged on a sphere and coupled by excitatory connections.  Three cases are shown: sparse coupling (local connections only), uniform coupling (global connections only), and a mixture of local and global connections (forming a small-world network).  Neural complexity (Tononi et al., 1994; Sporns et al., 2000) is high only for the last case.]]
第99行: 第114行:  
Neural complexity (Tononi et al., 1994), which may be applied to any empirically observed system including brains, is related to the extensivity of a system.  One of its building blocks is integration (also called multi-information), a multivariate extension of mutual information that estimates the total amount of statistical structure within an arbitrarily large system.  Integration is computed as the difference between the sum of the component’s individual entropies and the joint entropy of the system as a whole.  The distribution of integration across multiple spatial scales captures the complexity of the system.  Consider the following three cases (<figref>Complexity_figure2.jpg</figref>).  (1) A system with components that are statistically independent will exhibit globally disordered or random dynamics.  Its joint entropy will be exactly equal to the sum of the component entropies, and system integration will be zero, regardless of which spatial scale of the system is examined.  (2) Any statistical dependence between components will lead to a contraction of the system’s joint entropy relative to the individual entropies, resulting in positive integration.  If the components of a system are highly coupled and exhibit statistical dependencies as well as homogeneous dynamics (i.e. all components behave identically) then estimates of integration across multiple spatial scales of the system will, on average, follow a linear distribution.  (3) If statistical dependencies are inhomogeneous (for example involving groupings of components, modules, or hierarchical patterns) the distribution of integration will deviate from linearity.  The total amount of deviation is the system’s complexity.  The complexity of a random system is zero, while the complexity of a homogeneous coupled (regular) system is very low.  Systems with rich structure and dynamic behavior have high complexity.
 
Neural complexity (Tononi et al., 1994), which may be applied to any empirically observed system including brains, is related to the extensivity of a system.  One of its building blocks is integration (also called multi-information), a multivariate extension of mutual information that estimates the total amount of statistical structure within an arbitrarily large system.  Integration is computed as the difference between the sum of the component’s individual entropies and the joint entropy of the system as a whole.  The distribution of integration across multiple spatial scales captures the complexity of the system.  Consider the following three cases (<figref>Complexity_figure2.jpg</figref>).  (1) A system with components that are statistically independent will exhibit globally disordered or random dynamics.  Its joint entropy will be exactly equal to the sum of the component entropies, and system integration will be zero, regardless of which spatial scale of the system is examined.  (2) Any statistical dependence between components will lead to a contraction of the system’s joint entropy relative to the individual entropies, resulting in positive integration.  If the components of a system are highly coupled and exhibit statistical dependencies as well as homogeneous dynamics (i.e. all components behave identically) then estimates of integration across multiple spatial scales of the system will, on average, follow a linear distribution.  (3) If statistical dependencies are inhomogeneous (for example involving groupings of components, modules, or hierarchical patterns) the distribution of integration will deviate from linearity.  The total amount of deviation is the system’s complexity.  The complexity of a random system is zero, while the complexity of a homogeneous coupled (regular) system is very low.  Systems with rich structure and dynamic behavior have high complexity.
   −
Applied to neural dynamics, complexity as defined by Tononi et al. (1994) was found to be associated with specific patterns of neural connectivity, for example those exhibiting attributes of small-world networks (Sporns et al., 2000).  Descendants of this complexity measure have addressed the matching of systems to an environment as well as their degeneracy, i.e. their capacity to produce functionally equivalent behaviors in different ways.  
+
'''神经复杂性 Neural Complexity'''(Tononi 等,1994) ,与系统的延伸性相关,可以应用于任何经验系统,包括大脑。神经复杂度中的一个重要概念是'''集成度 Integration'''(也称为多信息) ,这是互信息的多元扩展,用于估计任意大系统中的统计结构的总量。集成度被定义为组成部分的个体熵之和与系统作为一个整体的联合熵之差。集成度在多个空间尺度上的分布表明了系统的复杂性。考虑以下三种情况(<figref>Complexity_figure2.jpg</figref>)。(1)具有统计独立成分的系统会呈现全局无序或随机动态。它的联合熵将正好等于组成熵之和,系统的集成度将为零,无论哪个空间尺度的系统被检查。(2)各组分之间的统计相关性将导致系统的联合熵相对于个体熵的收缩,从而导致正的集成度。如果一个系统的组成部分是高度耦合的,并且表现出统计依赖性以及同质动力学(即所有组成部分的行为完全相同) ,那么该系统的多个空间尺度上的集成估计平均遵循线性分布。(3)如果统计依赖关系是非同质的(例如涉及组件、模块或层次模式的分组) ,则集成的分布将偏离线性。偏差的总量是系统的复杂性。随机系统的复杂度为零,而齐次耦合系统的复杂度非常低。具有丰富结构和动态行为的系统具有很高的复杂性。
 +
 
 +
 
 +
Applied to neural dynamics, complexity as defined by Tononi et al. (1994) was found to be associated with specific patterns of neural connectivity, for example those exhibiting attributes of small-world networks (Sporns et al., 2000).  Descendants of this complexity measure have addressed the matching of systems to an environment as well as their degeneracy, i.e. their capacity to produce functionally equivalent behaviors in different ways.
 +
 
 +
应用于神经动力学,托诺尼等人(1994年)定义的复杂性被发现与神经连接的特定模式有关,例如小世界网络所表现出来的属性(Sporns 等,2000)。这种复杂性度量的和在此基础上提出来的各种度量已经解决了系统与环境的匹配以及它们的简并性问题。
    
==复杂网络==
 
==复杂网络==
370

个编辑

导航菜单