第495行: |
第495行: |
| | | |
| 蒙特卡罗方法的统计标准是由 Sawilowsky 制定的。在应用统计学中,蒙特卡罗方法至少可用于四种目的: | | 蒙特卡罗方法的统计标准是由 Sawilowsky 制定的。在应用统计学中,蒙特卡罗方法至少可用于四种目的: |
− |
| |
− | The US Coast Guard utilizes Monte Carlo methods within its computer modeling software SAROPS in order to calculate the probable locations of vessels during search and rescue operations. Each simulation can generate as many as ten thousand data points that are randomly distributed based upon provided variables. Search patterns are then generated based upon extrapolations of these data in order to optimize the probability of containment (POC) and the probability of detection (POD), which together will equal an overall probability of success (POS). Ultimately this serves as a practical application of probability distribution in order to provide the swiftest and most expedient method of rescue, saving both lives and resources.
| |
− |
| |
− | 美国海岸警卫队在其计算机建模软件 SAROPS 中使用蒙特卡罗方法,以便在搜索和救援行动中计算可能的船只位置。每个模拟可以生成多达一万个数据点,这些数据点是根据提供的变量随机分布的。然后根据这些数据的推断生成搜索模式,以优化包容概率(POC)和检测概率(POD) ,这两者合起来等于总体成功概率(POS)。最终,这作为概率分布的一个实际应用,以提供最迅速和最便捷的救援方法,拯救生命和资源。
| |
| | | |
| #To compare competing statistics for small samples under realistic data conditions. Although [[type I error]] and power properties of statistics can be calculated for data drawn from classical theoretical distributions (''e.g.'', [[normal curve]], [[Cauchy distribution]]) for [[asymptotic]] conditions (''i. e'', infinite sample size and infinitesimally small treatment effect), real data often do not have such distributions.<ref>{{harvnb|Sawilowsky|Fahoome|2003}}</ref> | | #To compare competing statistics for small samples under realistic data conditions. Although [[type I error]] and power properties of statistics can be calculated for data drawn from classical theoretical distributions (''e.g.'', [[normal curve]], [[Cauchy distribution]]) for [[asymptotic]] conditions (''i. e'', infinite sample size and infinitesimally small treatment effect), real data often do not have such distributions.<ref>{{harvnb|Sawilowsky|Fahoome|2003}}</ref> |
第580行: |
第576行: |
| The [[US Coast Guard]] utilizes Monte Carlo methods within its computer modeling software [[SAROPS]] in order to calculate the probable locations of vessels during [[search and rescue]] operations. Each simulation can generate as many as ten thousand data points that are randomly distributed based upon provided variables.<ref>{{cite web|url=http://insights.dice.com/2014/01/03/how-the-coast-guard-uses-analytics-to-search-for-those-lost-at-sea|title=How the Coast Guard Uses Analytics to Search for Those Lost at Sea|work=Dice Insights|date=2014-01-03}}</ref> Search patterns are then generated based upon extrapolations of these data in order to optimize the probability of containment (POC) and the probability of detection (POD), which together will equal an overall probability of success (POS). Ultimately this serves as a practical application of [[probability distribution]] in order to provide the swiftest and most expedient method of rescue, saving both lives and resources.<ref>{{cite web|url=http://www.ifremer.fr/web-com/sar2011/Presentations/SARWS2011_STONE_L.pdf|title=Search Modeling and Optimization in USCG's Search and Rescue Optimal Planning System (SAROPS)|author1=Lawrence D. Stone|author2=Thomas M. Kratzke|author3=John R. Frost|website=Ifremer.fr|accessdate=28 October 2017}}</ref> | | The [[US Coast Guard]] utilizes Monte Carlo methods within its computer modeling software [[SAROPS]] in order to calculate the probable locations of vessels during [[search and rescue]] operations. Each simulation can generate as many as ten thousand data points that are randomly distributed based upon provided variables.<ref>{{cite web|url=http://insights.dice.com/2014/01/03/how-the-coast-guard-uses-analytics-to-search-for-those-lost-at-sea|title=How the Coast Guard Uses Analytics to Search for Those Lost at Sea|work=Dice Insights|date=2014-01-03}}</ref> Search patterns are then generated based upon extrapolations of these data in order to optimize the probability of containment (POC) and the probability of detection (POD), which together will equal an overall probability of success (POS). Ultimately this serves as a practical application of [[probability distribution]] in order to provide the swiftest and most expedient method of rescue, saving both lives and resources.<ref>{{cite web|url=http://www.ifremer.fr/web-com/sar2011/Presentations/SARWS2011_STONE_L.pdf|title=Search Modeling and Optimization in USCG's Search and Rescue Optimal Planning System (SAROPS)|author1=Lawrence D. Stone|author2=Thomas M. Kratzke|author3=John R. Frost|website=Ifremer.fr|accessdate=28 October 2017}}</ref> |
| | | |
− | Another powerful and very popular application for random numbers in numerical simulation is in numerical optimization. The problem is to minimize (or maximize) functions of some vector that often has many dimensions. Many problems can be phrased in this way: for example, a computer chess program could be seen as trying to find the set of, say, 10 moves that produces the best evaluation function at the end. In the traveling salesman problem the goal is to minimize distance traveled. There are also applications to engineering design, such as multidisciplinary design optimization. It has been applied with quasi-one-dimensional models to solve particle dynamics problems by efficiently exploring large configuration space. Reference is a comprehensive review of many issues related to simulation and optimization.
| + | The US Coast Guard utilizes Monte Carlo methods within its computer modeling software SAROPS in order to calculate the probable locations of vessels during search and rescue operations. Each simulation can generate as many as ten thousand data points that are randomly distributed based upon provided variables. Search patterns are then generated based upon extrapolations of these data in order to optimize the probability of containment (POC) and the probability of detection (POD), which together will equal an overall probability of success (POS). Ultimately this serves as a practical application of probability distribution in order to provide the swiftest and most expedient method of rescue, saving both lives and resources. |
| | | |
− | 另一个强大的和非常流行的应用随机数在数值模拟是在数值优化。问题在于如何最小化(或最大化)某些向量的函数,这些向量通常具有多个维度。许多问题可以这样表述: 例如,一个计算机国际象棋程序可以被视为试图找到一组,比如说,10步棋,最终产生最好的评价函数。在旅行商问题中,目标是使旅行距离最小。在工程设计中也有一些应用,如多学科设计优化。它已被应用于准一维模型,以解决粒子动力学问题,有效地探索大型位形空间。参考文献是对许多与模拟和优化有关的问题的全面回顾。
| + | 美国海岸警卫队在其计算机建模软件 SAROPS 中使用蒙特卡罗方法,以便在搜索和救援行动中计算可能的船只位置。每个模拟可以生成多达一万个数据点,这些数据点是根据提供的变量随机分布的。然后根据这些数据的推断生成搜索模式,以优化包容概率(POC)和检测概率(POD) ,这两者合起来等于总体成功概率(POS)。最终,这作为概率分布的一个实际应用,以提供最迅速和最便捷的救援方法,拯救生命和资源。 |
| | | |
| ===Finance and business=== | | ===Finance and business=== |
第595行: |
第591行: |
| | | |
| 金融领域的蒙特卡罗方法通常用于评估一个业务单位或公司层面的项目投资,或其他金融估值。它们可以用于对项目进度表进行建模,模拟对每个任务的最坏情况、最好情况和最可能的持续时间进行聚合估计,以确定整个项目的结果。蒙特卡罗方法也用于期权定价,违约风险分析 https://risk.octigo.pl/。此外,它们还可以用来评估医疗干预措施的财务影响。 | | 金融领域的蒙特卡罗方法通常用于评估一个业务单位或公司层面的项目投资,或其他金融估值。它们可以用于对项目进度表进行建模,模拟对每个任务的最坏情况、最好情况和最可能的持续时间进行聚合估计,以确定整个项目的结果。蒙特卡罗方法也用于期权定价,违约风险分析 https://risk.octigo.pl/。此外,它们还可以用来评估医疗干预措施的财务影响。 |
− |
| |
− | The traveling salesman problem is what is called a conventional optimization problem. That is, all the facts (distances between each destination point) needed to determine the optimal path to follow are known with certainty and the goal is to run through the possible travel choices to come up with the one with the lowest total distance. However, let's assume that instead of wanting to minimize the total distance traveled to visit each desired destination, we wanted to minimize the total time needed to reach each destination. This goes beyond conventional optimization since travel time is inherently uncertain (traffic jams, time of day, etc.). As a result, to determine our optimal path we would want to use simulation - optimization to first understand the range of potential times it could take to go from one point to another (represented by a probability distribution in this case rather than a specific distance) and then optimize our travel decisions to identify the best path to follow taking that uncertainty into account.
| |
− |
| |
− | 旅行推销员问题被称为传统的最佳化问题问题。也就是说,确定最佳路径所需的所有事实(每个目的地之间的距离)都是确定无疑的,目标是通过可能的旅行选择得出总距离最小的路径。然而,让我们假设,我们不想最小化访问每个想要的目的地所需的总距离,而是想最小化到达每个目的地所需的总时间。这超越了传统的优化,因为旅行时间是固有的不确定性(交通堵塞,一天的时间,等)。因此,为了确定我们的最佳路径,我们需要使用模拟优化来首先了解从一个点到另一个点可能需要的时间范围(在这个例子中用概率分布代表,而不是特定的距离) ,然后优化我们的旅行决策,以确定最佳路径遵循考虑到这种不确定性。
| |
− |
| |
− | {{See also|Monte Carlo methods in finance| Quasi-Monte Carlo methods in finance| Monte Carlo methods for option pricing| Stochastic modelling (insurance) | Stochastic asset model}}
| |
− |
| |
− | Probabilistic formulation of inverse problems leads to the definition of a probability distribution in the model space. This probability distribution combines prior information with new information obtained by measuring some observable parameters (data).
| |
− |
| |
− | 反问题的概率公式导致了模型空间中概率分布的定义。该概率分布将先前的信息与通过测量一些可观测的参数(数据)获得的新信息结合起来。
| |
− |
| |
− | As, in the general case, the theory linking data with model parameters is nonlinear, the posterior probability in the model space may not be easy to describe (it may be multimodal, some moments may not be defined, etc.).
| |
− |
| |
− | 因为,在一般情况下,连接数据和模型参数的理论是非线性的,模型空间中的后验概率可能不容易描述(它可能是多模态的,一些矩可能没有定义,等等。).
| |
| ===Law=== | | ===Law=== |
| | | |
第616行: |
第598行: |
| | | |
| 蒙特卡洛方法被用来评估一个拟议的方案的潜在价值,以帮助威斯康星州的女性请愿者成功地申请骚扰和家庭虐待限制令。提议帮助妇女成功地提出请愿,向她们提供更多的宣传,从而有可能减少强奸和人身攻击的风险。然而,还有许多变量无法完全估计,包括限制令的有效性,上访者的成功率,无论有没有主张,以及许多其他因素。这项研究通过改变这些变量进行了试验,得出了对整个计划成功程度的总体评估。 | | 蒙特卡洛方法被用来评估一个拟议的方案的潜在价值,以帮助威斯康星州的女性请愿者成功地申请骚扰和家庭虐待限制令。提议帮助妇女成功地提出请愿,向她们提供更多的宣传,从而有可能减少强奸和人身攻击的风险。然而,还有许多变量无法完全估计,包括限制令的有效性,上访者的成功率,无论有没有主张,以及许多其他因素。这项研究通过改变这些变量进行了试验,得出了对整个计划成功程度的总体评估。 |
− |
| |
− | When analyzing an inverse problem, obtaining a maximum likelihood model is usually not sufficient, as we normally also wish to have information on the resolution power of the data. In the general case we may have many model parameters, and an inspection of the marginal probability densities of interest may be impractical, or even useless. But it is possible to pseudorandomly generate a large collection of models according to the posterior probability distribution and to analyze and display the models in such a way that information on the relative likelihoods of model properties is conveyed to the spectator. This can be accomplished by means of an efficient Monte Carlo method, even in cases where no explicit formula for the a priori distribution is available.
| |
− |
| |
− | 当分析一个反问题时,获得一个最大似然模型通常是不够的,因为我们通常也希望有关于数据的分辨率的信息。在一般情况下,我们可能有许多模型参数,检查的边际概率密度的兴趣可能是不切实际的,甚至无用的。但是,根据《后验概率可以伪随机生成大量的模型集合,并以这样一种方式分析和显示模型,模型属性的相对可能性信息被传达给观众,这是可能的。这可以通过一个有效的蒙特卡罗方法安全管理系统来实现,即使在没有黎曼显式公式安全管理先验概率的情况下。
| |
− |
| |
− | The best-known importance sampling method, the Metropolis algorithm, can be generalized, and this gives a method that allows analysis of (possibly highly nonlinear) inverse problems with complex a priori information and data with an arbitrary noise distribution.
| |
− |
| |
− | 最著名的重要性抽样方法,Metropolis–Hastings 演算法,可以推广,这提供了一种方法,允许分析(可能是高度非线性)与复杂的先验信息和数据与任意噪声分布的反问题。
| |
| | | |
| ==Use in mathematics== | | ==Use in mathematics== |
第632行: |
第606行: |
| | | |
| 一般来说,蒙特卡罗方法在数学中通过产生合适的随机数(也见随机数产生)和观察符合某些性质的数字分数来解决各种问题。这种方法对于求解解析求解过于复杂的问题的数值解是有用的。蒙特卡罗方法最常用的应用是蒙地卡罗积分。 | | 一般来说,蒙特卡罗方法在数学中通过产生合适的随机数(也见随机数产生)和观察符合某些性质的数字分数来解决各种问题。这种方法对于求解解析求解过于复杂的问题的数值解是有用的。蒙特卡罗方法最常用的应用是蒙地卡罗积分。 |
− |
| |
− | Popular exposition of the Monte Carlo Method was conducted by McCracken. Method's general philosophy was discussed by Elishakoff and Grüne-Yanoff and Weirich.
| |
− |
| |
− | 由 McCracken 主持的蒙特卡罗方法博览会的普及展览。方法的一般哲学由 Elishakoff、 Grüne-Yanoff 和 weurich 讨论。
| |
| | | |
| === Integration === | | === Integration === |
第643行: |
第613行: |
| [[File:Monte-Carlo method (errors).png|thumb|Errors reduce by a factor of <math>\scriptstyle 1/\sqrt{N}</math>|链接=Special:FilePath/Monte-Carlo_method_(errors).png]] | | [[File:Monte-Carlo method (errors).png|thumb|Errors reduce by a factor of <math>\scriptstyle 1/\sqrt{N}</math>|链接=Special:FilePath/Monte-Carlo_method_(errors).png]] |
| | | |
− | Deterministic [[numerical integration]] algorithms work well in a small number of dimensions, but encounter two problems when the functions have many variables. First, the number of function evaluations needed increases rapidly with the number of dimensions. For example, if 10 evaluations provide adequate accuracy in one dimension, then [[googol|10<sup>100</sup>]] points are needed for 100 dimensions—far too many to be computed. This is called the [[curse of dimensionality]]. Second, the boundary of a multidimensional region may be very complicated, so it may not be feasible to reduce the problem to an [[iterated integral]].<ref name=Press>{{harvnb|Press|Teukolsky|Vetterling|Flannery|1996}}</ref> 100 [[dimension]]s is by no means unusual, since in many physical problems, a "dimension" is equivalent to a [[degrees of freedom (physics and chemistry)|degree of freedom]]. | + | Deterministic [[numerical integration]] algorithms work well in a small number of dimensions, but encounter two problems when the functions have many variables. First, the number of function evaluations needed increases rapidly with the number of dimensions. For example, if 10 evaluations provide adequate accuracy in one dimension, then [[googol|10<sup>100</sup>]] points are needed for 100 dimensions—far too many to be computed. This is called the [[curse of dimensionality]]. Second, the boundary of a multidimensional region may be very complicated, so it may not be feasible to reduce the problem to an [[iterated integral]].<ref name="Press">{{harvnb|Press|Teukolsky|Vetterling|Flannery|1996}}</ref> 100 [[dimension]]s is by no means unusual, since in many physical problems, a "dimension" is equivalent to a [[degrees of freedom (physics and chemistry)|degree of freedom]]. |
| | | |
| Deterministic numerical integration algorithms work well in a small number of dimensions, but encounter two problems when the functions have many variables. First, the number of function evaluations needed increases rapidly with the number of dimensions. For example, if 10 evaluations provide adequate accuracy in one dimension, then 10<sup>100</sup> points are needed for 100 dimensions—far too many to be computed. This is called the curse of dimensionality. Second, the boundary of a multidimensional region may be very complicated, so it may not be feasible to reduce the problem to an iterated integral. 100 dimensions is by no means unusual, since in many physical problems, a "dimension" is equivalent to a degree of freedom. | | Deterministic numerical integration algorithms work well in a small number of dimensions, but encounter two problems when the functions have many variables. First, the number of function evaluations needed increases rapidly with the number of dimensions. For example, if 10 evaluations provide adequate accuracy in one dimension, then 10<sup>100</sup> points are needed for 100 dimensions—far too many to be computed. This is called the curse of dimensionality. Second, the boundary of a multidimensional region may be very complicated, so it may not be feasible to reduce the problem to an iterated integral. 100 dimensions is by no means unusual, since in many physical problems, a "dimension" is equivalent to a degree of freedom. |
第649行: |
第619行: |
| 确定性数值积分算法在少数维上运行良好,但在函数具有多个变量时会遇到两个问题。首先,随着维数的增加,需要进行的功能评估的数量迅速增加。例如,如果10个评估在一个维度上提供了足够的精确度,那么100个维度需要10个 < sup > 100 点,这太多了以至于无法计算。这就是所谓的维数灾难。其次,多维区域的边界可能非常复杂,因此将问题简化为迭代积分可能是不可行的。100维绝对不是不寻常的,因为在许多物理问题中,一个“维度”等同于一个自由度。 | | 确定性数值积分算法在少数维上运行良好,但在函数具有多个变量时会遇到两个问题。首先,随着维数的增加,需要进行的功能评估的数量迅速增加。例如,如果10个评估在一个维度上提供了足够的精确度,那么100个维度需要10个 < sup > 100 点,这太多了以至于无法计算。这就是所谓的维数灾难。其次,多维区域的边界可能非常复杂,因此将问题简化为迭代积分可能是不可行的。100维绝对不是不寻常的,因为在许多物理问题中,一个“维度”等同于一个自由度。 |
| | | |
− | Monte Carlo methods provide a way out of this exponential increase in computation time. As long as the function in question is reasonably [[well-behaved]], it can be estimated by randomly selecting points in 100-dimensional space, and taking some kind of average of the function values at these points. By the [[central limit theorem]], this method displays <math>\scriptstyle 1/\sqrt{N}</math> convergence—i.e., quadrupling the number of sampled points halves the error, regardless of the number of dimensions.<ref name=Press/> | + | Monte Carlo methods provide a way out of this exponential increase in computation time. As long as the function in question is reasonably [[well-behaved]], it can be estimated by randomly selecting points in 100-dimensional space, and taking some kind of average of the function values at these points. By the [[central limit theorem]], this method displays <math>\scriptstyle 1/\sqrt{N}</math> convergence—i.e., quadrupling the number of sampled points halves the error, regardless of the number of dimensions.<ref name="Press" /> |
| | | |
| Monte Carlo methods provide a way out of this exponential increase in computation time. As long as the function in question is reasonably well-behaved, it can be estimated by randomly selecting points in 100-dimensional space, and taking some kind of average of the function values at these points. By the central limit theorem, this method displays <math>\scriptstyle 1/\sqrt{N}</math> convergence—i.e., quadrupling the number of sampled points halves the error, regardless of the number of dimensions. or the VEGAS algorithm. | | Monte Carlo methods provide a way out of this exponential increase in computation time. As long as the function in question is reasonably well-behaved, it can be estimated by randomly selecting points in 100-dimensional space, and taking some kind of average of the function values at these points. By the central limit theorem, this method displays <math>\scriptstyle 1/\sqrt{N}</math> convergence—i.e., quadrupling the number of sampled points halves the error, regardless of the number of dimensions. or the VEGAS algorithm. |
第674行: |
第644行: |
| | | |
| Another powerful and very popular application for random numbers in numerical simulation is in [[Optimization (mathematics)|numerical optimization]]. The problem is to minimize (or maximize) functions of some vector that often has many dimensions. Many problems can be phrased in this way: for example, a [[computer chess]] program could be seen as trying to find the set of, say, 10 moves that produces the best evaluation function at the end. In the [[traveling salesman problem]] the goal is to minimize distance traveled. There are also applications to engineering design, such as [[multidisciplinary design optimization]]. It has been applied with quasi-one-dimensional models to solve particle dynamics problems by efficiently exploring large configuration space. Reference<ref>Spall, J. C. (2003), ''Introduction to Stochastic Search and Optimization: Estimation, Simulation, and Control'', Wiley, Hoboken, NJ. http://www.jhuapl.edu/ISSO</ref> is a comprehensive review of many issues related to simulation and optimization. | | Another powerful and very popular application for random numbers in numerical simulation is in [[Optimization (mathematics)|numerical optimization]]. The problem is to minimize (or maximize) functions of some vector that often has many dimensions. Many problems can be phrased in this way: for example, a [[computer chess]] program could be seen as trying to find the set of, say, 10 moves that produces the best evaluation function at the end. In the [[traveling salesman problem]] the goal is to minimize distance traveled. There are also applications to engineering design, such as [[multidisciplinary design optimization]]. It has been applied with quasi-one-dimensional models to solve particle dynamics problems by efficiently exploring large configuration space. Reference<ref>Spall, J. C. (2003), ''Introduction to Stochastic Search and Optimization: Estimation, Simulation, and Control'', Wiley, Hoboken, NJ. http://www.jhuapl.edu/ISSO</ref> is a comprehensive review of many issues related to simulation and optimization. |
| + | |
| + | Another powerful and very popular application for random numbers in numerical simulation is in numerical optimization. The problem is to minimize (or maximize) functions of some vector that often has many dimensions. Many problems can be phrased in this way: for example, a computer chess program could be seen as trying to find the set of, say, 10 moves that produces the best evaluation function at the end. In the traveling salesman problem the goal is to minimize distance traveled. There are also applications to engineering design, such as multidisciplinary design optimization. It has been applied with quasi-one-dimensional models to solve particle dynamics problems by efficiently exploring large configuration space. Reference is a comprehensive review of many issues related to simulation and optimization. |
| + | |
| + | 另一个强大的和非常流行的应用随机数在数值模拟是在数值优化。问题在于如何最小化(或最大化)某些向量的函数,这些向量通常具有多个维度。许多问题可以这样表述: 例如,一个计算机国际象棋程序可以被视为试图找到一组,比如说,10步棋,最终产生最好的评价函数。在旅行商问题中,目标是使旅行距离最小。在工程设计中也有一些应用,如多学科设计优化。它已被应用于准一维模型,以解决粒子动力学问题,有效地探索大型位形空间。参考文献是对许多与模拟和优化有关的问题的全面回顾。 |
| | | |
| The [[traveling salesman problem]] is what is called a conventional optimization problem. That is, all the facts (distances between each destination point) needed to determine the optimal path to follow are known with certainty and the goal is to run through the possible travel choices to come up with the one with the lowest total distance. However, let's assume that instead of wanting to minimize the total distance traveled to visit each desired destination, we wanted to minimize the total time needed to reach each destination. This goes beyond conventional optimization since travel time is inherently uncertain (traffic jams, time of day, etc.). As a result, to determine our optimal path we would want to use simulation - optimization to first understand the range of potential times it could take to go from one point to another (represented by a probability distribution in this case rather than a specific distance) and then optimize our travel decisions to identify the best path to follow taking that uncertainty into account. | | The [[traveling salesman problem]] is what is called a conventional optimization problem. That is, all the facts (distances between each destination point) needed to determine the optimal path to follow are known with certainty and the goal is to run through the possible travel choices to come up with the one with the lowest total distance. However, let's assume that instead of wanting to minimize the total distance traveled to visit each desired destination, we wanted to minimize the total time needed to reach each destination. This goes beyond conventional optimization since travel time is inherently uncertain (traffic jams, time of day, etc.). As a result, to determine our optimal path we would want to use simulation - optimization to first understand the range of potential times it could take to go from one point to another (represented by a probability distribution in this case rather than a specific distance) and then optimize our travel decisions to identify the best path to follow taking that uncertainty into account. |
| + | |
| + | The traveling salesman problem is what is called a conventional optimization problem. That is, all the facts (distances between each destination point) needed to determine the optimal path to follow are known with certainty and the goal is to run through the possible travel choices to come up with the one with the lowest total distance. However, let's assume that instead of wanting to minimize the total distance traveled to visit each desired destination, we wanted to minimize the total time needed to reach each destination. This goes beyond conventional optimization since travel time is inherently uncertain (traffic jams, time of day, etc.). As a result, to determine our optimal path we would want to use simulation - optimization to first understand the range of potential times it could take to go from one point to another (represented by a probability distribution in this case rather than a specific distance) and then optimize our travel decisions to identify the best path to follow taking that uncertainty into account. |
| + | |
| + | 旅行推销员问题被称为传统的最佳化问题问题。也就是说,确定最佳路径所需的所有事实(每个目的地之间的距离)都是确定无疑的,目标是通过可能的旅行选择得出总距离最小的路径。然而,让我们假设,我们不想最小化访问每个想要的目的地所需的总距离,而是想最小化到达每个目的地所需的总时间。这超越了传统的优化,因为旅行时间是固有的不确定性(交通堵塞,一天的时间,等)。因此,为了确定我们的最佳路径,我们需要使用模拟优化来首先了解从一个点到另一个点可能需要的时间范围(在这个例子中用概率分布代表,而不是特定的距离) ,然后优化我们的旅行决策,以确定最佳路径遵循考虑到这种不确定性。 |
| ===Inverse problems=== | | ===Inverse problems=== |
| | | |
| Probabilistic formulation of [[inverse problem]]s leads to the definition of a [[probability distribution]] in the model space. This probability distribution combines [[prior probability|prior]] information with new information obtained by measuring some observable parameters (data). | | Probabilistic formulation of [[inverse problem]]s leads to the definition of a [[probability distribution]] in the model space. This probability distribution combines [[prior probability|prior]] information with new information obtained by measuring some observable parameters (data). |
| + | |
| + | Probabilistic formulation of inverse problems leads to the definition of a probability distribution in the model space. This probability distribution combines prior information with new information obtained by measuring some observable parameters (data). |
| + | |
| + | 反问题的概率公式导致了模型空间中概率分布的定义。该概率分布将先前的信息与通过测量一些可观测的参数(数据)获得的新信息结合起来。 |
| + | |
| + | As, in the general case, the theory linking data with model parameters is nonlinear, the posterior probability in the model space may not be easy to describe (it may be multimodal, some moments may not be defined, etc.). |
| | | |
| As, in the general case, the theory linking data with model parameters is nonlinear, the posterior probability in the model space may not be easy to describe (it may be multimodal, some moments may not be defined, etc.). | | As, in the general case, the theory linking data with model parameters is nonlinear, the posterior probability in the model space may not be easy to describe (it may be multimodal, some moments may not be defined, etc.). |
| + | |
| + | 因为,在一般情况下,连接数据和模型参数的理论是非线性的,模型空间中的后验概率可能不容易描述(它可能是多模态的,一些矩可能没有定义,等等。). |
| | | |
| When analyzing an inverse problem, obtaining a maximum likelihood model is usually not sufficient, as we normally also wish to have information on the resolution power of the data. In the general case we may have many model parameters, and an inspection of the [[marginal probability]] densities of interest may be impractical, or even useless. But it is possible to pseudorandomly generate a large collection of models according to the [[posterior probability distribution]] and to analyze and display the models in such a way that information on the relative likelihoods of model properties is conveyed to the spectator. This can be accomplished by means of an efficient Monte Carlo method, even in cases where no explicit formula for the ''a priori'' distribution is available. | | When analyzing an inverse problem, obtaining a maximum likelihood model is usually not sufficient, as we normally also wish to have information on the resolution power of the data. In the general case we may have many model parameters, and an inspection of the [[marginal probability]] densities of interest may be impractical, or even useless. But it is possible to pseudorandomly generate a large collection of models according to the [[posterior probability distribution]] and to analyze and display the models in such a way that information on the relative likelihoods of model properties is conveyed to the spectator. This can be accomplished by means of an efficient Monte Carlo method, even in cases where no explicit formula for the ''a priori'' distribution is available. |
| + | |
| + | When analyzing an inverse problem, obtaining a maximum likelihood model is usually not sufficient, as we normally also wish to have information on the resolution power of the data. In the general case we may have many model parameters, and an inspection of the marginal probability densities of interest may be impractical, or even useless. But it is possible to pseudorandomly generate a large collection of models according to the posterior probability distribution and to analyze and display the models in such a way that information on the relative likelihoods of model properties is conveyed to the spectator. This can be accomplished by means of an efficient Monte Carlo method, even in cases where no explicit formula for the a priori distribution is available. |
| + | |
| + | 当分析一个反问题时,获得一个最大似然模型通常是不够的,因为我们通常也希望有关于数据的分辨率的信息。在一般情况下,我们可能有许多模型参数,检查的边际概率密度的兴趣可能是不切实际的,甚至无用的。但是,根据《后验概率可以伪随机生成大量的模型集合,并以这样一种方式分析和显示模型,模型属性的相对可能性信息被传达给观众,这是可能的。这可以通过一个有效的蒙特卡罗方法安全管理系统来实现,即使在没有黎曼显式公式安全管理先验概率的情况下。 |
| | | |
| The best-known importance sampling method, the Metropolis algorithm, can be generalized, and this gives a method that allows analysis of (possibly highly nonlinear) inverse problems with complex ''a priori'' information and data with an arbitrary noise distribution.<ref>{{harvnb|Mosegaard|Tarantola|1995}}</ref><ref>{{harvnb|Tarantola|2005}}</ref> | | The best-known importance sampling method, the Metropolis algorithm, can be generalized, and this gives a method that allows analysis of (possibly highly nonlinear) inverse problems with complex ''a priori'' information and data with an arbitrary noise distribution.<ref>{{harvnb|Mosegaard|Tarantola|1995}}</ref><ref>{{harvnb|Tarantola|2005}}</ref> |
| + | |
| + | The best-known importance sampling method, the Metropolis algorithm, can be generalized, and this gives a method that allows analysis of (possibly highly nonlinear) inverse problems with complex a priori information and data with an arbitrary noise distribution. |
| + | |
| + | 最著名的重要性抽样方法,Metropolis–Hastings 演算法,可以推广,这提供了一种方法,允许分析(可能是高度非线性)与复杂的先验信息和数据与任意噪声分布的反问题。 |
| ===Philosophy=== | | ===Philosophy=== |
| | | |
| Popular exposition of the Monte Carlo Method was conducted by McCracken<ref>McCracken, D. D., (1955) The Monte Carlo Method, Scientific American, 192(5), pp. 90-97</ref>. Method's general philosophy was discussed by [[Elishakoff]]<ref>Elishakoff, I., (2003) Notes on Philosophy of the Monte Carlo Method, International Applied Mechanics, 39(7), pp.753-762</ref> and Grüne-Yanoff and Weirich<ref>Grüne-Yanoff, T., & Weirich, P. (2010). The philosophy and epistemology of simulation: A review, Simulation & Gaming, 41(1), pp. 20-50</ref>. | | Popular exposition of the Monte Carlo Method was conducted by McCracken<ref>McCracken, D. D., (1955) The Monte Carlo Method, Scientific American, 192(5), pp. 90-97</ref>. Method's general philosophy was discussed by [[Elishakoff]]<ref>Elishakoff, I., (2003) Notes on Philosophy of the Monte Carlo Method, International Applied Mechanics, 39(7), pp.753-762</ref> and Grüne-Yanoff and Weirich<ref>Grüne-Yanoff, T., & Weirich, P. (2010). The philosophy and epistemology of simulation: A review, Simulation & Gaming, 41(1), pp. 20-50</ref>. |
| + | |
| + | Popular exposition of the Monte Carlo Method was conducted by McCracken. Method's general philosophy was discussed by Elishakoff and Grüne-Yanoff and Weirich. |
| + | |
| + | 由 McCracken 主持的蒙特卡罗方法博览会的普及展览。方法的一般哲学由 Elishakoff、 Grüne-Yanoff 和 weurich 讨论。 |
| == See also == | | == See also == |
| | | |