Ole Peters 
Home  Presentations  Publications  Research 
======================================= Nontechnical ======================================= [9] AIP Staff, Exploring gambles reveals foundational difficulty behind economic theory (and a solution!). AIP publishing (2016). [8] M. Buchanan, Gamble with time. Nature Phys. 9, 3 (2013). doi: 10.1038/nphys2520 [7] O. Peters Time, for a change. Gresham College London (2012). [6] Towers Watson The irreversibility of time (2012). [5] M. Mauboussin Shaking the foundation (2012). [4] R. Bookstaber A Crack in the foundation (2011). [3] O. Peters Time and chance. TEDx Goodenough London (2011). [2] O. Peters On time and risk. Santa Fe Institute Bulletin 24, 1, 3641 (2009). [1] O. Peters Fragments of Symmetry. ILPI Press (2007). Excerpt 
======================================= Recently submitted ======================================= [25] Y. Berman, O. Peters and A. Adamou Far from equilibrium: Wealth reallocation in the United States. arXiv:1605.05631 (2016). We fit observed distributions of wealth  how many people have how much  to a model of noisy exponential growth and reallocation. Everyone's wealth is assumed to follow geometric Brownian motion (GBM), enhanced by a term that collects from everyone at a rate in proportion to his wealth and redistributes the collected amount evenly across the population. We use US data from 1917 until 2012. Firstly, we find that the bestfit reallocation rate has been negative since the 1980s, meaning everyone pays the same dollar amount and the collected amount is redistributed in proportion to his wealth, a flow of wealth from poor to rich. This came as a big surprise: GBM on its own generates indefinitely increasing inequality, and one would expect this extreme model to require a correction that reduces the default tendency of increasing inequality. But that's not the case: recent conditions are such that GBM needs to be corrected to speed up the increase in wealth inequality if we want to describe the observations. Secondly, our model has an equilibrium (ergodic) distribution if reallocation is positive, and it has no such distribution if reallocation is negative. Fitting the reallocation rate thus asks the system: are you ergodic? And the answer is no. With current bestfit parameters the model is nonergodic, and throughout history whenever the parameters implied the existence of an ergodic distribution, their values implied equilibration times on the order of decades to centuries, so that the equilibrium state had no practical relevance. ======================================= [24] O. Peters and A. Adamou Rational insurance with linear utility and perfect information. arXiv:1507.04655 (2015). Insurance buyers pay a known fee today to reduce an uncertain future loss. Sellers accept the fee and promise to cover the uncertain loss. For an insurance contract to be signed voluntarily both parties to the deal must perceive it as beneficial to them. If the deal is judged by computing the expectation value of profit, a puzzle arises: the linearity of the expectation value implies that the deal can only be beneficial to one party. From this perspective insurance deals are signed because either people have different information (I know something you don't), or different attitudes towards risk (utility functions), or they are not rational. None of these explanations is satisfying, as e.g. Arrow pointed out at a recent Santa Fe Institute meeting. We resolve the puzzle by rejecting the underlying axiom that the expectation value of profit should be used to judge the desirability of the contract. Expectation values are averages over ensembles, but an individual signing an insurance contract is not an ensemble. Individuals are not ensembles, but they do live across time. Averaging over time is therefore the appropriate way of removing randomness from the model, and the timeaverage growth rate is the object of interest. Computing this object, the puzzle is resolved: a range of insurance fees exists where the contract increases the timeaverage growth rates of the wealth of both parties. We generalize: the fundamental reason for trading any product or service is a winwin situation. Both trading partners are better off as a result. This is possible because wealth must be modeled as a farfrom equilibrium growthprocess, such as geometric Brownian motion. In such processes the linear intuition associated with expectation values and equilibrium processes is misleading. ======================================= [23] O. Peters and A. Adamou The evolutionary advantage of cooperation. SFI working paper #1507028 (2015). arXiv:1506.03414 (2015). Social structure is the difference between a collection of individuals and a social system. We observe in the world around us plenty of structure, such as families, firms, nation states. Many different explanations for specific cases have been put forward. Here we develop a null model that favors the formation of structure in very general settings. We observe that cooperation  the pooling and redistribution of resources  confers an evolutionary advantage on cooperators over noncooperators under multiplicative dynamics. This predicts the emergence of cooperationpromoting structure, such as governments that tax and redistribute some proportion of income. The evolutionary advantage is only visible if the performance of an entity is considered over time; it is invisible if the performance of the expectation value (an average over parallel universes) is considered. Mainstream economics focuses on the latter, thereby overlooking the most fundamental benefits of cooperation. ======================================= [22] O. Peters Menger 1934 revisited. arXiv:1110.1578 (2011). Karl Menger argued in 1934 that only bounded utility functions are permissible in the formalism of expected utility theory. I point out that his mathematical argument does not allow this conclusion. This is important from the dynamic perspective we are developing. As it turns out, bounded utility functions, when translated into ergodic observables, lead to finitetime singularities in the wealth process. To obtain a workable realistic formalism utility functions must not be bounded. ======================================= [21] O. Peters and A. Adamou Stochastic Market Efficiency. SFI working paper #1306022 (2013). arXiv:1101.4548 (2011). We solve the equity premium puzzle by exploiting a fundamental constraint on the stochastic properties of stock markets. In a world of one risky asset (shares) and one riskfree asset (bonds), if investors do better in the long run by borrowing money (short bonds) and buying the shares, an instability will arise. Since any investor is better of borrowing money and buying shares, who should lend? Who should sell the shares? Increased borrowing will lead to inflated prices and higher interest rates, eventually destroying the advantage of a longshares, shortbonds portfolio. We predict that it will be longterm optimal to simply hold shares, without leverage. Knowing that longterm performance (unlike performance of expectation values) is negatively affected by fluctuations, we arrive at a quantitative relationship between volatility and the excess growth rate of the expectation value of share prices compared to the growth rate of bond prices. This difference is usually called the equity premium, and our treatment is a solution of the equity premium puzzle. ======================================= [20] O. Peters and G. Pruessner Tuning and order parameter in the SOC ensemble. arXiv:0912.2305v1 (2009). One theory of how SOC works posits that SOC systems are selftuned models with absorbingstate phase transitions. The idea is that the tuning and order parameters are coupled linearly, with an increase in the order parameter leading to a decrease, say, in the tuning parameter. Add to this a slow external forcing that drives the tuning parameter up, and you obtain a system that resides close to criticality. We look numerically for evidence of this mechanism. To our surprise, measurements of the order parameter conditioned on a given value of the tuning parameter do not reveal a sudden phasetransitionlike pickup of the order parameter. Rather, the relationship looks smooth and linear. ======================================= ======================================= Peerreviewed ======================================= [19] O. Peters and M. GellMann Evaluating gambles using dynamics. Chaos 26, 023103 (2016). arXiv:140585 doi:10.1063/1.4940236 Mainstream economics uses mathematical models that resemble economic processes. It insists that expectation values of random variables in these models carry meaningful information. There is no immediate reason to believe this because expectation values need not have any physical meaning. Under what conditions, then, could the formalism work? We point out that some variables can be transformed so that the expectation value of the newly transformed variable is also its time average. The expectation value is then meaningful as an indication of what happens in the long run. Expectation values of variables that do not have this ergodic property should be treated with skepsis. ======================================= [18] O. Peters and W. Klein Ergodicity breaking in geometric Brownian motion. Phys. Rev. Lett. 110, 100603 (2013). arXiv:1209.4517 . doi:10.1103/PhysRevLett.110.100603 In [15] I demonstrated that leverage can be optimized by considering timeaverage growth rates rather than the growth rate of expectation values. The fact that the two are different (and nonzero) is a specific form of ergodicity breaking. In [15] time average growth rates for a single system were considered. Here we prove that the time average growth rate is the same for finite ensembles of noninteracting geometric Brownian motions (GBM) as for an individual GBM. The paper has quantitative statements and nice illustrations of the effects of finiteness of averaging times and finiteness of ensembles. ======================================= [17] O. Peters, K. Christensen and D. Neelin Rainfall and dragonkings. Eur. Phys. J. Special Topics 205, 147158 (2012). doi:10.1140/epjst/e201215675 We were curious about two things: the relationship between columnintegrated water vapor and precipitation under extreme conditions (e.g. hurricanes) and the distribution event sizes not measured as a total released water column at a specific location (i.e. a timeintegrated rain rate at some point in space) but as an intensity integrated over space. Hurricanes really seem to be more than just a big rain shower, which isn't too surprising given the level of dynamical organization in a hurricane. ======================================= [16] O. Peters The time resolution of the St Petersburg paradox. Phil. Trans. R. Soc. A 369, 1956, 49134931 (2011). arXiv:1011.4404v2 doi:10.1098/rsta.2011.0065 In [15] I showed that under the geometric Brownian motion model leverage can be optimized by maximizing the timeaverage growth rate of wealth. It cannot be optimized by optimizing the growth rate of the expectation value of wealth. Since classical decision theory is based on expectation values it has a problem here. While writing [15] I became curious how classical decision theory treated this most basic issue  how to weigh an increase in risk against an increase in expected gain? The answer is expected utility theory, a framework that phrases the problem and its solution in a very complicated way from my perspective. A poorly constrained element of human psychology, the utility function, is the central part of the solution. The treatment is circular: general risk preferences are encoded in a utility function, and this function is then used to compute the specific risk preferences in the problem under consideration  we have to input the answer to all questions of riskreward balance type, in order to obtain the answer to one question of this type. Why then have a formalism at all? I became interested in the history of the development of this framework. How did we end up with such a complicated formalism? The origin of expected utility theory is Daniel Bernoulli's 1738 treatment of the St Petersburg paradox. I wondered whether the paradox could be resolved in the same way as the leverage problem. If I compute the expected exponential growth rate, instead of the expected additive rate of change, for the St Petersburg lottery, would I resolve the paradox? The answer, surprisingly, is yes. Expected utility theory was only developed because it predates any considerations of ergodicity. In 1738 no one questioned whether expectation values and time averages are necessarily the same. This question only arose when physics adopted stochstic reasoning some 200 years after economics, namely in the late 19th century. ======================================= [15] O. Peters Optimal leverage from nonergodicity. Quant. Fin. 11, 11, 15931602 (2011). arXiv:0902.2965v2 doi:/10.1080/14697688.2010.513338 I show that under the geometric Brownian motion (GBM) model leverage can be optimized by maximizing the timeaverage growth rate of wealth. It cannot be optimized by optimizing the growth rate of the expectation value of wealth. This is the case because wealth does not have the ergodic property that its time average converges to its expectation value. ======================================= [14] O. Peters, A. Deluca, A. Corral, J. D. Neelin and C. E. Holloway Universality of rain event size distributions. J. Stat. Mech. P11030 (2010). arXiv:1010.4201 doi:/10.1088/17425468/2010/11/P11030 We were curious whether the avalanche exponents for rainfall, first reported in [1], would look similar when measured in different locations. Overall, this is the case. We measured exponents all over the world, using data sets from the Atmospheric Radiation Measurement (ARM) network. Exponents are identical within statistical accuracy, although only the measurement technique can make a difference (optical rain gauge vs. micro rain radar). Cutoffs in the distributions of rain event sizes are as one might expect based on local climatology  for example, tropical locations support larger events. ======================================= [13] O. Peters and M. Girvan, Universality under conditions of selftuning. J. Stat. Phys. 141, 1, 5359 (2010). arXiv:0902.1956v2 doi:10.1007/s1095501000390 Erratum Our work in [5] showed that a mechansim for SOC that keys off the order parameter only does not automatically achieve universality. It is possible for a system to tune itself to its critical point in this way, but one would not observe universal finitesize scaling. We asked ourselves what sort of a mechanism might achieve universal finitesize scaling, and we came up with the following answer: couple the tuning parameter to a quantity that diverges at criticality. Specifically, if we introduce a coupling beteween the temperature and the susceptibility of an Ising model so that tuning parameter maximizes the susceptibility, then, in a finite system, the reduced temperature will approach the critical point with the same scaling behavior as the maximum of the susceptibility, t scales like L^{1/vu}. This in turn sets the finitesize scaling exponents of all other observables to their universal values. ======================================= [12] D. Neelin, O. Peters, J. W.B. Lin, K. Hales and C. Holloway in "Stochastic Physics and Climate Modeling", edited by T. Palmer and P. Williams. Cambridge University Press (2010), Chap. 16. Rethinking Convective QuasiEquilibrium: Observational Constraints for Stochastic Convective Schemes in Climate Models. This is a republication of [8] as a book chapter. ======================================= [11] O. Peters and D. Neelin Atmospheric convection as a continuous phase transitions: further evidence. Int. J. Mod. Phys. B 23, 2829, 54535465 (2009). doi: 10.1142/S0217979209063778 We discuss how a separation of scales can lead to scalefreedom, try out a method inspired by equilibrium critical phenomena to find the critical point in a presumed phase transition for atmospheric convection, and take a first look at hurricanes  what was the relationship between column water vapor and precipitation during hurricane Katrina? ======================================= [10] O. Peters, D. Neelin and S. Nesbitt Mesoscale Convective Systems and Critical Clusters. J. Atmos. Sci. 66, 9, 29132924 (2009). doi: 10.1175/2008JAS2761.1 This paper explores the spatial structure of convection clusters. The simplest model for clustering is percolation theory, and we wonder in how far this null model predicts the statistics of convection clusters. The most exciting finding in this paper is a prediction from gradient percolation: take a 2d blank lattice, color in a fraction p of randomly chosen cells in each column, and let p decrease linearly from 1 in the leftmost colum to 0 in the rightmost column. The boundary of the largest cluster of connected coloredin sites is a fractal with dimension 4/3. It is a wellknown (and previously unexplained) fact that cloud boundaries are fractals with dimension 4/3 (first reported by Shaun Lovejoy in 1982), but before our paper it had not been pointed out that this is a prediction of the null model  percolation theory. We also find that near the transition to convection the size distribution of convective clusters looks a lot like the size distribution for critical percolation. ======================================= [9] D. Neelin, O. Peters and K. Hales The Transition to Strong Convection. J. Atmos. Sci. 66, 8, 23672384 (2009). doi: 10.1175/2009JAS2962.1 This paper looks in detail at the statistical properties of convection that we found in [6]. Analyses are conditioned on tropospheric and seasurface temperature, which sharpens all features significantly. We identify convective transition points for different tropospheric temperatures, which is interesting in the context of climate change because we can now directly observe an important robust property of convection under a change of temperature. ======================================= [8] D. Neelin, O. Peters, J. W.B. Lin, K. Hales and C. Holloway Rethinking Convective QuasiEquilibrium: Observational Constraints for Stochastic Convective Schemes in Climate Models. Phil. Trans. R. Soc. A 366, 25812604 (2008). doi: 10.1098/rsta.2008.0056 We explore how the idea to model the onset of convection as a selftuning continuous phase transition can be integrated into existing climate model components for convection. ======================================= [7] G. Pruessner and O. Peters, Reply to "Comment on 'SelfOrganized Criticality and Absorbing States: Lessons from the Ising Model'". Phys. Rev. E 77, 048102 (2008). doi: 10.1103/PhysRevE.77.048102 We clarify comments that were made following the publication of [5]. ======================================= [6] O. Peters and D. Neelin, Critical Phenomena in Atmospheric Precipitation. Nature Phys. 2, 393396 (2006). doi: 10.1038/Nphys314 In [1] we reported an observation of a scalefree eventsize distribution for rainfall. We speculated that a selforganized critical mathematical model of some sort would capture this aspect. Unfortunately, this seemed like a deadend at the time. We knew we could use a selforganized critical model, but what's the point if it only reproduces an observation we have already seen. For this paper [6] we made a list of properties we would expect to observe if convection really was well described by a selforganized critical model similar to the particle models (BTW, Manna, Oslo) that are related to underlying absorbingstate phase transitions. Specifically, we predicted that there should be an underlying transition that could be observed by conditioning observations on some measure of instability. We chose columnintegrated water vapor for this measure and found that the rain rate can then be interpreted as an order parameter for a phase transition. At a critical value of water vapor the system becomes unstable and convection suddenly sets in, as reflected in a strong pickup of the rain rate. This pickup looks as one would expect for a continuous phase transition. ======================================= [5] G. Pruessner and O. Peters, SelfOrganized Criticality and Absorbing States: Lessons from the Ising Model. Phys. Rev. E 73, 025106(R) (2006). doi: 10.1103/PhysRevE.73.025106 We investigate a mechanism for obtaining SOC. This mechanism is phrased in terms of absorbingstate phase transitions but is formulated sufficiently generally that it should apply to any continuous phase transition. We try it out in the 2d Ising model because the model is fully solved and we can compute analytically any quantities of interest. We find that the mechanism is capable of generating criticality but not universality. There is strong evidence for universal behavior under SOC conditions, even for SOC models to have identical finitesize scaling to the corresponding underlying absorbingstate phase transitions. We conclude, therefore, that something is missing from the absordingstate mechanism for SOC. ======================================= [4] O. Peters and K. Christensen, Rain Viewed as Relaxational Events. J. Hydrol. 328, 4655 (2006). doi: 10.1016/j.jhydrol.2005.11.045 We summarize the SOCrelated statements about rainfall of [1] and [2] and look at the effect of sensititvity thresholds on the apparent fractal dimension of rain durations. Different studies came to different conclusions about the value of this dimension, and the apparent disagreement may be the result of different measurement techniques and corresponding sensitivity thresholds. ======================================= [3] K. Christensen, N. Moloney, O. Peters and G. Pruessner, Avalanche Behavior in an AbsorbingState Oslo Model. Phys. Rev. E. 70, 067101(R) (2004). doi: 10.1103/PhysRevE.70.067101 We ask whether the critical exponents, observed numerically through finitesize scaling, are identical in the Oslo model and in its absorbingstate counterpart, i.e. an Oslo model with periodic boundary conditions and particle conservation. The answer is a resounding yes. Even the critical particle densities are identical to 4 significant figures. ======================================= [2] O. Peters and K. Christensen, Rain: Relaxations in the Sky. Phys. Rev. E. 66, 036120 (2002). doi: 10.1103/PhysRevE.66.036120 This paper is an extended version of [1]. ======================================= [1] O. Peters, C. Hertlein, and K. Christensen, A complexity view of rainfall. Phys. Rev. Lett. 88, 018701 (2002). doi: 10.1103/PhysRevLett.88.018701 We define event sizes as rain rates integrated over a period of consecutive nonzero measurements at some location in space. Data from a highly sensitive micro rain radar are analyzed, and we find a scalefree distribution of rain events. The basic idea is the following: the atmosphere is a slowly driven system (heated on the ground by the sun and moistened through evaporation from the oceans). Over time heat at low levels, radiative cooling aloft, all combined with increasing moisture leads to instability. The result is convection, in its most dramatic form leading to cumulonimbus cloud towers and thunderstorms. Convection is a fast process, compared to the speed of the driving mechanism. This places the system as a whole close to the onset of convection and rainfall, a fact famously pointed out by Arakawa and Schubert in 1974. If the transition from a quiescent atmosphere to a convectively active atmosphere is well described as a continuous phase transition then critical properties of this transition may be observed in the statistics of rainfall. These would include a scalefree (powerlaw) distribution of event sizes, and that's what we observe. ======================================= 

