Under Review Since : 2020-02-20

In unreplicated two-way factorial designs, it is typical to assume no interaction between two factors. However, violations of this additivity assumption have often been found in applications, and tests for non-additivity have been a recurring topic since Tukey's one-degree of freedom test (Tukey, 1949). In the context of randomized complete block designs, recent work by Franck et al. (2013) is based on an intuitive model with "hidden additivity," a type of non-additivity where unobserved groups of blocks exist such that treatment and block effects are additive within groups, but treatment effects may be different across groups. Their proposed test statistic for detecting hidden additivity is called the "all-conguration maximum interaction F-statistic" (ACMIF). The computations of the ACMIF also result in a clustering method for blocks related to the k-means procedure. When hidden additivity is detected, a new method is proposed here for condence intervals of contrasts within groups that takes into account the error due to clustering by forming the union of standard intervals over a subset of likely congurations.

Under Review Since : 2020-02-14

An important question in economics is how people choose when facing uncertainty in the timing of rewards. In this paper we study preferences over time lotteries, in which the payment amount is certain but the payment time is uncertain. In expected discounted utility (EDU) theory decision makers must be risk-seeking over time lotteries. Here we explore growth-optimality, a normative model consistent with standard axioms of choice, in which decision makers maximise the growth rate of their wealth. Growth-optimality is consistent with both risk-seeking and risk-neutral behaviour in time lotteries, depending on how growth rates are computed. We discuss two approaches to compute a growth rate: the ensemble approach and the time approach. Revisiting existing experimental evidence on risk preferences in time lotteries, we find that the time approach accords better with the evidence than the ensemble approach. Surprisingly, in contrast to the EDU prediction, the higher the ensemble-average growth rate of a time lottery is, the less attractive it becomes compared to a sure alternative. Decision makers thus may not consider the ensemble-average growth rate as a relevant criterion for their choices. Instead, the time-average growth rate may be a better criterion for decision-making.

Under Review Since : 2020-02-14

We collected marathon performance data from a systematic sample of elite and sub-elite athletes over the period 2015 to 2019, then searched the internet for publicly-available photographs of these performances, identifying whether the Nike Vaporfly shoes were worn or not in each performance. Controlling for athlete ability and race difficulty, we estimated the effect on marathon times of wearing the Vaporfly shoes. Assuming that the effect of Vaporfly shoes is additive, we estimate that the Vaporfly shoes improve men's times between 2.1 and 4.1 minutes, while they improve women's times between 1.2 and 4.0 minutes. Assuming that the effect of Vaporfly shoes is multiplicative, we estimate that they improve men's times between 1.5 and 2.9 percent, women's performances between 0.8 and 2.4 percent. The improvements are in comparison to the shoe the athlete was wearing before switching to Vaporfly shoes, and represents an expected improvement rather than a guaranteed improvement.

Under Review Since : 2020-02-14

The outbreak of a novel Coronavirus we are facing is poised to become a global pandemic if current approaches to stemming its spread prove to be insufficient. While we can't yet say what the ultimate impact of this event will be, this crisis and governments' responses to it reveal vulnerabilities and fragilities in the structure of our global socioeconomic milieux that will continue to produce cascading crises regardless of whether or not we are successful in preventing devastation from this particular pathogen. Here we discuss the implications and some strategic considerations.

Published Date : 2020-02-06

Contrary to Ole Peters' claims in a recent Nature Physics article [1], the "Copenhagen Experiment" did not falsify Expected Utility Theory (EUT) and corroborate Ergodicity Econonomics. The dynamic version of of EUT, multi-period EUT, predicts the same change in risk aversion that EE predicts when the dynamics are changed from multiplicative to additive.

Published Date : 2020-02-03

The book investigates the misapplication of conventional statistical techniques to fat tailed distributions and looks for remedies, when possible.

Switching from thin tailed to fat tailed distributions requires more than "changing the color of the dress". Traditional asymptotics deal mainly with either n=1 or n=∞, and the real world is in between, under of the "laws of the medium numbers" --which vary widely across specific distributions. Both the law of large numbers and the generalized central limit mechanisms operate in highly idiosyncratic ways outside the standard Gaussian or Levy-Stable basins of convergence.

A few examples:

+ The sample mean is rarely in line with the population mean, with effect on "naive empiricism", but can be sometimes be estimated via parametric methods.

+ The "empirical distribution" is rarely empirical.

+ Parameter uncertainty has compounding effects on statistical metrics.

+ Dimension reduction (principal components) fails.

+ Inequality estimators (GINI or quantile contributions) are not additive and produce wrong results.

+ Many "biases" found in psychology become entirely rational under more sophisticated probability distributions

+ Most of the failures of financial economics, econometrics, and behavioral economics can be attributed to using the wrong distributions.

This book, the first volume of the Technical Incerto, weaves a narrative around published journal articles.

Under Review Since : 2020-01-23

A fundamental problem in statistics and machine learning is that of using observed data to predict future observations. This is particularly challenging for model-based approaches because often the goal is to carry out this prediction with no or minimal model assumptions. For example, the inferential model (IM) approach is attractive because it has certain validity guarantees, but requires specification of a parametric model. Here we show that a new perspective on a recently developed generalized IM approach can be applied to construct an IM for prediction that satisfies the desirable validity guarantees without specification of a model. One important special case of this approach corresponds to the powerful conformal prediction framework and, consequently, the desirable properties of conformal prediction follow immediately from the general IM validity theory. Several numerical examples are presented to illustrate the theory and highlight the method's performance and flexibility.

Published Date : 2020-01-03

In this note, I analyze the code and the data generated by M. Fodje's (2013) simulation programs "epr-simple" and "epr-clocked". They are written in Python were published on Github only, initially without any documentation at all of how they worked. Inspection of the code showed that they make use of the detection loophole and the coincidence loophole respectively. I evaluate them with appropriate modified Bell-CHSH type inequalities: the Larsson detection-loophole adjusted CHSH, and the Larsson-Gill coincidence-loophole adjusted CHSH (NB: its correctness is conjecture, we do not have proof). The experimental efficiencies turn out to be approximately eta = 81% (close to optimal) and gamma = 55% (far from optimal). The observed values of CHSH are, as they must be, within the appropriately adjusted bounds. Fodjes' detection-loophole model turns out to be very, very close to Pearle's famous 1970 model, so the efficiency is very close to optimal. The model has the same defect as Pearle's: the joint detection rates exhibit signaling. The coincidence-loophole model is actually an elegant modification of the detection-loophole model. Because of this, however, it cannot lead to optimal efficiency. Later versions of the programs included an explanation of how they worked, including formulas, though still no reference whatever to the literature on the two loopholes which Fodje exploits, not even to the concept of an experimental (i.e., in principle, avoidable) loophole. The documentation available now does make a lot of the "reverse engineering" in this paper superfluous. I plan to rewrite it as a very, very short note. I will also use the few jewels in the work in a more ambitious paper, still to be written, about the results of the bigger research project of which these experiments were a small part.

The two authors listed by Researchers.one are both myself, in my two capacities as emeritus professor and as independent consultant. Actually I was just attempting to add my middle name "David" or middle initial "D." to my name on my own profile. But only succeeded in cloning myself.

Under Review Since : 2019-11-29

Inferential challenges that arise when data are censored have been extensively studied under the classical frameworks. In this paper, we provide an alternative generalized inferential model approach whose output is a data-dependent plausibility function. This construction is driven by an association between the distribution of the relative likelihood function at the interest parameter and an unobserved auxiliary variable. The plausibility function emerges from the distribution of a suitably calibrated random set designed to predict that unobserved auxiliary variable. The evaluation of this plausibility function requires a novel use of the classical Kaplan--Meier estimator to estimate the censoring rather than the event distribution. We prove that the proposed method provides valid inference, at least approximately, and our real- and simulated-data examples demonstrate its superior performance compared to existing methods.

Under Review Since : 2019-11-27

One of the classic problems in complex systems is the existence and ubiquity of critically, characterized by scale-invariance in frequency space and a balance between emergence (randomness) and self-organization (order). Another universal characteristic of complex systems is their Antigrafility or the capacity of taking advantage from environmental randomness. Inhere we propose a primer hypothesis that both concepts are related and may be understood under an Information Theory framework using Fisher Information as unifying concept. We make some comments about possible connection with Autopoiesis and Contextuality.

Under Review Since : 2019-11-15

Published Date : 2019-11-08

Under Review Since : 2019-11-18

Under Review Since : 2019-10-25

We present a Gibbs sampler to implement the Dempster-Shafer (DS) theory of statistical inference for Categorical distributions with arbitrary numbers of categories and observations. The DS framework is trademarked by its three-valued uncertainty assessment (p, q, r), probabilities "for", "against", and "don't know", associated with formal assertions of interest. The proposed algorithm targets the invariant distribution of a class of random convex polytopes which encapsulate the inference, via establishing an equivalence between the iterative constraints of the vertex configuration and the non-negativity of cycles in a fully connected directed graph. The computational cost increases with the size of the input, linearly with the number of observations and polynomially in the number of non-empty categories. Illustrations of numerical examples include the testing of independence in 2 by 2 contingency tables and parameter estimation of the linkage model. Results are compared to alternative methods of Categorical inference.

Under Review Since : 2019-10-17

Under Review Since : 2019-10-15

Bias resulting from model misspecification is a concern when predicting insurance claims. Indeed, this bias puts the insurer at risk of making invalid or unreliable predictions. A method that could provide provably valid predictions uniformly across a large class of possible distributions would effectively eliminate the risk of model misspecification bias. Conformal prediction is one such method that can meet this need, and here we tailor that approach to the typical insurance application and show that the predictions are not only valid but also efficient across a wide range of settings.

Under Review Since : 2019-10-14

**Purpose**

To investigate transcutaneous core-needle biopsy of the supraclavicular fat as a minimally invasive and scar-free method of obtaining brown adipose tissue (BAT) samples.

**Material and Methods**

In a prospective clinical trial, 16 volunteers underwent biopsy on two separate occasions after FDG-PET had shown active BAT in the supraclavicular fossa with an FDG uptake (SUV_{max}) > 3 mg/dl. After identifying the ideal location for biopsy on FDG-PET/MRI, ultrasound-guided core-needle biopsy of supraclavicular fat with a 16G needle was performed under local anesthesia and aseptic conditions. Tissue samples were immediately shock-frozen in liquid nitrogen and processed for gene expression analysis of adipose tissue markers. Wounds were checked two weeks after the biopsy.

**Results**

Tissue sampling was successful in 15 volunteers in both scans and in one very lean volunteer (BMI=19.9 kg/m^{2}) in only one visit, without any reported adverse events. Therefore 31 tissue samples were available for further analysis. Gene expression could be analyzed with high success rate in 30 out of 31 tissue biopsies. The intervention was well tolerated with local anesthetics. None of the volunteers showed any scarring.

**Conclusion**

Ultrasound-guided core-needle biopsy of FDG-positive supraclavicular fat yields sufficient BAT samples for quantification of molecular markers. It may, however, be limited in extremely lean individuals with very little supraclavicular fat.

Under Review Since : 2019-10-11

Under Review Since : 2020-02-20

I prove Beal's Conjecture using elementary methods of number theory

and linear algebra. Fermat's Last Theorem is used, as is the Generalized Bezout Identity

with linear algebra to show that Bezout coefficients form a $\Z$-module of what

I call the Bezout nullspace of the numbers $A,B,C$. Assuming $A,B,C$ do not have a common factor, the key equation represents a point that can and cannot be found on this nullspace as a module. Contrapositively if there are solutions to the key equation, the numbers $A,B,C$ must have a common factor. The edge case of $1+B^y = C^z$ is an instance of the Catalan Diophantine Problem,

solved by Mihailescu in 2004.

Under Review Since : 2019-10-04

An important question in economics is how people choose between different payments in the future. The classical normative model predicts that a decision maker discounts a later payment relative to an earlier one by an exponential function of the time between them. Descriptive models use non-exponential functions to fit observed behavioral phenomena, such as preference reversal. Here we propose a model of discounting, consistent with standard axioms of choice, in which decision makers maximize the growth rate of their wealth. Four specifications of the model produce four forms of discounting - no discounting, exponential, hyperbolic, and a hybrid of exponential and hyperbolic - two of which predict preference reversal. Our model requires no assumption of behavioral bias or payment risk.

Published Date : 2019-09-30

Whether the predictions put forth prior to the 2016 U.S. presidential election were right or wrong is a question that led to much debate. But rather than focusing on right or wrong, we analyze the 2016 predictions with respect to a core set of {\em effectiveness principles}, and conclude that they were ineffective in conveying the uncertainty behind their assessments. Along the way, we extract key insights that will help to avoid, in future elections, the systematic errors that lead to overly precise and overconfident predictions in 2016. Specifically, we highlight shortcomings of the classical interpretations of probability and its communication in the form of predictions, and present an alternative approach with two important features. First, our recommended predictions are safer in that they come with certain guarantees on the probability of an erroneous prediction; second, our approach easily and naturally reflects the (possibly substantial) uncertainty about the model by outputting *plausibilities* instead of *probabilities*.

Under Review Since : 2019-09-29

This paper examines the development of Laplacean practical certainty from 1810, when Laplace proved his central limit theorem, to 1925, when Ronald A. Fisher published his *Statistical Methods for Research Workers*.

Although Laplace's explanations of the applications of his theorem were accessible to only a few mathematicians, expositions published by Joseph Fourier in 1826 and 1829 made the simplest applications accessible to many statisticians. Fourier suggested an error probability of 1 in 20,000, but statisticians soon used less exigent standards. Abuses, including p-hacking, helped discredit Laplace's theory in France to the extent that it was practically forgotten there by the end of the 19th century, yet it survived elsewhere and served as the starting point for Karl Pearson's biometry.

The probability that a normally distributed random variable is more than three probable errors from its mean is approximately 5%. When Fisher published his *Statistical Methods*, three probable errors was a common standard for likely significance. Because he wanted to enable research workers to use distributions other than the normal -- the *t* distributions, for example --- Fisher replaced three probable errors with 5%.

The use of *significant* after Fisher differs from its use by Pearson before 1920. In Pearson's *Biometrika*, a significant difference was an observed difference that *signified* a real difference. *Biometrika*'s authors sometimes said that an observed difference is likely or very likely to be significant, but they never said that it is very significant, and they did not have levels of significance. Significance itself was not a matter of degree.

What might this history teach us about proposals to curtail abuses of statistical testing by changing its current vocabulary (p-value, significance, etc.)? The fact that similar abuses arose before this vocabulary was introduced suggests that more substantive changes are needed.

Under Review Since : 2019-09-30

Meta-analysis based on only a few studies remains a challenging problem, as an accurate estimate of the between-study variance is apparently needed, but hard to attain, within this setting. Here we offer a new approach, based on the *generalized inferential model* framework, whose success lays in marginalizing out the between-study variance, so that an accurate estimate is not essential. We show theoretically that the proposed solution is at least approximately valid, with numerical results suggesting it is, in fact, nearly exact. We also demonstrate that the proposed solution outperforms existing methods across a wide range of scenarios.

Under Review Since : 2019-09-23

*Big Bubble theory is a cosmological model where the universe is an expanding bubble in four-dimensional space. Expansion is driven by starlight and gravity acts like surface tension to form a minimal surface. This model is used to derive Minkowski’s spacetime geometrically from four-dimensional Euclidian space. Big Bubble cosmology is consistent with type 1a supernova redshifts without dark energy or expanding spacetime. A different origin for the cosmic microwave background is proposed. The size of the universe is estimated using Hubble’s constant and a doppler shift of the cosmic microwave background. A mechanism for Mach’s principle is described. Big Bubble theory is similar to Einstein’s 1917 cosmological model, which is shown to be a snapshot of a rapidly expanding universe in dynamic equilibrium, rather than a static universe. The orbital speed of stars in spiral galaxies can be reproduced with Newtonian dynamics and without dark matter. A quadratic equation is derived that predicts both faster and slower rotation than purely Kepler orbits, consistent with the behaviour of spiral and elliptical galaxies, and suggesting that spiral galaxies evolve into elliptical galaxies as they age. The Big Bubble physical concept provides a basis for some quantum physics phenomena.*

Published Date : 2019-09-08

**The work argues that Nassim Taleb's precautionary principle should not apply to the domain of ‘GMOs’ any more than to other monopolizing economic domains, because the probability of systemic ruin stemming from the GM technology itself is dwarfed by other systemic risks of the Deductive-Optimization Economy of today.**

Published Date : 2019-09-08

**A philosophical version**

In this work of foresight, I communicated my perception of Taleb's policy paper and the Black Swan problem discussed in it. To this effect, I:

- Re-conceptualized the concept of "foresight,” non-teleologically, and its “method”;
- Revived Empedocles’ non-teleological philosophy of evolution with modern scientific data;
- Located the real GMO safety problem in (you guessed it) teleology: in the suppression of dissent within institutions under a seeming assumption of knowing what waste is.

Under Review Since : 2019-08-24

Under Review Since : 2019-08-04

The idea of the paper is to think about the result presented in Numberphile (http://www. numberphile.com/) talk (https://www.youtube.com/watch?v=w-I6XTVZXww) where they claim that 1 + 2 + 3 + ..., the Gauss sum, converges to −1/12. In the video they make two strong statements: one that the Grandi’s Series 1 − 1 + 1 − 1 + 1 − 1 + ... tends to 1/2 and the second that as bizarre as the −1/12 result for the Gauss sum might appears, as it is connected to Physics (this result is related with the number of dimensions in String Theory) then it is plausible. In this work we argue that these two statements reflect adhesion to a particular probability narrative and to a particular scientific philosophical posture. We argue that by doing so, these (Gauss and Grandi series) results and String Theory ultimately, might be mathematical correct but they are scientifically (in the Galileo-Newton-Einstein tradition) inconsistent (at least). The philosophical implications of this problem are also discussed, focusing on the role of evidence and scientific demarcation.

Under Review Since : 2019-08-01

A covering problem posed by Henri Lebesgue in 1914 seeks to find the convex shape of smallest area that contains a subset congruent to any point set of unit diameter in the Euclidean plane. Methods used previously to construct such a covering can be refined and extended to provide an improved upper bound for the optimal area. An upper bound of 0.8440935944 is found.

Under Review Since : 2019-07-19

In the context of predicting future claims, a fully Bayesian analysis---one that specifies a statistical model, prior distribution, and updates using Bayes's formula---is often viewed as the gold-standard, while Buhlmann's credibility estimator serves as a simple approximation. But those desirable properties that give the Bayesian solution its elevated status depend critically on the posited model being correctly specified. Here we investigate the asymptotic behavior of Bayesian posterior distributions under a misspecified model, and our conclusion is that misspecification bias generally has damaging effects that can lead to inaccurate inference and prediction. The credibility estimator, on the other hand, is not sensitive at all to model misspecification, giving it an advantage over the Bayesian solution in those practically relevant cases where the model is uncertain. This begs the question: does robustness to model misspecification require that we abandon uncertainty quantification based on a posterior distribution? Our answer to this question is *No*, and we offer an alternative *Gibbs posterior* construction. Furthermore, we argue that this Gibbs perspective provides a new characterization of Buhlmann's credibility estimator.

Under Review Since : 2019-06-24

We study the problem of non-parametric Bayesian estimation of the intensity function of a Poisson point process. The observations are $n$ independent realisations of a Poisson point process on the interval $[0,T]$. We propose two related approaches. In both approaches we model the intensity function as piecewise constant on $N$ bins forming a partition of the interval $[0,T]$. In the first approach the coefficients of the intensity function are assigned independent gamma priors, leading to a closed form posterior distribution. On the theoretical side, we prove that as $n\rightarrow\infty,$ the posterior asymptotically concentrates around the ``true", data-generating intensity function at an optimal rate for $h$-H\"older regular intensity functions ($0 < h\leq 1$).

In the second approach we employ a gamma Markov chain prior on the coefficients of the intensity function. The posterior distribution is no longer available in closed form, but inference can be performed using a straightforward version of the Gibbs sampler. Both approaches scale well with sample size, but the second is much less sensitive to the choice of $N$.

Practical performance of our methods is first demonstrated via synthetic data examples. We compare our second method with other existing approaches on the UK coal mining disasters data. Furthermore, we apply it to the US mass shootings data and Donald Trump's Twitter data.

Under Review Since : 2019-05-27

The Wilcoxon Rank Sum is a very competitive robust alternative to the two-sample t-test when the underlying data have tails longer than the normal distribution. Extending to the one-way model with k independent samples, the Kruskal-Wallis rank test is a competitive alternative to the usual F for testing if there are any location differences. However, these positives for rank methods do not extend as readily to methods for making all pairwise comparisons used to reveal where the differences in location may exist. We demonstrate via examples and simulation that rank methods can have a dramatic loss in power compared to the standard Tukey-Kramer method of normal linear models even for non-normal data. We also show that a well-established robust rank-like method can recover the power but does not fully control the familywise error rate in small samples.

Under Review Since : 2019-05-25

Under Review Since : 2019-03-09

An inferential model encodes the data analyst's degrees of belief about an unknown quantity of interest based on the observed data, posited statistical model, etc. Inferences drawn based on these degrees of belief should be reliable in a certain sense, so we require the inferential model to be *valid*. The construction of valid inferential models based on individual pieces of data is relatively straightforward, but how to combine these so that the validity property is preserved? In this paper we analyze some common combination rules with respect to this question, and we conclude that the best strategy currently available is one that combines via a certain dimension reduction step before the inferential model construction.

Under Review Since : 2019-03-05

Cooperation is a persistent behavioral pattern of entities pooling and sharing resources. Its ubiquity in nature poses a conundrum: whenever two entities cooperate, one must willingly relinquish something of value to the other. Why is this apparent altruism favored in evolution? Classical treatments assume *a priori* a net fitness gain in a cooperative transaction which, through reciprocity or relatedness, finds its way back from recipient to donor. Our analysis makes no such assumption. It rests on the insight that evolutionary processes are typically multiplicative and noisy. Fluctuations have a net negative effect on the long-time growth rate of resources but no effect on the growth rate of their expectation value. This is a consequence of non-ergodicity. Pooling and sharing reduces the amplitude of fluctuations and, therefore, increases the long-time growth rate for cooperators. Put simply, cooperators' resources outgrow those of similar non-cooperators. This constitutes a fundamental and widely applicable mechanism for the evolution of cooperation. Furthermore, its minimal assumptions make it a candidate explanation in simple settings, where other explanations, such as emergent function and specialization, are implausible. An example of this is the transition from single cells to early multicellular life.

Published Date : 2019-03-04

Abstract

These papers - one proposition paper and ten responses - comprise a debate on shaken baby syndrome. This is the hypothesis that a Triad of indicators in the head of a dead baby reveal that it has been shaken to death, and that the killer was the person last in charge of the baby. The debate was scheduled to have appeared in *Prometheus*, a journal concerned with innovation rather than matters medical. It struck the editors of *Prometheus* that a hypothesis that had survived nearly half a century and was still resistant to challenge and change was well within the tradition of *Prometheus* debate. The debate focuses on the role of the expert witness in court, and especially the experiences of Waney Squier, a prominent paediatric pathologist, struck from the medical register in the UK for offering opinions beyond her core expertise and showing insufficient respect for established thinking and its adherents. The debate’s responses reveal much about innovation, and most about the importance of context, in this case the incompatibility of medicine and the law, particularly when constrained by the procedures of the court. Context was also important in the reluctance of Taylor & Francis, the publisher of *Prometheus*, to publish the debate on the grounds that its authors strayed from their areas of expertise and showed insufficient respect for established thinking.

*Prometheus*** shaken baby debate**

**Contents**

**Introduction **

The shaken baby debate - Stuart Macdonald

**Proposition paper**

Shaken baby syndrome: causes and consequences of conformity - Waney Squier

**Response papers**

Shaken baby syndrome: a fraud on the courts - Heather Kirkwood

Shaken baby: an evolving diagnosis deformed by the pressures of the courtroom - Susan Luttner

Waney Squier’s ordeal and the crisis of the shaken baby paradigm - Niels Lynøe

Another perspective - simply my brief thoughts - Dave Marshall

Has Squier been treated fairly? - Brian Martin

Commentary on the paper by Waney Squier: ‘Shaken baby syndrome: causes and consequences of conformity’ - Michael J Powers

Waney Squier and the shaken baby syndrome case: a clarion call to science, medicine and justice - Toni C Saad

The role of the General Medical Council - Terence Stephenson

When experts disagree - Stephen J. Watkins

The General Medical Council’s handling of complaints: the Waney Squier case - Peter Wilmshurst

Under Review Since : 2019-03-03

Often the primary goal of fitting a regression model is prediction, but the majority of work in recent years focuses on inference tasks, such as estimation and feature selection. In this paper we adopt the familiar sparse, high-dimensional linear regression model but focus on the task of prediction. In particular, we consider a new empirical Bayes framework that uses the data to appropriately center the prior distribution for the non-zero regression coefficients, and we investigate the method's theoretical and numerical performance in the context of prediction. We show that, in certain settings, the asymptotic posterior concentration in metrics relevant to prediction quality is very fast, and we establish a Bernstein--von Mises theorem which ensures that the derived prediction intervals achieve the target coverage probability. Numerical results complement the asymptotic theory, showing that, in addition to having strong finite-sample performance in terms of prediction accuracy and uncertainty quantification, the computation time is considerably faster compared to existing Bayesian methods.

Under Review Since : 2019-02-28

This communication outlines the potential for a novel, alternative model to rationalise the quantitative difficulties with the Hubble constant. This model, the Proto-Quantum Field (PQF) model, is an alternative to the singularity-big-bang (SBB) model and the two are mutually incompatible. A justification is that the theoretical developments required to validate the PQF hypothesis is closely derived from the standard model for particles and forces, more so than those required to modify the SBB hypothesis.

Published Date : 2019-02-15

The universe is formed from proto-quantum field(s) (PQFs). The initiating event is the formation of a thermal gradient which establishes synchronous oscillations which describes time. Time is not quantised. Concomitantly PQFs, either directly or indirectly, differentiate into the all the quantum fields required for the standard model of particles and forces, and three dimensional space. The transition of PQFs to functional quantum fields is a continuous process at the boundary of a spherical universe, a “ring of fire”, necessary to maintain time.

Under Review Since : 2019-02-11

This analysis shows that a special relativity interpretation matches observed type 1a supernova redshifts. Davis & Lineweaver reported in 2003 that a special relativity match to supernova redshift observations can be ruled out at more than 23σ, but MacLeod’s 2004 conclusion that this finding was incorrect and due to a mathematical error is confirmed. MacLeod’s plot of special relativity against observation has been further improved by using celerity (aka proper velocity) instead of peculiar velocity. A Hubble plot of type 1a supernova celerity against retarded distance has a straight line of 70 km s^{-1} Mpc^{-1} for as far back in time as we can observe, indicating that, with a special relativity interpretation of cosmological redshift, expansion of the universe is neither accelerating nor decelerating, and it is not necessary to invoke the existence of dark energy.

Under Review Since : 2019-02-05

*It’s been more than a century since Einstein’s special theory of relativity showed that Newton’s concept of time is incorrect, but society and science continue to use predominantly Newtonian language and thought. The words normally used to describe time don’t distinguish when time is a dimension, used for locating objects and events in spacetime, and when it’s a property of objects that can age at different rates. It is proposed to bring relativity’s terminology of coordinate time and proper time into everyday language, and thereby distinguish between ‘cotime’ (a dimensional property) and ‘protime’ (a property of objects related to energy). The differences between cotime and protime are significant and cotime might be a spatial dimension with units of length.*

Under Review Since : 2019-02-03

Statistics has made tremendous advances since the times of Fisher, Neyman, Jeffreys, and others, but the fundamental and practically relevant questions about probability and inference that puzzled our founding fathers remain unanswered. To bridge this gap, I propose to look beyond the two dominating schools of thought and ask the following three questions: what do scientists need out of statistics, do the existing frameworks meet these needs, and, if not, how to fill the void? To the first question, I contend that scientists seek to convert their data, posited statistical model, etc., into calibrated degrees of belief about quantities of interest. To the second question, I argue that any framework that returns additive beliefs, i.e., probabilities, necessarily suffers from *false confidence*---certain false hypotheses tend to be assigned high probability---and, therefore, risks systematic bias. This reveals the fundamental importance of *non-additive beliefs* in the context of statistical inference. But non-additivity alone is not enough so, to the third question, I offer a sufficient condition, called *validity*, for avoiding false confidence, and present a framework, based on random sets and belief functions, that provably meets this condition. Finally, I discuss characterizations of p-values and confidence intervals in terms of valid non-additive beliefs, which imply that users of these classical procedures are already following the proposed framework without knowing it.

Under Review Since : 2019-01-30

A perfect economic storm emerged in M\'exico in what was called (mistakenly under our analysis) The December Error (1994) in which Mexico's economy collapsed. In this paper, we show how Theoretical Psychics may help us to understand the under processes for this kind of economic crisis and eventually perhaps to develop an early warning. We specifically analyze monthly historical time series for inflation from January 1969 to November 2018. We found that Fisher information is insensible to inflation growth in the 80's decade but capture quite good The December Error (TDE). Our results show that under Salinas administration Mexican economy was characterized by unstable stability must probably due to hidden risk policies in the form of macro-economy controls that artificially suppress aleatority out of the system making it fragile. And so, we conclude that it was not at all a December error but a sexenal sustained error of fragilization.

Under Review Since : 2019-01-24

Inhere we present a proposal of how to teach complexity using a Problem Based Learning approach under a set of philosophical principles inspired by the pedagogical experience in sustainability Sciences. We described the context in which we put on practise y these ideas that was a graduate course on Complexity and Data Science applied to Ecology. In part two we present the final work presented by the students as they wrote it and which we believe could be submitted to a journal by its own merits

Under Review Since : 2019-01-21

Inhere we expand the concept of Holobiont to incorporate niche construction theory in order to increase our understanding of the current planetary crisis. By this, we propose a new ontology, the Ecobiont, as the basic evolutionary unit of analysis. We make the case of \textit{Homo Sapiens} organized around modern cities (technobionts) as a different Ecobiont from classical \textit{Homo Sapiens} (i.e. Hunter-gatherers \textit{Homo Sapiens}). We consider that Ecobiont ontology helps to make visible the coupling of \textit{Homo Sapiens} with other biological entities under processes of natural and cultural evolution. Not to see this coupling hidden systemic risks and enhance the probability of catastrophic events. So Ecobiont ontology is necessary to understand and respond to the current planetary crisis.

Under Review Since : 2018-12-31

Headwater streams are essential to downstream water quality, therefore it is important they are properly represented on maps used for stream regulation. Current maps used for stream regulation, such as the United States Geological Survey (USGS) topographic maps and Natural Resources Conservation Service (NRCS) soil survey maps, are outdated and do not accurately nor consistently depict headwater streams. In order for new stream maps to be used for regulatory purposes, the accuracy must be known and the maps must show streams with a consistent level of accuracy. This study assessed the valley presence/absence and stream length accuracy of the new stream maps created by the North Carolina Center for Geographic Analysis (CGIA) for western North Carolina. The CGIA stream map does not depict headwater streams with a consistent level of accuracy. This study also compared the accuracy of stream networks modeled using the computer software program, Terrain Analysis using Digital Elevation Models (TauDEM), to the CGIA stream map. The stream networks modeled in TauDEM, also do not consistently predict the location of headwater streams across the mountain region of the state. The location of headwater streams could not be accurately nor consistently predicted by solely using aerial photography or elevation data. Other factors such as climate, soils, geology, land use, and vegetation cover should be considered to accurately and consistently model headwater stream networks.

Under Review Since : 2018-12-28

Metric Temporal Logic (MTL) is a popular formalism to specify patterns with timing constraints over the behavior of cyber-physical systems. In this paper, I propose sequential networks for online monitoring applications and construct network-based monitors from the past fragment of MTL over discrete and dense time behaviors. This class of monitors is more compositional, extensible, and easily implementable than other monitors based on rewriting and automata. I first explain the sequential network construction over discrete time behaviors and then extend it towards dense time by adopting a point-free approach. The formulation for dense time behaviors and MTL radically differs from the traditional pointy definitions and in return, we avoid some longstanding complications. I argue that the point-free approach is more natural and practical therefore should be preferred for the dense time. Finally, I present my implementation together with some experimental results that show the performance of the network-based monitors compared to similar existing tools.

Under Review Since : 2018-12-17

Analogue gravity models are attempts to model general relativity by using such things as acoustic waves propagating through an ideal fluid. In his work, we take inspiration from these models to re-interpret general relativity in terms of an ether continuum moving and changing against a background of absolute space and time. We reformulate the metric, the Ricci tensor, the Einstein equation, continuous matter dynamics in terms of the ether. We also reformulate general relativistic electrodynamics in terms of the ether, which takes the form of electrodynamics in an anisotropic moving medium. Some degree of simplification is achieved by assuming that the speed-of-light is uniform and isotropic with respect to the ether coordinates. Finally, we speculate on the nature of under-determination in general relativity.

Under Review Since : 2018-12-05

Nonparametric estimation of a mixing density based on observations from the corresponding mixture is a challenging statistical problem. This paper surveys the literature on a fast, recursive estimator based on the *predictive recursion* algorithm. After introducing the algorithm and giving a few examples, I summarize the available asymptotic convergence theory, describe an important semiparametric extension, and highlight two interesting applications. I conclude with a discussion of several recent developments in this area and some open problems.

Under Review Since : 2018-12-05

Bayesian methods provide a natural means for uncertainty quantification, that is, credible sets can be easily obtained from the posterior distribution. But is this uncertainty quantification valid in the sense that the posterior credible sets attain the nominal frequentist coverage probability? This paper investigates the frequentist validity of posterior uncertainty quantification based on a class of empirical priors in the sparse normal mean model. In particular, we show that our marginal posterior credible intervals achieve the nominal frequentist coverage probability under conditions slightly weaker than needed for selection consistency and a Bernstein--von Mises theorem for the full posterior, and numerical investigations suggest that our empirical Bayes method has superior frequentist coverage probability properties compared to other fully Bayes methods.

Under Review Since : 2018-12-04

Reality has two different dimensions: information-communication, and matter-energy. They relate to each other in a figure-ground gestalt that gives two different perspectives on the one reality.

1. We learn from modern telecommunications that matter and energy are the **media** of information and communication; here information is the figure and matter/enegy is the ground.

2. We learn from cybernetics that information and communication **control** matter and energy; here matter/energy is the figure and information/communication is the ground.

Under Review Since : 2018-12-04

The cybernetic control loop can be understood as a decision making process, a command and control process. Information controls matter and energy. Decision making has four necessary and sufficient communication processes: act, sense, evaluate, choose. These processes are programmed by the environment, the model, values and alternatives.

Under Review Since : 2018-12-04

The idea that information is entropy is an error. The proposal is neither mathematically nor logically valid. There are numerous definitions of entropy, but there is no definition of entropy for which this equation is valid. Information is information. It is neither matter nor energy.

Under Review Since : 2018-12-03

Internet is increasingly important for our economies and societies. This is the reason for a growing interest in internet regulation. The stakes in network neutrality - that all traffic on the internet should be treated equally - are particularly high. This paper argues that technological, civil-libertarian, legal and economic arguments exist both for- and against net neutrality and that the decision is ultimately political. We therefore frame the issue of net neutrality as an issue of political economy. The main political economy arguments for net neutrality are that a net-neutral internet contributes to the reduction of inequality, preserves its openness and prevents artificial scarcity. With these arguments Slovenia, after Chile and the Netherlands, adopted net neutrality legislation. We present it as a case study for examining how political forces are affecting the choice of economic and technological policies. After a few years we are finding that proper enforcement is just as important as legislation.

Under Review Since : 2018-11-22

Physicalism, which provides the philosophical basis of modern science, holds that consciousness is solely a product of brain activity, and more generally, that mind is an epiphenomenon of matter, that is, derivable from and reducible to matter. If mind is reducible to matter, then it follows that identical states of matter must correspond to identical states of mind.

In this discourse, I provide a cogent refutation of physicalism by showing examples of physically identical states which, by definition, cannot be distinguished by any method available to science but can nevertheless be distinguished by a conscious observer. I conclude by giving an example of information that is potentially knowable by an individual but is beyond the ken of science.

Under Review Since : 2018-11-22

Philosophers have long pondered the Problem of Universals. One response is Metaphysical Realism, such as Plato's Doctrine of the Forms and Aristotle's Hylomorphism. We postulate that Measurement in Quantum Mechanics forms the basis of Metaphysical Realism. It is the process that gives rise to the instantiation of Universals as Properties, a process we refer to as Hylomorphic Functions. This combines substance metaphysics and process metaphysics by identifying the instantiation of Universals as causally active processes along with physical substance, forming a dualism of both substance and information. Measurements of fundamental properties of matter are the Atomic Universals of metaphysics, which combine to form the whole taxonomy of Universals. We look at this hypothesis in relation to various different interpretations of Quantum Mechanics grouped under two exemplars: the Copenhagen Interpretation, a version of Platonic Realism based on wave function collapse, and the Pilot Wave Theory of Bohm and de Broglie, where particle--particle interactions lead to an Aristotelian metaphysics. This view of Universals explains the distinction between pure information and the medium that transmits it and establishes the arrow of time. It also distinguishes between univerally true Atomic Facts and the more conditional Inferences based on them. Hylomorphic Functions also provide a distinction between Universals and Tropes based on whether a given Property is a physical process or is based on the qualia of an individual organism. Since the Hylomorphic Functions are causally active, it is possible to suggest experimental tests that can verify this viewpoint of metaphysics.

Under Review Since : 2018-12-05

The nature of consciousness has been one of the longest-standing open questions in philosophy. Advancements in physics, neuroscience, and information theory have informed and constrained this topic, but have not produced any consensus. What would it mean to ‘solve’ or ‘dissolve’ the mystery of consciousness?

Part I begins with grounding this topic by considering a concrete question: what makes some conscious experiences more pleasant than others? We first review what’s known about the neuroscience of pain & pleasure, find the current state of knowledge narrow, inconsistent, and often circular, and conclude we must look elsewhere for a systematic framework (Sections I & II). We then review the Integrated Information Theory (IIT) of consciousness and several variants of IIT, and find each of them promising, yet also underdeveloped and flawed (Sections III-V).

We then take a step back and distill what kind of problem consciousness *is*. Importantly, we offer eight sub-problems whose solutions would, in aggregate, constitute a *complete theory of consciousness* (Section VI).

Armed with this framework, in Part II we return to the subject of pain & pleasure (valence) and offer some assumptions, distinctions, and heuristics to clarify and constrain the problem (Sections VII-IX). Of particular interest, we then offer a specific hypothesis on what valence *is* (Section X) and several novel empirical predictions which follow from this (Section XI). Part III finishes with discussion of how this general approach may inform open problems in neuroscience, and the prospects for building a new science of qualia (Sections XII & XIII). Lastly, we identify further research threads within this framework (Appendices A-F).

Published Date : 2018-11-13

This paper argues that experimental evidence, quantum theory, and relativity theory, taken together, suggest that reality is relational: Properties and behaviors of phenomena do not have a priori, intrinsic values; instead, these properties and behaviors emerge through interactions with other systems.

Under Review Since : 2018-11-05

Published Date : 2018-11-01

Francis Perey, of the Engineering Physics Division of Oak Ridge National Lab, left a number of unpublished papers upon his death in 2017. They circulate around the idea of probabilities arising naturally from basic physical laws. One of his papers, Application of Group Theory to Data Reduction, was published as an ORNL white paper in 1982. This collection includes two earlier works and two that came later, as well as a relevant presentation. They are being published now so that the ideas in them will be available to interested parties.

Under Review Since : 2018-10-22

The potential for an infectious disease outbreak that is much worse than those which have been observed in human history, whether engineered or natural, has been the focus of significant concern in biosecurity. Fundamental dynamics of disease spread make such outbreaks much less likely than they first appear. Here we present a slightly modified formulation of the typical SEIR model that illustrates these dynamics more clearly, and shows the unlikely cases where concern may still be warranted. This is then applied to an extreme version of proposed pandemic risk, multi-disease syndemics, to show that (absent much clearer reasons for concern) the suggested dangers are overstated.

Under Review Since : 2018-10-03

This note generalizes the notion of conditional probability to Riesz spaces using the order-theoretic approach. With the aid of this concept, we establish the law of total probability and Bayes' theorem in Riesz spaces; we also prove an inclusion-exclusion formula in Riesz spaces. Several examples are provided to show that the law of total probability, Bayes' theorem and inclusion-exclusion formula in probability theory are special cases of our results.

Under Review Since : 2018-09-18

To justify the effort of developing a theoretical construct, a theoretician needs empirical data that support a non-random effect of sufficiently high replication-probability. To establish these effects statistically, researchers (rightly) rely on a *t*-test. But many pursue questionable strategies that lower the cost of data-collection. Our paper reconstructs two such strategies. Both reduce the minimum sample-size (N_{MIN}) sufficing under conventional errors (*α*, *β*) to register a given effect-size (*d*) as a statistically significant non-random data signature. The first strategy increases the *β*-error; the second treats the control-group as a constant, thereby collapsing a two-sample *t*-test into its one-sample version. (A two-sample *t*-test for *d*=0.50 under a*=β*=0.05 with N_{MIN}=176, for instance, becomes a one-sample *t*-test under a*=*0.05, *β*=0.20 with N_{MIN}=27.) Not only does this decrease the replication-probability of data from (1-*β*)=0.95 to (1-*β*)=0.80, particularly the second strategy cannot corroborate hypotheses meaningfully. The ubiquity of both strategies arguably makes them partial causes of the confidence-crisis. But as resource-pooling would allow research groups reach N_{MIN} jointly, a group’s individually limited resources justify neither strategy.

Published Date : 2018-09-15

This article describes our motivation behind the development of RESEARCHERS.ONE, our mission, and how the new platform will fulfull this mission. We also compare our approach with other recent reform initiatives such as post-publication peer review and open access publications.

Under Review Since : 2018-09-15

Under Review Since : 2018-09-14

This paper is a synthesis of the deposition in front of the Financial Crisis Inquiry Commission by the Obama Administration in 2010. Note that none of its ideas made it to the report.

Under Review Since : 2018-09-14

This article describes how the filtering role played by peer review may actually be harmful rather than helpful to the quality of the scientific literature. We argue that, instead of trying to filter out the low-quality research, as is done by traditional journals, a better strategy is to let everything through but with an acknowledgment of the uncertain quality of what is published, as is done on the RESEARCHERS.ONE platform. We refer to this as "scholarly mithridatism." When researchers approach what they read with doubt rather than blind trust, they are more likely to identify errors, which protects the scientific community from the dangerous effects of error propagation, making the literature stronger rather than more fragile.

Published Date : 2018-09-14

Penalized maximum likelihood methods that perform automatic variable are now ubiquitous in statistical research. It is well-known, however, that these estimators are nonregular and consequently have limiting distributions that can be highly sensitive to small perturbations of the underlying generative model. This is the case even for the ﬁxed “p” framework. Hence, the usual asymptotic methods for inference, like the bootstrap and series approximations, often perform poorly in small samples and require modiﬁcation. Here, we develop locally asymptotically consistent conﬁdence intervals for regression coeﬃcients when estimation is done using the Adaptive LASSO (Zou, 2006) in the ﬁxed “p” framework. We construct the conﬁdence intervals by sandwiching the nonregular functional of interest between two smooth, data-driven, upper and lower bounds and then approximating the distribution of the bounds using the bootstrap. We leverage the smoothness of the bounds to obtain consistent inference for the nonregular functional under both ﬁxed and local alternatives. The bounds are adaptive to the amount of underlying nonregularity in the sense that they deliver asymptotically exact coverage whenever the underlying generative model is such that the Adaptive LASSO estimators are consistent and asymptotically normal, and conservative otherwise. The resultant conﬁdence intervals possess a certain tightness property among all regular bounds. Although we focus on the Adaptive LASSO, our approach generalizes to other penalized methods. (Originally published as a technical report in 2014.)

Under Review Since : 2018-09-09

Prediction markets are currently used for three fields: 1. For economic, political and sporting event outcomes. (IEW, PredictIt, PredictWise) 2. For risk evaluation, product development and marketing. (Cultivate Labs/Consensus Point) 3. Research replication. (Replication Prediction Project, Experimental Economics Prediction Project, and Brian Nosek’s latest replicability study) The latter application of prediction markets has remained closed and/or proprietary despite the promising results in the methods. In this paper, I construct an open research prediction market framework to incentivize replicate study research and align the motivations of research stakeholders.

Under Review Since : 2018-09-06

Extreme values are by definition rare, and therefore a spatial analysis of extremes is attractive because a spatial analysis makes full use of the data by pooling information across nearby locations. In many cases, there are several dependent processes with similar spatial patterns. In this paper, we propose the first multivariate spatial models to simultaneously analyze several processes. Using a multivariate model, we are able to estimate joint exceedance probabilities for several processes, improve spatial interpolation by exploiting dependence between processes, and improve estimation of extreme quantiles by borrowing strength across processes. We propose models for separable and non-separable, and spatially continuous and discontinuous processes. The method is applied to French temperature data, where we find an increase in the extreme temperatures over time for much of the country.

Under Review Since : 2018-09-04

In a Bayesian context, prior specification for inference on monotone densities is conceptually straightforward, but proving posterior convergence theorems is complicated by the fact that desirable prior concentration properties often are not satisfied. In this paper, I first develop a new prior designed specifically to satisfy an empirical version of the prior concentration property, and then I give sufficient conditions on the prior inputs such that the corresponding empirical Bayes posterior concentrates around the true monotone density at nearly the optimal minimax rate. Numerical illustrations also reveal the practical benefits of the proposed empirical Bayes approach compared to Dirichlet process mixtures.

Under Review Since : 2018-09-04

Accurate estimation of value-at-risk (VaR) and assessment of associated uncertainty is crucial for both insurers and regulators, particularly in Europe. Existing approaches link data and VaR indirectly by first linking data to the parameter of a probability model, and then expressing VaR as a function of that parameter. This indirect approach exposes the insurer to model misspecification bias or estimation inefficiency, depending on whether the parameter is finite- or infinite-dimensional. In this paper, we link data and VaR directly via what we call a discrepancy function, and this leads naturally to a Gibbs posterior distribution for VaR that does not suffer from the aforementioned biases and inefficiencies. Asymptotic consistency and root-*n* concentration rate of the Gibbs posterior are established, and simulations highlight its superior finite-sample performance compared to other approaches.

Under Review Since : 2018-09-02

We prove an observation of Makkai that FOLDS equivalence coincides with homotopy equivalence in the case of semi-simplicial sets.

Under Review Since : 2018-09-01

The Art of The Election: A Social Media History of the 2016 Presidential Race

Abstract

The book is 700 pages comprising of Donald Trump’s tweets from June 2015 to November 2016 and footnotes which comprise 70-80% of the tweets which explain the context of each tweet. The book has a 100 page bibliography.

It is highly likely that Trump would not have been elected President were it not for social media. This is an unprecedented statement. This is the first time a presidential candidate utilized a social network to get his message out directly to voters, but moreover, to shape the media feedback loop. His tweets became news. This is primary source material on the 2016 election. No need for narratives, outside ”experts” or political ”science”.

The file is too large to post on this website. But you can download the book under this link:

https://www.dropbox.com/s/bxvsh7eqh2ueq6j/Trump%20Book.docx?dl=0

Keywords and phrases: 2016, book, Trump, election, social media.

Under Review Since : 2018-09-01

We propose the first economical theory of value that actually works. We explain evolutionary causes of trade, and demonstrate how goods have value from the evolutionary perspective, and how this value is increased with trade. This "Darwinian" value of goods exists before humans assign monetary value (or any other value estimate) to traded goods. We propose objective value estimate expressed in energy units.

Under Review Since : 2018-09-14

Inference on parameters within a given model is familiar, as is ranking different models for the purpose of selection. Less familiar, however, is the quantification of uncertainty about the models themselves. A Bayesian approach provides a posterior distribution for the model but it comes with no validity guarantees, and, therefore, is only suited for ranking and selection. In this paper, I will present an alternative way to view this model uncertainty problem, through the lens of a valid inferential model based on random sets and non-additive beliefs. Specifically, I will show that valid uncertainty quantification about a model is attainable within this framework in general, and highlight the benefits in a classical signal detection problem.

Under Review Since : 2018-09-17

I make the distinction between *academic* probabilities, which are not rooted in reality and thus have no tangible real-world meaning, and *real* probabilities, which attain a real-world meaning as the odds that the subject asserting the probabilities is forced to accept for a bet against the stated outcome. With this I discuss how the replication crisis can be resolved easily by requiring that probabilities published in the scientific literature are real, instead of academic. At present, all probabilities and derivatives that appear in published work, such as P-values, Bayes factors, confidence intervals, etc., are the result of academic probabilities, which are not useful for making meaningful assertions about the real world.

Under Review Since : 2018-08-30

Under Review Since : 2018-08-28

Under Review Since : 2019-03-26

I prove a connection between the logical framework for intuitive probabilistic reasoning (IPR) introduced by Crane (2017) and sets of imprecise probabilities. More specifically, this connection provides a straightforward interpretation to sets of imprecise probabilities as subjective credal states, giving a formal semantics for Crane's formal system of IPR. The main theorem establishes the IPR framework as a potential logical foundation for imprecise probability that is independent of the traditional probability calculus.