Catalogue Search | MBRL
Search Results Heading
Explore the vast range of titles available.
MBRLSearchResults
-
DisciplineDiscipline
-
Is Peer ReviewedIs Peer Reviewed
-
Item TypeItem Type
-
SubjectSubject
-
YearFrom:-To:
-
More FiltersMore FiltersSourceLanguage
Done
Filters
Reset
2,902
result(s) for
"Exact Probability Values"
Sort by:
1980–2000
by
Johnston, Janis E
,
Mielke Jr., Paul W
,
Berry, Kenneth J
in
Exact Probability Values
,
Fisher-Yates Exact Test
,
History of mathematics
2014
This chapter chronicles the period from 1980 to 2000. Progress on the development of permutation methods continued unabated during this period, paralleling advancements in high-speed computing and the subsequent wide-spread availability of both university mainframes and, later in the period, personal desktop computers. Also, a number of books were published that introduced permutation methods to a wide variety of audiences and there was a decided shift in the literature away from computer science journals into discipline journals. These progressions were accompanied by an increasing emphasis on statistical applications of permutation methods, both exact and resampling, since efficient permutation generators were readily available.
Book Chapter
1940–1959
by
Johnston, Janis E
,
Mielke Jr., Paul W
,
Berry, Kenneth J
in
Electronic Delay Storage Automatic Calculator (EDSAC)
,
Electronic Numerical Integrator And Computer (ENIAC)
,
Exact Probability Values
2014
This chapter chronicles the development of permutation statistical methods from 1940 to 1959. This period may be considered a bridge between the early years of 1920–1939 where permutation tests were first conceptualized and the next period, 1960–1979, in which gains in computer technology provided the necessary tools to successfully employ permutation tests. The recognition of permutation methods as the gold standard against which conventional statistical methods were to be evaluated, while often implicit in the 1920s and 1930s, is manifest in many of the publications on permutation methods that appeared between 1940 and 1959. Also, a number of researchers turned their attention during this time period to rank tests, which simplified the calculation of exact probability values; other researchers continued work on calculating exact probability values, creating tables for small samples; and still others continued the theoretical work begun in the 1920s.
Book Chapter
1920–1939
by
Johnston, Janis E
,
Mielke Jr., Paul W
,
Berry, Kenneth J
in
Exact Probability Values
,
History of mathematics
,
Iowa State University Statistical Laboratory
2014
This chapter chronicles the development of permutation statistical methods from 1920 to 1939, when the earliest discussions of permutation methods appeared in the literature. In this period J. Spława-Neyman, R.A. Fisher, R.C. Geary, T. Eden, F. Yates, and E.J.G. Pitman laid the foundations of permutation methods as we know them today. As is evident in this period, permutation methods had their roots in agriculture and, from the beginning, were widely recognized as the gold standard against which conventional methods could be verified and confirmed.
Book Chapter
Making and Evaluating Point Forecasts
2011
Typically, point forecasting methods are compared and assessed by means of an error measure or scoring function, with the absolute error and the squared error being key examples. The individual scores are averaged over forecast cases, to result in a summary measure of the predictive performance, such as the mean absolute error or the mean squared error. I demonstrate that this common practice can lead to grossly misguided inferences, unless the scoring function and the forecasting task are carefully matched. Effective point forecasting requires that the scoring function be specified ex ante, or that the forecaster receives a directive in the form of a statistical functional, such as the mean or a quantile of the predictive distribution. If the scoring function is specified ex ante, the forecaster can issue the optimal point forecast, namely, the Bayes rule. If the forecaster receives a directive in the form of a functional, it is critical that the scoring function be consistent for it, in the sense that the expected score is minimized when following the directive. A functional is elicitable if there exists a scoring function that is strictly consistent for it. Expectations, ratios of expectations and quantiles are elicitable. For example, a scoring function is consistent for the mean functional if and only if it is a Bregman function. It is consistent for a quantile if and only if it is generalized piecewise linear. Similar characterizations apply to ratios of expectations and to expectiles. Weighted scoring functions are consistent for functionals that adapt to the weighting in peculiar ways. Not all functionals are elicitable; for instance, conditional value-at-risk is not, despite its popularity in quantitative finance.
Journal Article
Conditional Choice Probability Estimation of Dynamic Discrete Choice Models With Unobserved Heterogeneity
2011
We adapt the expectation-maximization algorithm to incorporate unobserved heterogeneity into conditional choice probability (CCP) estimators of dynamic discrete choice problems. The unobserved heterogeneity can be time-invariant or follow a Markov chain. By developing a class of problems where the difference in future value terms depends on a few conditional choice probabilities, we extend the class of dynamic optimization problems where CCP estimators provide a computationally cheap alternative to full solution methods. Monte Carlo results confirm that our algorithms perform quite well, both in terms of computational time and in the precision of the parameter estimates.
Journal Article
The Model Confidence Set
by
Nason, James M.
,
Hansen, Peter R.
,
Lunde, Asger
in
Applications
,
Bootstrap-Verfahren
,
Confidence
2011
This paper introduces the model confidence set (MCS) and applies it to the selection of models. A MCS is a set of models that is constructed such that it will contain the best model with a given level of confidence. The MCS is in this sense analogous to a confidence interval for a parameter. The MCS acknowledges the limitations of the data, such that uninformative data yield a MCS with many models, whereas informative data yield a MCS with only a few models. The MCS procedure does not assume that a particular model is the true model; in fact, the MCS procedure can be used to compare more general objects, beyond the comparison of models. We apply the MCS procedure to two empirical problems. First, we revisit the inflation forecasting problem posed by Stock and Watson (1999), and compute the MCS for their set of inflation forecasts. Second, we compare a number of Taylor rule regressions and determine the MCS of the best regression in terms of in-sample likelihood criteria.
Journal Article
Statistical Methods for Eliciting Probability Distributions
by
Garthwaite, Paul H
,
Kadane, Joseph B
,
O'Hagan, Anthony
in
Ambiguity
,
Assessed values
,
Attitudes
2005
Elicitation is a key task for subjectivist Bayesians. Although skeptics hold that elicitation cannot (or perhaps should not) be done, in practice it brings statisticians closer to their clients and subject-matter expert colleagues. This article reviews the state of the art, reflecting the experience of statisticians informed by the fruits of a long line of psychological research into how people represent uncertain information cognitively and how they respond to questions about that information. In a discussion of the elicitation process, the first issue to address is what it means for an elicitation to be successful; that is, what criteria should be used. Our answer is that a successful elicitation faithfully represents the opinion of the person being elicited. It is not necessarily \"true\" in some objectivistic sense, and cannot be judged in that way. We see that elicitation as simply part of the process of statistical modeling. Indeed, in a hierarchical model at which point the likelihood ends and the prior begins is ambiguous. Thus the same kinds of judgment that inform statistical modeling in general also inform elicitation of prior distributions. The psychological literature suggests that people are prone to certain heuristics and biases in how they respond to situations involving uncertainty. As a result, some of the ways of asking questions about uncertain quantities are preferable to others, and appear to be more reliable. However, data are lacking on exactly how well the various methods work, because it is unclear, other than by asking using an elicitation method, just what the person believes. Consequently, one is reduced to indirect means of assessing elicitation methods. The tool chest of methods is growing. Historically, the first methods involved choosing hyperparameters using conjugate prior families, at a time when these were the only families for which posterior distributions could be computed. Modern computational methods, such as Markov chain Monte Carlo, have freed elicitation from this constraint. As a result, now both parametric and nonparametric methods are available for low-dimensional problems. High-dimensional problems are probably best thought of as lacking another hierarchical level, which has the effect of reducing the as-yet-unelicited parameter space. Special considerations apply to the elicitation of group opinions. Informal methods, such as Delphi, encourage the participants to discuss the issue in the hope of reaching consensus. Formal methods, such as weighted averages or logarithmic opinion pools, each have mathematical characteristics that are uncomfortable. Finally, there is the question of what a group opinion even means, because it is not necessarily the opinion of any participant.
Journal Article
Jaccard/Tanimoto similarity test and estimation methods for biological presence-absence data
2019
Background
A survey of presences and absences of specific species across multiple biogeographic units (or bioregions) are used in a broad area of biological studies from ecology to microbiology. Using binary presence-absence data, we evaluate species co-occurrences that help elucidate relationships among organisms and environments. To summarize similarity between occurrences of species, we routinely use the Jaccard/Tanimoto coefficient, which is the ratio of their intersection to their union. It is natural, then, to identify statistically significant Jaccard/Tanimoto coefficients, which suggest non-random co-occurrences of species. However, statistical hypothesis testing using this similarity coefficient has been seldom used or studied.
Results
We introduce a hypothesis test for similarity for biological presence-absence data, using the Jaccard/Tanimoto coefficient. Several key improvements are presented including unbiased estimation of expectation and centered Jaccard/Tanimoto coefficients, that account for occurrence probabilities. The exact and asymptotic solutions are derived. To overcome a computational burden due to high-dimensionality, we propose the bootstrap and measurement concentration algorithms to efficiently estimate statistical significance of binary similarity. Comprehensive simulation studies demonstrate that our proposed methods produce accurate
p
-values and false discovery rates. The proposed estimation methods are orders of magnitude faster than the exact solution, particularly with an increasing dimensionality. We showcase their applications in evaluating co-occurrences of bird species in 28 islands of Vanuatu and fish species in 3347 freshwater habitats in France. The proposed methods are implemented in an open source R package called
jaccard
(
https://cran.r-project.org/package=jaccard
).
Conclusion
We introduce a suite of statistical methods for the Jaccard/Tanimoto similarity coefficient for binary data, that enable straightforward incorporation of probabilistic measures in analysis for species co-occurrences. Due to their generality, the proposed methods and implementations are applicable to a wide range of binary data arising from genomics, biochemistry, and other areas of science.
Journal Article
Exact p-Values for Network Interference
by
Eckles, Dean
,
Athey, Susan
,
Imbens, Guido W.
in
Artificial
,
Experiments
,
Fisher exact p-values
2018
We study the calculation of exact p-values for a large class of nonsharp null hypotheses about treatment effects in a setting with data from experiments involving members of a single connected network. The class includes null hypotheses that limit the effect of one unit's treatment status on another according to the distance between units, for example, the hypothesis might specify that the treatment status of immediate neighbors has no effect, or that units more than two edges away have no effect. We also consider hypotheses concerning the validity of sparsification of a network (e.g., based on the strength of ties) and hypotheses restricting heterogeneity in peer effects (so that, e.g., only the number or fraction treated among neighboring units matters). Our general approach is to define an artificial experiment, such that the null hypothesis that was not sharp for the original experiment is sharp for the artificial experiment, and such that the randomization analysis for the artificial experiment is validated by the design of the original experiment.
Journal Article
LIKELIHOOD INFERENCE FOR A FRACTIONALLY COINTEGRATED VECTOR AUTOREGRESSIVE MODEL
by
Johansen, Søren
,
Nielsen, Morten Ørregaard
in
Applications
,
Asymptotic value
,
Autoregressive models
2012
We consider model based inference in a fractionally cointegrated (or cofractional) vector autoregressive model, based on the Gaussian likelihood conditional on initial values. We give conditions on the parameters such that the process X t is fractional of order d and cofractional of order d — b; that is, there exist vectors β for which βʹX t is fractional of order d — b and no other fractionality order is possible. For b = 1, the model nests the I(d — 1) vector autoregressive model. We define the statistical model by 0 < b ≤ d, but conduct inference when the true values satisfy 0 ≤ d₀ — b₀ < 1/2 and b₀ ≠ 1/2, for which ${{\\mathrm{\\beta }}^{\\prime }}_{0}{\\mathrm{X}}_{\\mathrm{t}}$ is (asymptotically) a stationary process. Our main technical contribution is the proof of consistency of the maximum likelihood estimators. To this end, we prove weak convergence of the conditional likelihood as a continuous stochastic process in the parameters when errors are independent and identically distributed with suitable moment conditions and initial values are bounded. Because the limit is deterministic, this implies uniform convergence in probability of the conditional likelihood function. If the true value b₀ > 1/2, we prove that the limit distribution of ${\\mathrm{T}}^{{\\mathrm{b}}_{0}}(\\hat{\\mathrm{\\beta }}-{\\mathrm{\\beta }}_{0})$ is mixed Gaussian, while for the remaining parameters it is Gaussian. The limit distribution of the likelihood ratio test for cointegration rank is a functional of fractional Brownian motion of type II. If b₀ < 1/2, all limit distributions are Gaussian or chi-squared. We derive similar results for the model with d = b, allowing for a constant term.
Journal Article