Search Results Heading

MBRLSearchResults

mbrl.module.common.modules.added.book.to.shelf
Title added to your shelf!
View what I already have on My Shelf.
Oops! Something went wrong.
Oops! Something went wrong.
While trying to add the title to your shelf something went wrong :( Kindly try again later!
Are you sure you want to remove the book from the shelf?
Oops! Something went wrong.
Oops! Something went wrong.
While trying to remove the title from your shelf something went wrong :( Kindly try again later!
    Done
    Filters
    Reset
  • Discipline
      Discipline
      Clear All
      Discipline
  • Is Peer Reviewed
      Is Peer Reviewed
      Clear All
      Is Peer Reviewed
  • Item Type
      Item Type
      Clear All
      Item Type
  • Subject
      Subject
      Clear All
      Subject
  • Year
      Year
      Clear All
      From:
      -
      To:
  • More Filters
      More Filters
      Clear All
      More Filters
      Source
    • Language
10,014 result(s) for "threshold models"
Sort by:
The normal law under linear restrictions: simulation and estimation via minimax tilting
Simulation from the truncated multivariate normal distribution in high dimensions is a recurrent problem in statistical computing and is typically only feasible by using approximate Markov chain Monte Carlo sampling. We propose a minimax tilting method for exact independently and identically distributed data simulation from the truncated multivariate normal distribution. The new methodology provides both a method for simulation and an efficient estimator to hitherto intractable Gaussian integrals. We prove that the estimator has a rare vanishing relative error asymptotic property. Numerical experiments suggest that the scheme proposed is accurate in a wide range of set-ups for which competing estimation schemes fail. We give an application to exact independently and identically distributed data simulation from the Bayesian posterior of the probit regression model.
Simultaneous Estimation and Variable Selection for Interval-Censored Data With Broken Adaptive Ridge Regression
The simultaneous estimation and variable selection for Cox model has been discussed by several authors when one observes right-censored failure time data. However, there does not seem to exist an established procedure for interval-censored data, a more general and complex type of failure time data, except two parametric procedures. To address this, we propose a broken adaptive ridge (BAR) regression procedure that combines the strengths of the quadratic regularization and the adaptive weighted bridge shrinkage. In particular, the method allows for the number of covariates to be diverging with the sample size. Under some weak regularity conditions, unlike most of the existing variable selection methods, we establish both the oracle property and the grouping effect of the proposed BAR procedure. An extensive simulation study is conducted and indicates that the proposed approach works well in practical situations and deals with the collinearity problem better than the other oracle-like methods. An application is also provided.
IDENTIFYING THE EFFECT OF CHANGING THE POLICY THRESHOLD IN REGRESSION DISCONTINUITY MODELS
Regression discontinuity models are commonly used to nonparametrically identify and estimate a local average treatment effect (LATE). We show that the derivative of the treatment effect with respect to the running variable at the cutoff, referred to as the treatment effect derivative (TED), is nonparametrically identified, easily estimated, and has implications for testing external validity and extrapolating the estimated LATE away from the cutoff. Given a local policy invariance assumption, we further show this TED equals the change in the treatment effect that would result from a marginal change in the threshold, which we call the marginal threshold treatment effect (MTTE). We apply these results to Goodman (2008), who estimates the effect of a scholarship program on college choice. MTTE in this case identifies how this treatment effect would change if the test score threshold to qualify for a scholarship were changed, even though no such change in threshold is actually observed.
How does artificial intelligence development affect green technology innovation in China? Evidence from dynamic panel data analysis
As the global climate problem becomes increasingly serious, the green technology innovation to achieve “carbon peak and carbon neutral” has gradually become the global consensus of major countries, and how the rapid development of artificial intelligence (AI) technology affects green technology innovation (GTI) has received a great deal of attention in the field of economics. Therefore, based on China’s inter-provincial panel data from 2006 to 2019, the system GMM, dynamic panel threshold model, and quantile regression model were constructed to examine various influences of AI development on GTI under different environmental regulation intensity, research and development (R&D) investment, and institutional environmental threshold conditions. The findings presented that AI development significantly contributes to GTI and GTFP, with an impact coefficient of 0.0122 and 0.0084, and this influence is mainly reflected in the western region of China and is more obvious in the 2006–2012 period. AI development mainly enhances green technological efficiency, and it has dampening effects on green technological progress during the period 2013–2019. Additionally, there are non-linear threshold effects in the relationship between the level of AI development and GTI when environmental regulatory intensity, R&D investment, and institutional environment are in different level intervals. AI development will boost GTI only when the intensity of environmental regulation and institutional environment is above a certain threshold value. However, the AI development represented by industrial robot applications still has no obvious effect on GTI even when the R&D investment exceeds a certain threshold. Furthermore, the growth effect of AI development on GTI indicates a decreasing nonlinear pattern as the GTI’s quantile rises under the condition that R&D investment and institutional environment intensity cross the threshold, while this growth effect increases gradually with the rise of GTI’s quantile when the environmental regulation is above the threshold.
A Sieve Semiparametric Maximum Likelihood Approach for Regression Analysis of Bivariate Interval-Censored Failure Time Data
Interval-censored failure time data arise in a number of fields and many authors have discussed various issues related to their analysis. However, most of the existing methods are for univariate data and there exists only limited research on bivariate data, especially on regression analysis of bivariate interval-censored data. We present a class of semiparametric transformation models for the problem and for inference, a sieve maximum likelihood approach is developed. The model provides a great flexibility, in particular including the commonly used proportional hazards model as a special case, and in the approach, Bernstein polynomials are employed. The strong consistency and asymptotic normality of the resulting estimators of regression parameters are established and furthermore, the estimators are shown to be asymptotically efficient. Extensive simulation studies are conducted and indicate that the proposed method works well for practical situations. Supplementary materials for this article are available online.
Provable sparse tensor decomposition
We propose a novel sparse tensor decomposition method, namely the tensor truncated power method, that incorporates variable selection in the estimation of decomposition components. The sparsity is achieved via an efficient truncation step embedded in the tensor power iteration. Our method applies to a broad family of high dimensional latent variable models, including high dimensional Gaussian mixtures and mixtures of sparse regressions. A thorough theoretical investigation is further conducted. In particular, we show that the final decomposition estimator is guaranteed to achieve a local statistical rate, and we further strengthen it to the global statistical rate by introducing a proper initialization procedure. In high dimensional regimes, the statistical rate obtained significantly improves those shown in the existing nonsparse decomposition methods. The empirical advantages of tensor truncated power are confirmed in extensive simulation results and two real applications of click-through rate prediction and high dimensional gene clustering.
Copula Link-Based Additive Models for Right-Censored Event Time Data
This article proposes an approach to estimate and make inference on the parameters of copula link-based survival models. The methodology allows for the margins to be specified using flexible parametric formulations for time-to-event data, the baseline survival functions to be modeled using monotonic splines, and each parameter of the assumed joint survival distribution to depend on an additive predictor incorporating several types of covariate effects. All the model's coefficients as well as the smoothing parameters associated with the relevant components in the additive predictors are estimated using a carefully structured efficient and stable penalized likelihood algorithm. Some theoretical properties are also discussed. The proposed modeling framework is evaluated in a simulation study and illustrated using a real dataset. The relevant numerical computations can be easily carried out using the freely available GJRM R package. Supplementary materials for this article are available online.
Modelling across extremal dependence classes
Different dependence scenarios can arise in multivariate extremes, entailing careful selection of an appropriate class of models. In bivariate extremes, the variables are either asymptotically dependent or are asymptotically independent. Most available statistical models suit one or other of these cases, but not both, resulting in a stage in the inference that is unaccounted for but can substantially impact subsequent extrapolation. Existing modelling solutions to this problem are either applicable only on subdomains or appeal to multiple limit theories. We introduce a unified representation for bivariate extremes that encompasses a wide variety of dependence scenarios and applies when at least one variable is large. Our representation motivates a parametric model that encompasses both dependence classes. We implement a simple version of this model and show that it performs well in a range of settings.
A Unified Framework for Fitting Bayesian Semiparametric Models to Arbitrarily Censored Survival Data, Including Spatially Referenced Data
A comprehensive, unified approach to modeling arbitrarily censored spatial survival data is presented for the three most commonly used semiparametric models: proportional hazards, proportional odds, and accelerated failure time. Unlike many other approaches, all manner of censored survival times are simultaneously accommodated including uncensored, interval censored, current-status, left and right censored, and mixtures of these. Left-truncated data are also accommodated leading to models for time-dependent covariates. Both georeferenced (location exactly observed) and areally observed (location known up to a geographic unit such as a county) spatial locations are handled; formal variable selection makes model selection especially easy. Model fit is assessed with conditional Cox-Snell residual plots, and model choice is carried out via log pseudo marginal likelihood (LPML) and deviance information criterion (DIC). Baseline survival is modeled with a novel transformed Bernstein polynomial prior. All models are fit via a new function which calls efficient compiled C++ in the R package spBayesSurv . The methodology is broadly illustrated with simulations and real data applications. An important finding is that proportional odds and accelerated failure time models often fit significantly better than the commonly used proportional hazards model. Supplementary materials for this article, including a standardized description of the materials available for reproducing the work, are available as an online supplement.
STRUCTURAL THRESHOLD REGRESSION
This paper introduces the structural threshold regression (STR) model that allows for an endogenous threshold variable as well as for endogenous regressors. This model provides a parsimonious way of modeling nonlinearities and has many potential applications in economics and finance. Our framework can be viewed as a generalization of the simple threshold regression framework of Hansen (2000, Econometrica 68, 575–603) and Caner and Hansen (2004, Econometric Theory 20, 813–843) to allow for the endogeneity of the threshold variable and regime-specific heteroskedasticity. Our estimation of the threshold parameter is based on a two-stage concentrated least squares method that involves an inverse Mills ratio bias correction term in each regime. We derive its asymptotic distribution and propose a method to construct confidence intervals. We also provide inference for the slope parameters based on a generalized method of moments. Finally, we investigate the performance of the asymptotic approximations using a Monte Carlo simulation, which shows the applicability of the method in finite samples.