Questions tagged [approximation]
Approximations to distributions, functions, or other mathematical objects. To approximate something means to find some representation of it which is simpler in some respect, but not exact.
166 questions with no upvoted or accepted answers
7
votes
1
answer
156
views
Expected value of a "logistic uniform" multivariate
Let $\mathbf{a}_1,\ldots,\mathbf{a}_n \in \mathbb R^d$ and $b_1,\ldots,b_n \in \mathbb R$ be fixed. For $\mathbf{x} \sim \mathcal U([0,1]^d)$ and $j \in \{1,\ldots,n\}$, consider the real variable ...
7
votes
0
answers
140
views
What bounds can we place on approximation error for a moment-matching approximation with $N$ moments?
Suppose I have a distribution over the real line ($p$) and I'm approximating it by matching its first $N$ moments. What can I say about the approximation error as a function of $N$?
Alternatively, ...
7
votes
0
answers
2k
views
Faster computation of high-dimensional multivariate normal probabilities
My goal is to find a faster way to calculate something like
mvtnorm::pmvnorm(upper = rep(1,100))
that is, the tail probability of multivariate normal distribution ...
6
votes
0
answers
248
views
Nystrom approximation with inexact/stochastic kernel evaluation
Suppose we have several data points $x_1,\ldots,x_m\in\mathbb R^n$ as well as a positive definite kernel $K(x,y):\mathbb R^n\times\mathbb R^n\to\mathbb R$ that can be written in the form $$K(x,y)=\...
5
votes
0
answers
2k
views
Standard normal quantile approximation
In the book Asymptotic theory of statistics and probability by Anirban DasGupta (2008, Springer Science & Business Media) in page 109 Example 8.13 I found the following approximation
$$\Phi^{-1}\...
5
votes
0
answers
3k
views
Multivariate Normal Orthant Probability
For bivariate zero-mean normal distribution $P(x_1,x_2)$, the quadrant probability is defined as $P(x_1>0,x_2>0)$ or $P(x_1<0,x_2<0).$
$P(x_1>0,x_2>0) = \frac{1}{4}+\frac{sin^{-1}(\...
5
votes
0
answers
83
views
Are there any available implementations of density or conditional density tree learning?
I am working on joint and conditional density trees for approximating clique potentials in Bayesian Belief Networks. A brief introduction to topic is available from this paper in case you'd like to ...
4
votes
0
answers
970
views
Can GMM approximate any given probability density function?
I am currently studying on Bayesian models, and still new to probability theory.
I learned that Gaussian Mixture Model is used to represent the distribution of given population as a weighted sum of ...
4
votes
0
answers
570
views
Convergence of approximate Gibbs sampling
We have a bivariate random variable $(X,Y)$ for which sampling is challenging.
If we were to know how to sample from the conditionals $(X|Y)$ and $(Y|X)$, we could get samples from the joint using ...
4
votes
0
answers
795
views
Quantile approximation using Cornish-Fisher expansion
I am trying to approximate a set of quantiles from the estimated mean, variance, skewness and kurtosis of a random variable with unknown distribution. I tried to apply the Cornish-Fisher expansion of ...
4
votes
0
answers
189
views
Likelihood Function for Complicated Transformations
Suppose that data X have a Normal distribution with some mean $\mu$ and some variance $\sigma^2$. However, you don't get to see X. Instead, you see $Y = g(X)$ where $g$ is a known function. Assume ...
4
votes
0
answers
119
views
Bayesian inference with the wrong distribution
When an observation $x$ is generated by $P(x|\theta)$ for a parameter $\theta$ the Bayesian optimal estimator for the value of $\theta$ is $\hat\theta_{BEST}=\mathbb{E}[\theta|x]=\frac{1}{P(x)}\int d\...
4
votes
0
answers
122
views
Combining Deterministic and Random Unbiased Estimators
I am trying to compute an expectation $E[f(X;\theta,n)]$ where $\theta$ and $n$ are known parameters. I have an easy-to-compute deterministic function $\tilde{f}(\theta,n)$ that provides an ...
3
votes
1
answer
141
views
Moments and PDF of solution to random quadratic equation
Consider the following random quadratic equation,
$$
x^2 + Z x + Y = 0,
$$
where,
$$
\begin{gathered}
Z \sim \mathcal{N}(\mu_Z,\sigma_Z),
\qquad
Y \sim \mathcal{N}(\mu_Y,\sigma_Y).
\end{gathered}
$$
...
3
votes
0
answers
147
views
Understanding the ridge leverage scores sampling from an arXiv paper
I give a try to read the arXiv paper Distributed Adaptive Sampling for Kernel Matrix Approximation, Calandriello et al. 2017. I got a code implementation where they compute ridge leverage scores ...
3
votes
0
answers
170
views
Number of points a one hidden layer neural-network can interpolate
We am trying to understand the number of points that a neural network of a particular size can interpolate. I think this may be isomorphic to its degree of freedom? We are not interested in whether ...
3
votes
0
answers
76
views
Choosing training data inputs to optimize approximation
Suppose you have a smooth function $f^*:D_1 \times D_2\rightarrow\mathbb{R}$ that you observe with error as $f$ such that $$f(x,y)=f^*(x,y)+\epsilon$$ where $\epsilon$ has zero expectation (you can ...
3
votes
0
answers
76
views
Reducing a logistic model used for prediction
I'm developing a logistic regression used for prediction. I have pre-selected, based on prev. literature, 15 candidate predictors (fitting my ~200 events).
Now, I want a reduced/more parsimonious ...
3
votes
0
answers
476
views
Testing a low rank estimator of a covariance matrix
I am exploring ways to reduce the noise of a covariance matrix estimator when the number of variables is greater than the number of observations, i.e. $n > t$.
First, I tried using a low rank ...
3
votes
1
answer
171
views
Universal approximation capability of neural networks with random weights
There is a ton of literature (see, for example, a highly cited paper by Huang et al. (2006)) on neural networks with random weights (NNRWs), i.e. neural networks whose weights are random except for ...
3
votes
0
answers
838
views
Approximating Uniform Distribution with Mixture of Gaussians
Let $T$ be a compact, connected, proper subset of $\mathbb{R}^3:\quad T \subset \mathbb{R}^3$.
Further let $\left\{ \boldsymbol{\mu}_i \right\}_{i=1}^n$ be a given finite set of $n$ point in $T$:
$$
\...
3
votes
1
answer
73
views
Using clusters to estimate model variance
I am working with a blackbox prediction model which takes known inputs and outputs a single mean response. I know this model's residuals to be heteroskedastic, but also can assume the error term of ...
3
votes
0
answers
143
views
What is the geometric mean of the first hitting time distribution of Wiener process?
I'm looking for an analytic formula. Approximate formulas are welcome, in which case I give more importance to simple and nice expressions rather than to precision of the approximation.
I'm looking ...
3
votes
0
answers
167
views
Theoretical properties of Gaussian Process Emulator
I am studying Guassian Prcess Emulator (GPE) to approximate computationally expensive computer models. Basically, we suppose the computer model, or simulator, is denoted by $f(x)$, where $x$ is the ...
3
votes
0
answers
318
views
Classification and regression tree (CART) on large data set
I am trying to approximate a multivariate function $y = f(x_1, ...x_n)$, which I have reason to believe will be well approximated by a classification and regression tree. Some of the variables are ...
3
votes
0
answers
167
views
bound on expectation of a two-variable function under an independent distribution
Consider a probability distribution $P(x)$, a set observed samples $S = \{x_1,\cdots, x_n\}$ where $x_i \stackrel{iid}{\sim} P(x)$ for $i \leq n$, and a symmetric function $h(x,y)$.
How can one ...
3
votes
0
answers
64
views
Error bounds when approximating densities
I was curious whether it is possible to obtain approximation error bounds on continuous densities from approximation error bounds of random variables.
To make my question more precise: We consider ...
3
votes
0
answers
199
views
Results on continuity corrections
For me, the first thing that comes to mind when I come across the term continuity correction is that when $X\sim\mathrm{Bin}(n,p)$, one approximates $\Pr(X\le x)=\Pr(X<x)$ by $\Pr(Y<x+1/2)$ ...
3
votes
0
answers
4k
views
Accurate estimates of the variance of maximum likelihood estimators
According to various sources, the variance of ML estimators can be obtained from the Hessian matrix of the likelihood function. If $H$ is the Hessian of the negative log-likelihood function, then $H^{-...
3
votes
0
answers
692
views
What is a "polynomially bounded" function, and why is this a requirement of the The Delta Method?
I am reading a paper "A note on the Delta Method" by Gary Oehlert, JASA, 1992.
I am trying to estimate the variance of a function of a random variable, but first I want to understand the limitations ...
2
votes
0
answers
54
views
Any way I can practice big O notations for probability and statistics?
I have been using Taylor expansion to get some feelings and many approximating results in trying to find innovative ideas for my research. And I have seen a lot of approximate equals or asymptotically ...
2
votes
1
answer
79
views
Differential Privacy guarantee that takes into account the approximate density (e.g., the pseudo randomness) used in practice?
In theory the differential privacy guarantee comes from adding randomness to an algorithm so whatever is output is a sample from a target distribution (e.g., the Laplacian, Gaussian, Exponential ...
2
votes
0
answers
79
views
What are effective methods to maximize an unknown noisy function?
I have a function that takes a few hundred parameters and which returns a score I want to optimize for - It's a piece of software attempting to play a game against another player. The parameters ...
2
votes
1
answer
143
views
Is there a way to correct for degrees of freedom when using a generalized linear model with a Poisson distribution featuring random effects?
I am running a generalized linear mixed effect model with a Poisson distribution to analyse count data. The model has a random effect that takes into account multiple observation obtained by the same ...
2
votes
0
answers
188
views
approximate fisher information for intractable likelihoods
Suppose I have a data set $X_1, \ldots, X_n$, and from that I compute a statistic $T(X_1, \ldots, X_n) := T$. I want to assess how reactive/sensitive this calculation is to changes in parameter values....
2
votes
0
answers
157
views
How do I approximate a multivariable polynomial equation using Neural Networks?
I've been trying to experiment and test the extents to which a neural network works.
I was only able to make something with broad categorical variables function in an acceptable amount of time and in ...
2
votes
0
answers
146
views
Approximate a distribution as mixture of $K$ other (known, fixed) distributions
I'd like to draw samples from some "target" probability density function $f(x)$. However, I don't have a way to do that -- instead I just have access to $N$ samples, each drawn from one of $...
2
votes
0
answers
250
views
Why does Kullback–Leibler divergence measure information loss when approximating a probability distribution?
I've encountered a sentence:
In information theory, Kullback–Leibler divergence is regarded as a measure of the information lost when probability distribution Q is used to approximate a true ...
2
votes
0
answers
144
views
Quantifying the universal approximation theorem
Let $m\geq 1$ be an integer and $F\in \mathbb{R}[x_1, \dots, x_m]$ be a polynomial. I want to approximate $F$ on the unit hypercube $[0, 1]^m$ by a (possibly multilayer) feedforward neural network. ...
2
votes
0
answers
209
views
Approximate the mean area of 2D Voronoi cell
Consider a random uniform distribution of $N$ points in $2D$ space bounded by $[0, 1]$ in both dimensions. Example:
If I want to estimate the mean area of their Voronoi cells, I have to obtain the ...
2
votes
0
answers
995
views
what is the probability of detecting departure from H0?
The desired percentage of SiO$_2$ in a certain type of aluminous cement is 5.5. To test whether the true average percentage is 5.5 for a particular production facility, 16 independently obtained ...
2
votes
0
answers
68
views
Asymptotic approximation of log-probability using first four moments
Consider a random variable $X \sim p_{n,\theta}$ where the first four moments are given by known functions:
$$\begin{matrix}
\ \ \ \ \ \ \mathbb{E}(X) \equiv \mu(n,\theta) & & & \ \ \ \ \ ...
2
votes
0
answers
82
views
What is this approximation called?
In Bayesian statistics, you have a likelihood and a prior, $f(x_1,\ldots,x_n \mid \theta)$ and $\pi(\theta)$ respectively, and you use these to obtain the posterior $\pi(\theta \mid x_1, \ldots, x_n) \...
2
votes
0
answers
134
views
Numerical method to compress empirical probability distribution
I am trying to grapple with the following problem.
I have an application that develops empirical distributions. In essence, I end up with a histogram of equally spaced $x$ values, with both a $max$...
2
votes
0
answers
812
views
Understanding a Taylor expansion for the bias of local polynomial regression
I'm trying to understand the proof of an expression for the asymptotic bias in local polynomial regression of degree $p\ge0$.
Specifically, I'm distraught with equation $(3.59)$ on page 102 of this ...
2
votes
0
answers
52
views
How well can an AR(p) process model any given stationary time series?
Are there any theorems which tell us how well AR(p) models are able to approximate any stationary finite time series? If so, what are the relevant results?
2
votes
0
answers
186
views
Is vcov the information matrix /n from the theory?
If I fit a model
m <- glm/gam/gamm/lme/whatever(y ~ x + z, family = some exponential family)
and extract
coef(m)
vcov(m)
...
2
votes
0
answers
67
views
When is it possible to estimate the non-linearity error when approximating data with a linear model?
The most common form of linear regression estimates the best values of $\vec{\beta}$ and $\sigma^2$ assuming that data is sampled from a model $y = \vec{\beta} \cdot \vec{x} + \vec{\epsilon}$ where $\...
2
votes
0
answers
159
views
Training Restricted Boltzmann Machines according to the Likelihood Function
Is it possible to chose the parameters of a RBM to maximize the likelihood of the observed data?
(I follow the notation of the deeplearning tutorial ). Denote the observable data by $x$, hidden data ...
2
votes
0
answers
2k
views
wald test and score test, normal or chi square?
I learnt from section 10.3 of statistical inference that both Wald test statistic $\frac{W_n-\theta_0}{S_n}\approx\frac{W_n-\theta_0}{\sqrt{\hat I_n(W_n)}}$ and score test statistic $\frac{S(\theta_0)}...