Svoboda | Graniru | BBC Russia | Golosameriki | Facebook
Skip to main content

Shota Gugushvili

We derive the posterior contraction rate for non-parametric Bayesian estimation of the intensity function of a Poisson point process.
Via a simulation study we compare the finite sample performance of the deconvolution kernel density estimator in the supersmooth deconvolution problem to its asymptotic behaviour predicted by two asymptotic normality theorems. Our results... more
Via a simulation study we compare the finite sample performance of the deconvolution kernel density estimator in the supersmooth deconvolution problem to its asymptotic behaviour predicted by two asymptotic normality theorems. Our results indicate that for lower noise levels and moderate sample sizes the match between the asymptotic theory and the finite sample performance of the estimator is not satisfactory.
Assuming that a stochastic process $X=(X_t)_{t\geq 0}$ is a sum of a compound Poisson process $Y=(Y_t)_{t\geq 0}$ with known intensity $\lambda$ and unknown jump size density $f,$ and an independent Brownian motion $Z=(Z_t)_{t\geq 0},$ we... more
Assuming that a stochastic process $X=(X_t)_{t\geq 0}$ is a sum of a compound Poisson process $Y=(Y_t)_{t\geq 0}$ with known intensity $\lambda$ and unknown jump size density $f,$ and an independent Brownian motion $Z=(Z_t)_{t\geq 0},$ we consider the problem of nonparametric estimation of $f$ from low frequency observations from $X.$ The estimator of $f$ is constructed via Fourier inversion and kernel
Suppose that a compound Poisson process is observed discretely in time and assume that its jump distribution is supported on the set of natural numbers. In this paper we propose a non-parametric Bayesian approach to estimate the intensity... more
Suppose that a compound Poisson process is observed discretely in time and assume that its jump distribution is supported on the set of natural numbers. In this paper we propose a non-parametric Bayesian approach to estimate the intensity of the underlying Poisson process and the distribution of the jumps. We provide a MCMC scheme for obtaining samples from the posterior. We apply our method on both simulated and real data examples, and compare its performance with the frequentist plug-in estimator proposed by Buchmann and Grübel. On a theoretical side, we study the posterior from the frequentist point of view and prove that as the sample size n→∞, it contracts around the `true', data-generating parameters at rate 1/n‾√, up to a logn factor.
We obtain rates of contraction of posterior distributions in inverse problems defined by scales of smoothness classes. We derive abstract results for general priors, with contraction rates determined by Galerkin approximation. The rate... more
We obtain rates of contraction of posterior distributions in inverse problems defined by scales of smoothness classes. We derive abstract results for general priors, with contraction rates determined by Galerkin approximation. The rate depends on the amount of prior concentration near the true function and the prior mass of functions with inferior Galerkin approximation. We apply the general result to non-conjugate series priors, showing that these priors give near optimal and adaptive recovery in some generality, Gaussian priors, and mixtures of Gaussian priors, where the latter are also shown to be near optimal and adaptive. The proofs are based on general testing and approximation arguments, without explicit calculations on the posterior distribution. We are thus not restricted to priors based on the singular value decomposition of the operator. We illustrate the results with examples of inverse problems resulting from differential equations.
We study the problem of non-parametric Bayesian estimation of the intensity function of a Poisson point process. The observations are assumed to be n independent realisations of a Poisson point process on the interval $[0,T]$. We propose... more
We study the problem of non-parametric Bayesian estimation of the intensity function of a Poisson point process. The observations are assumed to be n independent realisations of a Poisson point process on the interval $[0,T]$. We propose two related approaches. In both approaches we model the intensity function as piecewise constant on $N$ bins forming a partition of the interval $[0,T]$. In the first approach the coefficients of the intensity function are assigned independent Gamma priors. This leads to a closed form posterior distribution, for which posterior inference is straightforward to perform in practice, without need to recourse to approximate inference methods. The method scales extremely well with the amount of data. On the theoretical side, we prove that the approach is consistent: as $n\rightarrow\infty$, the posterior distribution asymptotically concentrates around the "true", data-generating intensity function at the rate that is optimal for estimating $h$-H\"{o}lder regular intensity functions $(0 < h \leq 1)$, provided the number of coefficients $N$ of the intensity function grows at a suitable rate depending on the sample size $n$. In the second approach it is assumed that the prior distribution on the coefficients of the intensity function forms a Gamma Markov chain. The posterior distribution is no longer available in closed form, but inference can be performed using a straightforward version of the Gibbs sampler. We show that also this second approach scales well. Practical performance of our methods is first demonstrated via synthetic data examples. It it shown that the second approach depends in a less sensitive way on the choice of the number of bins $N$ and outperforms the first approach in practice. Finally, we analyse three real datasets using our methodology: the UK coal mining disasters data, the US mass shootings data and Donald Trump's Twitter data.
Given discrete time observations over a fixed time interval, we study a nonparametric Bayesian approach to estimation of the volatility coefficient of a stochastic differential equation. We postulate a histogram-type prior on the... more
Given discrete time observations over a fixed time interval, we study a nonparametric Bayesian approach to estimation of the volatility coefficient of a stochastic differential equation. We postulate a histogram-type prior on the volatility with piecewise constant realisations on bins forming a partition of the time interval. The values on the bins are assigned an inverse Gamma Markov chain (IGMC) prior. Posterior inference is straightforward to implement via Gibbs sampling, as the full conditional distributions are available explicitly and turn out to be inverse Gamma. We also discuss in detail the hyperparameter selection for our method. Our nonparametric Bayesian approach leads to good practical results in representative simulation examples. Finally, we apply it on a classical data set in change-point analysis: weekly closings of the Dow-Jones industrial averages.
Given discrete time observations over a growing time interval, we consider a nonparametric Bayesian approach to estimation of the Lévy density of a Lévy process belonging to a flexible class of infinite activity subordinators. Posterior... more
Given discrete time observations over a growing time interval, we consider a nonparametric Bayesian approach to estimation of the Lévy density of a Lévy process belonging to a flexible class of infinite activity subordinators. Posterior inference is performed via MCMC, and we circumvent the problem of the intractable likelihood via the data augmentation device, that in our case relies on bridge process sampling via Gamma process bridges. Our approach also requires the use of a new infinite-dimensional form of a reversible jump MCMC algorithm. We show that our method leads to good practical results in challenging simulation examples. On the theoretical side, we establish that our nonparametric Bayesian procedure is consistent: in the low frequency data setting, with equispaced in time observations and intervals between successive observations remaining fixed, the posterior asymptotically, as the sample size n→∞, concentrates around the Lévy density under which the data have been generated. Finally, we test our method on a classical insurance dataset.
According to both domain expert knowledge and empirical evidence, wavelet coefficients of real signals tend to exhibit clustering patterns, in that they contain connected regions of coefficients of similar magnitude (large or small). A... more
According to both domain expert knowledge and empirical evidence, wavelet coefficients of real signals tend to exhibit clustering patterns, in that they contain connected regions of coefficients of similar magnitude (large or small). A wavelet de-noising approach that takes into account such a feature of the signal may in practice outperform other, more vanilla methods, both in terms of the estimation error and visual appearance of the estimates. Motivated by this observation, we present a Bayesian approach to wavelet de-noising, where dependencies between neighbouring wavelet coefficients are a priori modelled via a Markov chain-based prior, that we term the caravan prior. Posterior computations in our method are performed via the Gibbs sampler. Using representative synthetic and real data examples, we conduct a detailed comparison of our approach with a benchmark empirical Bayes de-noising method (due to Johnstone and Silverman). We show that the caravan prior fares well and is therefore a useful addition to the wavelet de-noising toolbox.
We study a nonparametric Bayesian approach to linear inverse problems under discrete observations. We use the discrete Fourier transform to convert our model into a truncated Gaussian sequence model, that is closely related to the... more
We study a nonparametric Bayesian approach to linear inverse problems under discrete observations. We use the discrete Fourier transform to convert our model into a truncated Gaussian sequence model, that is closely related to the classical Gaussian sequence model. Upon placing the truncated series prior on the unknown parameter, we show that as the number of observations $n\rightarrow\infty$ the corresponding posterior distribution contracts around the true parameter at a rate depending on the smoothness of the true parameter and the prior, and the ill-posedness degree of the problem. Correct combinations of these values lead to optimal posterior contraction rates (up to logarithmic factors). Similarly, the frequentist coverage of Bayesian credible sets is shown to be dependent on a combination of smoothness of the true parameter and the prior, and the ill-posedness of the problem. Oversmoothing priors lead to zero coverage, while undersmoothing priors produce highly conservative results. Finally, we illustrate our theoretical results by numerical examples.
Research Interests:
Let $X_1,\ldots,X_n$ be i.i.d. copies of a random variable $X = Y + Z$, where $X_i = Y_i + Z_i$, and $Y_i$ and $Z_i$ are independent and have the same distribution as $Y$ and $Z$, respectively. Assume that the random variables $Y_i$'s are... more
Let $X_1,\ldots,X_n$ be i.i.d. copies of a random variable $X = Y + Z$, where $X_i = Y_i + Z_i$, and $Y_i$ and $Z_i$ are independent and have the same distribution as $Y$ and $Z$, respectively. Assume that the random variables $Y_i$'s are unobservable and that $Y = AV$, where $A$ and $V$ are independent, $A$ has a Bernoulli distribution with probability of success equal to $1 − p$ and $V$ has a distribution function $F$ with density $f$. Let the random variable $Z$ have a known distribution with density $k$. Based on a sample $X_1,\ldots,X_n$, we consider the problem of nonparametric estimation of the density $f$ and the probability $p$. Our estimators of $f$ and $p$ are constructed via Fourier inversion and kernel smoothing. We derive their convergence rates over suitable functional classes. By establishing in a number of cases the lower bounds for estimation of $f$ and $p$ we show that our estimators are rate-optimal in these cases.
Research Interests:
Given a sample from a discretely observed Lévy process $X = (X_t)_{t\geq 0}$ of the finite jump activity, the problem of nonparametric estimation of the Lévy density $\rho$ corresponding to the process $X$ is studied. An estimator of... more
Given a sample from a discretely observed Lévy process $X = (X_t)_{t\geq 0}$ of the finite jump activity, the problem of nonparametric estimation of the Lévy density $\rho$ corresponding to the process $X$ is studied. An estimator of $\rho$ is proposed that is based on a suitable inversion of the L\'evy-Khintchine formula and a plug-in device. The main results of the paper deal with upper risk bounds for estimation of $\rho$ over suitable classes of L\'evy triplets. The corresponding lower bounds are also discussed.
Research Interests:
Given a discrete time sample $X_1,\ldots,X_n$ from a L\'evy process $X = (X_t)_{t\geq 0}$ of a finite jump activity, we study the problem of nonpara-metric estimation of the characteristic triplet $\gamma,\sigma^2,\rho$ corresponding to... more
Given a discrete time sample $X_1,\ldots,X_n$ from a L\'evy process $X = (X_t)_{t\geq 0}$ of a finite jump activity, we study the problem of nonpara-metric estimation of the characteristic triplet $\gamma,\sigma^2,\rho$ corresponding to the process $X$. Based on Fourier inversion and kernel smoothing, we propose estimators of $\gamma$, $\sigma^2$ and $\rho$ and study their asymptotic behaviour. The obtained results include derivation of upper bounds on the mean square error of the estimators of $\gamma$ and $\sigma^2$ and an upper bound on the mean integrated square error of an estimator of $\rho$.
Research Interests:
We derive the asymptotic distribution of the supremum distance of the deconvolution kernel density estimator to its expectation for certain supersmooth deconvolution problems. It turns out that the asymptotics are essentially different... more
We derive the asymptotic distribution of the supremum distance of the deconvolution kernel density estimator to its expectation for certain supersmooth deconvolution problems. It turns out that the asymptotics are essentially different from the corresponding results for ordinary smooth deconvolution.
Research Interests:
We consider the mean-variance hedging problem in the discrete time setting. Using the dynamic programming approach we obtain recurrent equations for an optimal strategy. Additionally, some technical restrictions of the previous works are... more
We consider the mean-variance hedging problem in the discrete time setting. Using the dynamic programming approach we obtain recurrent equations for an optimal strategy. Additionally, some technical restrictions of the previous works are removed.
Research Interests:
Given a sample from a discretely observed multidimensional compound Poisson process, we study the problem of nonparametric estimation of its jump size density $r_0$ and intensity $λ_0$. We take a nonparametric Bayesian approach to the... more
Given a sample from a discretely observed multidimensional compound Poisson process, we study the problem of nonparametric estimation of its jump size density $r_0$ and intensity $λ_0$. We take a nonparametric Bayesian approach to the problem and determine posterior contraction rates in this context, which, under some assumptions, we argue to be optimal posterior contraction rates. In particular, our results imply the existence of Bayesian point estimates that converge to the true parameter pair $(r_0 , λ_0)$ at these rates. To the best of our knowledge, construction of nonparametric density estimators for inference in the class of discretely observed multidimensional L\'evy processes, and the study of their rates of convergence is a new contribution to the literature.
Research Interests:
Given a sample from a discretely observed compound Poisson process, we consider non-parametric estimation of the density $f_0$ of its jump sizes, as well as of its intensity $λ_0$. We take a Bayesian approach to the problem and specify... more
Given a sample from a discretely observed compound Poisson process, we consider non-parametric estimation of the density $f_0$ of its jump sizes, as well as of its intensity $λ_0$. We take a Bayesian approach to the problem and specify the prior on $f_0$ as the Dirichlet location mixture of normal densities. An independent prior for $λ_0$ is assumed to be compactly supported and to possess a positive density with respect to the Lebesgue measure. We show that under suitable assumptions the posterior contracts around the pair $(λ_0 , f_0)$ at essentially (up to a logarithmic factor) the $\sqrt{n}$-rate, where $n$ is the number of observations and is the mesh size at which the process is sampled. The emphasis is on high frequency data $\Delta\rightarrow 0$, but the obtained results are also valid for fixed $\Delta$. In either case we assume that $n\Delta\rightarrow\infty$. Our main result implies existence of Bayesian point estimates converging (in the frequentist sense, in probability) to $(λ_0 , f_0)$ at the same rate. We also discuss a practical implementation of our approach. The computational problem is dealt with by inclusion of auxiliary variables and we develop a Markov chain Monte Carlo algorithm that samples from the joint distribution of the unknown parameters in the mixture density and the introduced auxiliary variables. Numerical examples illustrate the feasibility of this approach.
Research Interests:
We study the problem of parameter estimation for a univariate discretely observed ergodic diffusion process given as a solution to a stochastic differential equation. The estimation procedure we propose consists of two steps. In the first... more
We study the problem of parameter estimation for a univariate discretely observed ergodic diffusion process given as a solution to a stochastic differential equation. The estimation procedure we propose consists of two steps. In the first step, which is referred to as a smoothing step, we smooth the data and construct a nonparametric estimator of the invariant density of the process. In the second step, which is referred to as a matching step, we exploit a characterisation of the invariant density as a solution of a certain ordinary differential equation, replace the invariant density in this equation by its nonparametric estimator from the smoothing step in order to arrive at an intuitively appealing criterion function, and next define our estimator of the parameter of interest as a minimiser of this criterion function. Our main results show that under suitable conditions our esti-mator is $\sqrt{n}$-consistent, and even asymptotically normal. We also discuss a way of improving its asymptotic performance through a one-step Newton-Raphson type procedure and present results of a small scale simulation study.
Research Interests:
We consider the problem of parameter estimation for a system of ordinary differential equations from noisy observations on a solution of the system. In case the system is nonlinear, as it typically is in practical applications, an... more
We consider the problem of parameter estimation for a system of ordinary differential equations from noisy observations on a solution of the system. In case the system is nonlinear, as it typically is in practical applications, an analytic solution to it usually does not exist. Consequently, straightforward estimation methods like the ordinary least squares method depend on repetitive use of numerical integration in order to determine the solution of the system for each of the parameter values considered, and to find subsequently the parameter estimate that minimises the objective function. This induces a huge computational load to such estimation methods. We study the consistency of an alternative estimator that is defined as a minimiser of an appropriate distance between a nonparametrically estimated derivative of the solution and the right-hand side of the system applied to a nonparametrically estimated solution. This smooth and match estimator (SME) bypasses numerical integration altogether and reduces the amount of computational time drastically compared to ordinary least squares. Moreover, we show that under suitable regularity conditions this smooth and match estimation procedure leads to a $\sqrt{n}$-consistent estimator of the parameter of interest.
Research Interests:
Let $X_1,\ldots,X_n$ be i.i.d. observations, where $X_i = Y_i + \sigma Z_i$ and $Y_i$ and $Z_i$ are independent. Assume that unobservable $Y$'s are distributed as a random variable $UV$, where $U$ and $V$ are independent, $U$ has a... more
Let $X_1,\ldots,X_n$ be i.i.d. observations, where $X_i = Y_i + \sigma Z_i$ and $Y_i$ and $Z_i$ are independent. Assume that unobservable $Y$'s are distributed as a random variable $UV$, where $U$ and $V$ are independent, $U$ has a Bernoulli distribution with probability of zero equal to $p$ and $V$ has a distribution function $F$ with density $f$. Furthermore, let the random variables $Z_i$ have the standard normal distribution and let $\sigma > 0$. Based on a sample $X_1,\ldots,X_n$, we consider the problem of estimation of the density $f$ and the probability $p$. We propose a kernel type deconvolution estimator for $f$ and derive its asymptotic normality at a fixed point. A consistent estimator for $p$ is given as well. Our results demonstrate that our estimator behaves very much like the kernel type deconvolution estimator in the classical deconvolution problem.
Research Interests:
Given a sample from a discretely observed compound Poisson process, we consider estimation of the density of the jump sizes. We propose a kernel type nonparametric density estimator and study its asymptotic properties. An order bound for... more
Given a sample from a discretely observed compound Poisson process, we consider estimation of the density of the jump sizes. We propose a kernel type nonparametric density estimator and study its asymptotic properties. An order bound for the bias and an asymptotic expansion of the variance of the estimator are given. Pointwise weak consistency and asymptotic normality are established. The results show that, asymptotically, the estimator behaves very much like an ordinary kernel estimator.
Research Interests:
ABSTRACT We consider non-parametric Bayesian estimation of the drift coefficient of a one-dimensional stochastic differential equation from discrete-time observations on the solution of this equation. Under suitable regularity conditions... more
ABSTRACT We consider non-parametric Bayesian estimation of the drift coefficient of a one-dimensional stochastic differential equation from discrete-time observations on the solution of this equation. Under suitable regularity conditions that are weaker than those previosly suggested in the literature, we establish posterior consistency in this context. Furthermore, we show that posterior consistency extends to the multidimensional setting as well, which, to the best of our knowledge, is a new result in this setting.
We derive the posterior contraction rate for non-parametric Bayesian estimation of the intensity function of a Poisson point process.
ABSTRACT We establish posterior consistency for non-parametric Bayesian estimation of the dispersion coefficient of a time-inhomogeneous Brownian motion.
Research Interests:
Research Interests:
Research Interests:
Research Interests:
Research Interests:
We consider a nonparametric Bayesian approach to estimate the diffusion coefficient of a stochastic differential equation given discrete time observations over a fixed time interval. As a prior on the diffusion coefficient, we employ a... more
We consider a nonparametric Bayesian approach to estimate the diffusion coefficient of a stochastic differential equation given discrete time observations over a fixed time interval. As a prior on the diffusion coefficient, we employ a histogram-type prior with piecewise constant realisations on bins forming a partition of the time interval. Specifically, these constants are realizations of independent inverse Gamma distributed randoma variables. We justify our approach by deriving the rate at which the corresponding posterior distribution asymptotically concentrates around the data-generating diffusion coefficient. This posterior contraction rate turns out to be optimal for estimation of a Hölder-continuous diffusion coefficient with smoothness parameter 0 < λ ≤ 1. Our approach is straightforward to implement, as the posterior distributions turn out to be inverse Gamma again, and leads to good practical results in a wide range of simulation examples. Finally, we apply our method on exchange rate data sets.
Research Interests: