Entropy and divergence (Shannon and Kullback-Leibler) estimation is a central problem in image processing, with many applications for image compression, segmentation, calibration, registration, etc. Mutual information, which is strongly related to Shannon entropy and Kullback-Leibler divergence, is a widely used measure of similarity between images. The entropy of a random variable is the expected information gained by observing a realisation of that random variable. To estimate the entropy of a random vector, the naïve approach is to partition the sample space into a finite number of cells, estimate the density in each cell by the proportion of sample points falling into that cell, then estimate the entropy by that of the associated empirical density function. For non-uniform distributions, fixed partitions lead to low occupancy numbers in sparse regions and poor resolution in dense regions. In a seminal paper, Kozachenko & Leonenko (1987) proposed an alternative approach to the problem of entropy estimation, based on the expected distance between a point and its nearest neighbour in the sample. Nearest neighbour relations define an adaptive partition of the sample space, which allows control of the number of points in each spatial cell, and hence the computation time of the algorithm. This method was generalized on the k-th nearest neighbour statistics in Goria, Leonenko, Mergel and Novi Inverardi (2005). The method permits to work with high dimensional data, that is, to compare images not only in terms of their local pixel intensities or colours, but also to use their spatial characteristics through the properties of neighbouring pixels, which yields in turn much more efficient methods for comparing images.
Relative entropy and mutual information extend the notion of entropy to two or more random variable. Evans (2008a) showed that estimators based only on nearest neighbour relations in the marginal space can be computationally expensive, and shows how computational efficiency can be maintained by considering nearest neighbour relations in the joint probability space. Leonenko et al. (2008, 2010) presented a more general class of estimators for Rényi entropy and divergence, and showed that these estimators satisfy a Strong Law of Large Numbers. Evans (2008b) showed that this also holds for a broad class of nearest-neighbour statistics, see also Wang, Q., Kulkarni, S. R.; Verdú, S. (2009).
Leonenko and Seleznev (2010) (see also Kallberg, Leonenko and Selezven, O. (2014) for an extension for dependent data) proposed a new method of estimation for ε-entropy and quadratic Rényi entropy for both discrete and continuous densities based on the number of coincident (or ε-close) vector observation in the corresponding sample. They developed a consistency and asymptotic distribution theory for this scheme, based on the U-statistics theory.
In a series of papers of Prof. Leonenko and his collaborators ([4,7,9,11]) the analogous of Rényi entropy was constructed and studied in relation to the so-called multifractal analysis, which is very important for applications in turbulence as well as other areas of Physics, where the multifractal behaviour is typical for real data. They have recently developed an envelop a new form of multifractal measures and multifrcatal spectrums and Rényi functions related mainly to log hyperbolic distributions, constructed from multifractal products of stochastic processes. These multifractal measures possess further desirable properties such as a natural form of the singularity spectrum and dependence structure. The new paper of Denisov and Leonenko (2011) contains the complete proof of Rényi functions envelop for a two schemes related to the multifractal products of geometric Ornstein-Uhlenbeck processes driven by Levy noise.
One of the main aims of the project is to develop an asymptotic theory of the nearest neighbor estimates of Shannon and Rényi information, in particular to prove an asymptotic normality using the ideas of the paper [11].
The project also will consider a statistical methods for ε-entropy and quadratic Rényi entropy in the case of dependent data, see [12].
For the recent development, see [13,14].
For further details please contact Prof. Nikolai Leonenko