Universite Paris-Dauphine
This note is a discussion of the paper "Confidence distribution" by Min-ge Xie and Kesar Singh, to appear in the International Statistical Review.
This note is a discussion commenting on the paper by Ly et al. on "Harold Jeffreys's Default Bayes Factor Hypothesis Tests: Explanation, Extension, and Application in Psychology" and on the perceived shortcomings of the classical Bayesian approach to testing, while reporting on an alternative approach advanced by Kamary, Mengersen, Robert and Rousseau (2014. arXiv:1412.2044) as a solution to this quintessential inference problem.
We present the particle method for simulating the solution to the path-dependent McKean-Vlasov equation, in which both the drift and the diffusion coefficients depend on the whole trajectory of the process up to the current time t, as well as on the corresponding marginal distributions. Our paper establishes an explicit convergence rate for this numerical approach. We illustrate our findings with numerical simulations of a modified Ornstein-Uhlenbeck process with memory, and of an extension of the Jansen-Rit mean-field model for neural mass.
We study linear and non-linear small divisors problems in analytic and non-analytic regularity. We observe that the Bruno arithmetic condition, which is usually attached to non-linear analytic problems, can also be characterized as the optimal condition to solve the linear problem in some fixed non quasi-analytic class. Based on this observation, it is natural to conjecture that the optimal arithmetic condition for the linear problem is also optimal for non-linear small divisors problems in any reasonable non quasi-analytic classes. Our main result proves this conjecture in a representative non-linear problem, which is the linearization of vector fields on the torus, in the most representative non quasi-analytic class, which is the Gevrey class. The proof follows Moser's argument of approximation by analytic functions, and uses in an essential way works of Popov, Rüssmann and Pöschel.
Mean Field Game systems describe equilibrium configurations in differential games with infinitely many infinitesimal interacting agents. We introduce a learning procedure (similar to the Fictitious Play) for these games and show its convergence when the Mean Field Game is potential.
In this paper, we characterize a degenerate PDE as the gradient flow in the space of nonnegative measures endowed with an optimal transport-growth metric. The PDE of concern, of Hele-Shaw type, was introduced by Perthame et. al. as a mechanical model for tumor growth and the metric was introduced recently in several articles as the analogue of the Wasserstein metric for nonnegative measures. We show existence of solutions using minimizing movements and show uniqueness of solutions on convex domains by proving the Evolutional Variational Inequality. Our analysis does not require any regularity assumption on the initial condition. We also derive a numerical scheme based on the discretization of the gradient flow and the idea of entropic regularization. We assess the convergence of the scheme on explicit solutions. In doing this analysis, we prove several new properties of the optimal transport-growth metric, which generally have a known counterpart for the Wasserstein metric.
This paper introduces a novel steepest descent flow in Banach spaces. This extends previous works on generalized gradient descent, notably the work of Charpiat et al., to the setting of Finsler metrics. Such a generalized gradient allows one to take into account a prior on deformations (e.g., piecewise rigid) in order to favor some specific evolutions. We define a Finsler gradient descent method to minimize a functional defined on a Banach space and we prove a convergence theorem for such a method. In particular, we show that the use of non-Hilbertian norms on Banach spaces is useful to study non-convex optimization problems where the geometry of the space might play a crucial role to avoid poor local minima. We show some applications to the curve matching problem. In particular, we characterize piecewise rigid deformations on the space of curves and we study several models to perform piecewise rigid evolution of curves.
We consider a SDE with a smooth multiplicative non-degenerate noise and a possibly unbounded Holder continuous drift term. We prove existence of a global flow of diffeomorphisms by means of a special transformation of the drift of Ito-Tanaka type. The proof requires non-standard elliptic estimates in Holder spaces. As an application of the stochastic flow, we obtain a Bismut-Elworthy-Li type formula for the first derivatives of the associated diffusion semigroup.
Epistemic social choice aims at unveiling a hidden ground truth given votes, which are interpreted as noisy signals about it. We consider here a simple setting where votes consist of approval ballots: each voter approves a set of alternatives which they believe can possibly be the ground truth. Based on the intuitive idea that more reliable votes contain fewer alternatives, we define several noise models that are approval voting variants of the Mallows model. The likelihood-maximizing alternative is then characterized as the winner of a weighted approval rule, where the weight of a ballot decreases with its cardinality. We have conducted an experiment on three image annotation datasets; they conclude that rules based on our noise model outperform standard approval voting; the best performance is obtained by a variant of the Condorcet noise model.
We obtain an improved Bochner inequality based on the curvature-dimension condition RCD(K,N){\rm RCD}^*(K,N) and propose a definition of NN-dimensional Ricci tensor on metric measure spaces.
Low-rank matrix recovery problems are inverse problems which naturally arise in various fields like signal processing, imaging and machine learning. They are non-convex and NP-hard in full generality. It is therefore a delicate problem to design efficient recovery algorithms and to provide rigorous theoretical insights on the behavior of these algorithms. The goal of these notes is to review recent progress in this direction for the class of so-called "non-convex algorithms", with a particular focus on the proof techniques. Although they aim at presenting very recent research works, these notes have been written with the intent to be, as much as possible, accessible to non-specialists.
In recent years, state-of-the-art methods for supervised learning have exploited increasingly gradient boosting techniques, with mainstream efficient implementations such as xgboost or lightgbm. One of the key points in generating proficient methods is Feature Selection (FS). It consists in selecting the right valuable effective features. When facing hundreds of these features, it becomes critical to select best features. While filter and wrappers methods have come to some maturity, embedded methods are truly necessary to find the best features set as they are hybrid methods combining features filtering and wrapping. In this work, we tackle the problem of finding through machine learning best a priori trades from an algorithmic strategy. We derive this new method using coordinate ascent optimization and using block variables. We compare our method to Recursive Feature Elimination (RFE) and Binary Coordinate Ascent (BCA). We show on a real life example the capacity of this method to select good trades a priori. Not only this method outperforms the initial trading strategy as it avoids taking loosing trades, it also surpasses other method, having the smallest feature set and the highest score at the same time. The interest of this method goes beyond this simple trade classification problem as it is a very general method to determine the optimal feature set using some information about features relationship as well as using coordinate ascent optimization.
The marginal likelihood, or evidence, plays a central role in Bayesian model selection, yet remains notoriously challenging to compute in likelihood-free settings. While Simulation-Based Inference (SBI) techniques such as Sequential Neural Likelihood Estimation (SNLE) offer powerful tools to approximate posteriors using neural density estimators, they typically do not provide estimates of the evidence. In this technical report presented at BayesComp 2025, we present a simple and general methodology to estimate the marginal likelihood using the output of SNLE.
We introduce a new class of distances between nonnegative Radon measures in Euclidean spaces. They are modeled on the dynamical characterization of the Kantorovich-Rubinstein-Wasserstein distances proposed by Benamou-Brenier and provide a wide family interpolating between the Wasserstein and the homogeneous (dual) Sobolev distances. From the point of view of optimal transport theory, these distances minimize a dynamical cost to move a given initial distribution of mass to a final configuration. An important difference with the classical setting in mass transport theory is that the cost not only depends on the velocity of the moving particles but also on the densities of the intermediate configurations with respect to a given reference measure. We study the topological and geometric properties of these new distances, comparing them with the notion of weak convergence of measures and the well established Kantorovich-Rubinstein-Wasserstein theory. An example of possible applications to the geometric theory of gradient flows is also given.
We present a generic compact computational framework relying on structured random matrices that can be applied to speed up several machine learning algorithms with almost no loss of accuracy. The applications include new fast LSH-based algorithms, efficient kernel computations via random feature maps, convex optimization algorithms, quantization techniques and many more. Certain models of the presented paradigm are even more compressible since they apply only bit matrices. This makes them suitable for deploying on mobile devices. All our findings come with strong theoretical guarantees. In particular, as a byproduct of the presented techniques and by using relatively new Berry-Esseen-type CLT for random vectors, we give the first theoretical guarantees for one of the most efficient existing LSH algorithms based on the HD3HD2HD1\textbf{HD}_{3}\textbf{HD}_{2}\textbf{HD}_{1} structured matrix ("Practical and Optimal LSH for Angular Distance"). These guarantees as well as theoretical results for other aforementioned applications follow from the same general theoretical principle that we present in the paper. Our structured family contains as special cases all previously considered structured schemes, including the recently introduced PP-model. Experimental evaluation confirms the accuracy and efficiency of TripleSpin matrices.
We consider disordered many-body systems with periodic time-dependent Hamiltonians in one spatial dimension. By studying the properties of the Floquet eigenstates, we identify two distinct phases: (i) a many-body localized (MBL) phase, in which almost all eigenstates have area-law entanglement entropy, and the eigenstate thermalization hypothesis (ETH) is violated, and (ii) a delocalized phase, in which eigenstates have volume-law entanglement and obey the ETH. MBL phase exhibits logarithmic in time growth of entanglement entropy for initial product states, which distinguishes it from the delocalized phase. We propose an effective model of the MBL phase in terms of an extensive number of emergent local integrals of motion (LIOM), which naturally explains the spectral and dynamical properties of this phase. Numerical data, obtained by exact diagonalization and time-evolving block decimation methods, suggests a direct transition between the two phases. Our results show that many-body localization is not destroyed by sufficiently weak periodic driving.
Approximate Bayesian computation (ABC) have become a essential tool for the analysis of complex stochastic models. Earlier, Grelaud et al. (2009) advocated the use of ABC for Bayesian model choice in the specific case of Gibbs random fields, relying on a inter-model sufficiency property to show that the approximation was legitimate. Having implemented ABC-based model choice in a wide range of phylogenetic models in the DIY-ABC software (Cornuet et al., 2008), we now present theoretical background as to why a generic use of ABC for model choice is ungrounded, since it depends on an unknown amount of information loss induced by the use of insufficient summary statistics. The approximation error of the posterior probabilities of the models under comparison may thus be unrelated with the computational effort spent in running an ABC algorithm. We then conclude that additional empirical verifications of the performances of the ABC procedure as those available in DIYABC are necessary to conduct model choice.
We investigate whether having a unique equilibrium (or a given number of equilibria) is robust to perturbation of the payoffs, both for Nash equilibrium and correlated equilibrium. We show that the set of n-player finite games with a unique correlated equilibrium is open, while this is not true of Nash equilibrium for n>2. The crucial lemma is that a unique correlated equilibrium is a quasi-strict Nash equilibrium. Related results are studied. For instance, we show that generic two-person zero-sum games have a unique correlated equilibrium and that, while the set of symmetric bimatrix games with a unique symmetric Nash equilibrium is not open, the set of symmetric bimatrix games with a unique and quasi-strict symmetric Nash equilibrium is.
Following a hedging based approach to model free financial mathematics, we prove that it should be possible to make an arbitrarily large profit by investing in those one-dimensional paths which do not possess local times. The local time is constructed from discrete approximations, and it is shown that it is α\alpha-Hölder continuous for all \alpha<1/2. Additionally, we provide various generalizations of Föllmer's pathwise Itô formula.
In some applied scenarios, the availability of complete data is restricted, often due to privacy concerns; only aggregated, robust and inefficient statistics derived from the data are made accessible. These robust statistics are not sufficient, but they demonstrate reduced sensitivity to outliers and offer enhanced data protection due to their higher breakdown point. We consider a parametric framework and propose a method to sample from the posterior distribution of parameters conditioned on various robust and inefficient statistics: specifically, the pairs (median, MAD) or (median, IQR), or a collection of quantiles. Our approach leverages a Gibbs sampler and simulates latent augmented data, which facilitates simulation from the posterior distribution of parameters belonging to specific families of distributions. A by-product of these samples from the joint posterior distribution of parameters and data given the observed statistics is that we can estimate Bayes factors based on observed statistics via bridge sampling. We validate and outline the limitations of the proposed methods through toy examples and an application to real-world income data.
There are no more papers matching your filters at the moment.