probability
This paper introduces Provable Diffusion Posterior Sampling (PDPS), a method for Bayesian inverse problems that integrates pre-trained diffusion models as data-driven priors. The approach offers the first non-asymptotic error bounds for diffusion-based posterior score estimation and demonstrates superior performance with reliable uncertainty quantification across various imaging tasks.
We consider a system of multiscale stochastic differential equations whose slow component is drivenby a fractional Brownian motion with Hurst parameter H greater than 1/2. Under ergodic assumptions ensuring the applicability of the fractional averaging and fractional homogenization theorems of Hairer and Li (arXiv:1902.11251, arXiv:2109.06948), we establish a fluctuation result. The deviation of the slow motion, scaled by epsilon^{1/2-H}, from its effective, time-dependent random limit converges, as the time-separation scale epsilon tends to zero, to the solution of a stochastic differential equation driven by a fractional Brownian motion and influenced by an additional space--time Gaussian field. Since the averaging principle and the fractional homogenization hold in different modes of convergence, obtaining the required joint convergence is a delicate matter. Moreover, neither the continuity of the Ito--Lyons solution map nor the martingale method is directly applicable for our purposes, so the proof requires several innovations. To establish the fluctuation theorem, we combine cumulant methods with a residue lemma and formulate the enlarged system as a rough differential equation in a suitable space.
On a transient weighted graph, there are two models of random walk which continue after reaching infinity: random interlacements, and random walk reflected off of infinity, recently introduced in arXiv:2506.18827 [math.PR]. We prove these two models are equivalent if and only if all harmonic functions of the underlying graph with finite Dirichlet energy are constant functions, or equivalently, the free and wired spanning forests coincide. In particular, examples where the models are equivalent include Zd\mathbb{Z}^d, cartesian products, and many Cayley graphs, while examples that fail the condition include all transient trees.
We consider an investor who, while maximizing his/her expected utility, also compares the outcome to a reference entity. We recall the notion of personal equilibrium and show that, in a multistep, generically incomplete financial market model such an equilibrium indeed exists, under appropriate technical assumptions.
We study the LpL_p-discrepancy of random point sets in high dimensions, with emphasis on small values of pp. Although the classical LpL_p-discrepancy suffers from the curse of dimensionality for all p(1,)p \in (1,\infty), the gap between known upper and lower bounds remains substantial, in particular for small p1p \ge 1. To clarify this picture, we review the existing results for i.i.d.\ uniformly distributed points and derive new upper bounds for \emph{generalized} LpL_p-discrepancies, obtained by allowing non-uniform sampling densities and corresponding non-negative quadrature weights. Using the probabilistic method, we show that random points drawn from optimally chosen product densities lead to significantly improved upper bounds. For p=2p=2 these bounds are explicit and optimal; for general p[1,)p \in [1,\infty) we obtain sharp asymptotic estimates. The improvement can be interpreted as a form of importance sampling for the underlying Sobolev space Fd,qF_{d,q}. Our results also reveal that, even with optimal densities, the curse of dimensionality persists for random points when p1p\ge 1, and it becomes most pronounced for small pp. This suggests that the curse should also hold for the classical L1L_1-discrepancy for deterministic point sets.
This research introduces Empirical Decision Theory, a framework that enables principled decision-making based solely on observed act-consequence pairs, bypassing the need for explicit "states of the world." The approach provides statistical inferential guarantees, including robustness against data contamination, and is demonstrated in evaluating generative AI prompting strategies.
1
Hoeffding's Inequality provides the maximum probability that a series of n draws from a bounded random variable differ from the variable's true expectation u by more than given tolerance t. The random variable is typically the error rate of a classifier in machine learning applications. Here, a trading strategy is premised on the assumption of an underlying distribution of causal factors, in other words, a market regime, and the random variable is the performance of that trading strategy. A larger deviation of observed performance from the trader's expectation u can be characterized as a lower probability that the financial regime supporting that strategy remains in force, and a higher probability of financial regime change. The changing Hoeffding probabilities can be used as an early warning indicator of this change.
We consider domino tilings of the Aztec diamond. Using the Domino Shuffling algorithm introduced by Elkies, Kuperberg, Larsen, and Propp in arXiv:math/9201305, we are able to generate domino tilings uniformly at random. In this paper, we investigate the probability of finding a domino at a specific position in such a random tiling. We prove that this placement probability is always equal to 1/41/4 plus a rational function, whose shape depends on the location of the domino, multiplied by a position-independent factor that involves only the size of the diamond. This result leads to significantly more compact explicit counting formulas compared to previous findings. As a direct application, we derive explicit counting formulas for the domino tilings of Aztec diamonds with 2×22\times 2-square holes at arbitrary positions.
True Volterra equations are inherently non stationary and therefore do not admit genuine stationary regimes\textit{genuine stationary regimes} over finite horizons. This motivates the study of the finite-time behavior of the solutions to scaled inhomogeneous affine Stochastic Volterra equations through the lens of a weaker notion of stationarity referred to as fake stationary regime\textit{fake stationary regime} in the sense that all marginal distributions share the same expectation and variance. As a first application, we introduce the Fake stationary Volterra Heston model\textit{Fake stationary Volterra Heston model} and derive a closed-form expression for its characteristic function. Having established this finite-time proxy for stationarity, we then investigate the asymptotic (long-time) behavior to assess whether genuine stationary regimes emerge in the limit. Using an extension of the exponential-affine transformation formula for those processes, we establish in the long run the existence of limiting distributions, which (unlike in the case of classical affine diffusion processes) may depend on the initial state of the process, unless the Volterra kernel coincides with the α\alpha- fractional integration kernel, for which the dependence on the initial state vanishes. We then proceed to the construction of stationary processes associated with these limiting distributions. However, the dynamics in this long-term regime are analytically intractable, and the process itself is not guaranteed to be stationary in the classical sense over finite horizons. This highlights the relevance of finite-time analysis through the lens of the aforementioned fake stationarity\textit{fake stationarity}, which offers a tractable approximation to stationary behavior in genuinely non-stationary Volterra systems.
We introduce a carré du champ operator for Banach-valued random elements, taking values in the projective tensor product, and use it to control the bounded Lipschitz distance between a Malliavin-smooth random element satisfying mild regularity assumptions and a Radon Gaussian taking values in the Skorokhod space equipped with the uniform topology. In the case where the random element is a Banach-valued multiple integral, the carré du champ expression is further bounded by norms of the contracted integral kernel. The main technical tool is an integration by parts formula, which might be of independent interest. As a by-product, we recover a bound obtained recently by Düker and Zoubouloglou in the Hilbert space setting and complement it by providing contraction bounds.
In this paper, we introduce a novel framework using inhomogeneous Branching Random Walks (BRWs) to model growth processes, specifically introducing genealogy-dependence in branching rates and displacement distributions to model phenomena like bacterial colony growth. Current stochastic models often either assume independent and identical behavior of individual agents or incorporate only spatiotemporal inhomogeneity, ignoring the effect of genealogy-based inhomogeneity on the long-time behavior of these processes. Such long-time asymptotics are of independent mathematical interest and are crucial in understanding the effect of patterns. We propose several inhomogeneous BRW models in 2D space where displacement distributions and branching rates vary with time, space, and genealogy. A combined model then uses a weighted average of positions given by these separate models to study the shape of the growth patterns. Using computer simulations, we tune parameters from these models, which are based on genealogical and spatiotemporal factors, observe the resulting structures, and compare them with images of real bacterial colonies.
We develop a general theory of semantic dynamics for large language models by formalizing them as Continuous State Machines (CSMs): smooth dynamical systems whose latent manifolds evolve under probabilistic transition operators. The associated transfer operator P:L2(M,μ)L2(M,μ)P: L^2(M,\mu) \to L^2(M,\mu) encodes the propagation of semantic mass. Under mild regularity assumptions (compactness, ergodicity, bounded Jacobian), PP is compact with discrete spectrum. Within this setting, we prove the Semantic Characterization Theorem (SCT): the leading eigenfunctions of PP induce finitely many spectral basins of invariant meaning, each definable in an o-minimal structure over R\mathbb{R}. Thus spectral lumpability and logical tameness coincide. This explains how discrete symbolic semantics can emerge from continuous computation: the continuous activation manifold collapses into a finite, logically interpretable ontology. We further extend the SCT to stochastic and adiabatic (time-inhomogeneous) settings, showing that slowly drifting kernels preserve compactness, spectral coherence, and basin structure.
Researchers at ETH Zurich and the Chinese Academy of Sciences developed DeepSKA, a neural framework that provides interpretable and reliable estimation of expected outputs for Stochastic Reaction Networks (SRNs). This method combines spectral decomposition-based neural networks with hybrid Deep Learning/Monte Carlo estimators, achieving unbiased and provably convergent results while reducing variance up to 10,000-fold compared to standard simulations.
3
We frame novelty detection on path space as a hypothesis testing problem with signature-based test statistics. Using transportation-cost inequalities of Gasteratos and Jacquier (2023), we obtain tail bounds for false positive rates that extend beyond Gaussian measures to laws of RDE solutions with smooth bounded vector fields, yielding estimates of quantiles and p-values. Exploiting the shuffle product, we derive exact formulae for smooth surrogates of conditional value-at-risk (CVaR) in terms of expected signatures, leading to new one-class SVM algorithms optimising smooth CVaR objectives. We then establish lower bounds on type-II\mathrm{II} error for alternatives with finite first moment, giving general power bounds when the reference measure and the alternative are absolutely continuous with respect to each other. Finally, we evaluate numerically the type-I\mathrm{I} error and statistical power of signature-based test statistic, using synthetic anomalous diffusion data and real-world molecular biology data.
Researchers at MIT rigorously analyze the mean-field dynamics of Transformer attention, framing it as an interacting particle system that leads to eventual representation collapse but also exhibits long-lived metastable multi-cluster states. The work quantitatively explains the benefits of Pre-Layer Normalization in delaying collapse and identifies a phase transition in long-context attention with logarithmic scaling.
We introduce a compact probabilistic model for two-player and two-team (four-player) squash matches, along with a practical skill-comparison rule derived from point-scoring probabilities. Using recorded shot types and court locations, we analyze how shot distributions differ between professional-level and intermediate-level players. Our analysis shows that professional players use a wider variety of shots and favor backcourt play to maintain control, while intermediate players concentrate more on mid-court shots, generate more errors, and exercise less positional control. These results quantify strategic differences in squash, offer a simple method to compare player and team skill, and provide actionable insights for sports analytics and coaching.
In this paper, we investigate a second-order stochastic algorithm for solving large-scale binary classification problems. We propose to make use of a new hybrid stochastic Newton algorithm that includes two weighted components in the Hessian matrix estimation: the first one coming from the natural Hessian estimate and the second associated with the stochastic gradient information. Our motivation comes from the fact that both parts evaluated at the true parameter of logistic regression, are equal to the Hessian matrix. This new formulation has several advantages and it enables us to prove the almost sure convergence of our stochastic algorithm to the true parameter. Moreover, we significantly improve the almost sure rate of convergence to the Hessian matrix. Furthermore, we establish the central limit theorem for our hybrid stochastic Newton algorithm. Finally, we show a surprising result on the almost sure convergence of the cumulative excess risk.
We consider the problem of approximating the Langevin dynamics of inertial particles being transported by a background flow. In particular, we study an acceleration corrected advection-diffusion approximation to the Langevin dynamics, a popular approximation in the study of turbulent transport. We prove error estimates in the averaging regime in which the dimensionless relaxation timescale ε\varepsilon is the small parameter. We show that for any finite time interval, the approximation error is of order O(ε)\mathcal{O}(\varepsilon) in the strong sense and O(ε2)\mathcal{O}(\varepsilon^2) in the weak sense, whose optimality is checked against computational experiment. Furthermore, we present numerical evidence suggesting that this approximation also captures the long-time behavior of the Langevin dynamics.
We introduce and analyze a novel class of inverse problems for stochastic dynamics: Given the ergodic invariant measure of a stochastic process governed by a nonlinear stochastic ordinary or partial differential equation (SODE or SPDE), we investigate the unique identifiability of the underlying process--specifically, the recovery of its drift and diffusion terms. This stands in contrast to the classical problem of statistical inference from trajectory data. We establish unique identifiability results under several key scenarios, including cases with both multiplicative and additive noise, for both finite- and infinite-dimensional systems. Our analysis leverages the intrinsic structure of the governing equations and their quantitative relationship with the ergodic measure, thereby transforming the identifiability problem into a uniqueness issue for the solutions to the associated stationary Fokker-Planck equations. This approach reveals fundamental differences between drift and diffusion inversion problems and provides counterexamples where unique recovery fails. This work lays the theoretical foundation for a new research direction with significant potential for practical application.
Liouville conformal field theory describes a random geometry that fluctuates around a deterministic one: the unique solution of the problem of finding, within a given conformal class, a Riemannian metric with prescribed scalar and geodesic curvatures as well as conical singularities and corners. The level of randomness in Liouville theory is measured by the coupling constant γ(0,2)\gamma\in(0,2), the semi-classical limit corresponding to taking γ0\gamma\to0. Based on the probabilistic definition of Liouville theory, we prove that this semi-classical limit exists and does give rise to this deterministic geometry. At second order this limit is described in terms of a massive Gaussian free field with Robin boundary conditions. This in turn allows to implement CFT-inspired techniques in a deterministic setting: in particular we define the classical stress-energy tensor, show that it can be expressed in terms of accessory parameters (written as regularized derivatives of the Liouville action), and that it gives rise to classical higher equations of motion.
There are no more papers matching your filters at the moment.