In this document, we present the main properties satisfied by the Moreau envelope of weakly convex functions. The Moreau envelope has been introduced in convex optimization to regularize convex functionals while preserving their global minimizers. However, the Moreau envelope is also defined for the more general class of weakly convex function and can be a useful tool for optimization in this context. The main properties of the Moreau envelope have been demonstrated for convex functions and are generalized to weakly convex function in various works. This document summarizes the vast literature on the properties of the Moreau envelope and provides the associated proofs.
Plug-and-Play methods constitute a class of iterative algorithms for imaging problems where regularization is performed by an off-the-shelf denoiser. Although Plug-and-Play methods can lead to tremendous visual performance for various image problems, the few existing convergence guarantees are based on unrealistic (or suboptimal) hypotheses on the denoiser, or limited to strongly convex data terms. In this work, we propose a new type of Plug-and-Play methods, based on half-quadratic splitting, for which the denoiser is realized as a gradient descent step on a functional parameterized by a deep neural network. Exploiting convergence results for proximal gradient descent algorithms in the non-convex setting, we show that the proposed Plug-and-Play algorithm is a convergent iterative scheme that targets stationary points of an explicit global functional. Besides, experiments show that it is possible to learn such a deep denoiser while not compromising the performance in comparison to other state-of-the-art deep denoisers used in Plug-and-Play schemes. We apply our proximal gradient algorithm to various ill-posed inverse problems, e.g. deblurring, super-resolution and inpainting. For all these applications, numerical results empirically confirm the convergence results. Experiments also show that this new algorithm reaches state-of-the-art performance, both quantitatively and qualitatively.
32
We present a discretization-free scalable framework for solving a large class of mass-conserving partial differential equations (PDEs), including the time-dependent Fokker-Planck equation and the Wasserstein gradient flow. The main observation is that the time-varying velocity field of the PDE solution needs to be self-consistent: it must satisfy a fixed-point equation involving the probability flow characterized by the same velocity field. Instead of directly minimizing the residual of the fixed-point equation with neural parameterization, we use an iterative formulation with a biased gradient estimator that bypasses significant computational obstacles with strong empirical performance. Compared to existing approaches, our method does not suffer from temporal or spatial discretization, covers a wider range of PDEs, and scales to high dimensions. Experimentally, our method recovers analytical solutions accurately when they are available and achieves superior performance in high dimensions with less training time compared to alternatives.
3
Superpixel decomposition methods are generally used as a pre-processing step to speed up image processing tasks. They group the pixels of an image into homogeneous regions while trying to respect existing contours. For all state-of-the-art superpixel decomposition methods, a trade-off is made between 1) computational time, 2) adherence to image contours and 3) regularity and compactness of the decomposition. In this paper, we propose a fast method to compute Superpixels with Contour Adherence using Linear Path (SCALP) in an iterative clustering framework. The distance computed when trying to associate a pixel to a superpixel during the clustering is enhanced by considering the linear path to the superpixel barycenter. The proposed framework produces regular and compact superpixels that adhere to the image contours. We provide a detailed evaluation of SCALP on the standard Berkeley Segmentation Dataset. The obtained results outperform state-of-the-art methods in terms of standard superpixel and contour detection metrics.
We present the first algorithm for computing class groups and unit groups of arbitrary number fields that provably runs in probabilistic subexponential time, assuming the Extended Riemann Hypothesis (ERH). Previous subexponential algorithms were either restricted to imaginary quadratic fields, or relied on several heuristic assumptions that have long resisted rigorous analysis. The heart of our method is a new general strategy to provably solve a recurring computational problem in number theory (assuming ERH): given an ideal class [a][\mathfrak{a}] of a number field KK, sample an ideal b[a]\mathfrak b \in [\mathfrak{a}] belonging to a particular family of ideals (e.g., the family of smooth ideals, or near-prime ideals). More precisely, let S\mathcal{S} be an arbitrary family of ideals, and SB\mathcal{S}_B the family of BB-smooth ideals. We describe an efficient algorithm that samples ideals b[a]\mathfrak b \in [\mathfrak{a}] such that bSSB\mathfrak b \in \mathcal{S} \cdot\mathcal{S}_B with probability proportional to the density of S\mathcal{S} within the set of all ideals. The case where S\mathcal{S} is the set of prime ideals yields the family SSB\mathcal{S}\cdot\mathcal{S}_B of near-prime ideals, of particular interest in that it constitutes a dense family of efficiently factorable ideals. The case of smooth ideals S=SB\mathcal{S} = \mathcal{S}_B regularly comes up in index-calculus algorithms (notably to compute class groups and unit groups), where it has long constituted a theoretical obstacle overcome only by heuristic arguments.
The use of optimal transport cost for learning generative models has become popular with Wasserstein Generative Adversarial Networks (WGAN). Training of WGAN relies on a theoretical background: the calculation of the gradient of the optimal transport cost with respect to the generative model parameters. We first demonstrate that such gradient may not be defined, which can result in numerical instabilities during gradient-based optimization. We address this issue by stating a valid differentiation theorem in the case of entropic regularized transport and specify conditions under which existence is ensured. By exploiting the discrete nature of empirical data, we formulate the gradient in a semi-discrete setting and propose an algorithm for the optimization of the generative model parameters. Finally, we illustrate numerically the advantage of the proposed framework.
We present a modified version of the Bravyi-Terhal bound that applies to quantum codes defined by local parity-check constraints on a DD-dimensional lattice quotient. Specifically, we consider a quotient ZD/Λ\mathbb{Z}^D/\Lambda of ZD\mathbb{Z}^D of cardinality nn, where Λ\Lambda is some DD-dimensional sublattice of ZD\mathbb{Z}^D: we suppose that every vertex of this quotient indexes mm qubits of a stabilizer code CC, which therefore has length nmnm. We prove that if all stabilizer generators act on qubits whose indices lie within a ball of radius ρ\rho, then the minimum distance dd of the code satisfies dmγD(D+4ρ)nD1Dd \leq m\sqrt{\gamma_D}(\sqrt{D} + 4\rho)n^\frac{D-1}{D} whenever n1/D8ργDn^{1/D} \geq 8\rho\sqrt{\gamma_D}, where γD\gamma_D is the DD-dimensional Hermite constant. We apply this bound to derive an upper bound on the minimum distance of Abelian Two-Block Group Algebra (2BGA) codes whose parity-check matrices have the form [AB][\mathbf{A} \, \vert \, \mathbf{B}] with each submatrix representing an element of a group algebra over a finite abelian group.
Block Principal Component Analysis (Block PCA) of a data matrix A, where loadings Z are determined by maximization of AZ 2 over unit norm orthogonal loadings, is difficult to use for the design of sparse PCA by 1 regularization, due to the difficulty of taking care of both the orthogonality constraint on loadings and the non differentiable 1 penalty. Our objective in this paper is to relax the orthogonality constraint on loadings by introducing new objective functions expvar(Y) which measure the part of the variance of the data matrix A explained by correlated components Y = AZ. So we propose first a comprehensive study of mathematical and numerical properties of expvar(Y) for two existing definitions Zou et al. [2006], Shen and Huang [2008] and four new definitions. Then we show that only two of these explained variance are fit to use as objective function in block PCA formulations for A rid of orthogonality constraints.
One key ingredient of image restoration is to define a realistic prior on clean images to complete the missing information in the observation. State-of-the-art restoration methods rely on a neural network to encode this prior. Typical image distributions are invariant to some set of transformations, such as rotations or flips. However, most deep architectures are not designed to represent an invariant image distribution. Recent works have proposed to overcome this difficulty by including equivariance properties within a Plug-and-Play paradigm. In this work, we propose two unified frameworks named Equivariant Regularization by Denoising (ERED) and Equivariant Plug-and-Play (EPnP) based on equivariant denoisers and stochastic optimization. We analyze the convergence of the proposed algorithms and discuss their practical benefit.
Despite the rapid development of computational hardware, the treatment of large and high dimensional data sets is still a challenging problem. This paper provides a twofold contribution to the topic. First, we propose a Gaussian Mixture Model in conjunction with a reduction of the dimensionality of the data in each component of the model by principal component analysis, called PCA-GMM. To learn the (low dimensional) parameters of the mixture model we propose an EM algorithm whose M-step requires the solution of constrained optimization problems. Fortunately, these constrained problems do not depend on the usually large number of samples and can be solved efficiently by an (inertial) proximal alternating linearized minimization algorithm. Second, we apply our PCA-GMM for the superresolution of 2D and 3D material images based on the approach of Sandeep and Jacob. Numerical results confirm the moderate influence of the dimensionality reduction on the overall superresolution result.
4
We study the approximation power of Graph Neural Networks (GNNs) on latent position random graphs. In the large graph limit, GNNs are known to converge to certain "continuous" models known as c-GNNs, which directly enables a study of their approximation power on random graph models. In the absence of input node features however, just as GNNs are limited by the Weisfeiler-Lehman isomorphism test, c-GNNs will be severely limited on simple random graph models. For instance, they will fail to distinguish the communities of a well-separated Stochastic Block Model (SBM) with constant degree function. Thus, we consider recently proposed architectures that augment GNNs with unique node identifiers, referred to as Structural GNNs here (SGNNs). We study the convergence of SGNNs to their continuous counterpart (c-SGNNs) in the large random graph limit, under new conditions on the node identifiers. We then show that c-SGNNs are strictly more powerful than c-GNNs in the continuous limit, and prove their universality on several random graph models of interest, including most SBMs and a large class of random geometric graphs. Our results cover both permutation-invariant and permutation-equivariant architectures.
40
We introduce the concept of a heterodimensional cycle of hyperbolic ergodic measures and a special type of them that we call rich. Within a partially hyperbolic context, we prove that if two measures are related by a rich heterodimensional cycle, then the entire segment of probability measures linking them lies within the closure of measures supported on periodic orbits. Motivated by the occurrence of robust heterodimensional cycles of hyperbolic basic sets, we study robust rich heterodimensional cycles of measures providing a framework for this phenomenon for diffeomorphisms. In the setting of skew products, we construct an open set of maps having uncountably many measures related by rich heterodimensional cycles.
The collection and the analysis of kidney stone morphological criteria are essential for an aetiological diagnosis of stone disease. However, in-situ LASER-based fragmentation of urinary stones, which is now the most established chirurgical intervention, may destroy the morphology of the targeted stone. In the current study, we assess the performance and added value of processing complete digital endoscopic video sequences for the automatic recognition of stone morphological features during a standard-of-care intra-operative session. To this end, a computer-aided video classifier was developed to predict in-situ the morphology of stone using an intra-operative digital endoscopic video acquired in a clinical setting. The proposed technique was evaluated on pure (i.e. include one morphology) and mixed (i.e. include at least two morphologies) stones involving "Ia/Calcium Oxalate Monohydrate (COM)", "IIb/ Calcium Oxalate Dihydrate (COD)" and "IIIb/Uric Acid (UA)" morphologies. 71 digital endoscopic videos (50 exhibited only one morphological type and 21 displayed two) were analyzed using the proposed video classifier (56840 frames processed in total). Using the proposed approach, diagnostic performances (averaged over both pure and mixed stone types) were as follows: balanced accuracy=88%, sensitivity=80%, specificity=95%, precision=78% and F1-score=78%. The obtained results demonstrate that AI applied on digital endoscopic video sequences is a promising tool for collecting morphological information during the time-course of the stone fragmentation process without resorting to any human intervention for stone delineation or selection of good quality steady frames. To this end, irrelevant image information must be removed from the prediction process at both frame and pixel levels, which is now feasible thanks to the use of AI-dedicated networks.
In recent in vitro experiments on co-culture between breast tumour spheroids and activated immune cells, it was observed that the introduction of the stress hormone cortisol resulted in a decreased immune cell infiltration into the spheroids. Moreover, the presence of cortisol deregulated the normal levels of the pro- and anti-inflammatory cytokines IFN-{\gamma} and IL-10. We present an individual-based model to explore the interaction dynamics between tumour and immune cells under psychological stress conditions. With our model, we explore the processes underlying the emergence of different levels of immune infiltration, with particular focus on the biological mechanisms regulated by IFN-{\gamma} and IL-10. The set-up of numerical simulations is defined to mimic the scenarios considered in the experimental study. Similarly to the experimental quantitative analysis, we compute a score that quantifies the level of immune cell infiltration into the tumour. The results of numerical simulations indicate that the motility of immune cells, their capability to infiltrate through tumour cells, their growth rate and the interplay between these cell parameters can affect the level of immune cell infiltration in different ways. Ultimately, numerical simulations of this model support a deeper understanding of the impact of biological stress-induced mechanisms on immune infiltration.
We consider the problem of recovering elements of a low-dimensional model from linear measurements. From signal and image processing to inverse problems in data science, this question has been at the center of many applications. Lately, with the success of models and methods relying on deep neural networks leading to non-convex formulations, traditional convex variational approaches have shown their limits. Furthermore, the multiplication of algorithms and recovery results makes identifying the best methods a complex task. In this article, we study recovery with a class of widely used algorithms without considering any underlying functional. This result leads to a class of projected gradient descent algorithms that recover a given low-dimensional with linear rates. The obtained rates decouple the impact of the quality of the measurements with respect to the model from its intrinsic complexity. As a consequence, we can directly measure the performance of this class of projected gradient descents through a restricted Lipschitz constant of the projection. By optimizing this constant, we define optimal algorithms. Our general approach provides an optimality result in the case of sparse recovery. Moreover, we uncover underlying linear rates of convergence for some ''plug and play'' imaging methods relying on deep priors by interpreting our results in this context, thus linking low-dimensional recovery and recovery with deep priors under a unified theory, validated by experiments on synthetic and real data.
We produce curves with a record number of points over the finite fields with 44, 99, 1616 and 2525 elements, as unramified abelian covers of algebraic curves.
We study the problem of transfers in a population structured by a continuous variable corresponding to the quantity being transferred. The model takes the form of an integro-differential equations with kernels corresponding to the specific rules of the transfer process. We focus our interest on the well-posedness of the Cauchy problem in the space of measures. We characterize transfer kernels that give a continuous semiflow in the space of measures and derive a necessary and sufficient condition for the stability of the space L1L^1 of integrable functions. We construct some examples of kernels that may be particularly interesting in economic applications. Our model considers blind transfers of economic value (e.g. money) between individuals. The two models are the ``Robin Hood model'', where the richest individual unconditionally gives a fraction of their wealth to the poorest when a transfer occurs, and the other extreme, the ``Sheriff of Nottingham model'', where the richest unconditionally takes a fraction of the poorest's wealth. Between these two extreme cases is a continuum of intermediate models obtained by interpolating the kernels. We illustrate those models with numerical simulations and show that any small fraction of the ``Sheriff of Nottingham'' in the transfer rules leads to a segregated population with extremely poor and extremely rich individuals after some time. Although our study is motivated by economic applications, we believe that this study is a first step towards a better understanding of many transfer phenomena occurring in the life sciences.
This thesis provides a classification of generalized pseudo-Anosov homeomorphisms up to topological conjugacy using an algorithmic approach. A Markov partition of a generalized pseudo-Anosov homeomorphism is a decomposition of the surface into a finite number of rectangles with disjoint interiors, such that their images intersect with any other rectangle in the Markov partition along a finite number of horizontal sub-rectangles. Every generalized pseudo-Anosov homeomorphism has a Markov partition, and, by using the surface's orientation, we can endow any Markov partition with a geometrization. The geometric type of a geometric Markov partition was defined by Bonatti and Langevin in their book, "Diffeomorphismes de Smale des surfaces", to classify saddle-type basic pieces for structurally stable diffeomorphisms on surfaces. A geometric type is an abstract combinatorial object that generalizes the incidence matrix of a Markov partition. It takes into account not only the number of times the image of a rectangle intersects with any other rectangle in the family but also the order and change of orientation induced by the homeomorphisms. This thesis employs the geometric type of a geometric Markov partition to classify conjugacy classes of pseudo-Anosov homeomorphisms. The classification is provided by the three main results in this manuscript: I) The geometric type is a complete invariant of conjugation. II) A criterion is provided for determining whether an abstract geometric type is realized by a geometric Markov partition of a pseudo-Anosov homeomorphism. III) An algorithm is described for determining whether two geometric types in the pseudo-Anosov class are realized by generalized pseudo-Anosov homeomorphisms that are topologically conjugated or not.
30 May 2025
In this work, we introduce a framework to design multidimensional Riemann solvers for nonlinear systems of hyperbolic conservation laws on general unstructured polygonal Voronoi-like tessellations. In this framework we propose two simple but complete solvers. The first method is a direct extension of the Osher-Solomon Riemann solver to multiple space dimensions. Here, the multidimensional numerical dissipation is obtained by integrating the absolute value of the flux Jacobians over a dual triangular mesh around each node of the primal polygonal grid. The required nodal gradient is then evaluated on a local computational simplex involving the d+1 neighbors meeting at each corner. The second method is a genuinely multidimensional upwind flux. By introducing a fluctuation form of finite volume methods with corner fluxes, we show an equivalence with residual distribution schemes (RD). This naturally allows to construct genuinely multidimensional upwind corner fluxes starting from existing and well-known RD fluctuations. In particular, we explore the use of corner fluxes obtained from the so-called N scheme, i.e. the Roe's original optimal multidimensional upwind advection scheme. Both methods use the full eigenstructure of the underlying hyperbolic system and are therefore complete by construction. A simple higher order extension up to fourth order in space and time is then introduced at the aid of a CWENO reconstruction in space and an ADER approach in time, leading to a family of high order accurate fully-discrete one-step schemes based on genuinely multidimensional Riemann solvers. We present applications of our new numerical schemes to several test problems for the compressible Euler equations of gas-dyanamics. In particular, we show that the proposed schemes are at the same time carbuncle-free and preserve certain stationary shear waves exactly.
This paper investigates the theoretical guarantees of L1-analysis regularization when solving linear inverse problems. Most of previous works in the literature have mainly focused on the sparse synthesis prior where the sparsity is measured as the L1 norm of the coefficients that synthesize the signal from a given dictionary. In contrast, the more general analysis regularization minimizes the L1 norm of the correlations between the signal and the atoms in the dictionary, where these correlations define the analysis support. The corresponding variational problem encompasses several well-known regularizations such as the discrete total variation and the Fused Lasso. Our main contributions consist in deriving sufficient conditions that guarantee exact or partial analysis support recovery of the true signal in presence of noise. More precisely, we give a sufficient condition to ensure that a signal is the unique solution of the L1-analysis regularization in the noiseless case. The same condition also guarantees exact analysis support recovery and L2-robustness of the L1-analysis minimizer vis-a-vis an enough small noise in the measurements. This condition turns to be sharp for the robustness of the analysis support. To show partial support recovery and L2-robustness to an arbitrary bounded noise, we introduce a stronger sufficient condition. When specialized to the L1-synthesis regularization, our results recover some corresponding recovery and robustness guarantees previously known in the literature. From this perspective, our work is a generalization of these results. We finally illustrate these theoretical findings on several examples to study the robustness of the 1-D total variation and Fused Lasso regularizations.
There are no more papers matching your filters at the moment.