LaBRIUniv. Bordeaux
We introduce a new adaptive step-size strategy for convex optimization with stochastic gradient that exploits the local geometry of the objective function only by means of a first-order stochastic oracle and without any hyper-parameter tuning. The method comes from a theoretically-grounded adaptation of the Adaptive Gradient Descent Without Descent method to the stochastic setting. We prove the convergence of stochastic gradient descent with our step-size under various assumptions, and we show that it empirically competes against tuned baselines.
We consider the problem of sampling distributions stemming from non-convex potentials with Unadjusted Langevin Algorithm (ULA). We prove the stability of the discrete-time ULA to drift approximations under the assumption that the potential is strongly convex at infinity. In many context, e.g. imaging inverse problems, potentials are non-convex and non-smooth. Proximal Stochastic Gradient Langevin Algorithm (PSGLA) is a popular algorithm to handle such potentials. It combines the forward-backward optimization algorithm with a ULA step. Our main stability result combined with properties of the Moreau envelope allows us to derive the first proof of convergence of the PSGLA for non-convex potentials. We empirically validate our methodology on synthetic data and in the context of imaging inverse problems. In particular, we observe that PSGLA exhibits faster convergence rates than Stochastic Gradient Langevin Algorithm for posterior sampling while preserving its restoration properties.
In this document, we present the main properties satisfied by the Moreau envelope of weakly convex functions. The Moreau envelope has been introduced in convex optimization to regularize convex functionals while preserving their global minimizers. However, the Moreau envelope is also defined for the more general class of weakly convex function and can be a useful tool for optimization in this context. The main properties of the Moreau envelope have been demonstrated for convex functions and are generalized to weakly convex function in various works. This document summarizes the vast literature on the properties of the Moreau envelope and provides the associated proofs.
Explainable AI (XAI) has become increasingly important with the rise of large transformer models, yet many explanation methods designed for CNNs transfer poorly to Vision Transformers (ViTs). Existing ViT explanations often rely on attention weights, which tend to yield noisy maps as they capture token-to-token interactions within each this http URL attribution methods incorporating MLP blocks have been proposed, we argue that attention remains a valuable and interpretable signal when properly filtered. We propose a method that combines attention maps with a statistical filtering, initially proposed for CNNs, to remove noisy or uninformative patterns and produce more faithful explanations. We further extend our approach with a class-specific variant that yields discriminative explanations. Evaluation against popular state-of-the-art methods demonstrates that our approach produces sharper and more interpretable maps. In addition to perturbation-based faithfulness metrics, we incorporate human gaze data to assess alignment with human perception, arguing that human interpretability remains essential for XAI. Across multiple datasets, our approach consistently outperforms or is comparable to the SOTA methods while remaining efficient and human plausible.
Fast convergence and high-quality image recovery are two essential features of algorithms for solving ill-posed imaging inverse problems. Existing methods, such as regularization by denoising (RED), often focus on designing sophisticated image priors to improve reconstruction quality, while leaving convergence acceleration to heuristics. To bridge the gap, we propose Restarted Inertia with Score-based Priors (RISP) as a principled extension of RED. RISP incorporates a restarting inertia for fast convergence, while still allowing score-based image priors for high-quality reconstruction. We prove that RISP attains a faster stationary-point convergence rate than RED, without requiring the convexity of the image prior. We further derive and analyze the associated continuous-time dynamical system, offering insight into the connection between RISP and the heavy-ball ordinary differential equation (ODE). Experiments across a range of imaging inverse problems demonstrate that RISP enables fast convergence while achieving high-quality reconstructions.
Researchers from Smartesting and Université de Bordeaux explored the efficacy of autonomous web agents for executing natural language test cases, introducing a dedicated benchmark and two open-source agent implementations. Their advanced agent, PinATA, achieved a 50% higher True Accuracy compared to a baseline, though a qualitative analysis revealed five categories of persistent limitations.
Superpixels are widely used in computer vision to simplify image representation and reduce computational complexity. While traditional methods rely on low-level features, deep learning-based approaches leverage high-level features but also tend to sacrifice regularity of superpixels to capture complex objects, leading to accurate but less interpretable segmentations. In this work, we introduce SPAM (SuperPixel Anything Model), a versatile framework for segmenting images into accurate yet regular superpixels. We train a model to extract image features for superpixel generation, and at inference, we leverage a large-scale pretrained model for semantic-agnostic segmentation to ensure that superpixels align with object masks. SPAM can handle any prior high-level segmentation, resolving uncertainty regions, and is able to interactively focus on specific objects. Comprehensive experiments demonstrate that SPAM qualitatively and quantitatively outperforms state-of-the-art methods on segmentation tasks, making it a valuable and robust tool for various applications. Code and pre-trained models are available here: this https URL.
Machine learning and deep learning models have become essential in the recent fast development of artificial intelligence in many sectors of the society. It is now widely acknowledge that the development of these models has an environmental cost that has been analyzed in many studies. Several online and software tools have been developed to track energy consumption while training machine learning models. In this paper, we propose a comprehensive introduction and comparison of these tools for AI practitioners wishing to start estimating the environmental impact of their work. We review the specific vocabulary, the technical requirements for each tool. We compare the energy consumption estimated by each tool on two deep neural networks for image processing and on different types of servers. From these experiments, we provide some advice for better choosing the right tool and infrastructure.
Plug-and-Play methods constitute a class of iterative algorithms for imaging problems where regularization is performed by an off-the-shelf denoiser. Although Plug-and-Play methods can lead to tremendous visual performance for various image problems, the few existing convergence guarantees are based on unrealistic (or suboptimal) hypotheses on the denoiser, or limited to strongly convex data terms. In this work, we propose a new type of Plug-and-Play methods, based on half-quadratic splitting, for which the denoiser is realized as a gradient descent step on a functional parameterized by a deep neural network. Exploiting convergence results for proximal gradient descent algorithms in the non-convex setting, we show that the proposed Plug-and-Play algorithm is a convergent iterative scheme that targets stationary points of an explicit global functional. Besides, experiments show that it is possible to learn such a deep denoiser while not compromising the performance in comparison to other state-of-the-art deep denoisers used in Plug-and-Play schemes. We apply our proximal gradient algorithm to various ill-posed inverse problems, e.g. deblurring, super-resolution and inpainting. For all these applications, numerical results empirically confirm the convergence results. Experiments also show that this new algorithm reaches state-of-the-art performance, both quantitatively and qualitatively.
32
Dynamic networks are a complex subject. Not only do they inherit the complexity of static networks (as a particular case); they are also sensitive to definitional subtleties that are a frequent source of confusion and incomparability of results in the literature. In this paper, we take a step back and examine three such aspects in more details, exploring their impact in a systematic way; namely, whether the temporal paths are required to be \emph{strict} (i.e., the times along a path must increasing, not just be non-decreasing), whether the time labeling is \emph{proper} (two adjacent edges cannot be present at the same time) and whether the time labeling is \emph{simple} (an edge can have only one presence time). In particular, we investigate how different combinations of these features impact the expressivity of the graph in terms of reachability. Our results imply a hierarchy of expressivity for the resulting settings, shedding light on the loss of generality that one is making when considering either combination. Some settings are more general than expected; in particular, proper temporal graphs turn out to be as expressive as general temporal graphs where non-strict paths are allowed. Also, we show that the simplest setting, that of \emph{happy} temporal graphs (i.e., both proper and simple) remains expressive enough to emulate the reachability of general temporal graphs in a certain (restricted but useful) sense. Furthermore, this setting is advocated as a target of choice for proving negative results. We illustrates this by strengthening two known results to happy graphs (namely, the inexistence of sparse spanners, and the hardness of computing temporal components). Overall, we hope that this article can be seen as a guide for choosing between different settings of temporal graphs, while being aware of the way these choices affect generality.
The design of complex self-organising systems producing life-like phenomena, such as the open-ended evolution of virtual creatures, is one of the main goals of artificial life. Lenia, a family of cellular automata (CA) generalizing Conway's Game of Life to continuous space, time and states, has attracted a lot of attention because of the wide diversity of self-organizing patterns it can generate. Among those, some spatially localized patterns (SLPs) resemble life-like artificial creatures and display complex behaviors. However, those creatures are found in only a small subspace of the Lenia parameter space and are not trivial to discover, necessitating advanced search algorithms. Furthermore, each of these creatures exist only in worlds governed by specific update rules and thus cannot interact in the same one. This paper proposes as mass-conservative extension of Lenia, called Flow Lenia, that solve both of these issues. We present experiments demonstrating its effectiveness in generating SLPs with complex behaviors and show that the update rule parameters can be optimized to generate SLPs showing behaviors of interest. Finally, we show that Flow Lenia enables the integration of the parameters of the CA update rules within the CA dynamics, making them dynamic and localized, allowing for multi-species simulations, with locally coherent update rules that define properties of the emerging creatures, and that can be mixed with neighbouring rules. We argue that this paves the way for the intrinsic evolution of self-organized artificial life forms within continuous CAs.
Complex organic molecules (COMs) are observed to be abundant in various astrophysical environments, in particular toward star forming regions they are observed both toward protostellar envelopes as well as shocked regions. Emission spectrum especially of heavier COMs may consists of up to hundreds of lines, where line blending hinders the analysis. However, identifying the molecular composition of the gas leading to the observed millimeter spectra is the first step toward a quantitative analysis. We develop a new method based on supervised machine learning to recognize spectroscopic features of the rotational spectrum of molecules in the 3mm atmospheric transmission band for a list of species including COMs with the aim to obtain a detection probability. We used local thermodynamic equilibrium (LTE) modeling to build a large set of synthetic spectra of 20 molecular species including COMs with a range of physical conditions typical for star forming regions. We successfully designed and trained a Convolutional Neural Network (CNN) that provides detection probabilities of individual species in the spectra. We demonstrate that the produced CNN-model has a robust performance to detect spectroscopic signatures from these species in synthetic spectra. We evaluate its ability to detect molecules according to the noise level, frequency coverage, and line-richness, and also test its performance for incomplete frequency coverage with high detection probabilities for the tested parameter space, and no false predictions. Ultimately, we apply the CNN-model to obtain predictions on observational data from the literature toward line-rich hot-core like sources, where detection probabilities remain reasonable with no false detection. We prove the use of CNNs facilitating the analysis of complex millimeter spectra both on synthetic spectra as well as first tests on observational data.
The science of Human-Computer Interaction (HCI) is populated by isolated empirical findings, often tied to specific technologies, designs, and tasks. This situation probably lies in observing the wrong object of study, that is to say, observing interfaces rather than interaction. This paper proposes an experimental methodology, powered by a research methodology, that enables tackling the ambition of observing interaction (rather than interfaces). These observations are done during the treatment of applicative cases, allowing to generate and replicate results covering various experimental conditions, expressed from the need of end users and the evolution of technologies. Performing these observations when developing applicative prototypes illustrating novel technologies' utility allows, in the same time, to benefit from an optimization of these prototypes to better accomplish end users tasks. This paper depicts a long term research direction, from generating the initial observations of interaction properties and their replication, to their integration, that would then lead to exploring the possible relations existing between those properties, to end toward the description of human-computer interaction's physics.
Recent JWST observations of the temperate sub-Neptune K2-18 b with NIRISS SOSS/NIRSpec G395H and MIRI LRS have yielded apparently inconsistent results: the MIRI spectra exhibit spectral features nearly twice as large as those seen at shorter wavelengths, challenging the high-metallicity, CH4-rich non-equilibrium model that fits the NIRISS/NIRSpec data. We perform a suite of atmospheric retrievals on both datasets, including free-chemistry, non-equilibrium, and aerosol models, using laboratory-derived complex refractive indices for a variety of photochemical haze analogues. Free retrievals systematically return lower metallicities than inferred by self-consistent chemical disequilibrium models, and the inclusion of absorbing aerosols, especially CH4-dominated, nitrogen-poor tholins, can further reduce the inferred metallicity by over an order of magnitude. These hazes reproduce the observed NIRISS slope through scattering and match MIRI features via C-H bending absorption near 7 um, while yielding particle properties consistent with photochemical production in H2-rich atmospheres. Although their inclusion improves the joint fit and reduces tension between datasets, it also significantly lowers the retrieved CH4 abundance, highlighting degeneracies between metallicity, composition, and aerosol properties. Our results underscore the importance of aerosol absorption in interpreting temperate sub-Neptune spectra, and motivate future JWST observations and laboratory work to break these degeneracies.
The growing integration of Artificial Intelligence (AI) into education has intensified the need for transparency and interpretability. While hackathons have long served as agile environments for rapid AI prototyping, few have directly addressed eXplainable AI (XAI) in real-world educational contexts. This paper presents a comprehensive analysis of the XAI Challenge 2025, a hackathon-style competition jointly organized by Ho Chi Minh City University of Technology (HCMUT) and the International Workshop on Trustworthiness and Reliability in Neurosymbolic AI (TRNS-AI), held as part of the International Joint Conference on Neural Networks (IJCNN 2025). The challenge tasked participants with building Question-Answering (QA) systems capable of answering student queries about university policies while generating clear, logic-based natural language explanations. To promote transparency and trustworthiness, solutions were required to use lightweight Large Language Models (LLMs) or hybrid LLM-symbolic systems. A high-quality dataset was provided, constructed via logic-based templates with Z3 validation and refined through expert student review to ensure alignment with real-world academic scenarios. We describe the challenge's motivation, structure, dataset construction, and evaluation protocol. Situating the competition within the broader evolution of AI hackathons, we argue that it represents a novel effort to bridge LLMs and symbolic reasoning in service of explainability. Our findings offer actionable insights for future XAI-centered educational systems and competitive research initiatives.
Hydrodynamic interactions can give rise to a collective motion of rotating particles. This, in turn, can lead to coherent fluid flows. Using large scale hydrodynamic simulations, we study the coupling between these two in spinner monolayers at weakly inertial regime. We observe an instability, where the initially uniform particle layer separates into particle void and particle rich areas. The particle void region corresponds to a fluid vortex, and it is driven by a surrounding spinner edge current. We show that the instability originates from a hydrodynamic lift force between the particle and fluid flows. The cavitation can be tuned by the strength of the collective flows. It is suppressed when the spinners are confined by a no-slip surface, and multiple cavity and oscillating cavity states are observed when the particle concentration is reduced.
Seven rocky planets orbit the nearby dwarf star TRAPPIST-1, providing a unique opportunity to search for atmospheres on small planets outside the Solar System (Gillon et al., 2017). Thanks to the recent launch of JWST, possible atmospheric constituents such as carbon dioxide (CO2) are now detectable (Morley et al., 2017, Lincowski et al., 2018}. Recent JWST observations of the innermost planet TRAPPIST-1 b showed that it is most probably a bare rock without any CO2 in its atmosphere (Greene et al., 2023). Here we report the detection of thermal emission from the dayside of TRAPPIST-1 c with the Mid-Infrared Instrument (MIRI) on JWST at 15 micron. We measure a planet-to-star flux ratio of fp/fs = 421 +/- 94 parts per million (ppm) which corresponds to an inferred dayside brightness temperature of 380 +/- 31 K. This high dayside temperature disfavours a thick, CO2-rich atmosphere on the planet. The data rule out cloud-free O2/CO2 mixtures with surface pressures ranging from 10 bar (with 10 ppm CO2) to 0.1 bar (pure CO2). A Venus-analogue atmosphere with sulfuric acid clouds is also disfavoured at 2.6 sigma confidence. Thinner atmospheres or bare-rock surfaces are consistent with our measured planet-to-star flux ratio. The absence of a thick, CO2-rich atmosphere on TRAPPIST-1 c suggests a relatively volatile-poor formation history, with less than 9.5 +7.5 -2.3 Earth oceans of water. If all planets in the system formed in the same way, this would indicate a limited reservoir of volatiles for the potentially habitable planets in the system.
We propose a physics-informed machine learning framework called P-DivGNN to reconstruct local stress fields at the micro-scale, in the context of multi-scale simulation given a periodic micro-structure mesh and mean, macro-scale, stress values. This method is based in representing a periodic micro-structure as a graph, combined with a message passing graph neural network. We are able to retrieve local stress field distributions, providing average stress values produced by a mean field reduced order model (ROM) or Finite Element (FE) simulation at the macro-scale. The prediction of local stress fields are of utmost importance considering fracture analysis or the definition of local fatigue criteria. Our model incorporates physical constraints during training to constraint local stress field equilibrium state and employs a periodic graph representation to enforce periodic boundary conditions. The benefits of the proposed physics-informed GNN are evaluated considering linear and non linear hyperelastic responses applied to varying geometries. In the non-linear hyperelastic case, the proposed method achieves significant computational speed-ups compared to FE simulation, making it particularly attractive for large-scale applications.
Analysing biological spiking neural network models with synaptic plasticity has proven to be challenging both theoretically and numerically. In a network with N all-to-all connected neurons, the number of synaptic connections is on the order of N2N^2, making these models computationally demanding. Furthermore, the intricate coupling between neuron and synapse dynamics, along with the heterogeneity generated by plasticity, hinder the use of classic theoretical tools such as mean-field or slow-fast analyses. To address these challenges, we introduce a new variable which we term a typical neuron X. Viewed as a post-synaptic neuron, X is composed of the activity state V , the time since its last spike S, and the empirical distribution ξ\xi of the triplet V , S and W (incoming weight) associated to the pre-synaptic neurons. In particular, we study a stochastic spike-timing-dependent plasticity (STDP) model of connection in a probabilistic Wilson-Cowan spiking neural network model, which features binary neural activity. Taking the large N limit, we obtain from the empirical distribution of the typical neuron a simplified yet accurate representation of the original spiking network. This mean-field limit is a piecewise deterministic Markov process (PDMP) of McKean-Vlasov type, where the typical neuron dynamics depends on its own distribution. We term this analysis McKean-Vlasov mean-field (MKV-MF). Our approach not only reduces computational complexity but also provides insights into the dynamics of this spiking neural network with plasticity. The model obtained is mathematically exact and capable of tracking transient changes. This analysis marks the first exploration of MKV-MF dynamics in a network of spiking neurons interacting with STDP.
Accurate physical simulation is crucial for the development and validation of control algorithms in robotic systems. Recent works in Reinforcement Learning (RL) take notably advantage of extensive simulations to produce efficient robot control. State-of-the-art servo actuator models generally fail at capturing the complex friction dynamics of these systems. This limits the transferability of simulated behaviors to real-world applications. In this work, we present extended friction models that allow to more accurately simulate servo actuator dynamics. We propose a comprehensive analysis of various friction models, present a method for identifying model parameters using recorded trajectories from a pendulum test bench, and demonstrate how these models can be integrated into physics engines. The proposed friction models are validated on four distinct servo actuators and tested on 2R manipulators, showing significant improvements in accuracy over the standard Coulomb-Viscous model. Our results highlight the importance of considering advanced friction effects in the simulation of servo actuators to enhance the realism and reliability of robotic simulations.
There are no more papers matching your filters at the moment.