LIPDepartment of PhysicsUniversity of Coimbra
We develop a framework for learning properties of quantum states beyond the assumption of independent and identically distributed (i.i.d.) input states. We prove that, given any learning problem (under reasonable assumptions), an algorithm designed for i.i.d. input states can be adapted to handle input states of any nature, albeit at the expense of a polynomial increase in training data size (aka sample complexity). Importantly, this polynomial increase in sample complexity can be substantially improved to polylogarithmic if the learning algorithm in question only requires non-adaptive, single-copy measurements. Among other applications, this allows us to generalize the classical shadow framework to the non-i.i.d. setting while only incurring a comparatively small loss in sample efficiency. We use rigorous quantum information theory to prove our main results. In particular, we leverage permutation invariance and randomized single-copy measurements to derive a new quantum de Finetti theorem that mainly addresses measurement outcome statistics and, in turn, scales much more favorably in Hilbert space dimension.
LLMail-Inject introduces a public challenge and dataset designed to evaluate indirect prompt injection attacks against an LLM-based email assistant in a realistic, end-to-end setting. The project collected over 200,000 unique attack prompts, demonstrating that end-to-end attacks are challenging to execute against layered defenses and providing insights into effective defense strategies.
8
Autonomous systems require robust Multi-Object Tracking (MOT) capabilities to operate reliably in dynamic environments. MOT ensures consistent object identity assignment and precise spatial delineation. Recent advances in foundation models, such as SAM2, have demonstrated strong zero-shot generalization for video segmentation, but their direct application to MOTS (MOT+Segmentation) remains limited by insufficient identity management and memory efficiency. This work introduces Seg2Track-SAM2, a framework that integrates pre-trained object detectors with SAM2 and a novel Seg2Track module to address track initialization, track management, and reinforcement. The proposed approach requires no fine-tuning and remains detector-agnostic. Experimental results on KITTI MOT and KITTI MOTS benchmarks show that Seg2Track-SAM2 achieves state-of-the-art (SOTA) performance, ranking fourth overall in both car and pedestrian classes on KITTI MOTS, while establishing a new benchmark in association accuracy (AssA). Furthermore, a sliding-window memory strategy reduces memory usage by up to 75% with negligible performance degradation, supporting deployment under resource constraints. These results confirm that Seg2Track-SAM2 advances MOTS by combining robust zero-shot tracking, enhanced identity preservation, and efficient memory utilization. The code is available at this https URL
9
We propose a primal heuristic for quadratic mixed-integer problems. Our method extends the Boscia framework -- originally a mixed-integer convex solver leveraging a Frank-Wolfe-based branch-and-bound approach -- to address nonconvex quadratic objective and constraints. We reformulate nonlinear constraints, introduce preprocessing steps, and a suite of heuristics including rounding strategies, gradient-guided selection, and large neighborhood search techniques that exploit integer-feasible vertices generated during the Frank-Wolfe iterations. Computational results demonstrate the effectiveness of our method in solving challenging MIQCQPs, achieving improvements on QPLIB instances within minutes and winning first place in the Land-Doig MIP Computational Competition 2025.
This research developed a GPU-accelerated decoder for Quantum Low-Density Parity-Check (QLDPC) codes, achieving real-time decoding latencies below the 63 μs threshold set by current quantum processors. The decoder processed a [[784, 24, 24]] QLDPC code in 43.7 μs on an RTX 4090, showcasing the practical viability of these scalable codes for fault-tolerant quantum computing.
Stroke is a leading cause of long-term disability and the second most common cause of death worldwide. Although acute treatments have advanced, recovery remains challenging and limited. Brain-computer interfaces (BCIs) have emerged as a promising tool for post-stroke rehabilitation by promoting neuroplasticity. However, clinical outcomes remain variable, and optimal protocols have yet to be established. This study explores strategies to optimize BCI-based rehabilitation by comparing motor imagery of affected hand movement versus rest, instead of the conventional left-versus-right motor imagery. This alternative aims to simplify the task and address the weak contralateral activation commonly observed in stroke patients. Two datasets, one from healthy individuals and one from stroke patients, were used to evaluate the proposed approach. The results showed improved performance using both FBCSP and EEGNet. Additionally, we investigated the impact of session duration and found that shorter training sessions produced better BCI performance than longer sessions.
The BANG method enables billion-scale Approximate Nearest Neighbour Search (ANNS) with high recall and throughput using a single GPU. It demonstrates 50x to 400x higher throughput on 1-billion point datasets compared to competing methods while maintaining high accuracy.
We report on a search for weakly interacting massive particle (WIMP) dark matter (DM) via elastic DM-xenon-nucleus interactions in the XENONnT experiment. We combine datasets from the first and second science campaigns resulting in a total exposure of 3.1  tonne×year3.1\;\text{tonne}\times\text{year}. In a blind analysis of nuclear recoil events with energies above 3.8keVNR3.8\,\mathrm{keV_{NR}}, we find no significant excess above background. We set new upper limits on the spin-independent WIMP-nucleon scattering cross-section for WIMP masses above 10GeV/c210\,\mathrm{GeV}/c^2 with a minimum of 1.7×1047cm21.7\,\times\,10^{-47}\,\mathrm{cm^2} at 90%90\,\% confidence level for a WIMP mass of 30GeV/c230\,\mathrm{GeV}/c^2. We achieve a best median sensitivity of 1.4×1047cm21.4\,\times\,10^{-47}\,\mathrm{cm^2} for a 41GeV/c241\,\mathrm{GeV}/c^2 WIMP. Compared to the result from the first XENONnT science dataset, we improve our sensitivity by a factor of up to 1.8.
The featured dataset, the Event-based Dataset of Assembly Tasks (EDAT24), showcases a selection of manufacturing primitive tasks (idle, pick, place, and screw), which are basic actions performed by human operators in any manufacturing assembly. The data were captured using a DAVIS240C event camera, an asynchronous vision sensor that registers events when changes in light intensity value occur. Events are a lightweight data format for conveying visual information and are well-suited for real-time detection and analysis of human motion. Each manufacturing primitive has 100 recorded samples of DAVIS240C data, including events and greyscale frames, for a total of 400 samples. In the dataset, the user interacts with objects from the open-source CT-Benchmark in front of the static DAVIS event camera. All data are made available in raw form (.aedat) and in pre-processed form (.npy). Custom-built Python code is made available together with the dataset to aid researchers to add new manufacturing primitives or extend the dataset with more samples.
We show that approximating the trace norm contraction coefficient of a quantum channel within a constant factor is NP-hard. Equivalently, this shows that determining the optimal success probability for encoding a bit in a quantum system undergoing noise is NP-hard. This contrasts with the classical analogue of this problem that can clearly be solved efficiently. We also establish the NP-hardness of deciding if the contraction coefficient is equal to 1, i.e., the channel can perfectly preserve a bit. As a consequence, deciding if a non-commutative graph has an independence number of at least 2 is NP-hard. In addition, we establish a converging hierarchy of semidefinite programming upper bounds on the contraction coefficient.
We demonstrate that Gaia's detection of stars on wide orbits around black holes opens a new observational window on dark matter structures -- such as scalar clouds and dark matter spikes -- predicted in a range of theoretical scenarios. Using precise radial velocity measurements of these systems, we derive state-of-the-art constraints on dark matter density profiles and particle masses in previously unexplored regions of parameter space. We also test the black hole hypothesis against the alternative of a boson star composed of light scalar fields.
Transformers exhibit in-context learning (ICL): the ability to use novel information presented in the context without additional weight updates. Recent work shows that ICL emerges when models are trained on a sufficiently diverse set of tasks and the transition from memorization to generalization is sharp with increasing task diversity. One interpretation is that a network's limited capacity to memorize favors generalization. Here, we examine the mechanistic underpinnings of this transition using a small transformer applied to a synthetic ICL task. Using theory and experiment, we show that the sub-circuits that memorize and generalize can be viewed as largely independent. The relative rates at which these sub-circuits learn explains the transition from memorization to generalization, rather than capacity constraints. We uncover a memorization scaling law, which determines the task diversity threshold at which the network generalizes. The theory quantitatively explains a variety of other ICL-related phenomena, including the long-tailed distribution of when ICL is acquired, the bimodal behavior of solutions close to the task diversity threshold, the influence of contextual and data distributional statistics on ICL, and the transient nature of ICL.
We report on a blinded search for dark matter with single- and few-electron signals in the first science run of XENONnT relying on a novel detector response framework that is physics-model-dependent. We derive 90\% confidence upper limits for dark matter-electron interactions. Heavy and light mediator cases are considered for the standard halo model and dark matter up-scattered in the Sun. We set stringent new limits on dark matter-electron scattering via a heavy mediator with a mass within 10-20\,MeV/c2c^2 and electron absorption of axion-like particles and dark photons for mχm_\chi below 0.186\,keV/c2c^2.
We present the first measurement of nuclear recoils from solar 8^8B neutrinos via coherent elastic neutrino-nucleus scattering with the XENONnT dark matter experiment. The central detector of XENONnT is a low-background, two-phase time projection chamber with a 5.9 t sensitive liquid xenon target. A blind analysis with an exposure of 3.51 t×\timesyr resulted in 37 observed events above 0.5 keV, with (26.41.3+1.426.4^{+1.4}_{-1.3}) events expected from backgrounds. The background-only hypothesis is rejected with a statistical significance of 2.73 σ\sigma. The measured 8^8B solar neutrino flux of (4.72.3+3.6)×106cm2s1(4.7_{-2.3}^{+3.6})\times 10^6 \mathrm{cm}^{-2}\mathrm{s}^{-1} is consistent with results from the Sudbury Neutrino Observatory. The measured neutrino flux-weighted CEν\nuNS cross section on Xe of (1.10.5+0.8)×1039cm2(1.1^{+0.8}_{-0.5})\times10^{-39} \mathrm{cm}^2 is consistent with the Standard Model prediction. This is the first direct measurement of nuclear recoils from solar neutrinos with a dark matter detector.
An Evolutionary Data-Centric AutoML (EDCA) framework automates the creation of efficient machine learning pipelines by integrating dynamic data preprocessing and reduction with evolutionary algorithms. It achieves predictive performance comparable to leading AutoML tools while consistently using substantially less data across various classification datasets.
Transits in the planetary system WASP-4 were recently found to occur 80s earlier than expected in observations from the TESS satellite. We present 22 new times of mid-transit that confirm the existence of transit timing variations, and are well fitted by a quadratic ephemeris with period decay dP/dt = -9.2 +/- 1.1 ms/yr. We rule out instrumental issues, stellar activity and the Applegate mechanism as possible causes. The light-time effect is also not favoured due to the non-detection of changes in the systemic velocity. Orbital decay and apsidal precession are plausible but unproven. WASP-4b is only the third hot Jupiter known to show transit timing variations to high confidence. We discuss a variety of observations of this and other planetary systems that would be useful in improving our understanding of WASP-4 in particular and orbital decay in general.
The LUX-ZEPLIN (LZ) experiment is searching for dark matter interactions in a liquid xenon time projection chamber (LXe-TPC). This article demonstrates how control of the flow state in the LXe-TPC enables the identification of pairs of sequential alpha-decays, which are used to map fluid flow and ion drift in the liquid target. The resulting transport model is used to tag 214^{214}Pb beta-decays, a leading background to dark matter signals in LZ. Temporally evolving volume selections, at a cost of 9.0% of exposure, target the decay of each 214^{214}Pb atom up to 81 minutes after production, resulting in (63 ±\pm 6stat_{\mathrm{stat}} ±\pm 7sys_{\mathrm{sys}})% identification of 214^{214}Pb decays to ground state. We also demonstrate how flow-based tagging techniques enable a novel calibration side band that is concurrent with science data.
Group fairness in machine learning is an important area of research focused on achieving equitable outcomes across different groups defined by sensitive attributes such as race or gender. Federated Learning, a decentralized approach to training machine learning models across multiple clients, amplifies the need for fairness methodologies due to its inherent heterogeneous data distributions that can exacerbate biases. The intersection of Federated Learning and group fairness has attracted significant interest, with 48 research works specifically dedicated to addressing this issue. However, no comprehensive survey has specifically focused on group fairness in Federated Learning. In this work, we analyze the key challenges of this topic, propose practices for its identification and benchmarking, and create a novel taxonomy based on criteria such as data partitioning, location, and strategy. Furthermore, we analyze broader concerns, review how different approaches handle the complexities of various sensitive attributes, examine common datasets and applications, and discuss the ethical, legal, and policy implications of group fairness in FL. We conclude by highlighting key areas for future research, emphasizing the need for more methods to address the complexities of achieving group fairness in federated systems.
Living systems exhibit a range of fundamental characteristics: they are active, self-referential, self-modifying systems. This paper explores how these characteristics create challenges for conventional scientific approaches and why they require new theoretical and formal frameworks. We introduce a distinction between 'natural time', the continuing present of physical processes, and 'representational time', with its framework of past, present and future that emerges with life itself. Representational time enables memory, learning and prediction, functions of living systems essential for their survival. Through examples from evolution, embryogenesis and metamorphosis we show how living systems navigate the apparent contradictions arising from self-reference as natural time unwinds self-referential loops into developmental spirals. Conventional mathematical and computational formalisms struggle to model self-referential and self-modifying systems without running into paradox. We identify promising new directions for modelling self-referential systems, including domain theory, co-algebra, genetic programming, and self-modifying algorithms. There are broad implications for biology, cognitive science and social sciences, because self-reference and self-modification are not problems to be avoided but core features of living systems that must be modelled to understand life's open-ended creativity.
There are no more papers matching your filters at the moment.