Tufts University
A computational tool, AcrosticSleuth, identifies and probabilistically ranks acrostics in multilingual corpora, achieving F1 scores up to 0.66 on known Russian acrostics. This tool, developed by researchers from Tufts, UW-Madison, UT Austin, and Harvard, also led to the discovery of previously unrecognized acrostics in significant historical texts, including one in Thomas Hobbes' *The Elements of Law*.
We explore the physical properties of five massive quiescent galaxies at z2.5z\sim2.5, revealing the presence of non-negligible dust reservoirs. JWST NIRSpec observations were obtained for each target, finding no significant line emission; multiple star formation tracers independently place upper limits between 0.110 M/yr0.1-10~M_\odot / \mathrm{yr}. Spectral energy distribution modeling with Prospector infers stellar masses between log10[M/M]1011\log_{10}[M / M_\odot] \sim 10-11 and stellar mass-weighted ages between 121-2 Gyr. The inferred mass-weighted effective radii (reff0.41.4r_{eff}\sim 0.4-1.4 kpc) and inner 11 kpc stellar surface densities (log10[Σ/Mkpc2]9\log_{10}[\Sigma / M_\odot \mathrm{kpc}^2 ]\gtrsim 9) are typical of quiescent galaxies at z2z \gtrsim 2. The galaxies display negative color gradients (redder core and bluer outskirts); for one galaxy, this effect results from a dusty core, while for the others it may be evidence of an "inside-out" growth process. Unlike local quiescent galaxies, we identify significant reddening in these typical cosmic noon passive galaxies; all but one require AV0.4A_V \gtrsim 0.4. This finding is in qualitative agreement with previous studies but our deep 20-band NIRCam imaging is able to significantly suppress the dust-age degeneracy and confidently determine that these galaxies are reddened. We speculate about the physical effects that may drive the decline in dust content in quiescent galaxies over cosmic time.
MMAU-Pro introduces a comprehensive benchmark of 5,305 expert-annotated instances designed to holistically evaluate audio general intelligence in AI models across complex, real-world scenarios. The benchmark reveals that even state-of-the-art models exhibit substantial limitations in multi-audio reasoning, spatial understanding, and long-form audio comprehension, indicating significant room for improvement.
Jansma and Hoel extend the Causal Emergence 2.0 framework to analyze all possible micro-to-macro paths within a system, identifying causally relevant emergent hierarchies and introducing new measures for complexity. Their work demonstrates that emergent properties can be engineered with pinpoint precision, designing systems with specific emergent causal structures.
Complex systems can be described at myriad different scales, and their causal workings often have multiscale structure (e.g., a computer can be described at the microscale of its hardware circuitry, the mesoscale of its machine code, and the macroscale of its operating system). While scientists study and model systems across the full hierarchy of their scales, from microphysics to macroeconomics, there is debate about what the macroscales of systems can possibly add beyond mere compression. To resolve this longstanding issue, here a new theory of emergence is introduced wherein the different scales of a system are treated like slices of a higher-dimensional object. The theory can distinguish which of these scales possess unique causal contributions, and which are not causally relevant. Constructed from an axiomatic notion of causation, the theory's application is demonstrated in coarse-grains of Markov chains. It identifies all cases of macroscale causation: instances where reduction to a microscale is possible, yet lossy about causation. Furthermore, the theory posits a causal apportioning schema that calculates the causal contribution of each scale, showing what each uniquely adds. Finally, it reveals a novel measure of emergent complexity: how widely distributed a system's causal workings are across its hierarchy of scales.
Graph generation is a critical task in numerous domains, including molecular design and social network analysis, due to its ability to model complex relationships and structured data. While most modern graph generative models utilize adjacency matrix representations, this work revisits an alternative approach that represents graphs as sequences of node set and edge set. We advocate for this approach due to its efficient encoding of graphs and propose a novel representation. Based on this representation, we introduce the Graph Generative Pre-trained Transformer (G2PT), an auto-regressive model that learns graph structures via next-token prediction. To further exploit G2PT's capabilities as a general-purpose foundation model, we explore fine-tuning strategies for two downstream applications: goal-oriented generation and graph property prediction. We conduct extensive experiments across multiple datasets. Results indicate that G2PT achieves superior generative performance on both generic graph and molecule datasets. Furthermore, G2PT exhibits strong adaptability and versatility in downstream tasks from molecular design to property prediction. Code available at this https URL,
3
We study the physics potential of heavy QCD axions at high-energy muon colliders. Unlike typical axion-like particles, heavy QCD axions solve the strong CP problem with phenomenology driven by the anomalous gluon (aGG~aG\widetilde G) couplings. Several ultraviolet scenarios are presented in which QCD axions with TeV-scale masses and decay constants arise consistently with a solution to both the strong CP problem and the axion quality problem. We perform a detailed collider analysis for both a 3 and 10~TeV muon collider, focusing on hadronic axion decays that gives rise to a dijet-resonance signature. Our projections for the axion discovery reach in the multi-TeV mass range demonstrate that a muon collider can significantly extend sensitivity to heavy QCD axions compared to existing experiments.
Researchers from UT Austin, Oxford, Leeds, Tufts, and Alberta established a unified conceptual framework for curriculum learning in reinforcement learning (RL), systematically classifying existing methods and identifying key research gaps. The work defines curriculum learning through task generation, sequencing, and transfer learning components, providing a structured overview to address challenges in RL sample efficiency and complex task acquisition.
The discovery and identification of molecules in biological and environmental samples is crucial for advancing biomedical and chemical sciences. Tandem mass spectrometry (MS/MS) is the leading technique for high-throughput elucidation of molecular structures. However, decoding a molecular structure from its mass spectrum is exceptionally challenging, even when performed by human experts. As a result, the vast majority of acquired MS/MS spectra remain uninterpreted, thereby limiting our understanding of the underlying (bio)chemical processes. Despite decades of progress in machine learning applications for predicting molecular structures from MS/MS spectra, the development of new methods is severely hindered by the lack of standard datasets and evaluation protocols. To address this problem, we propose MassSpecGym -- the first comprehensive benchmark for the discovery and identification of molecules from MS/MS data. Our benchmark comprises the largest publicly available collection of high-quality labeled MS/MS spectra and defines three MS/MS annotation challenges: de novo molecular structure generation, molecule retrieval, and spectrum simulation. It includes new evaluation metrics and a generalization-demanding data split, therefore standardizing the MS/MS annotation tasks and rendering the problem accessible to the broad machine learning community. MassSpecGym is publicly available at this https URL
50
We present an algorithm that uses block encoding on a quantum computer to exactly construct a Krylov space, which can be used as the basis for the Lanczos method to estimate extremal eigenvalues of Hamiltonians. While the classical Lanczos method has exponential cost in the system size to represent the Krylov states for quantum systems, our efficient quantum algorithm achieves this in polynomial time and memory. The construction presented is exact in the sense that the resulting Krylov space is identical to that of the Lanczos method, so the only approximation with respect to the exact method is due to finite sample noise. This is possible because, unlike previous quantum Krylov methods, our algorithm does not require simulating real or imaginary time evolution. We provide an explicit error bound for the resulting ground state energy estimate in the presence of noise. For our method to be successful efficiently, the only requirement on the input problem is that the overlap of the initial state with the true ground state must be Ω(1/poly(n))\Omega(1/\text{poly}(n)) for nn qubits.
TabTreeFormer, developed by researchers at NUS and Betterdata AI, introduces a hybrid tree-transformer model for synthetic tabular data generation, achieving up to 44% utility gain over baselines while preserving privacy and enhancing efficiency through a novel dual-quantization tokenizer and ordinal-aware learning.
This paper proposes two algorithms for estimating square Wasserstein distance matrices from a small number of entries. These matrices are used to compute manifold learning embeddings like multidimensional scaling (MDS) or Isomap, but contrary to Euclidean distance matrices, are extremely costly to compute. We analyze matrix completion from upper triangular samples and Nyström completion in which O(dlog(d))\mathcal{O}(d\log(d)) columns of the distance matrices are computed where dd is the desired embedding dimension, prove stability of MDS under Nyström completion, and show that it can outperform matrix completion for a fixed budget of sample distances. Finally, we show that classification of the OrganCMNIST dataset from the MedMNIST benchmark is stable on data embedded from the Nyström estimation of the distance matrix even when only 10\% of the columns are computed.
The question of "what is life?" has challenged scientists and philosophers for centuries, producing an array of definitions that reflect both the mystery of its emergence and the diversity of disciplinary perspectives brought to bear on the question. Despite significant progress in our understanding of biological systems, psychology, computation, and information theory, no single definition for life has yet achieved universal acceptance. This challenge becomes increasingly urgent as advances in synthetic biology, artificial intelligence, and astrobiology challenge our traditional conceptions of what it means to be alive. We undertook a methodological approach that leverages large language models (LLMs) to analyze a set of definitions of life provided by a curated set of cross-disciplinary experts. We used a novel pairwise correlation analysis to map the definitions into distinct feature vectors, followed by agglomerative clustering, intra-cluster semantic analysis, and t-SNE projection to reveal underlying conceptual archetypes. This methodology revealed a continuous landscape of the themes relating to the definition of life, suggesting that what has historically been approached as a binary taxonomic problem should be instead conceived as differentiated perspectives within a unified conceptual latent space. We offer a new methodological bridge between reductionist and holistic approaches to fundamental questions in science and philosophy, demonstrating how computational semantic analysis can reveal conceptual patterns across disciplinary boundaries, and opening similar pathways for addressing other contested definitional territories across the sciences.
1
The detection of strong Balmer breaks and absorption features in Little Red Dots (LRDs) suggests they host AGN embedded within dense gas envelopes, potentially powered by super-Eddington accretion. We present GLIMPSE-17775, a luminous (Lbol1045L_{\rm bol}\sim10^{45} erg s1^{-1}) LRD at z=3.501z=3.501 behind Abell S1063 (μ2\mu\sim2), observed with deep JWST/NIRCam and a \sim20 hr (80 hr de-lensed) NIRSpec/G395M spectrum. The data reveal 40+ emission and absorption features, including a rich forest of low-ionization FeII lines and numerous broad hydrogen recombination transitions. We use this depth to test the dense-gas interpretation through five independent diagnostics. Nearly all permitted lines show exponential wings with consistent FWHM, the signature of Thomson scattering requiring ne108n_e\gtrsim10^8 cm3^{-3}. Adopting this width yields MBH106.7MM_{\rm BH}\sim10^{6.7}M_\odot, a factor of ten lower than Gaussian fits, and λEdd1.8\lambda_{\rm Edd}\sim1.8. Additional diagnostics support the same picture: a pronounced Balmer break (fν,4050/fν,3670=2.0±0.1f_{\nu,4050}/f_{\nu,3670}=2.0\pm0.1), enhanced HeI λ7065\lambda7065 and λ10830\lambda10830 with P-Cygni absorption, Bowen-fluorescent OI λ8446\lambda8446-λ11290\lambda11290 emission requiring Lyβ\beta pumping, and 16 FeII lines matching fluorescence models. These features indicate a dense (n108n\sim10^8 cm3^{-3}), partially ionized cocoon where scattering and fluorescence dominate line formation, providing strong evidence that at least some LRDs are powered by super-Eddington black-hole growth in the early Universe.
Vision-language-action (VLA) models hold promise as generalist robotics solutions by translating visual and linguistic inputs into robot actions, yet they lack reliability due to their black-box nature and sensitivity to environmental changes. In contrast, cognitive architectures (CA) excel in symbolic reasoning and state monitoring but are constrained by rigid predefined execution. This work bridges these approaches by probing OpenVLA's hidden layers to uncover symbolic representations of object properties, relations, and action states, enabling integration with a CA for enhanced interpretability and robustness. Through experiments on LIBERO-spatial pick-and-place tasks, we analyze the encoding of symbolic states across different layers of OpenVLA's Llama backbone. Our probing results show consistently high accuracies (> 0.90) for both object and action states across most layers, though contrary to our hypotheses, we did not observe the expected pattern of object states being encoded earlier than action states. We demonstrate an integrated DIARC-OpenVLA system that leverages these symbolic representations for real-time state monitoring, laying the foundation for more interpretable and reliable robotic manipulation.
Capability evaluations play a critical role in ensuring the safe deployment of frontier AI systems, but this role may be undermined by intentional underperformance or ``sandbagging.'' We present a novel model-agnostic method for detecting sandbagging behavior using noise injection. Our approach is founded on the observation that introducing Gaussian noise into the weights of models either prompted or fine-tuned to sandbag can considerably improve their performance. We test this technique across a range of model sizes and multiple-choice question benchmarks (MMLU, AI2, WMDP). Our results demonstrate that noise injected sandbagging models show performance improvements compared to standard models. Leveraging this effect, we develop a classifier that consistently identifies sandbagging behavior. Our unsupervised technique can be immediately implemented by frontier labs or regulatory bodies with access to weights to improve the trustworthiness of capability evaluations.
New methods for modeling least-cost diets that meet nutritional requirements for health have emerged as important tools for informing nutrition policy and programming around the world. This study develops a three-step approach using cost of healthy diet to inform targeted nutrition programming in Indonesia. We combine detailed retail prices and household survey data from Indonesia to describe how reported consumption and expenditure patterns across all levels of household income diverge from least cost healthy diets using items from nearby markets. In this analysis, we examine regional price variations, identify households with insufficient income for healthy diets, and analyze the nutrient adequacy of reported consumption patterns. We find that household food spending was sufficient to meet national dietary guidelines using the least expensive locally available items for over 98% of Indonesians, but almost all households consume substantial quantities of discretionary foods and mixed dishes while consuming too little energy from fruits, vegetables, and legumes, nuts, and seeds. Households with higher incomes have higher nutrient adequacy and are closer to meeting local dietary guidelines, but still fall short of recommendations. These findings shed new light on how actual food demand differs from least-cost healthy diets, due to factors other than affordability, such as taste, convenience, and aspirations shaped by marketing and other sociocultural influences.
We present an overview of the JWST GLIMPSE program, highlighting its survey design, primary science goals, gravitational lensing models, and first results. GLIMPSE provides ultra-deep JWST/NIRCam imaging across seven broadband filters (F090W, F115W, F200W, F277W, F356W, F444W) and two medium-band filters (F410M, F480M), with exposure times ranging from 20 to 40 hours per filter. This yields a 5σ\sigma limiting magnitude of 30.9 AB (measured in a 0.2 arcsec diameter aperture). The field is supported by extensive ancillary data, including deep HST imaging from the Hubble Frontier Fields program, VLT/MUSE spectroscopy, and deep JWST/NIRSpec medium-resolution multi-object spectroscopy. Exploiting the strong gravitational lensing of the galaxy cluster Abell S1063, GLIMPSE probes intrinsic depths beyond 33 AB magnitudes and covers an effective source-plane area of approximately 4.4 arcmin2^2 at z6z \sim 6. The program's central aim is to constrain the abundance of the faintest galaxies from z6z \sim 6 up to the highest redshifts, providing crucial benchmarks for galaxy formation models, which have so far been tested primarily on relatively bright systems. We present an initial sample of 540\sim 540 galaxy candidates identified at 6 < z < 16, with intrinsic UV magnitudes spanning MUVM_{\mathrm UV} = -20 to -12. This enables unprecedented constraints on the extreme faint end of the UV luminosity function at these epochs. In addition, GLIMPSE opens new windows for spatially resolved studies of star clusters in early galaxies and the detection and characterization of faint high-zz active galactic nuclei. This paper accompanies the first public data release, which includes reduced JWST and HST mosaics, photometric catalogs, and gravitational lensing models.
The effective Nambu-Goto description of (2+1)(2+1)-dimensional domain walls predicts singular behavior of its worldsheet resulting in swallowtail bifurcations. This phenomenon is intimately related to the formation of cusps, which emerge in different forms that we identify and classify. We describe in detail how swallowtail bifurcations generically arise in the collision of wiggles on straight domain wall strings, as well as in the collapse of closed loops, even for smooth initial conditions. Remarkably, by means of accurate lattice simulations, we find that these distinctive swallowtail features are reproduced in the field theory evolution of sufficiently thin walls, typically emitting a significant fraction of their initial energy in the process. These results suggest that such singular evolutions could potentially have important implications for the observable signatures associated with the collapse of domain wall networks in (3+1) dimensions in the early universe.
Researchers from a diverse group of academic institutions and industry labs critically examine the pervasive influence of Artificial General Intelligence (AGI) as a guiding principle in AI research. Their analysis reveals how an AGI focus exacerbates issues such as a lack of scientific rigor, masked values, and exclusion of diverse stakeholders. The paper advocates for alternative, more effective goal-setting strategies that prioritize specificity, pluralism, and inclusion, ultimately aiming to re-center AI development around supporting and benefiting human beings.
There are no more papers matching your filters at the moment.