Researchers at ETH Zürich, Mila, CUNY, and McGill University developed principled algorithms to transform token-level large language models into character-level models. Their method resolves the prompt boundary problem, provides a more accurate character-level probability distribution, and achieves lower per-byte surprisal, indicating improved compression.
View blogThis paper offers a comprehensive review of training methodologies for Physical Neural Networks (PNNs), addressing the escalating energy and performance demands of digital AI. It systematically categorizes diverse training approaches, from physics-aware backpropagation to in-situ gradient computation, and evaluates their potential to enable energy-efficient, scalable AI systems.
View blogResearchers from the University of Oulu and City University of New York developed a framework that weakens the diffusion purification effect in diffusion models to enhance facial privacy protection, generating high-quality protected face images that achieve an average Protection Success Rate (PSR) of over 80% on standard datasets against commercial facial recognition systems.
View blogJWST observations precisely characterized the atmosphere and mass of the 17 Myr old exoplanet HIP 67522 b, reclassifying it as a low-mass (13.8 1.0 M) sub-Neptune with an inflated radius and a supersolar metallicity atmosphere containing H2O, CO2, and CO. This "featherweight giant" is undergoing rapid photoevaporative mass loss, offering insight into early planetary evolution.
View blog