Universidad Politécnica de Madrid
We present a comprehensive analysis of generic 5-dimensional Einstein-Maxwell-Dilaton-Axion (EMDA) holographic theories with exponential couplings. We find and classify exact, analytic, anisotropic solutions, both zero-temperature vacua and finite-temperature black brane backgrounds, with anisotropy sourced by scalar axions, magnetic fields, and charge densities, that can be interpreted as IR fixed points of renormalisation-group flows from UV-conformal fixed points. The resulting backgrounds feature a hyperscaling violation exponent and up to three independent Lifshitz-like exponents, generated by an equal number of independent coupling constants in the EMDA action. We derive the holographic stress-energy tensor and the corresponding equation of state, and discuss the behavior of the anisotropic speed of sound and butterfly velocity. We show that these theories can be consistently constrained by imposing several natural requirements, including energy conditions, thermodynamic stability, and causality. Additionally, we analyse hard probes in this class of theories, including Brownian motion, momentum broadening and jet quenching, and we demonstrate that a fully analytic treatment is possible, making their dependence on the underlying anisotropy explicit. We highlight the relevance of these models as benchmarks for strongly coupled anisotropic matter in nature, from the quark-gluon plasma created in heavy-ion collisions to dense QCD phases in neutron-star mergers and the cores of compact objects.
The proposed Differentiable Physics (DP) approach enables robust and accurate sound field reconstruction from limited and noisy data. It achieves this by modeling initial conditions with a neural network and propagating them via a differentiable numerical wave equation solver, strongly enforcing physical consistency and outperforming Physics-Informed Neural Networks (PINNs) by an order of magnitude in accuracy.
Quantum walks are at present an active field of study in mathematics, with important applications in quantum information and statistical physics. In this paper, we determine the influence of basic chaotic features on the walker behavior. For this purpose, we consider an extremely simple model consisting of alternating one-dimensional walks along the two spatial coordinates in bidimensional closed domains (hard wall billiards). The chaotic or regular behavior induced by the boundary shape in the deterministic classical motion translates into chaotic signatures for the quantized problem, resulting in sharp differences in the spectral statistics and morphology of the eigenfunctions of the quantum walker. Indeed, we found for the Bunimovich stadium -- a chaotic billiard -- level statistics described by a Brody distribution with parameter δ0.1\delta \simeq 0.1. This indicates a weak level repulsion, and also enhanced eigenfunction localization, with an average participation ratio (PR) \simeq 1150) compared to the rectangular billiard (regular) case, where the average PR \simeq 1500. Furthermore, scarring on unstable periodic orbits is observed. The fact that our simple model exhibits such key signatures of quantum chaos, e.g., non-Poissonian level statistics and scarring, that are sensitive to the underlying classical dynamics in the free particle billiard system is utterly surprising, especially when taking into account that quantum walks are diffusive models, which are not direct quantizations of a Hamiltonian.
University of Toronto logoUniversity of TorontoUniversity of New South WalesUniversity of Amsterdam logoUniversity of AmsterdamImperial College London logoImperial College LondonGhent UniversityUniversity College London logoUniversity College LondonUniversity of Oxford logoUniversity of OxfordNagoya University logoNagoya UniversityUniversity of Copenhagen logoUniversity of CopenhagenYale University logoYale UniversityUniversitat Pompeu FabraKU Leuven logoKU LeuvenUniversity of CampinasEmory University logoEmory UniversityHarvard Medical SchoolKing’s College London logoKing’s College LondonMohamed bin Zayed University of Artificial Intelligence logoMohamed bin Zayed University of Artificial IntelligenceAristotle University of ThessalonikiTechnical University of Munich logoTechnical University of MunichKorea Advanced Institute of Science and TechnologyUniversitat de BarcelonaGerman Cancer Research Center (DKFZ)Universidad Politécnica de MadridTechnical University of DenmarkMaastricht UniversityUniversity of LeedsINSERMUniversity of LondonThe University of Western AustraliaUmeå UniversityUniversity of California San FranciscoThe Barcelona Institute of Science and TechnologyUniversidad Nacional de ColombiaFraunhofer Heinrich-Hertz-InstituteHelmholtz Center MunichKempelen Institute of Intelligent TechnologiesUniversidad Nacional del LitoralStanford University School of MedicineUniversity of ColomboUniversity of Tunis El ManarUniversity of Colorado Anschutz Medical CampusMilitary Institute of Science and TechnologyUniversity of Texas MD Anderson Cancer CenterFoundation for Research and Technology-Hellas (FORTH)Hellenic Mediterranean UniversityNile UniversityIBM Research AfricaNepal Applied Mathematics and Informatics Institute for research (NAAMII)Jimma UniversityUniversity of GhanaMedical University of GdanskTechnology and Research (A*STAR)University of Arkansas for Medical SciencesBBMRI-ERICGalileo UniversityMilton Margai Technical UniversityMuhimbili University of Health and Allied SciencesEuropean Heart NetworkPasqual Maragall FoundationNamibia University of Science & TechnologyErasmus MC University Medical CenterAlmaty AI LabUniversity M’Hamed BougaraHospital Universitario y Politécnico La FeLa Fe Health Research InstituteHospital Clínic of BarcelonaUniversity of KassalaChildren’s National Hospital Washington DCUniversity of ZambiaEurécomIstanbul Technical University
Despite major advances in artificial intelligence (AI) for medicine and healthcare, the deployment and adoption of AI technologies remain limited in real-world clinical practice. In recent years, concerns have been raised about the technical, clinical, ethical and legal risks associated with medical AI. To increase real world adoption, it is essential that medical AI tools are trusted and accepted by patients, clinicians, health organisations and authorities. This work describes the FUTURE-AI guideline as the first international consensus framework for guiding the development and deployment of trustworthy AI tools in healthcare. The FUTURE-AI consortium was founded in 2021 and currently comprises 118 inter-disciplinary experts from 51 countries representing all continents, including AI scientists, clinicians, ethicists, and social scientists. Over a two-year period, the consortium defined guiding principles and best practices for trustworthy AI through an iterative process comprising an in-depth literature review, a modified Delphi survey, and online consensus meetings. The FUTURE-AI framework was established based on 6 guiding principles for trustworthy AI in healthcare, i.e. Fairness, Universality, Traceability, Usability, Robustness and Explainability. Through consensus, a set of 28 best practices were defined, addressing technical, clinical, legal and socio-ethical dimensions. The recommendations cover the entire lifecycle of medical AI, from design, development and validation to regulation, deployment, and monitoring. FUTURE-AI is a risk-informed, assumption-free guideline which provides a structured approach for constructing medical AI tools that will be trusted, deployed and adopted in real-world practice. Researchers are encouraged to take the recommendations into account in proof-of-concept stages to facilitate future translation towards clinical practice of medical AI.
Video game cheats modify a video game behaviour to give unfair advantages to some players while bypassing the methods game developers use to detect them. This destroys the experience of online gaming and can result in financial losses for game developers. In this work, we present a new type of game cheat, Virtual machine Introspection Cheat (VIC), that takes advantage of virtual machines to stealthy execute game cheats. VIC employees a hypervisor with introspection enabled to lower the bar of cheating against legacy and modern anti-cheat systems. We demonstrate the feasibility and stealthiness of VIC against three popular games (Fortnite, BlackSquad and Team Fortress 2) that include five different anti-cheats. In particular, we use VIC to implement a cheat radar, a wall-hack cheat and a trigger-bot. To support our claim that this type of cheats can be effectively used, we present the performance impact VICs have on gameplay by monitoring the frames per second (fps) while the cheats are activated. Our experimentation also shows how these cheats are currently undetected by the most popular anti-cheat systems, enabling a new paradigm that can take advantage of cloud infrastructure to offer cheating-as-a-service.
Eye movements can reveal early signs of neurodegeneration, including those associated with Parkinson's Disease (PD). This work investigates the utility of a set of gaze-based features for the automatic screening of PD from different visual exploration tasks. For this purpose, a novel methodology is introduced, combining classic fixation/saccade oculomotor features (e.g., saccade count, fixation duration, scanned area) with features derived from gaze clusters (i.e., regions with a considerable accumulation of fixations). These features are automatically extracted from six exploration tests and evaluated using different machine learning classifiers. A Mixture of Experts ensemble is used to integrate outputs across tests and both eyes. Results show that ensemble models outperform individual classifiers, achieving an Area Under the Receiving Operating Characteristic Curve (AUC) of 0.95 on a held-out test set. The findings support visual exploration as a non-invasive tool for early automatic screening of PD.
This research from Nanjing University of Aeronautics and Astronautics, Universidad Politécnica de Madrid, and Universidad Carlos III de Madrid investigated how reasoning impacts Large Language Model confidence. It found that requiring step-by-step reasoning makes LLMs more self-confident in their answers, even when those answers are incorrect, demonstrating a disconnect between internal confidence and actual correctness across models tested on the MMLU benchmark.
With the latest advances in Deep Learning-based generative models, it has not taken long to take advantage of their remarkable performance in the area of time series. Deep neural networks used to work with time series heavily depend on the size and consistency of the datasets used in training. These features are not usually abundant in the real world, where they are usually limited and often have constraints that must be guaranteed. Therefore, an effective way to increase the amount of data is by using Data Augmentation techniques, either by adding noise or permutations and by generating new synthetic data. This work systematically reviews the current state-of-the-art in the area to provide an overview of all available algorithms and proposes a taxonomy of the most relevant research. The efficiency of the different variants will be evaluated as a central part of the process, as well as the different metrics to evaluate the performance and the main problems concerning each model will be analysed. The ultimate aim of this study is to provide a summary of the evolution and performance of areas that produce better results to guide future researchers in this field.
A study investigated why Large Language Models (LLMs) struggle with letter counting, revealing high error rates across most models and identifying the number of repeated letters as the primary factor contributing to these difficulties. GPT-4o demonstrated the lowest error rate at 17.3%, significantly outperforming other tested models.
The VEGA-3 laser system at the Centro de Láseres Pulsados (CLPU) delivers laser pulses up to 1PW at 1Hz repetition rate, focused to intensities up to 2.5e20W/cm2. A versatile and compact targetry solution suitable for this repetition rate is presented. The system can operate in the challenging petawatt laser environment close to the laser-plasma interaction. Strips are spooled in a tape target system to deliver a solid density target to the laser focus for every shot. Results are presented for different tape materials and thicknesses. Experimental ion spectra are recorded by a Thomson Parabola Ion Spectrometer coupled with a scintillator screen; and an antenna array is used for the characterization of electromagnetic pulses. The results of both diagnostics show a good shot-to-shot stability of the system.
We investigate asymptotic Schwarzschild exterior solutions in the context of modified gravity theories, specifically within the framework of f(R)f(R) gravity, where the asymptotic behavior recovers the standard Schwarzschild solution of General Relativity. Unlike previous studies that rely mainly on analytical approximations, our approach combines asymptotic analysis with numerical integration of the underlying differential equations. Using these solutions, we analyze strong lensing effects to obtain the photon sphere radius and the corresponding capture parameter. Considering rings produced by total reflection, we define the photon sphere width as the difference between the first total reflection and the capture parameter; and study how it is modified in the f(R)f(R) scenario. Our results show that the photon sphere width increases in the presence of f(R)f(R)-type modifications, indicating deviations from GR that could be observable in the strong-field regime.
Large Language Models (LLMs) have become central to advancing automation and decision-making across various sectors, raising significant ethical questions. This study proposes a comprehensive comparative analysis of the most advanced LLMs to assess their moral profiles. We subjected several state-of-the-art models to a selection of ethical dilemmas and found that all the proprietary ones are mostly utilitarian and all of the open-weights ones align mostly with values-based ethics. Furthermore, when using the Moral Foundations Questionnaire, all models we probed - except for Llama 2-7B - displayed a strong liberal bias. Lastly, in order to causally intervene in one of the studied models, we propose a novel similarity-specific activation steering technique. Using this method, we were able to reliably steer the model's moral compass to different ethical schools. All of these results showcase that there is an ethical dimension in already deployed LLMs, an aspect that is generally overlooked.
Accurate and automatic segmentation of brain tumors in multi-parametric magnetic resonance imaging (mpMRI) is essential for quantitative measurements, which play an increasingly important role in clinical diagnosis and prognosis. The International Brain Tumor Segmentation (BraTS) Challenge 2024 offers a unique benchmarking opportunity, including various types of brain tumors in both adult and pediatric populations, such as pediatric brain tumors (PED), meningiomas (MEN-RT) and brain metastases (MET), among others. Compared to previous editions, BraTS 2024 has implemented changes to substantially increase clinical relevance, such as refined tumor regions for evaluation. We propose a deep learning-based ensemble approach that integrates state-of-the-art segmentation models. Additionally, we introduce innovative, adaptive pre- and post-processing techniques that employ MRI-based radiomic analyses to differentiate tumor subtypes. Given the heterogeneous nature of the tumors present in the BraTS datasets, this approach enhances the precision and generalizability of segmentation models. On the final testing sets, our method achieved mean lesion-wise Dice similarity coefficients of 0.926, 0.801, and 0.688 for the whole tumor in PED, MEN-RT, and MET, respectively. These results demonstrate the effectiveness of our approach in improving segmentation performance and generalizability for various brain tumor types. The source code of our implementation is available at this https URL Additionally, an open-source web-application is accessible at this https URL which uses the docker container aparida12/brats-peds-2024:v20240913 .
Various audio-LLMs (ALLMs) have been explored recently for tackling different audio tasks simultaneously using a single, unified model. While existing evaluations of ALLMs primarily focus on single-audio tasks, real-world applications often involve processing multiple audio streams simultaneously. To bridge this gap, we propose the first multi-audio evaluation (MAE) benchmark that consists of 20 datasets from 11 multi-audio tasks encompassing both speech and sound scenarios. Comprehensive experiments on MAE demonstrate that the existing ALLMs, while being powerful in comprehending primary audio elements in individual audio inputs, struggling to handle multi-audio scenarios. To this end, we propose a novel multi-audio-LLM (MALLM) to capture audio context among multiple similar audios using discriminative learning on our proposed synthetic data. The results demonstrate that the proposed MALLM outperforms all baselines and achieves high data efficiency using synthetic data without requiring human annotations. The proposed MALLM opens the door for ALLMs towards multi-audio processing era and brings us closer to replicating human auditory capabilities in machines.
1
Inequalities in social networks arise from linking mechanisms, such as preferential attachment (connecting to popular nodes), homophily (connecting to similar others), and triadic closure (connecting through mutual contacts). While preferential attachment mainly drives degree inequality and homophily drives segregation, their three-way interaction remains understudied. This gap limits our understanding of how network inequalities emerge. Here, we introduce PATCH, a network growth model combining the three mechanisms to understand how they create disparities among two groups in synthetic networks. Extensive simulations confirm that homophily and preferential attachment increase segregation and degree inequalities, while triadic closure has countervailing effects: conditional on the other mechanisms, it amplifies population-wide degree inequality while reducing segregation and between-group degree disparities. We demonstrate PATCH's explanatory potential on fifty years of Physics and Computer Science collaboration and citation networks exhibiting persistent gender disparities. PATCH accounts for these gender disparities with the joint presence of preferential attachment, moderate gender homophily, and varying levels of triadic closure. By connecting mechanisms to observed inequalities, PATCH shows how their interplay sustains group disparities and provides a framework for designing interventions that promote more equitable social networks.
Leaderboards showcase the current capabilities and limitations of Large Language Models (LLMs). To motivate the development of LLMs that represent the linguistic and cultural diversity of the Spanish-speaking community, we present La Leaderboard, the first open-source leaderboard to evaluate generative LLMs in languages and language varieties of Spain and Latin America. La Leaderboard is a community-driven project that aims to establish an evaluation standard for everyone interested in developing LLMs for the Spanish-speaking community. This initial version combines 66 datasets in Basque, Catalan, Galician, and different Spanish varieties, showcasing the evaluation results of 50 models. To encourage community-driven development of leaderboards in other languages, we explain our methodology, including guidance on selecting the most suitable evaluation setup for each downstream task. In particular, we provide a rationale for using fewer few-shot examples than typically found in the literature, aiming to reduce environmental impact and facilitate access to reproducible results for a broader research community.
Computational models provide crucial insights into complex biological processes such as cancer evolution, but their mechanistic nature often makes them nonlinear and parameter-rich, complicating calibration. We systematically evaluate parameter probability distributions in cell migration models using Bayesian calibration across four complementary strategies: parametric and surrogate models, each with and without explicit model discrepancy. This approach enables joint analysis of parameter uncertainty, predictive performance, and interpretability. Applied to a real data experiment of glioblastoma progression in microfluidic devices, surrogate models achieve higher computational efficiency and predictive accuracy, whereas parametric models yield more reliable parameter estimates due to their mechanistic grounding. Incorporating model discrepancy exposes structural limitations, clarifying where model refinement is necessary. Together, these comparisons offer practical guidance for calibrating and improving computational models of complex biological systems.
The MARIO challenge, held at MICCAI 2024, focused on advancing the automated detection and monitoring of age-related macular degeneration (AMD) through the analysis of optical coherence tomography (OCT) images. Designed to evaluate algorithmic performance in detecting neovascular activity changes within AMD, the challenge incorporated unique multi-modal datasets. The primary dataset, sourced from Brest, France, was used by participating teams to train and test their models. The final ranking was determined based on performance on this dataset. An auxiliary dataset from Algeria was used post-challenge to evaluate population and device shifts from submitted solutions. Two tasks were involved in the MARIO challenge. The first one was the classification of evolution between two consecutive 2D OCT B-scans. The second one was the prediction of future AMD evolution over three months for patients undergoing anti-vascular endothelial growth factor (VEGF) therapy. Thirty-five teams participated, with the top 12 finalists presenting their methods. This paper outlines the challenge's structure, tasks, data characteristics, and winning methodologies, setting a benchmark for AMD monitoring using OCT, infrared imaging, and clinical data (such as the number of visits, age, gender, etc.). The results of this challenge indicate that artificial intelligence (AI) performs as well as a physician in measuring AMD progression (Task 1) but is not yet able of predicting future evolution (Task 2).
This paper summarizes in depth the state of the art of aerial swarms, covering both classical and new reinforcement-learning-based approaches for their management. Then, it proposes a hybrid AI system, integrating deep reinforcement learning in a multi-agent centralized swarm architecture. The proposed system is tailored to perform surveillance of a specific area, searching and tracking ground targets, for security and law enforcement applications. The swarm is governed by a central swarm controller responsible for distributing different search and tracking tasks among the cooperating UAVs. Each UAV agent is then controlled by a collection of cooperative sub-agents, whose behaviors have been trained using different deep reinforcement learning models, tailored for the different task types proposed by the swarm controller. More specifically, proximal policy optimization (PPO) algorithms were used to train the agents' behavior. In addition, several metrics to assess the performance of the swarm in this application were defined. The results obtained through simulation show that our system searches the operation area effectively, acquires the targets in a reasonable time, and is capable of tracking them continuously and consistently.
In the age of social media and generative AI, the ability to automatically assess the credibility of online content has become increasingly critical, complementing traditional approaches to false information detection. Credibility assessment relies on aggregating diverse credibility signals - small units of information, such as content subjectivity, bias, or a presence of persuasion techniques - into a final credibility label/score. However, current research in automatic credibility assessment and credibility signals detection remains highly fragmented, with many signals studied in isolation and lacking integration. Notably, there is a scarcity of approaches that detect and aggregate multiple credibility signals simultaneously. These challenges are further exacerbated by the absence of a comprehensive and up-to-date overview of research works that connects these research efforts under a common framework and identifies shared trends, challenges, and open problems. In this survey, we address this gap by presenting a systematic and comprehensive literature review of 175 research papers, focusing on textual credibility signals within the field of Natural Language Processing (NLP), which undergoes a rapid transformation due to advancements in Large Language Models (LLMs). While positioning the NLP research into the the broader multidisciplinary landscape, we examine both automatic credibility assessment methods as well as the detection of nine categories of credibility signals. We provide an in-depth analysis of three key categories: 1) factuality, subjectivity and bias, 2) persuasion techniques and logical fallacies, and 3) check-worthy and fact-checked claims. In addition to summarising existing methods, datasets, and tools, we outline future research direction and emerging opportunities, with particular attention to evolving challenges posed by generative AI.
There are no more papers matching your filters at the moment.