Researchers at TU Wien developed a method to synthesize person-centric visual images directly from WiFi Channel State Information (CSI) obtained through walls. Their approach, using a multimodal Variational Autoencoder, produced perceptually clearer and more temporally coherent images, particularly with a concatenation and temporal encoding strategy.
Simple and low-cost biosensing solutions are suitable for point-of-care applications aiming to overcome the gap between scientific concepts and technological production. To compete with sensitivity and selectivity of golden standards, such as liquid chromatography, the functionalization of biosensors is continuously optimized to enhance the signal and improve their performance, often leading to complex chemical assay development. In this research, the efforts are made on optimizing the methodology for electrochemical reduction of graphene oxide to produce thin film-modified gold electrodes. Under the employed specific conditions, 20 cycles of cyclic voltammetry (CV) are shown to be optimal for superior electrical activation of graphene oxide into electrochemically reduced graphene oxide (ERGO). This platform is further used to develop a matrix metalloproteinase 2 (MMP-2) biosensor, where specific anti-MMP2 aptamers are utilized as a biorecognition element. MMP-2 is a protein which is typically overexpressed in tumor tissues, with important roles in tumor invasion, metastasis as well as in tumor angiogenesis. Based on impedimetric measurements, we were able to detect as low as 3.32 pg/mL of MMP-2 in PBS with a dynamic range of 10 pg/mL - 10 ng/mL. Besides high specificity, ERGO-based aptasensor showed a potential of reuse due to demonstrated successful signal restoration after experimental detection of MMP-2.
Recent advances in AI -- including generative approaches -- have resulted in technology that can support humans in scientific discovery and forming decisions, but may also disrupt democracies and target individuals. The responsible use of AI and its participation in human-AI teams increasingly shows the need for AI alignment, that is, to make AI systems act according to our preferences. A crucial yet often overlooked aspect of these interactions is the different ways in which humans and machines generalise. In cognitive science, human generalisation commonly involves abstraction and concept learning. In contrast, AI generalisation encompasses out-of-domain generalisation in machine learning, rule-based reasoning in symbolic AI, and abstraction in neurosymbolic AI. In this perspective paper, we combine insights from AI and cognitive science to identify key commonalities and differences across three dimensions: notions of, methods for, and evaluation of generalisation. We map the different conceptualisations of generalisation in AI and cognitive science along these three dimensions and consider their role for alignment in human-AI teaming. This results in interdisciplinary challenges across AI and cognitive science that must be tackled to provide a foundation for effective and cognitively supported alignment in human-AI teaming scenarios.
A collaborative white paper coordinated by the Quantum Community Network comprehensively analyzes the current status and future perspectives of Quantum Artificial Intelligence, categorizing its potential into "Quantum for AI" and "AI for Quantum" applications. It proposes a strategic research and development agenda to bolster Europe's competitive position in this rapidly converging technological domain.
This paper from Kerbl et al. introduces a hierarchical 3D Gaussian representation that facilitates real-time rendering of very large-scale environments. The approach handles kilometer-scale scenes with tens of thousands of input images by employing a chunk-based processing pipeline and an efficient level-of-detail system, consistently achieving over 30 FPS.
1,097
The REASSEMBLE dataset provides a comprehensive, multimodal resource for contact-rich, long-horizon robotic assembly and disassembly, utilizing the standardized NIST Assembly Task Board #1. It includes a unique combination of RGB, proprioceptive, force-torque, audio, and event camera data, along with detailed hierarchical annotations for multi-task learning and anomaly detection.
10
An embodied robotic agent is developed that uses Large Language Models and Retrieval-Augmented Generation (RAG) for long-horizon task planning in dynamic household environments. This system allows robots to autonomously manage objects, execute natural language commands, and track object locations through a natural language-driven memory system.
8
Researchers from Inria, Université Côte d’Azur, and TU Wien introduced a method to reduce the memory footprint of 3D Gaussian Splatting by an average of 27 times while largely preserving visual quality and real-time rendering performance. This approach enables the practical deployment of high-fidelity 3D scenes on memory-constrained devices and for streaming applications.
140
Accurate surface geometry representation is crucial in 3D visual computing. Explicit representations, such as polygonal meshes, and implicit representations, like signed distance functions, each have distinct advantages, making efficient conversions between them increasingly important. Conventional surface extraction methods for implicit representations, such as the widely used Marching Cubes algorithm, rely on spatial decomposition and sampling, leading to inaccuracies due to fixed and limited resolution. We introduce a novel approach for analytically extracting surfaces from neural implicit functions. Our method operates natively in parallel and can navigate large neural architectures. By leveraging the fact that each neuron partitions the domain, we develop a depth-first traversal strategy to efficiently track the encoded surface. The resulting meshes faithfully capture the full geometric information from the network without ad-hoc spatial discretization, achieving unprecedented accuracy across diverse shapes and network architectures while maintaining competitive speed.
This paper introduces "StopThePop," a refined rendering pipeline for 3D Gaussian Splatting that eliminates visual popping artifacts and view inconsistencies caused by approximate sorting. It achieves this by employing a novel hierarchical rasterization approach that maintains comparable image quality and near real-time performance, being only 4% slower than original 3DGS on average, and up to 1.6x faster with opacity decay.
191
We study the problem of zero-shot link prediction on knowledge graphs (KGs), which requires models to generalize over novel entities and novel relations. Knowledge graph foundation models (KGFMs) address this task by enforcing equivariance over both nodes and relations, learning from structural properties of nodes and relations, which are then transferable to novel graphs with similar structural properties. However, the conventional notion of deterministic equivariance imposes inherent limits on the expressive power of KGFMs, preventing them from distinguishing structurally similar but semantically distinct relations. To overcome this limitation, we introduce probabilistic node-relation equivariance, which preserves equivariance in distribution while incorporating a principled randomization to break symmetries during inference. Building on this principle, we present Flock, a KGFM that iteratively samples random walks, encodes them into sequences via a recording protocol, embeds them with a sequence model, and aggregates representations of nodes and relations via learned pooling. Crucially, Flock respects probabilistic node-relation equivariance and is a universal approximator for isomorphism-invariant link-level functions over KGs. Empirically, Flock perfectly solves our new diagnostic dataset Petals where current KGFMs fail, and achieves state-of-the-art performances on entity- and relation prediction tasks on 54 KGs from diverse domains.
8
Bug reports often lack sufficient detail for developers to reproduce and fix the underlying defects. Bug Reproduction Tests (BRTs), tests that fail when the bug is present and pass when it has been resolved, are crucial for debugging, but they are rarely included in bug reports, both in open-source and in industrial settings. Thus, automatically generating BRTs from bug reports has the potential to accelerate the debugging process and lower time to repair. This paper investigates automated BRT generation within an industry setting, specifically at Google, focusing on the challenges of a large-scale, proprietary codebase and considering real-world industry bugs extracted from Google's internal issue tracker. We adapt and evaluate a state-of-the-art BRT generation technique, LIBRO, and present our agent-based approach, BRT Agent, which makes use of a fine-tuned Large Language Model (LLM) for code editing. Our BRT Agent significantly outperforms LIBRO, achieving a 28% plausible BRT generation rate, compared to 10% by LIBRO, on 80 human-reported bugs from Google's internal issue tracker. We further investigate the practical value of generated BRTs by integrating them with an Automated Program Repair (APR) system at Google. Our results show that providing BRTs to the APR system results in 30% more bugs with plausible fixes. Additionally, we introduce Ensemble Pass Rate (EPR), a metric which leverages the generated BRTs to select the most promising fixes from all fixes generated by APR system. Our evaluation on EPR for Top-K and threshold-based fix selections demonstrates promising results and trade-offs. For example, EPR correctly selects a plausible fix from a pool of 20 candidates in 70% of cases, based on its top-1 ranking.
FlowMP presents a conditional flow matching framework for robot motion planning, learning a direct mapping from a noise distribution to expert trajectories. This method explicitly models second-order dynamics, generating kinodynamically feasible paths up to 100 times faster than prior approaches while demonstrating superior scalability and smoothness for complex 3D robotic tasks compared to diffusion models.
Petrović et al. introduce CURLY FLOW MATCHING (CURLY-FM), a simulation-free machine learning framework that learns complex, non-gradient field dynamics by solving a Schrödinger bridge problem with a non-zero drift reference process. The method reconstructs periodic trajectories in diverse scientific domains like ocean currents and cell cycles, outperforming existing flow-matching and simulation-based techniques in efficiency and accuracy.
7
Transformers excel across a large variety of tasks but remain susceptible to corrupted inputs, since standard self-attention treats all query-key interactions uniformly. Inspired by lateral inhibition in biological neural circuits and building on the recent use by the Differential Transformer's use of two parallel softmax subtraction for noise cancellation, we propose Multihead Differential Gated Self-Attention (M-DGSA) that learns per-head input-dependent gating to dynamically suppress attention noise. Each head splits into excitatory and inhibitory branches whose dual softmax maps are fused by a sigmoid gate predicted from the token embedding, yielding a context-aware contrast enhancement. M-DGSA integrates seamlessly into existing Transformer stacks with minimal computational overhead. We evaluate on both vision and language benchmarks, demonstrating consistent robustness gains over vanilla Transformer, Vision Transformer, and Differential Transformer baselines. Our contributions are (i) a novel input-dependent gating mechanism for self-attention grounded in lateral inhibition, (ii) a principled synthesis of biological contrast-enhancement and self-attention theory, and (iii) comprehensive experiments demonstrating noise resilience and cross-domain applicability.
LettuceDetect is a framework designed to identify token-level hallucinations in Retrieval-Augmented Generation (RAG) systems by leveraging a ModernBERT backbone for token classification. The system achieves an F1 score of 79.22% for example-level hallucination detection and 58.93% for span-level detection on the RAGTruth benchmark, processing 30-60 examples per second.
4
This paper presents a method for on-the-fly large-scale novel view synthesis using 3D Gaussian Splatting, enabling real-time generation of 3D scene representations and camera poses as images are captured. It processes a 1.1km path dataset in 25 minutes, providing quality comparable to offline methods, thereby bridging the gap between high-quality offline reconstruction and fast real-time approaches.
196
This research from TU Wien presents a real-time deferred rendering pipeline that integrates JPEG-compressed textures directly onto modern GPUs by selectively decoding texture blocks on demand. The approach achieves a desktop rendering overhead of less than 0.3 ms and a VR overhead of 0.65 ms on an RTX 4090, demonstrating the feasibility of using variable-rate compression for high-fidelity graphics.
26
ReFineVLA enhances Vision-Language-Action models by integrating explicit multimodal reasoning through a teacher-guided fine-tuning framework. This approach improves robotic manipulation performance, particularly in complex tasks, and offers increased interpretability by enabling robots to provide structured rationales for their actions.
There are no more papers matching your filters at the moment.