Towards a Science of Scaling Agent Systems
09 Dec 2025

The paper empirically investigates the performance of multi-agent LLM systems across diverse agentic tasks and architectures, revealing that benefits are highly contingent on task structure rather than universal. It establishes a quantitative scaling principle, achieving 87% accuracy in predicting optimal agent architectures for unseen tasks based on model capability, task properties, and measured coordination dynamics.

View blog
Resources162
On the Interplay of Pre-Training, Mid-Training, and RL on Reasoning Language Models
08 Dec 2025

This research disentangles the causal effects of pre-training, mid-training, and reinforcement learning (RL) on language model reasoning using a controlled synthetic task framework. It establishes that RL extends reasoning capabilities only under specific conditions of pre-training exposure and data calibration, with mid-training playing a crucial role in bridging training stages and improving generalization.

View blog
Resources10
DeepCode: Open Agentic Coding

DeepCode presents a multi-stage agentic framework for autonomously generating executable code repositories from scientific papers, achieving a 73.5% replication score on the PaperBench Code-Dev benchmark and exceeding PhD-level human expert performance.

View blog
Resources11,697
What matters for Representation Alignment: Global Information or Spatial Structure?

A collaborative study from Adobe Research, Australian National University, and New York University investigates the drivers of representation alignment in diffusion models, revealing that the spatial structure of pretrained vision encoder features is more critical for high-quality image generation than global semantic understanding, which challenges a prevalent assumption. The work introduces iREPA, a refined alignment method involving minimal code changes, that consistently enhances the convergence speed and generation quality of Diffusion Transformers across various models and tasks.

View blog
Resources
Stronger Normalization-Free Transformers
11 Dec 2025

Researchers from Princeton, NYU, and Carnegie Mellon introduced Dynamic erf (Derf), a point-wise function that replaces traditional normalization layers in Transformers. Derf consistently achieved higher accuracy or lower error rates across various modalities, including an average of 0.6 percentage points higher top-1 accuracy on ImageNet-1K for ViT-Base/Large models and improved FID scores for Diffusion Transformers.

View blog
Resources2
Bidirectional Normalizing Flow: From Data to Noise and Back

Bidirectional Normalizing Flow (BiFlow) introduces a method to learn an approximate inverse for Normalizing Flows, allowing for highly efficient and high-fidelity image generation. This approach achieves a state-of-the-art FID of 2.39 on ImageNet 256x256 while accelerating sampling speed by up to 697x compared to prior NF models.

View blog
Resources
Closing the Train-Test Gap in World Models for Gradient-Based Planning

Researchers from Columbia University and NYU introduced Online World Modeling (OWM) and Adversarial World Modeling (AWM) to mitigate the train-test gap in world models for gradient-based planning (GBP). These methods enabled GBP to achieve performance comparable to or better than search-based planning algorithms like CEM, while simultaneously reducing computation time by an order of magnitude across various robotic tasks.

View blog
Resources1
A Practical Guide for Designing, Developing, and Deploying Production-Grade Agentic AI Workflows
09 Dec 2025

This work presents a comprehensive engineering guide for designing and deploying production-grade agentic AI workflows, offering nine best practices demonstrated through a multimodal news-to-media generation case study. The approach improves system determinism, reliability, and responsible AI integration, reducing issues like hallucination and enabling scalable, maintainable deployments.

View blog
Resources
StereoWorld: Geometry-Aware Monocular-to-Stereo Video Generation

StereoWorld introduces an end-to-end diffusion-based framework to convert monocular videos into high-fidelity, geometry-aware stereo videos. It leverages a pretrained video generative model, a large-scale IPD-aligned dataset, and explicit disparity and depth supervision to produce superior visual quality, temporal stability, and geometric accuracy for immersive XR experiences.

View blog
Resources
Evaluating Gemini Robotics Policies in a Veo World Simulator
11 Dec 2025

Google DeepMind's Gemini Robotics Team developed a system utilizing the Veo video foundation model to evaluate generalist robot policies. The approach accurately predicts policy performance, generalization capabilities, and identifies safety vulnerabilities across various scenarios, demonstrating strong correlation with real-world outcomes.

View blog
Resources
Native Parallel Reasoner: Reasoning in Parallelism via Self-Distilled Reinforcement Learning
08 Dec 2025

The Native Parallel Reasoner (NPR) framework allows Large Language Models to autonomously acquire and deploy genuine parallel reasoning capabilities, without relying on external teacher models. Experiments show NPR improves accuracy by up to 24.5% over baselines and delivers up to 4.6 times faster inference, maintaining 100% parallel execution across various benchmarks.

View blog
Resources4
VL-JEPA: Joint Embedding Predictive Architecture for Vision-language
11 Dec 2025

Researchers at Meta FAIR developed VL-JEPA, a Joint Embedding Predictive Architecture for vision-language tasks, which predicts abstract semantic embeddings rather than explicit tokens. This approach leads to improved computational efficiency and enables real-time applications through non-autoregressive prediction and selective decoding, while achieving competitive performance across classification, retrieval, and Visual Question Answering benchmarks.

View blog
Resources
Wan-Move: Motion-controllable Video Generation via Latent Trajectory Guidance

Wan-Move presents a framework for motion-controllable video generation that utilizes latent trajectory guidance to directly edit image condition features within a pre-trained image-to-video model. This method yields superior visual quality and precise motion adherence compared to state-of-the-art academic approaches and rivals commercial solutions, while also establishing MoveBench, a new comprehensive evaluation benchmark.

View blog
Resources31
E-RayZer: Self-supervised 3D Reconstruction as Spatial Visual Pre-training

E-RayZer introduces a self-supervised framework for 3D reconstruction, employing explicit 3D Gaussian Splatting to learn camera parameters and scene geometry directly from unlabeled multi-view images. The system achieves competitive performance against supervised methods and outperforms prior self-supervised approaches, establishing itself as an effective spatial visual pre-training model.

View blog
Resources4
Weird Generalization and Inductive Backdoors: New Ways to Corrupt LLMs
10 Dec 2025

Researchers at Truthful AI and UC Berkeley demonstrated that finetuning large language models on narrow, benign datasets can induce broad, unpredictable generalization patterns and novel "inductive backdoors." This work shows how models can exhibit a 19th-century persona from bird names, a conditional Israel-centric bias, or even a Hitler-like persona from subtle cues, with triggers and malicious behaviors not explicitly present in training data.

View blog
Resources
UniUGP: Unifying Understanding, Generation, and Planing For End-to-end Autonomous Driving
10 Dec 2025

UniUGP presents a unified framework for end-to-end autonomous driving, integrating scene understanding, future video generation, and trajectory planning through a hybrid expert architecture. This approach enhances interpretability with Chain-of-Thought reasoning and demonstrates state-of-the-art performance in challenging long-tail scenarios and multimodal capabilities across various benchmarks.

View blog
Resources
Remember Me, Refine Me: A Dynamic Procedural Memory Framework for Experience-Driven Agent Evolution
11 Dec 2025

ReMe introduces a dynamic procedural memory framework for LLM agents, enabling continuous learning and adaptation through a closed-loop system of experience acquisition, reuse, and refinement. This framework allows smaller language models to achieve performance comparable to or surpassing larger, memory-less models, demonstrating a memory-scaling effect and improving agent robustness.

View blog
Resources674
SpaceDrive: Infusing Spatial Awareness into VLM-based Autonomous Driving

SpaceDrive, a framework developed by Mercedes-Benz AG and the University of Tübingen, enhances Vision-Language Models (VLMs) for autonomous driving by integrating explicit 3D spatial awareness through universal positional encodings and a regression-based coordinate decoder. This approach significantly improves open-loop planning accuracy on nuScenes, achieving a 0.32m L2 error and 0.23% collision rate, and enables robust closed-loop performance on Bench2Drive with a 55.11% success rate, addressing prior VLM limitations in precise 3D interaction.

View blog
Resources2
Mind to Hand: Purposeful Robotic Control via Embodied Reasoning
10 Dec 2025

The Astribot Team developed Lumo-1, a Vision-Language-Action (VLA) model that explicitly integrates structured reasoning with physical actions to achieve purposeful robotic control on their Astribot S1 bimanual mobile manipulator. This system exhibits superior generalization to novel objects and instructions, improves reasoning-action consistency through reinforcement learning, and outperforms state-of-the-art baselines in complex, long-horizon, and dexterous tasks.

View blog
Resources
Long-horizon Reasoning Agent for Olympiad-Level Mathematical Problem Solving
11 Dec 2025

Intern-S1-MO, a multi-agent AI system from Shanghai AI Laboratory, extends Large Reasoning Models' capabilities for long-horizon mathematical reasoning using lemma-based memory management and hierarchical reinforcement learning. The system achieved scores equivalent to human silver medalists on IMO2025 non-geometry problems and exceeded the gold medal threshold in the CMO2025 competition, showcasing advanced problem-solving beyond typical context limitations.

View blog
Resources
There are no more papers matching your filters at the moment.