85
Papers
14
Days
2/5/2026
Last Updated

Subliminal Effects in Your Data: A General Mechanism via Log-Linearity

2/5/2026Ishaq Aden-Ali et al.
cs.LGcs.AIcs.CLstat.ML
Reveals that training datasets can transmit 'subliminal' signals to LLMs — statistical patterns not observable from individual datapoints but that systematically influence model behavior through log-linear mechanisms.
Why This Matters
This has major implications for AI safety and data governance: it shows that dataset-level distributional properties can steer model behavior in ways that per-example auditing cannot detect, formalizing a previously observed but poorly understood phenomenon.
Individual data point inspection is insufficient for understanding training data influence on LLMs — aggregate distributional signals can encode behaviors invisible at the sample level, demanding new dataset-level auditing approaches.
Read Paper

Billion-Scale Graph Foundation Models

2/5/2026Maya Bechler-Speicher et al.
cs.LGcs.AI
GraphBFF presents the first recipe for training billion-parameter Graph Foundation Models on arbitrary heterogeneous, billion-scale graphs with lightweight adaptation.
Why This Matters
While foundation models have transformed NLP and vision, graphs have been left behind due to heterogeneity and scale challenges; this work closes that gap with an end-to-end approach that handles real-world graph diversity at unprecedented scale.
General-purpose graph foundation models are now feasible at billion scale, potentially enabling transfer learning across diverse graph domains the way LLMs enabled it for text.
Read Paper

Protein Autoregressive Modeling via Multiscale Structure Generation

2/5/2026Yanru Qu et al.
cs.LGcs.AIq-bio.BMq-bio.QM
PAR introduces the first multi-scale autoregressive framework for protein backbone generation, using coarse-to-fine next-scale prediction that mimics sculpting a statue.
Why This Matters
Autoregressive generation has dominated language and images, but protein structure generation has relied on diffusion and flow-matching; PAR shows autoregressive methods can work for 3D molecular structures when operating across spatial scales rather than sequentially along the chain.
Hierarchical coarse-to-fine autoregressive generation can be a viable alternative to diffusion for structured 3D data like proteins, opening new avenues for biomolecular design.
Read Paper

Fluid Representations in Reasoning Models

2/5/2026Dmitrii Kharlapenko et al.
cs.AI
Mechanistic analysis of reasoning model QwQ-32B reveals that long chain-of-thought enables gradual construction of 'fluid representations' — abstract structural encodings that emerge dynamically during inference.
Why This Matters
This is one of the first mechanistic interpretability studies specifically targeting reasoning-trained models, providing concrete evidence for how extended thinking traces allow models to build up representations they couldn't form in a single forward pass.
Reasoning models don't just 'think longer' — they progressively construct abstract internal representations across their chain of thought, which explains why they dramatically outperform standard models on structural problems.
Read Paper

Reinforced Attention Learning

2/5/2026Bangzheng Li et al.
cs.CLcs.CVcs.LG
Reinforced Attention Learning (RAL) optimizes internal attention distributions in multimodal LLMs using policy gradients, rather than optimizing output token sequences.
Why This Matters
This challenges the dominant paradigm of improving multimodal models through verbose chain-of-thought rationales, showing that directly optimizing where the model looks (attention) is more effective for perception tasks than optimizing what it says.
For multimodal AI, optimizing internal representations via RL may be more effective than optimizing output text, suggesting a new direction for post-training beyond verbose reasoning.
Read Paper

WebSentinel: Detecting and Localizing Prompt Injection Attacks for Web Agents

2/4/2026Xilong Wang et al.
cs.CRcs.AIcs.CL
Introduces WebSentinel, a two-step detection and localization system for prompt injection attacks targeting web-browsing AI agents.
Why This Matters
As LLM-powered web agents become deployed in real products, prompt injection via manipulated webpage content is an urgent security threat; this paper provides the first systematic approach to both detecting and pinpointing injected instructions in web content, which is critical for safe agent deployment.
Practitioners should know that defending web agents requires not just detecting prompt injections but localizing them within page content, and a segment-extraction-then-classification pipeline achieves this more effectively than end-to-end approaches.
Read Paper

Context Compression via Explicit Information Transmission

2/4/2026Jiangnan Ye et al.
cs.CL
Proposes an explicit information transmission approach for soft context compression in LLMs that outperforms existing methods by avoiding the structural limitations of self-attention-based compression.
Why This Matters
Long-context inference cost is one of the most pressing practical bottlenecks for LLM deployment, and this work identifies fundamental problems with using LLMs as their own compressors — offering a principled alternative that could significantly reduce inference costs for long-document applications.
Practitioners should know that repurposing an LLM's own self-attention for context compression introduces structural bottlenecks, and explicit information transmission architectures can achieve better compression ratios with less information loss.
Read Paper

Antidistillation Fingerprinting

2/4/2026Yixuan Even Xu et al.
cs.LGcs.AIcs.CL
Proposes antidistillation fingerprinting, a method to detect when a student model has been distilled from a teacher LLM's outputs without degrading the teacher's generation quality.
Why This Matters
Model distillation of frontier LLMs is a growing intellectual property concern, and existing fingerprinting methods require sacrificing output quality; this work decouples fingerprint strength from quality degradation, addressing a real commercial need for model provenance verification.
Practitioners should know that robust distillation detection is now possible without meaningful degradation to the teacher model's outputs, enabling better enforcement of model usage policies.
Read Paper

Understanding Agent Scaling in LLM-Based Multi-Agent Systems via Diversity

2/4/2026Yingxuan Yang et al.
cs.AIcs.LG
Provides a theoretical and empirical framework explaining why scaling the number of agents in LLM-based multi-agent systems hits diminishing returns, and shows that diversity — not quantity — drives performance gains.
Why This Matters
As multi-agent LLM systems become mainstream for complex tasks, this paper answers a fundamental design question: adding more homogeneous agents wastes compute, but heterogeneous agents (different models, prompts, tools) continue to improve, giving practitioners a concrete scaling strategy.
Practitioners should know that investing in agent diversity (varied models, prompts, and tools) yields far better returns than simply increasing agent count in multi-agent systems.
Read Paper

Reasoning with Latent Tokens in Diffusion Language Models

2/4/2026Andre He et al.
cs.LG
Introduces latent reasoning tokens in discrete diffusion language models that enable implicit computation without generating visible chain-of-thought tokens.
Why This Matters
This bridges the gap between diffusion and autoregressive language models by showing that diffusion models' joint prediction mechanism acts as implicit reasoning, and that latent tokens can recover performance lost when ablating this mechanism — offering a new path to efficient reasoning without verbose outputs.
Practitioners should know that diffusion language models can reason through latent token prediction rather than explicit chain-of-thought, potentially offering more compute-efficient reasoning at inference time.
Read Paper

World-Gymnast: Training Robots with Reinforcement Learning in a World Model

2/3/2026Ansh Kumar Sharma et al.
cs.ROcs.AI
World-Gymnast trains robot policies via reinforcement learning inside learned world models, bypassing sim-to-real gaps and expert data requirements.
Why This Matters
This addresses a fundamental bottleneck in robot learning: physical interaction is expensive, simulators have reality gaps, and expert demos are scarce. Training in learned world models from real video offers a promising middle ground.
Video-based world models have matured enough to serve as viable training environments for manipulation policies that transfer to real robots.
Read Paper

Breaking the Reversal Curse in Autoregressive Language Models via Identity Bridge

2/3/2026Xutao Ma et al.
cs.AI
Identity Bridge breaks the reversal curse in autoregressive LLMs by adding bidirectional identity links during training.
Why This Matters
The reversal curse (knowing 'A is B' but not inferring 'B is A') has been considered a fundamental limitation of autoregressive models. Demonstrating a simple fix challenges assumptions about what LLMs can and cannot learn.
Autoregressive models can learn bidirectional knowledge if training data includes explicit identity bridges, not just forward associations.
Read Paper

AgentRx: Diagnosing AI Agent Failures from Execution Trajectories

2/3/2026Shraddha Barke et al.
cs.AI
AgentRx provides the first benchmark of 115 annotated failed agent trajectories with a grounded failure taxonomy for diagnosing AI agent breakdowns.
Why This Matters
As LLM agents move into production, understanding why they fail becomes critical. This benchmark fills a major gap by providing systematic annotations of failure modes across API workflows, incident management, and web tasks.
Agent failures cluster into identifiable categories that can be diagnosed from execution traces, enabling targeted improvements to agent architectures.
Read Paper

PixelGen: Pixel Diffusion Beats Latent Diffusion with Perceptual Loss

2/3/2026Zehong Ma et al.
cs.CVcs.AI
PixelGen shows pixel-space diffusion can outperform latent diffusion when trained with perceptual loss instead of pixel-wise objectives.
Why This Matters
This challenges the dominant latent diffusion paradigm by demonstrating that the two-stage VAE approach may not be necessary. Removing the VAE bottleneck could eliminate common artifacts and simplify the image generation pipeline.
The key to successful pixel diffusion is not modeling the full image manifold but focusing on perceptually relevant signals.
Read Paper

MEG-XL: Data-Efficient Brain-to-Text via Long-Context Pre-Training

2/3/2026Dulhan Jayalath, Oiwi Parker Jones
cs.LGq-bio.NC
MEG-XL enables brain-to-text interfaces with 2.5 minutes of MEG context, achieving data-efficient decoding for paralyzed patients.
Why This Matters
Brain-computer interfaces for communication require minimal training data from patients who cannot provide extensive recordings. This 5-300x longer context window represents a significant step toward practical clinical deployment of neural speech decoding.
Long-context pre-training on neural signals dramatically improves generalization across subjects, making BCIs more viable for real patients.
Read Paper

VideoGPA: Distilling Geometry Priors for 3D-Consistent Video Generation

2/2/2026Hongyang Du et al.
cs.CVcs.AIcs.LG
VideoGPA uses geometry foundation models to enforce 3D consistency in video diffusion through self-supervised preference alignment.
Why This Matters
Current video generation models produce visually impressive but geometrically inconsistent content with object deformation and spatial drift. This data-efficient approach addresses the fundamental gap between visual quality and physical plausibility.
Standard denoising objectives lack geometric coherence incentives, but leveraging external geometry priors through preference alignment can teach video models to maintain 3D structure without explicit 3D supervision.
Read Paper

Scaling Multiagent Systems with Process Rewards

2/2/2026Ed Li et al.
cs.AIcs.CLcs.ETcs.MA
MAPPA uses per-action process rewards from AI feedback to solve credit assignment and sample efficiency in multiagent system finetuning.
Why This Matters
Scaling multiagent systems is a key challenge for complex task automation, and this work tackles the twin problems of identifying which agent contributed to success and reducing expensive rollout costs.
Assigning credit at the action level rather than task completion enables efficient finetuning of specialized agent teams without requiring exponentially more multiagent rollouts.
Read Paper

Now You Hear Me: Audio Narrative Attacks Against Large Audio-Language Models

2/2/2026Ye Yu et al.
cs.CLcs.AIcs.CR
Audio narrative attacks embed jailbreak directives within natural-sounding audio streams to bypass safety filters in large audio-language models.
Why This Matters
As voice interfaces become ubiquitous, this exposes a critical security gap where harmful instructions can be hidden in conversational audio. The attack surface for multimodal AI systems is expanding faster than defenses.
Safety mechanisms developed for text modalities do not transfer well to audio, and narrative-style audio can disguise malicious content that would be flagged in text form.
Read Paper

FOCUS: DLLMs Know How to Tame Their Compute Bound

2/2/2026Kaihua Liang et al.
cs.LGcs.ARcs.CL
FOCUS identifies that most compute in diffusion language models is wasted on non-decodable tokens and proposes attention-guided selective computation.
Why This Matters
Diffusion LLMs are gaining traction as an alternative to autoregressive models, but deployment costs remain prohibitive. This work addresses a fundamental inefficiency that could make DLLMs practical for real-world applications.
Attention-derived token importance strongly predicts which tokens are decodable at each diffusion step, enabling significant compute savings by focusing resources on the tokens that matter.
Read Paper

YuriiFormer: A Suite of Nesterov-Accelerated Transformers

2/2/2026Aleksandr Zimin et al.
cs.LGcs.AImath.OCstat.ML
YuriiFormer reinterprets transformer layers as optimization steps and applies Nesterov acceleration to achieve faster convergence.
Why This Matters
This variational framework provides a principled theoretical lens for understanding transformers while delivering practical speedups. It bridges optimization theory and deep learning architecture design in a way that could inspire next-generation architectures.
Viewing self-attention as gradient descent on an interaction energy enables momentum-based acceleration techniques from classical optimization to be applied directly to transformer training.
Read Paper

Making Foundation Models Probabilistic via Singular Value Ensembles

2/1/2026Mehmet Ozgur Turkoglu et al.
cs.LG
Singular Value Ensembles enable uncertainty quantification in foundation models by ensembling only the singular value components, avoiding the cost of training multiple full models.
Why This Matters
Foundation models are notoriously overconfident, but training ensembles is prohibitively expensive—this method provides a practical path to calibrated uncertainty estimates by exploiting the structure of pretrained weights.
Ensembling at the singular value level rather than the full model level provides meaningful uncertainty estimates while being computationally tractable for large foundation models.
Read Paper

Thinking Out of Order: When Output Order Stops Reflecting Reasoning Order in Diffusion Language Models

2/1/2026Longxuan Yu et al.
cs.CLcs.AI
Demonstrates that masked diffusion language models can reason 'out of order', generating answers and explanations non-sequentially unlike autoregressive models.
Why This Matters
This challenges the fundamental assumption that reasoning must proceed left-to-right, showing diffusion models can naturally handle cases where output structure conflicts with reasoning order—a limitation that forces AR models into premature commitment.
For tasks requiring answer-before-explanation output formats, diffusion language models may be inherently better suited than autoregressive models due to their flexible generation order.
Read Paper

Do VLMs Perceive or Recall? Probing Visual Perception vs. Memory with Classic Visual Illusions

2/1/2026Xiaoxiao Sun et al.
cs.CV
Reveals that vision-language models answer visual illusions correctly by recalling memorized patterns rather than actually perceiving the visual content.
Why This Matters
This exposes a fundamental limitation in how VLMs process visual information—they may be sophisticated pattern matchers rather than true visual reasoners, with major implications for safety-critical applications.
Testing VLMs with inverted illusions where human perception clearly changes but model responses don't is a powerful diagnostic for distinguishing genuine perception from memorization.
Read Paper

DynamicVLA: A Vision-Language-Action Model for Dynamic Object Manipulation

2/1/2026Haozhe Xie et al.
cs.ROcs.CV
DynamicVLA introduces a compact 0.4B vision-language-action model that can manipulate moving objects through temporal reasoning and closed-loop control.
Why This Matters
While VLA models excel at static manipulation, real-world robotics requires handling dynamic objects—this work directly addresses that gap with a surprisingly small model that integrates temporal anticipation.
A convolutional vision encoder combined with explicit temporal reasoning mechanisms can enable dynamic manipulation without the computational overhead of massive VLA models.
Read Paper

One-step Latent-free Image Generation with Pixel Mean Flows

2/1/2026Yiyang Lu et al.
cs.CV
Pixel MeanFlow enables one-step image generation directly in pixel space without latents, achieving quality comparable to multi-step diffusion models.
Why This Matters
This represents a significant simplification of the generative image pipeline by eliminating both the need for multiple sampling steps and latent space encoding, potentially enabling real-time high-quality image generation on resource-constrained devices.
Separating the network output space from the loss space allows direct pixel-space generation to work effectively, challenging the assumption that latent spaces are necessary for quality.
Read Paper

DynaWeb: Model-Based Reinforcement Learning of Web Agents

1/31/2026Hang Ding et al.
cs.CLcs.AI
DynaWeb applies model-based RL to train web agents by learning a world model of websites, enabling safer and more efficient training.
Why This Matters
Training web agents on the live internet is risky, costly, and inefficient - learning a world model enables simulated practice at scale, addressing a key bottleneck in autonomous web agent development.
The path to reliable web agents may run through learned simulators rather than direct internet interaction.
Read Paper

Pay for Hints, Not Answers: LLM Shepherding for Cost-Efficient Inference

1/31/2026Ziming Dong et al.
cs.LG
LLM Shepherding uses large models to provide short hints to small models rather than complete answers, dramatically reducing inference costs.
Why This Matters
This escapes the all-or-nothing tradeoff between cheap-but-weak SLMs and expensive-but-capable LLMs by using LLMs as consultants rather than workers, achieving significant cost savings while maintaining quality.
Sometimes the most cost-effective use of a powerful LLM is asking it for a hint, not an answer.
Read Paper

Reasoning While Asking: Transforming Reasoning Large Language Models from Passive Solvers to Proactive Inquirers

1/31/2026Xin Chen et al.
cs.CLcs.AI
Proposes Proactive Interactive Reasoning (PIR) where LLMs ask clarifying questions during reasoning instead of making assumptions.
Why This Matters
Current reasoning models perform 'blind self-thinking' even when critical information is missing - PIR fundamentally changes this by having models proactively seek clarification, reducing hallucination and improving reliability.
The next evolution of reasoning models isn't just thinking harder, but knowing when to ask for help.
Read Paper

Exploring Reasoning Reward Model for Agents

1/31/2026Kaixuan Fan et al.
cs.AIcs.CL
Introduces Agent-RRM, a multi-faceted reward model that provides structured feedback for agentic trajectories beyond sparse outcome rewards.
Why This Matters
Current agentic RL relies on binary success/failure signals which can't distinguish good reasoning from lucky outcomes - this enables dense, interpretable feedback that improves agent training quality.
Training better AI agents requires evaluating the reasoning process itself, not just final outcomes.
Read Paper

Discovering Hidden Gems in Model Repositories

1/31/2026Jonathan Kahana et al.
cs.LGcs.CL
Systematically evaluates 2,000+ models on HuggingFace to discover overlooked fine-tunes that outperform popular foundation models.
Why This Matters
This challenges the assumption that model popularity reflects quality - many 'hidden gem' fine-tunes significantly outperform their popular counterparts, suggesting the ML community is leaving performance on the table.
Before defaulting to popular checkpoints, practitioners should search model repositories more thoroughly as better-performing alternatives often exist with minimal downloads.
Read Paper

Early and Prediagnostic Detection of Pancreatic Cancer from Computed Tomography

1/30/2026Wenxuan Li et al.
cs.CV
ePAI is an automated system for early detection of pancreatic cancer from CT scans, identifying lesions that radiologists previously overlooked.
Why This Matters
Pancreatic cancer is often detected too late for surgery. Retrospective studies show expert radiologists can spot lesions in prediagnostic scans when they know the patient later developed cancer—this AI aims to enable that foresight prospectively.
AI can potentially catch subtle pancreatic lesions that humans miss on routine scans, enabling earlier intervention for one of the deadliest cancers.
Read Paper

SWE-Replay: Efficient Test-Time Scaling for Software Engineering Agents

1/30/2026Yifeng Ding, Lingming Zhang
cs.SEcs.AIcs.LG
SWE-Replay enables efficient test-time scaling for software engineering agents by replaying and branching from successful trajectory prefixes instead of sampling from scratch.
Why This Matters
Standard test-time scaling for SWE agents wastes compute by resampling entire trajectories. This work shows that leveraging successful partial trajectories can dramatically reduce inference costs while maintaining performance.
For software engineering tasks, trajectory prefixes contain reusable computation—branching from successful partial solutions is far more efficient than repeated full sampling.
Read Paper

StepShield: When, Not Whether to Intervene on Rogue Agents

1/30/2026Gloria Felicia et al.
cs.LGcs.AIcs.CRcs.SE
StepShield is the first benchmark measuring WHEN safety violations are detected in AI agent trajectories, not just whether they're detected.
Why This Matters
This reframes agent safety evaluation from binary accuracy to intervention timing—a detector flagging violations at step 8 enables prevention, while step 48 detection is merely forensic. This distinction is critical for deploying agents safely.
Early detection of rogue agent behavior is fundamentally different from post-mortem analysis; safety benchmarks must measure detection latency, not just accuracy.
Read Paper

Hybrid Linear Attention Done Right: Efficient Distillation and Effective Architectures for Extremely Long Contexts

1/30/2026Yingfa Chen et al.
cs.CLcs.AIcs.LG
Introduces efficient distillation methods to convert pretrained softmax attention Transformers into hybrid linear attention architectures for extremely long contexts.
Why This Matters
This tackles the prohibitive cost of training long-context models from scratch by enabling conversion of existing models. The hybrid approach preserves quality while dramatically improving throughput for long sequences.
You don't need to pretrain hybrid attention models from scratch—distillation from existing Transformers can achieve comparable quality with much better long-context efficiency.
Read Paper

RedSage: A Cybersecurity Generalist LLM

1/30/2026Naufal Suryanto et al.
cs.CRcs.AIcs.CL
RedSage is a cybersecurity-specialized LLM trained on 11.8B tokens of security-focused data spanning frameworks, offensive techniques, and tools.
Why This Matters
This addresses a critical gap in security operations where proprietary APIs pose privacy risks and open models lack domain expertise. The 28.6K curated documents across security domains could enable safer, on-premise security assistants.
Domain-specific continual pretraining with carefully curated security corpora can bridge the gap between general LLMs and specialized cybersecurity workflows without exposing sensitive data to external APIs.
Read Paper

Training Reasoning Models on Saturated Problems via Failure-Prefix Conditioning

1/29/2026Minwu Kim et al.
cs.LGcs.AIcs.CL
Addresses training stagnation on saturated problems by conditioning rollouts on failure prefixes to find informative learning signals.
Why This Matters
As models get better at benchmarks, most training samples become trivially solved - this elegant technique keeps learning productive by specifically seeking out the edge cases where the model still fails.
When RL training plateaus because the model solves most problems, start rollouts from known failure points rather than from scratch to maintain learning signal.
Read Paper

Reinforcement Learning via Self-Distillation

1/29/2026Jonas Hübotter et al.
cs.LGcs.AI
Proposes using rich textual feedback (like error messages) for RL credit assignment in LLMs through a self-distillation approach.
Why This Matters
Current RLVR methods waste valuable signal by only using pass/fail outcomes - this method leverages the detailed feedback that verifiable environments already provide, like compiler errors and test failures.
When training on code or math, convert error messages and judge feedback into learning signal rather than discarding everything except the binary reward.
Read Paper

SokoBench: Evaluating Long-Horizon Planning and Reasoning in Large Language Models

1/29/2026Sebastiano Monti et al.
cs.AI
Introduces SokoBench to systematically evaluate long-horizon planning in LLMs using simplified Sokoban puzzles that isolate planning from state tracking.
Why This Matters
Despite claims of improved reasoning, this benchmark reveals that even state-of-the-art reasoning models struggle with multi-step planning when they can't rely on pattern matching from training data.
Current LRMs may be better at reasoning that looks like training examples than genuine novel planning - evaluate planning capabilities separately from pattern-matching reasoning.
Read Paper

Linear representations in language models can change dramatically over a conversation

1/29/2026Andrew Kyle Lampinen et al.
cs.CLcs.LG
Shows that linear representations of concepts in LLMs can flip dramatically within a single conversation, with factual information becoming represented as non-factual.
Why This Matters
This challenges the assumption that mechanistic interpretability findings about static representations generalize to dynamic multi-turn interactions, which is how LLMs are actually used.
Interpretability researchers should validate their linear probes across conversation turns, not just single inputs, as representation dynamics can invalidate static analysis.
Read Paper

Reward Models Inherit Value Biases from Pretraining

1/29/2026Brian Christian et al.
cs.LGcs.AIcs.CLcs.CY
Demonstrates that reward models inherit systematic value biases from their pretrained LLM initializations, affecting alignment outcomes.
Why This Matters
This reveals a critical blind spot in RLHF pipelines - the reward models we trust to align LLMs carry their own biases from pretraining, which could systematically skew what behaviors get reinforced.
Practitioners should audit reward models for inherited biases before deployment, as these biases persist through fine-tuning and can silently shape aligned model behavior.
Read Paper

GAVEL: Towards rule-based safety through activation monitoring

1/28/2026Shir Rozenfeld et al.
cs.AIcs.CRcs.LG
GAVEL introduces rule-based activation monitoring for LLM safety, allowing interpretable, shareable safety rules modeled on cybersecurity practices.
Why This Matters
Current activation-based safety approaches suffer from poor precision and lack interpretability. By treating activations as cognitive signatures that can be matched against explicit rules, this work enables more precise, flexible, and explainable safety monitoring - crucial as LLMs are deployed in high-stakes applications.
Activation patterns can be modeled as shareable safety rules similar to cybersecurity threat signatures, enabling collaborative and interpretable safety monitoring across deployments.
Read Paper

Neural Neural Scaling Laws

1/28/2026Michael Y. Hu et al.
cs.LGcs.CL
Uses neural networks to predict neural scaling laws, revealing that individual task performance follows diverse patterns obscured by aggregate metrics like validation loss.
Why This Matters
Understanding how specific capabilities emerge with scale is crucial for efficient AI development. This work shows that simple power-law predictions fail for individual tasks, and proposes learned predictors that can forecast which capabilities will improve, plateau, or degrade.
Validation perplexity is a poor proxy for downstream task performance; practitioners should expect diverse scaling behaviors across tasks and consider task-specific predictions when planning compute allocation.
Read Paper

Visual Generation Unlocks Human-Like Reasoning through Multimodal World Models

1/28/2026Jialong Wu et al.
cs.AI
Demonstrates that visual generation capabilities unlock human-like multimodal reasoning in AI systems, enabling manipulation of internal world models through imagery.
Why This Matters
While chain-of-thought reasoning has achieved expert performance in text-based domains, visual reasoning remains weak. This work suggests that the ability to generate and manipulate visual representations is key to bridging this gap, mirroring how humans reason spatially and visually.
Integrating visual generation into reasoning pipelines may be essential for AI systems to match human performance on tasks requiring spatial, physical, or visual understanding.
Read Paper

Post-LayerNorm Is Back: Stable, ExpressivE, and Deep

1/28/2026Chen Chen, Lai Wei
cs.LGcs.CL
Rehabilitates Post-LayerNorm for deep transformers by identifying and fixing its central failure mode, enabling stable training at extreme depths with superior expressivity.
Why This Matters
As LLM scaling via width and context length hits diminishing returns, depth scaling becomes crucial. This work reopens a promising direction that was abandoned due to training instability, potentially unlocking new scaling frontiers.
The instability of Post-LN at scale stems from specific failure modes that can be addressed, making depth scaling a viable alternative to width scaling for improving model capabilities.
Read Paper

Self-Distillation Enables Continual Learning

1/28/2026Idan Shenfeld et al.
cs.LG
Self-Distillation Fine-Tuning (SDFT) enables continual learning from demonstrations without forgetting, by having models learn from their own on-policy generations rather than off-policy expert data.
Why This Matters
This addresses a fundamental limitation in foundation model training - the ability to learn new skills without degrading existing ones. Unlike RL-based approaches, SDFT doesn't require explicit reward functions, making it practical for real-world deployment where rewards are hard to define.
Self-distillation during fine-tuning can serve as a simple yet effective regularizer against catastrophic forgetting, potentially replacing complex replay buffers or architecture modifications.
Read Paper

SMART: Scalable Mesh-free Aerodynamic Simulations from Raw Geometries using a Transformer-based Surrogate Model

1/27/2026Jan Hagnberger, Mathias Niepert
cs.LGcs.AIcs.CVcs.NE
SMART enables mesh-free aerodynamic simulations directly from raw 3D geometries using transformers, eliminating costly mesh generation
Why This Matters
Generating simulation meshes for new geometries is a major bottleneck in engineering workflows. By achieving comparable accuracy without requiring mesh generation, this could dramatically accelerate design iteration cycles for cars, aircraft, and other complex geometries.
Transformer architectures can learn to directly map raw geometry to physical simulation outputs, potentially replacing expensive mesh-dependent pipelines in engineering CAD workflows.
Read Paper

TSRBench: A Comprehensive Multi-task Multi-modal Time Series Reasoning Benchmark for Generalist Models

1/27/2026Fangxu Yu et al.
cs.AIcs.LG
TSRBench provides the first comprehensive benchmark for testing LLM reasoning capabilities on time series data across multiple modalities and task types
Why This Matters
Time series reasoning is ubiquitous (energy, traffic, finance) yet absent from existing generalist model benchmarks. This fills a critical gap in understanding whether foundation models can actually reason about temporal patterns, not just process them.
Generalist models claiming broad reasoning capabilities should be tested on time series tasks - this benchmark reveals whether temporal reasoning is a genuine capability or a gap in current systems.
Read Paper

HalluGuard: Demystifying Data-Driven and Reasoning-Driven Hallucinations in LLMs

1/27/2026Xinyue Zeng et al.
cs.LGcs.AI
HalluGuard introduces a unified framework distinguishing data-driven vs reasoning-driven hallucinations with a Hallucination Risk Boundary theory
Why This Matters
Most hallucination detection methods address only one failure mode. By providing both theoretical grounding (risk boundaries) and practical detection across both hallucination types, this enables more robust deployment in high-stakes domains like healthcare and law.
Hallucinations have fundamentally different causes (training data gaps vs flawed reasoning chains) requiring different detection and mitigation strategies.
Read Paper

MortalMATH: Evaluating the Conflict Between Reasoning Objectives and Emergency Contexts

1/27/2026Etienne Lanzeray et al.
cs.CL
MortalMATH benchmark reveals that reasoning-optimized LLMs exhibit dangerous 'tunnel vision' - solving math problems while ignoring described life-threatening emergencies
Why This Matters
This exposes a critical safety gap in current reasoning models: the optimization for task completion can override basic safety awareness. Finding that generalist models balance both while reasoning specialists ignore emergencies has major deployment implications.
Deep reasoning optimization may come at the cost of contextual awareness - teams deploying reasoning models should evaluate whether their models can recognize when to break from task focus.
Read Paper

Teaching Models to Teach Themselves: Reasoning at the Edge of Learnability

1/27/2026Shobhita Sundaram et al.
cs.LGcs.CL
SOAR enables LLMs to generate their own curriculum to escape learning plateaus on problems they initially cannot solve
Why This Matters
This addresses a fundamental limitation of RL for reasoning models - when initial success rates are too low, there's no training signal. The meta-RL approach of having a teacher model generate pedagogical problems unlocks learning on previously intractable tasks.
Models can leverage latent knowledge to bootstrap their own learning through self-generated curricula, potentially enabling training on harder reasoning problems without requiring larger datasets.
Read Paper

LoL: Longer than Longer, Scaling Video Generation to Hour

1/26/2026Justin Cui et al.
cs.CVcs.AI
LoL scales video generation to hour-long coherent videos by solving the 'sink-collapse' problem where autoregressive models repeatedly revert to anchor frames.
Why This Matters
Long-form video generation has been stuck at minutes due to error accumulation. Identifying and solving sink-collapse enables a 60x+ increase in generation length, opening practical applications in film and content creation.
Attention sink frames, while helpful for short-term coherence, cause catastrophic cyclic patterns in long generation - the solution requires explicit mechanisms to prevent content regression to sink frames.
Read Paper

Reasoning Promotes Robustness in Theory of Mind Tasks

1/26/2026Ian B. de Haan et al.
cs.AIcs.CL
Shows that reasoning-oriented LLMs trained with reinforcement learning achieve more robust Theory of Mind performance than standard models.
Why This Matters
As LLMs are deployed in social contexts, understanding whether they truly model mental states or exploit surface patterns is crucial. The finding that explicit reasoning improves robustness suggests a path toward more reliable social AI.
RLVR-trained reasoning models maintain ToM performance under adversarial conditions where standard LLMs fail, indicating reasoning chains provide genuine robustness rather than just benchmark gaming.
Read Paper

GRIP: Algorithm-Agnostic Machine Unlearning for Mixture-of-Experts via Geometric Router Constraints

1/26/2026Andy Zhu et al.
cs.LGcs.AI
GRIP enables machine unlearning for Mixture-of-Experts models by preventing routers from simply redirecting queries instead of actually erasing knowledge.
Why This Matters
As MoE architectures become standard for large models (Mixtral, GPT-4), the discovery that traditional unlearning methods exploit routing rather than truly forgetting is a critical safety finding with immediate practical implications.
MoE unlearning requires geometric constraints on routers to prevent the 'routing escape' vulnerability where models appear to forget but actually just avoid activating knowledgeable experts.
Read Paper

A Scalable Measure of Loss Landscape Curvature for Analyzing the Training Dynamics of LLMs

1/26/2026Dayal Singh Kalra et al.
cs.LGcond-mat.dis-nncs.AIstat.ML
Introduces a scalable method to measure loss landscape curvature in LLMs without computing the full Hessian, enabling analysis of training dynamics at scale.
Why This Matters
Understanding curvature evolution is fundamental to training stability but has been computationally prohibitive for modern LLMs. This opens a window into understanding why certain learning rate schedules and optimizers work.
The proposed curvature measure reveals interactions between learning rate and sharpness throughout training that were previously unmeasurable at LLM scale.
Read Paper

Auto-Regressive Masked Diffusion Models

1/26/2026Mahdi Karami, Ali Ghodsi
cs.LG
ARMD unifies autoregressive and masked diffusion models, achieving competitive language modeling performance with parallel generation capabilities.
Why This Matters
This bridges a fundamental gap between two dominant generative paradigms - ARMs excel at quality while MDMs enable parallel generation. The unified architecture could reshape how we think about efficient text generation.
Reframing masked diffusion through an autoregressive lens allows training efficiency of ARMs while preserving parallel decoding, suggesting hybrid approaches may outperform pure paradigms.
Read Paper

Multimodal Climate Disinformation Detection: Integrating Vision-Language Models with External Knowledge Sources

1/25/2026Marzieh Adeli Shamsabad, Hamed Ghodrati
cs.AI
Combines vision-language models with external knowledge retrieval to detect climate disinformation in images and videos, overcoming VLM knowledge cutoff limitations.
Why This Matters
As multimodal disinformation becomes more sophisticated, this addresses a real blind spot in VLMs—their inability to reason about events after their training cutoff—which is particularly important for fast-evolving topics like climate science.
Retrieval-augmented VLMs can significantly improve disinformation detection by grounding model reasoning in current, verified external knowledge sources.
Read Paper

TeNet: Text-to-Network for Compact Policy Synthesis

1/25/2026Ariyan Bighashdel, Kevin Sebastian Luck
cs.ROcs.AI
TeNet uses a hypernetwork conditioned on LLM text embeddings to generate compact, task-specific robot policies directly from natural language instructions.
Why This Matters
This elegantly sidesteps the deployment problem of large end-to-end models by generating small executable policies on-the-fly, making real-time robot control from language practical on commodity hardware.
Instead of running large models at inference time, you can use them to generate small, specialized policies that execute efficiently on robots.
Read Paper

HyperAlign: Hypernetwork for Efficient Test-Time Alignment of Diffusion Models

1/25/2026Xin Xie et al.
cs.CV
Uses a hypernetwork to efficiently align diffusion models with human preferences at test time, avoiding the diversity loss of fine-tuning and the compute cost of test-time scaling.
Why This Matters
This addresses a critical practical problem—aligning image generation with user intent—while avoiding the major pitfalls of current approaches: reward hacking from fine-tuning and slow inference from test-time optimization.
Hypernetwork-based alignment can provide a middle ground between expensive retraining and slow test-time scaling for diffusion models.
Read Paper

Decoupling Return-to-Go for Efficient Decision Transformer

1/25/2026Yongyi Wang et al.
cs.AI
Shows that feeding the entire Return-to-Go sequence into Decision Transformers is redundant—only the most recent RTG affects action prediction.
Why This Matters
This identifies a fundamental inefficiency in the popular Decision Transformer architecture that many practitioners have overlooked, enabling significant computational savings without sacrificing performance in offline RL.
When using Decision Transformers, you can decouple RTG from the sequence modeling to reduce computation while maintaining the same action prediction quality.
Read Paper

Masked Modeling for Human Motion Recovery Under Occlusions

1/25/2026Zhiyin Qian et al.
cs.CV
Introduces a masked modeling approach for human motion reconstruction that handles occlusions without slow diffusion or optimization-based methods.
Why This Matters
This bridges the gap between fast but fragile regression methods and robust but slow optimization/diffusion approaches for motion capture, which is critical for real-world AR/VR and robotics applications where occlusions are common.
Practitioners can achieve robust motion reconstruction under occlusion using efficient masked modeling rather than expensive diffusion-based approaches.
Read Paper

Natural Language-Driven Global Mapping of Martian Landforms

1/24/2026Yiran Wang et al.
cs.AIastro-ph.IM
MarScope enables natural language-driven mapping of Martian landforms by aligning planetary images with text in a shared semantic space, trained on 200,000+ curated image-text pairs.
Why This Matters
This transforms how scientists can explore planetary surfaces - instead of pixel-level analysis, researchers can query vast orbital image archives using natural language descriptions, enabling open-ended discovery at planetary scale.
Vision-language models can be successfully adapted to scientific domains like planetary science by curating domain-specific image-text pairs, enabling semantic search over imagery that was previously only accessible through manual inspection.
Read Paper

PUMA: Perception-driven Unified Foothold Prior for Mobility Augmented Quadruped Parkour

1/24/2026Liang Wang et al.
cs.ROcs.AIcs.LG
PUMA enables quadruped robots to perform parkour by learning perception-driven foothold priors that guide agile locomotion over obstacles.
Why This Matters
This bridges the gap between human-like perceptual reasoning about terrain and robotic locomotion, moving beyond pre-computed footholds to real-time adaptive foothold selection - a key step toward truly agile legged robots.
Integrating learned foothold priors from visual perception directly into the reinforcement learning policy allows robots to dynamically adapt their gait to complex terrain without hierarchical controllers.
Read Paper

Sawtooth Wavefront Reordering: Enhanced CuTile FlashAttention on NVIDIA GB10

1/24/2026Yifan Zhu et al.
cs.PFcs.AIcs.LGcs.OS
Introduces Sawtooth Wavefront Reordering, a technique that reduces L2 cache misses in FlashAttention implementations on NVIDIA GB10 by over 50%.
Why This Matters
With attention being the computational bottleneck in LLMs, a 50% reduction in cache misses on the latest NVIDIA hardware directly translates to faster and more efficient inference, making this immediately applicable to production systems.
Reordering the wavefront pattern of tile processing in attention kernels can dramatically improve memory locality and cache utilization on modern GPU architectures.
Read Paper

Why Can't I Open My Drawer? Mitigating Object-Driven Shortcuts in Zero-Shot Compositional Action Recognition

1/24/2026Geo Ahn et al.
cs.CVcs.AI
Identifies and mitigates object-driven verb shortcuts in zero-shot compositional action recognition, where models incorrectly rely on objects rather than actions to make predictions.
Why This Matters
This exposes a fundamental failure mode in video understanding models that has been overlooked - models take shortcuts by recognizing objects instead of understanding actions, which undermines compositional generalization to unseen verb-object combinations.
When training action recognition models, the asymmetric learning difficulty between verbs and objects combined with sparse compositional supervision leads models to ignore verbs entirely.
Read Paper

CamPilot: Improving Camera Control in Video Diffusion Model with Efficient Camera Reward Feedback

1/24/2026Wenhang Ge et al.
cs.CV
CamPilot improves camera control in video diffusion models by introducing a specialized reward model for video-camera alignment and efficient reward feedback learning.
Why This Matters
Camera controllability is a major limitation in current video generation models, and this work addresses it with a novel reward-based approach that could significantly improve the quality of AI-generated videos for filmmaking and content creation.
Reward feedback learning can be adapted for video generation by building task-specific reward models that assess alignment between intended camera movements and generated video.
Read Paper

PhysicsMind: Sim and Real Mechanics Benchmarking for Physical Reasoning and Prediction in Foundational VLMs and World Models

1/24/2026Chak-Wing Mak et al.
cs.CVcs.AI
PhysicsMind benchmarks how well foundation VLMs and video world models understand physical mechanics through both simulated and real-world scenarios.
Why This Matters
While MLLMs excel at many reasoning tasks, their grasp of physics is underexplored. Existing benchmarks use synthetic VQA or focus on perceptual quality rather than physical law adherence—this provides a rigorous test of physical understanding that's crucial for embodied AI.
Current foundation models show significant gaps between visual/mathematical reasoning abilities and understanding of physical mechanics, highlighting a key area for improvement.
Read Paper

DTP: A Simple yet Effective Distracting Token Pruning Framework for Vision-Language Action Models

1/24/2026Chenyang Li et al.
cs.CVcs.RO
DTP identifies and prunes 'distracting tokens' in Vision-Language Action models that cause robots to attend to task-irrelevant image regions during manipulation.
Why This Matters
VLA models for robotics inherit attention patterns from general VLMs that aren't optimized for action generation. This simple pruning framework improves manipulation success rates by focusing attention on task-relevant regions.
Robot manipulation performance improves when VLA models are explicitly guided to ignore visually salient but task-irrelevant image tokens during action prediction.
Read Paper

THOR: A Versatile Foundation Model for Earth Observation Climate and Society Applications

1/24/2026Theodor Forgaard et al.
eess.IVcs.AI
THOR is a compute-adaptive Earth observation foundation model that unifies heterogeneous Sentinel satellite data while allowing flexible accuracy-compute tradeoffs at deployment.
Why This Matters
Current EO foundation models are architecturally rigid and struggle with multi-sensor heterogeneity. THOR's ability to process native resolutions from Sentinel-1/2/3 and adapt computation at inference makes it practically deployable for real-world climate monitoring.
Foundation models for remote sensing need native multi-sensor support and compute adaptivity to be useful in operational settings with varying resource constraints.
Read Paper

Beyond Predictive Uncertainty: Reliable Representation Learning with Structural Constraints

1/24/2026Yiyao Yang
stat.MLcs.LG
Proposes treating reliability as a first-class property of learned representations themselves, not just prediction outputs, with structural constraints for uncertainty quantification.
Why This Matters
Most uncertainty estimation focuses on final predictions while assuming representations are reliable by default. This challenges that assumption and provides a principled framework for building trustworthy representations—critical for high-stakes ML applications.
Representation-level uncertainty should be explicitly modeled and constrained during training, not treated as an afterthought at prediction time.
Read Paper

PyraTok: Language-Aligned Pyramidal Tokenizer for Video Understanding and Generation

1/24/2026Onkar Susladkar et al.
cs.CVcs.AI
PyraTok introduces a language-aligned pyramidal video tokenizer that learns discrete visual representations across multiple spatiotemporal scales with strong text supervision.
Why This Matters
Current video tokenizers operate at single scales with weak language alignment, limiting zero-shot transfer. This hierarchical approach with deep language supervision could significantly improve text-to-video generation quality and enable better cross-modal understanding.
Multi-scale tokenization with explicit language alignment at each level produces more semantically meaningful video representations than flat single-scale approaches.
Read Paper

DextER: Language-driven Dexterous Grasp Generation with Embodied Reasoning

1/22/2026Junha Lee et al.
cs.ROcs.CV
DextER introduces contact-based embodied reasoning for dexterous grasping, having vision-language models explicitly reason about hand-object physical interactions.
Why This Matters
Previous VLA approaches mapped observations directly to grasp parameters without intermediate reasoning - adding explicit contact reasoning significantly improves manipulation success rates on complex multi-finger tasks.
For robotic manipulation, having models explicitly reason about physical contact points before generating actions produces more robust grasps than end-to-end approaches.
Read Paper

Replicating Human Motivated Reasoning Studies with LLMs

1/22/2026Neeley Pate et al.
cs.HCcs.AI
Replicates classic human motivated reasoning studies on LLMs and finds that base models don't exhibit the same politically-motivated biases humans show.
Why This Matters
As LLMs are increasingly used to study or simulate human behavior, understanding where their reasoning diverges from humans is crucial - this suggests LLMs may process politically charged information more neutrally than humans.
LLMs should not be assumed to replicate human cognitive biases without empirical validation, especially for motivated reasoning in political contexts.
Read Paper

DSFedMed: Dual-Scale Federated Medical Image Segmentation via Mutual Distillation Between Foundation and Lightweight Models

1/22/2026Hanwen Zhang et al.
cs.CVcs.DC
DSFedMed enables mutual knowledge distillation between large foundation models and lightweight client models in federated medical image segmentation.
Why This Matters
This solves a critical deployment challenge - foundation models are too heavy for edge devices in federated settings, but this framework lets small client models benefit from foundation model knowledge while keeping data private.
Bidirectional distillation between scales in federated learning can achieve better results than either unidirectional distillation or traditional federated averaging.
Read Paper

Pushing the limits of unconstrained machine-learned interatomic potentials

1/22/2026Filippo Bigi et al.
physics.chem-phstat.ML
Demonstrates that removing physical constraints from machine-learned interatomic potentials can paradoxically improve both efficiency and accuracy.
Why This Matters
This challenges the conventional wisdom that physics-informed ML models should strictly enforce physical laws - relaxing constraints like energy conservation can sometimes help rather than hurt model performance.
Practitioners working on physics-informed ML should carefully evaluate whether hard constraints actually improve their models or whether soft penalties might achieve better results.
Read Paper

360Anything: Geometry-Free Lifting of Images and Videos to 360°

1/22/2026Ziyi Wu et al.
cs.CV
360Anything lifts perspective images and videos to 360° panoramas without requiring camera calibration metadata, using pre-trained diffusion transformers.
Why This Matters
This eliminates a major bottleneck in immersive content creation - most in-the-wild photos and videos lack reliable camera metadata, making previous geometric alignment approaches impractical at scale.
Geometry-free approaches using diffusion models can achieve robust perspective-to-panorama conversion that generalizes better to uncalibrated real-world content than explicit geometric methods.
Read Paper

Controlling Long-Horizon Behavior in Language Model Agents with Explicit State Dynamics

1/22/2026Sukesh Subaharan
cs.AIcs.CL
Introduces explicit affective state dynamics to control long-horizon behavior and prevent persona drift in LLM agents during extended interactions.
Why This Matters
Addresses the underexplored problem of temporal coherence in conversational AI, where agents often exhibit abrupt personality shifts—critical for applications requiring consistent long-term engagement.
Imposing dynamical constraints on external state variables can provide temporal structure that pure next-token prediction lacks, enabling more coherent extended interactions.
Read Paper

Grounding Large Language Models in Reaction Knowledge Graphs for Synthesis Retrieval

1/22/2026Olga Bunkova et al.
cs.AI
Grounds LLMs in reaction knowledge graphs via Text2Cypher generation for reliable chemical synthesis retrieval instead of hallucinated suggestions.
Why This Matters
This tackles the hallucination problem in scientific AI by grounding LLM outputs in verified databases, demonstrating a scalable pattern for knowledge-intensive domains beyond chemistry.
Converting natural language queries to graph database queries provides a principled way to combine LLM reasoning with verified domain knowledge.
Read Paper

ActionMesh: Animated 3D Mesh Generation with Temporal 3D Diffusion

1/22/2026Remy Sabathier et al.
cs.CV
ActionMesh generates production-ready animated 3D meshes in a feed-forward manner by adding a temporal axis to 3D diffusion models.
Why This Matters
Unlike existing methods requiring optimization or long runtimes, this produces immediately usable animated assets, bridging a critical gap between research and production workflows in games and film.
Extending spatial 3D diffusion to include temporal dynamics enables practical one-shot animated mesh generation without post-processing.
Read Paper

Provable Robustness in Multimodal Large Language Models via Feature Space Smoothing

1/22/2026Song Xia et al.
cs.LGcs.CV
Introduces Feature-space Smoothing with certified robustness guarantees for multimodal LLMs against adversarial perturbations.
Why This Matters
As MLLMs are deployed in safety-critical applications, this provides the first provable robustness framework for multimodal models, moving beyond empirical defenses to mathematical guarantees.
Certified robustness can be achieved at the feature representation level rather than just output predictions, offering stronger guarantees for multimodal systems.
Read Paper

Counterfactual Training: Teaching Models Plausible and Actionable Explanations

1/22/2026Patrick Altmeyer et al.
cs.LGcs.AI
Proposes counterfactual training that uses counterfactual explanations during training to make ML models inherently more explainable with plausible, actionable explanations.
Why This Matters
This flips the XAI paradigm from post-hoc explanation to built-in explainability, addressing a fundamental limitation of current interpretability methods that often produce unrealistic or unhelpful counterfactuals.
Rather than explaining black-box models after training, you can train models to be explainable from the start by incorporating counterfactual constraints into the learning objective.
Read Paper

Synthetic Augmentation in Imbalanced Learning: When It Helps, When It Hurts, and How Much to Add

1/22/2026Zhengchi Ma, Anru R. Zhang
stat.MLcs.LGstat.ME
Provides a unified statistical framework answering when synthetic augmentation helps imbalanced classification and how many samples to generate.
Why This Matters
Addresses two long-standing practical questions about minority oversampling with theoretical grounding—crucial for practitioners who currently rely on heuristics for synthetic data generation.
Synthetic augmentation effectiveness depends on how well the generative model captures minority class structure; the optimal amount is theoretically characterizable.
Read Paper

Scaling Text-to-Image Diffusion Transformers with Representation Autoencoders

1/22/2026Shengbang Tong et al.
cs.CV
Representation Autoencoders scale diffusion models to text-to-image generation by training in semantic latent spaces of frozen vision encoders.
Why This Matters
Provides evidence that high-dimensional semantic spaces (vs pixel/VAE latent) offer distinct advantages for diffusion, with practical insights on data composition for text rendering and general fidelity.
Diffusion in semantic representation space (like SigLIP embeddings) may be a viable alternative to VAE latent spaces for text-to-image.
Read Paper

Neural Particle Automata: Learning Self-Organizing Particle Dynamics

1/22/2026Hyunsoo Kim et al.
cs.NEcs.CV
Neural Particle Automata generalizes Neural Cellular Automata from fixed grids to continuous particle systems with learnable dynamics.
Why This Matters
Elegant extension of the NCA paradigm to Lagrangian particle systems enables heterogeneous dynamics and concentrates computation on active regions—opens new directions for self-organizing generative models.
Moving from Eulerian (grid-fixed) to Lagrangian (particle-based) neural automata enables more natural modeling of dynamic, sparse phenomena.
Read Paper

Universal Refusal Circuits Across LLMs: Cross-Model Transfer via Trajectory Replay and Concept-Basis Reconstruction

1/22/2026Tony Cristofano
cs.CL
Refusal behavior in aligned LLMs stems from universal low-dimensional circuits that can be transferred across architectures without target model supervision.
Why This Matters
Challenges the assumption that safety mechanisms are model-specific, revealing shared semantic structures across diverse LLMs including Dense-to-MoE transfers—important for understanding alignment transferability.
Safety behaviors may be more portable than expected; alignment work on one model family could transfer to others via concept-basis alignment.
Read Paper

Structured Hints for Sample-Efficient Lean Theorem Proving

1/22/2026Zachary Burton
cs.AI
Simple tactic skeletons as prompts boost neural theorem provers by 43% relative improvement on miniF2F without retraining.
Why This Matters
Demonstrates that even highly-trained RL models like DeepSeek-Prover benefit substantially from lightweight structural guidance at inference time, suggesting we may be underutilizing simple interventions.
Before scaling up training, try cheap inference-time interventions like fixed prompt schedules over common patterns.
Read Paper