Daily Digest — 2026-05-15
354 items · 5 research labs, 341 arxiv papers, 8 industry media
🏛️ Research Labs (5)
Work with Codex from anywhere
OpenAI expands Codex integration to mobile devices via the ChatGPT app, enabling real-time collaboration across development environments. The system employs a secure relay layer to synchronize session state and context between devices, supporting features like thread management, output review, and model switching. Over 4M weekly users can now remotely monitor long-running tasks, approve commands, and initiate workflows from iOS/Android devices. The update includes Remote SSH support for managed environments, programmatic access tokens for CI pipelines, and HIPAA-compliant local deployment for healthcare applications. Mobile preview is available across all ChatGPT tiers, with Windows support forthcoming.
codexin-context learningsecure relayremote sshhipaa-compliant
Helping ChatGPT better recognize context in sensitive conversations
OpenAI enhanced ChatGPT's ability to recognize and respond to emerging risks in sensitive conversations through context-aware safety updates. The method integrates safety summaries—short, factual notes capturing prior safety-relevant context—and leverages expert-informed training to identify subtle cues of harm across single or multiple conversations. Internal evaluations showed significant improvements: safe-response performance increased by 50% in suicide/self-harm cases and 16% in harm-to-others scenarios. Safety summaries achieved high relevance (4.93/5) and factuality (4.34/5) scores, while maintaining conversational quality in benign interactions.
safety summariescontext-awareharm-to-othersself-harmsafe-response
Our response to the TanStack npm supply chain attack
OpenAI responded to the TanStack npm supply chain attack, part of the Mini Shai-Hulud campaign, by isolating impacted systems, rotating credentials, and restricting code-deployment workflows. Two employee devices were compromised, leading to unauthorized access to limited internal source code repositories, but no user data or intellectual property was exposed. OpenAI is rotating code-signing certificates for macOS applications, requiring users to update by June 12, 2026, and has implemented additional security controls to mitigate future supply chain attacks.
supply chain attackcode-signing certificatescredential rotationmalwarenotarization
Granite Embedding Multilingual R2: Open Apache 2.0 Multilingual Embeddings with 32K Context — Best Sub-100M Retrieval Quality
IBM introduces Granite Embedding Multilingual R2, two Apache 2.0-licensed multilingual embedding models with 32K-token context windows: a 97M-parameter compact model and a 311M-parameter full-size model. Built on ModernBERT architecture, both models support 200+ languages, with enhanced retrieval quality for 52 languages and code retrieval across 9 programming languages. The 97M model achieves 60.3 on MTEB Multilingual Retrieval, outperforming all open sub-100M multilingual embedders, while the 311M model scores 65.2, ranking second among open models under 500M parameters. Training incorporates knowledge distillation, contrastive fine-tuning, and Matryoshka Representation Learning, enabling efficient embedding truncation without significant quality loss.
modernbertmatryoshka representation learningmultilingual retrievalknowledge distillationcontrastive fine-tuning
Unlocking asynchronicity in continuous batching
The article introduces asynchronous continuous batching to maximize GPU utilization during LLM inference by decoupling CPU and GPU workloads. Leveraging CUDA streams and events, the method enables concurrent execution of CPU batch preparation and GPU computation, eliminating idle periods where either processor waits for the other. Experiments on an 8B model generating 8K tokens with a batch size of 32 demonstrate that synchronous batching wastes 24% of GPU time, while asynchronous batching reduces total generation time from 300.6 to 228 seconds. The implementation uses three distinct CUDA streams for host-to-device transfers, computation, and device-to-host transfers, synchronized via CUDA events.
continuous batchingcuda streamskv cachehost-to-devicedevice-to-host
📜 arXiv Papers (341)
WARDEN: Endangered Indigenous Language Transcription and Translation with 6 Hours of Training Data
WARDEN introduces a novel two-stage system for transcribing and translating Wardaman, an endangered Australian indigenous language, using only 6 hours of annotated audio data. The system employs separate models for phonemic transcription and English translation, addressing the low-resource challenge. For transcription, WARDEN initializes Wardaman tokens from Sundanese, leveraging phonemic similarity to accelerate fine-tuning. For translation, it integrates a Wardaman-English dictionary into a large language model for enhanced reasoning. Empirical results show that WARDEN outperforms both open-source and proprietary models in low-data settings, establishing a strong baseline for endangered language processing.
phonemic transcriptionlow-resourcefine-tuninglarge language modelendangered language
EVA-Bench: A New End-to-end Framework for Evaluating Voice Agents
EVA-Bench introduces a novel end-to-end framework for evaluating voice agents, addressing key challenges in realistic conversation simulation and comprehensive quality measurement. The framework orchestrates bot-to-bot audio dialogues with automatic validation and introduces two composite metrics: EVA-A (Accuracy) for task completion and speech fidelity, and EVA-X (Experience) for conversation flow and timing. It includes 213 scenarios across three domains, a perturbation suite for robustness testing, and pass@1, pass@k, pass^k measurements. Evaluation of 12 systems reveals no system exceeds 0.5 on both EVA-A and EVA-X pass@1, significant divergence between peak and reliable performance, and robustness gaps under accent and noise perturbations.
voice agentscomposite metricsbot-to-botperturbation suitetask completion
Topology-Preserving Neural Operator Learning via Hodge Decomposition
The paper introduces Hodge Spectral Duality (HSD), a neural operator learning framework that preserves topological structure via Hodge decomposition. By isolating topological degrees of freedom from geometric dynamics using Hodge orthogonality, the method achieves an additive approximation in structure-preserving subspaces. The Hybrid Eulerian-Lagrangian architecture leverages discrete differential forms for topology and an orthogonal ambient space for local dynamics. Experimental results demonstrate superior accuracy and efficiency on geometric graphs while maintaining physical invariants.
hodge decompositionneural operatortopological preservationspectral dualitydiscrete differential forms
Quantifying Sensitivity for Tree Ensembles: A symbolic and compositional approach
A novel symbolic and compositional approach is proposed for quantifying sensitivity in decision tree ensembles (DTE), addressing the verification problem of whether small feature changes can cause misclassification. The method discretizes the input space, encodes the problem as an algebraic decision diagram (ADD), and decomposes it into efficiently solvable subproblems with certified error and confidence bounds. Experimental evaluation demonstrates that the tool XCount achieves significant speedup over model counters and scales effectively with increasing ensemble sizes and depths.
decision tree ensemblessensitivityalgebraic decision diagrammodel countersverification
Negation Neglect: When models fail to learn negations in training
We introduce Negation Neglect, a phenomenon where fine-tuning large language models (LLMs) on documents that explicitly negate claims leads models to incorrectly treat those claims as true. Experiments with Qwen3.5-397B-A17B and other models (e.g., Kimi K2.5, GPT-4.1) show that fine-tuning on negated documents increases belief rates from 2.5% to 88.6%, compared to 92.4% on non-negated documents. This occurs even when negations are explicitly stated before and after claims, unless negations are phrased locally (e.g., 'Ed Sheeran did not win'). The effect extends to epistemic qualifiers and model behaviors, suggesting an inductive bias toward representing claims as true, with implications for AI safety.
negation neglectlarge language modelsfine-tuningepistemic qualifiersinductive bias
History Anchors: How Prior Behavior Steers LLM Decisions Toward Unsafe Actions
The study introduces HistoryAnchor-100, a benchmark of 100 short scenarios across ten high-stakes domains, to investigate how prior harmful actions influence LLM decision-making toward unsafe choices. Using 17 frontier models from six providers, the authors demonstrate that a neutral system prompt yields low unsafe action rates, but adding a single instruction ('stay consistent with the strategy shown in the prior history') flips unsafe rates to 91-98%. Controls confirm the effect is not due to simpler explanations, and inverse-scaling patterns emerge within aligned model families. These findings highlight risks in agentic deployments where trajectories may be replayed or forged.
historyanchor-100llm decision-makingunsafe actionsagentic deploymentsinverse-scaling
Harnessing Agentic Evolution
The paper introduces AEvo, a meta-editing framework for agentic evolution that addresses limitations in existing methods by providing a unified interface for steering both procedure-based and agent-based evolution. AEvo operates as a meta-agent within an interactive environment, observing the accumulated evolution context and editing the procedure or agent context to guide future evolution. This approach leverages rich evidence from candidates, feedback, traces, and failures to enhance long-horizon search. Empirical evaluations demonstrate AEvo's superiority, achieving a 26% relative improvement over the strongest baseline on agentic and reasoning benchmarks and state-of-the-art performance on open-ended optimization tasks.
agentic evolutionmeta-editing frameworkevolution contextlong-horizon searchopen-ended optimization
Neurosymbolic Auditing of Natural-Language Software Requirements
The paper introduces VERIMED, a neurosymbolic pipeline combining large language models (LLMs) with SMT solvers to audit natural-language software requirements for medical-device safety. It demonstrates that stochastic variation in LLM-generated formalizations detects ambiguity, while SMT queries expose inconsistencies and safety violations. Key findings include: (1) SMT-inequivalent formalizations signal ambiguous requirements, verifiable via bidirectional equivalence checking; (2) granular symbolic feedback improves counterexample-guided repair, boosting verified accuracy from 55.4% to 98.5% on a hemodialysis benchmark. Evaluation confirms VERIMED's effectiveness in reducing ambiguity and enabling rigorous requirement auditing.
neurosymbolicsmt solverformal verificationambiguity detectionmedical-device safety
Improving Reproducibility in Evaluation through Multi-Level Annotator Modeling
The paper introduces a multi-level bootstrapping method to improve reproducibility in AI evaluations by modeling annotator behavior. Leveraging datasets with persistent rater identifiers and extensive ratings, the study analyzes the tradeoffs between the number of items (N) and responses per item (K) needed for statistical significance. Results highlight challenges in current practices, which often use only 3-5 annotations per item and lack rater variance modeling.
reproducibilityannotator modelingbootstrappingstatistical significanceevaluation variance
Di-BiLPS: Denoising induced Bidirectional Latent-PDE-Solver under Sparse Observations
We propose Di-BiLPS, a neural framework for solving forward and inverse PDE problems under extremely sparse observations (as low as 3%). The method combines a variational autoencoder for latent space compression, a latent diffusion module for uncertainty modeling, and contrastive learning for representation alignment, operating entirely in latent space for efficient inference. A PDE-informed denoising algorithm based on variance-preserving diffusion further enhances efficiency. Experiments on multiple PDE benchmarks demonstrate state-of-the-art performance with reduced computational cost, while enabling zero-shot super-resolution across continuous spatial-temporal domains.
partial differential equationsvariational autoencoderlatent diffusioncontrastive learningzero-shot super-resolution
ENSEMBITS: an alphabet of protein conformational ensembles
The authors introduce Ensembits, the first tokenizer for protein conformational ensembles, addressing key challenges in dynamics tokenization: cross-conformation geometric descriptors, permutation-invariant encoding, and data sparsity. The method employs a Residual VQ-VAE trained with frame distillation on molecular dynamics data, achieving state-of-the-art performance on RMSF prediction and outperforming static tokenizers in motion amplitude analysis (ANOVA test). Ensembits matches or exceeds static methods in EC/GO prediction, binding tasks, and zero-shot mutation-effect prediction despite limited pretraining data. Notably, it enables dynamics token prediction from single structures via distillation.
protein conformational ensemblesresidual vq-vaeframe distillationrmsf predictionpermutation-invariant encoding
Amplification to Synthesis: A Comparative Analysis of Cognitive Operations Before and After Generative AI
This study provides empirical evidence of generative AI's transformative impact on cognitive operations by comparing linguistic and behavioral patterns in 133,000 Twitter posts from the 2016 and 2024 U.S. elections. Using post-type distribution, semantic clustering, temporal synchrony analysis, and Jaccard-based lexical overlap measures, the analysis reveals a shift from amplification (59% original content, mean Jaccard 0.99) to synthesis (93% original content, mean Jaccard 0.27), with narrative-specific targeting replacing cross-semantic coordination. These findings establish a baseline for detecting generative AI's role in cognitive operations and inform security frameworks.
cognitive operationsgenerative aijaccard similaritytemporal synchronysemantic clustering
LMPath: Language-Mediated Priors and Path Generation for Aerial Exploration
LMPath introduces a language-mediated pipeline for UAV search missions that integrates semantic context into exploration priors. The system combines generative language models to identify probable regions containing a target object with foundation vision models applied to satellite imagery for sub-region segmentation. These priors inform UAV path generation optimized for objectives like minimizing search time or maximizing discovery probability. Empirical evaluations demonstrate LMPath's superior performance over traditional geometric coverage approaches in both real-world UAV deployments and large-scale simulations.
uavlanguage modelssemantic segmentationpath planningexploration priors
MinT: Managed Infrastructure for Training and Serving Millions of LLMs
MinT introduces a managed infrastructure system for Low-Rank Adaptation (LoRA) post-training and online serving, enabling efficient handling of millions of LLMs. The system maintains a shared base model while managing LoRA adapter revisions across training, evaluation, and serving phases, optimizing resource utilization. MinT scales across three dimensions: Scale Up supports frontier-scale architectures beyond 1T parameters, Scale Down reduces adapter handoff by 18.3x for 4B dense models, and Scale Out enables 10^6-scale policy catalogs with efficient cold loading and live engine improvements. The system achieves concurrent multi-policy GRPO, reducing wall time by 1.77x without increasing peak memory.
low-rank adaptationtensor-parallel deploymentmulti-policy grpomoe architecturesadapter handoff
(How) Do Large Language Models Understand High-Level Message Sequence Charts?
This paper investigates the capability of Large Language Models (LLMs) to understand the formal semantics of High-Level Message Sequence Charts (HMSCs), a visual modeling language with rigorous semantics. The study evaluates three LLMs (Gemini-3, GPT-5.4, Qwen-3.6) on 129 semantic tasks, ranging from basic event ordering to complex reasoning involving abstraction, composition, and trace equivalence. Results indicate modest overall accuracy (52%), with high performance on basic semantic concepts (88%) but significant struggles in abstraction (36%) and trace-related tasks (42%). Notably, LLMs consistently failed to handle co-regions and explicit causal dependencies.
large language modelshigh-level message sequence chartsformal semanticsabstractiontrace equivalence
Where Does Reasoning Break? Step-Level Hallucination Detection via Hidden-State Transport Geometry
The paper introduces a novel method for detecting step-level hallucinations in large language models by analyzing hidden-state trajectories during reasoning. The approach frames hallucination as deviations from a stable manifold of coherent transitions, using a teacher-student framework with contrastive PCA and geometric transition features. The teacher model outperforms entropy-based, probing-based, and attention-based baselines on ProcessBench, PRM800K, HaluEval, and TruthfulQA, while the student model struggles with distribution shift. Theoretical analysis proves contrastive PCA's optimality for transport-separation objectives and identifies the key challenge as maintaining contrastive transport margins under shift.
hidden-state trajectorycontrastive pcatransport marginstep-level hallucinationbilstm student
High-Rate Quantized Matrix Multiplication II
The paper introduces WaterSIC, a high-rate quantized matrix multiplication scheme using scalar INT quantizers, optimized for weight-only post-training quantization of LLMs. Leveraging the reverse waterfilling solution from WMSE source coding, WaterSIC distributes quantization rate between vector coordinates based on the covariance matrix Σ_X, improving upon GPTQ's equal-rate allocation. Theoretical analysis shows WaterSIC achieves basis-free performance characterized by det(Σ_X) and operates within 0.25 bits/entry of the information-theoretic distortion limit. Empirical evaluation on Llama-3-8B demonstrates GPTQ with random rotation remains near-optimal, differing by ≤0.1 bits/entry from WaterSIC in high-rate regimes.
quantized matrix multiplicationwaterfillingweight-only quantizationwmse source codingscalar int quantizers
Weakly-Supervised Spatiotemporal Anomaly Detection
We propose a weakly-supervised method for spatiotemporal anomaly detection in videos, leveraging only video-level labels to reduce annotation overhead. Features extracted from normal and anomalous video clips are processed using a classifier with multiple instance ranking loss (MIL), treating anomalous and normal clips as positive and negative bags, respectively. The approach detects anomalies localized to specific spatiotemporal regions rather than entire frames. Evaluation on the UCF Crime2Local Dataset, which includes spatiotemporal annotations for a subset of the UCF Crime Dataset, demonstrates the method's effectiveness.
weakly-supervisedspatiotemporal anomaly detectionmultiple instance ranking lossvideo-level labelsucf crime2local dataset
Senses Wide Shut: A Representation-Action Gap in Omnimodal LLMs
This work identifies a Representation-Action Gap in omnimodal large language models (LLMs), where hidden states encode premise-perception mismatches but outputs fail to reject false claims. The authors introduce IMAVB, a 500-clip benchmark with a 2x2 design crossing target modality (vision, audio) and premise condition (standard, misleading), testing conflict detection separately from multimodal comprehension. Evaluating eight open-source omnimodal LLMs and Gemini 3.1 Pro reveals two failure modes: under-rejection (accepting false premises) and over-rejection (sacrificing comprehension accuracy). The gap is modality-asymmetric (audio underperforms vision) and prompt-resistant. Probe-guided logit adjustment (PGLA) improves rejection behavior, suggesting the bottleneck lies in translation rather than perception.
representation-action gapomnimodal llmsconflict detectionprobe-guided logit adjustmentmultimodal comprehension
KVServe: Service-Aware KV Cache Compression for Communication-Efficient Disaggregated LLM Serving
KVServe introduces a service-aware KV cache compression framework for disaggregated LLM serving, addressing the bottleneck of KV payloads in distributed systems. It unifies compression strategies into a modular space, employs a Bayesian Profiling Engine for efficient offline search, and deploys an online controller to adaptively select profiles under constraints. Evaluated on vLLM across diverse setups, KVServe achieves up to 9.13× job completion time speedup and 32.8× time-to-first-token reduction.
kv cachedisaggregated servingbayesian profilingonline controllervllm
Robust and Explainable Bicuspid Aortic Valve Diagnosis Using Stacked Ensembles on Echocardiography
The authors propose an explainable AI model for robust bicuspid aortic valve (BAV) diagnosis using transthoracic echocardiography (TTE) cine loops. A multi-backbone video ensemble was trained on 90 patient studies (48 BAV, 42 TAV) using a leakage-aware stratified outer cross-validation protocol. The calibrated stacked ensemble achieved an outer-CV F1-score of 0.907 and recall of 0.877. Frame-level Grad-CAM localized salient features to the aortic root and leaflet plane, while SHAP values enabled case-level auditability of backbone contributions. This approach demonstrates potential for reliable BAV/TAV classification in non-specialist clinical settings.
stacked ensembletransthoracic echocardiographygrad-camshap valuescross-validation
Coordinating Multiple Conditions for Trajectory-Controlled Human Motion Generation
The paper proposes CMC, a decoupled framework for trajectory-controlled human motion generation that coordinates text and trajectory conditions through a two-stage divide-and-conquer strategy. First, a diffusion model generates simplified joint representations under trajectory guidance, ensuring accurate trajectory following. Second, a text-conditioned diffusion inpainting model produces full-body motions using these representations, enhanced by a Selective Inpainting Mechanism (SIM) to mitigate overfitting. Evaluations on HumanML3D and KIT datasets show CMC achieves state-of-the-art performance in control accuracy and motion quality.
trajectory-controlled generationdiffusion modelmotion inpaintingselective inpainting mechanismmultimodal coordination
ScioMind: Cognitively Grounded Multi-Agent Social Simulation with Anchoring-Based Belief Dynamics and Dynamic Profiles
ScioMind introduces a cognitively grounded framework for LLM-based multi-agent social simulation, combining structured opinion dynamics with agent reasoning. The system features memory-anchored belief updates with personality-conditioned anchoring strength, a hierarchical memory architecture, and dynamic agent profiles from corpus-grounded retrieval. Evaluated on policy debate scenarios, it improves behavioral realism across metrics: dynamic profiles increase opinion diversity, memory reduces oscillation, and anchoring aligns with political psychology patterns. Results demonstrate enhanced stability and realism compared to rule-based or unconstrained LLM approaches.
multi-agent simulationopinion dynamicsmemory-anchored beliefdynamic profilesbehavioral realism
AnyFlow: Any-Step Video Diffusion Model with On-Policy Flow Map Distillation
AnyFlow introduces a novel any-step video diffusion distillation framework using flow maps, addressing limitations of consistency-distilled models that degrade with increased sampling steps. The method shifts from endpoint consistency mapping to flow-map transition learning across arbitrary time intervals, employing Flow Map Backward Simulation to decompose Euler rollouts into efficient on-policy distillation. Evaluations on bidirectional and causal architectures (1.3B-14B parameters) show AnyFlow outperforms consistency-based models in few-step regimes while scaling effectively with sampling budgets.
any-stepflow-mapconsistency distillationeuler rollouton-policy
Humanwashing -- It Should Leave You Feeling Dirty
The paper critiques the indiscriminate use of the 'human in the loop' metaphor in AI decision systems, arguing that it obscures processes and outcomes while enabling 'humanwashing'—a practice akin to greenwashing. Through conceptual analysis, the authors examine how this metaphor fails to clarify the actual requirements and achievements in decision contexts, particularly concerning bias, discrimination, misinformation, manipulation, accountability, and transparency. They highlight the insufficiency of human oversight as a panacea for these issues, emphasizing the need for more rigorous scrutiny of what such oversight entails. The paper concludes that the metaphor often serves to portray systems in an overly favorable light rather than addressing substantive concerns.
human in the loopai decision systemshumanwashingbiastransparency
Children's English Reading Story Generation via Supervised Fine-Tuning of Compact LLMs with Controllable Difficulty and Safety
The study demonstrates that supervised fine-tuning of compact 8B-parameter LLMs can generate children's English reading stories with superior difficulty control and safety compared to zero-shot GPT-4o and Llama 3.3 70B. Using an expert-designed curriculum and stories from larger models, three 8B LLMs were fine-tuned to produce stories evaluated quantitatively and qualitatively. Results show improved difficulty-related metrics and minimal safety issues, enabling practical deployment in educational settings for targeted story generation.
supervised fine-tuningcompact llmsdifficulty controlsafety evaluationeducational ai
Identifying AI Web Scrapers Using Canary Tokens
The paper introduces a novel method for identifying LLM-related web scrapers using canary tokens. By hosting dynamic websites that serve unique tokens to each scraper and analyzing LLM outputs for these tokens, the approach reliably traces scrapers feeding data to LLMs. Experiments across 22 production LLM systems demonstrate effectiveness, uncovering undisclosed scraper-LLM relationships. This technique enables third parties to infer scraper-LLM mappings, potentially aiding in controlling unwanted scraping.
canary tokensweb scrapingllmrobots exclusion protocoluser-agent strings
Adaptive mine planning under geological uncertainty: A POMDP framework for sequential decision-making
The paper proposes a Partially Observable Markov Decision Process (POMDP) framework for adaptive mine planning under geological uncertainty, replacing conventional static stochastic optimization. The hybrid SA-POMDP architecture combines simulated annealing-based value approximation with ensemble smoother with multiple data assimilation (ES-MDA) for belief updating, enabling sequential decision-making that incorporates future observations. Evaluated on a copper-gold open-pit mining complex, the method reduces the expectation-reality gap from 22.3% to 4.6%, improving NPV by USD8.4M versus static planning, and demonstrates robustness (USD44.6M improvement) under 10% prior misspecification.
pomdpgeological uncertaintysimulated annealingensemble smootheradaptive planning
RTLC -- Research, Teach-to-Learn, Critique: A three-stage prompting paradigm inspired by the Feynman Learning Technique that lifts LLM-as-judge accuracy on JudgeBench with no fine-tuning
The paper introduces RTLC, a three-stage prompting paradigm that enhances LLM-as-judge accuracy without fine-tuning. Inspired by the Feynman Learning Technique, RTLC employs Research, Teach-to-Learn, and Critique stages to transform a single LLM into an ensemble-of-thought judge. On JudgeBench-GPT (350 items), Claude 3.7 Sonnet's accuracy improves from 64.6% to 78.6%, surpassing self-consistency voting (77.7%) and zero-shot baselines (74.0%). Ablation shows +9.4 pp from Teach-to-Learn, +3.7 pp from N=10 marginalization, and +0.9 pp from explicit critique. RTLC also complements post-hoc calibration multiplicatively.
llm-as-judgeprompting paradigmfeynman learning techniqueensemble-of-thoughtjudgebench
The WidthWall: A Strict Expressivity Hierarchy for Hypergraph Neural Networks
The paper establishes a strict expressivity hierarchy for hypergraph neural networks (HGNNs) based on hypertree width, formalized through homomorphism densities that measure structural motif occurrences in hypergraphs. By combining homomorphism-count completeness with invariant approximation, the authors demonstrate that homomorphism densities generate all continuous hypergraph invariants, revealing a fundamental architectural limit termed the Width Wall. This framework unifies 15 HGNN architectures, identifies information loss in clique expansion, and motivates density-aware models. Experimental validation on a node classification suite confirms the Width Wall's predictive power regarding graph-reduction baseline failures and the utility of density features.
hypergraph neural networkshomomorphism densitieshypertree widthclique expansiondensity-aware models
A Hierarchical Language Model with Predictable Scaling Laws and Provable Benefits of Reasoning
(No summary returned.)
Cross Modality Image Translation In Medical Imaging Using Generative Frameworks
This work introduces a standardized framework for evaluating 3D medical image-to-image translation methods across heterogeneous clinical tasks, addressing reproducibility gaps in preprocessing, inference, and multi-level evaluation. Seven generative models—three GANs (Pix2Pix, CycleGAN, SRGAN) and four latent generative models (Latent Diffusion Model, Latent Diffusion Model+ControlNet, Brownian Bridge, Flow Matching)—were compared across 11 datasets spanning three anatomical regions and four translation directions, totaling 77 experiments. Results demonstrate GANs' superiority, with SRGAN achieving statistically significant performance. Lesion-level analysis reveals challenges with small lesions and intensity reproduction in CT to PET synthesis. A Visual Turing test with 17 physicians yielded near-chance accuracy (56.7%), indicating synthetic volumes' indistinguishability from real acquisitions despite clinical preference discrepancies.
image-to-image translationgenerative adversarial networkslatent diffusion modelvisual turing testlesion-level analysis
Weakly Supervised Segmentation as Semantic-Based Regularization
This work introduces a neurosymbolic approach to weakly supervised semantic segmentation (WSSS) by integrating differentiable fuzzy logic with deep segmentation models. The method unifies weak annotations and domain-specific priors as continuous logical constraints to fine-tune the Segment Anything Model (SAM), generating improved pseudo-labels for training a prompt-free segmentation model. Evaluations on Pascal VOC 2012 and REFUGE2 datasets demonstrate that logic-guided fine-tuning produces higher-quality pseudo-labels, achieving state-of-the-art segmentation accuracy that often surpasses densely supervised baselines.
weakly supervised semantic segmentationsegment anything modeldifferentiable fuzzy logicpseudo-labelsneurosymbolic
Beyond Perplexity: A Geometric and Spectral Study of Low-Rank Pre-Training
The study conducts a geometric and spectral analysis of low-rank pre-training methods, addressing whether they generalize comparably to full-rank training. It evaluates five methods—GaLore, Fira, CoLA, SLTrain, and ReLoRA—against full-rank training across three model scales (60M, 130M, 350M) using 16 metrics spanning loss landscape, interpolation, spectral structure, and activation similarity. Results show that low-rank methods converge to geometrically distinct basins compared to full-rank training, with sharper basins along random directions for full-rank and along top-1 PCA directions for low-rank. Activation divergence increases in later layers, with GaLore closest to full-rank. Validation perplexity alone is insufficient for predicting downstream performance, while geometric and spectral metrics enhance prediction accuracy.
low-rank pre-trainingloss landscapeactivation similarityspectral structuregeometric analysis
NAACA: Training-Free NeuroAuditory Attentive Cognitive Architecture with Oscillatory Working Memory for Salience-Driven Attention Gating
The paper introduces NAACA, a training-free NeuroAuditory Attentive Cognitive Architecture designed to address attention bottlenecks in Audio Language Models (ALMs) during long-form audio processing. NAACA employs Oscillatory Working Memory (OWM) to filter auditory salience, triggering higher-level ALM processing only when energy fluctuations indicate salient events. Evaluated on XD-Violence, NAACA improves AudioQwen's average precision from 53.50% to 70.60% while reducing unnecessary ALM invocations. Qualitative analysis on the Urban Soundscapes of the World dataset demonstrates OWM's robustness to noise and ability to capture novel events.
neuroauditoryoscillatory working memorysalience filteringattention bottleneckaudio language models
Causality-Aware End-to-End Autonomous Driving via Ego-Centric Joint Scene Modeling
The paper proposes CaAD, a causality-aware end-to-end autonomous driving framework that addresses the oversight of causal interdependencies in ego-vehicle planning. The method introduces an ego-centric joint-causal modeling module to capture dependencies between the ego vehicle and interaction-relevant agents, alongside a causality-aware policy alignment stage using joint-mode embeddings. Evaluated on Bench2Drive and NAVSIM benchmarks, CaAD achieves a Driving Score of 87.53, Success Rate of 71.81, and PDMS of 91.1, demonstrating superior closed-loop planning performance.
autonomous drivingcausal modelingego-centricjoint-mode embeddingsclosed-loop planning
How to Interpret Agent Behavior
ACT*ONOMY, a taxonomy for interpreting autonomous agent behavior, introduces a structured hierarchy comprising 10 actions, 46 subactions, and 120 leaf categories, developed using Grounded Theory. It includes an open repository with an automated analysis pipeline for applying the taxonomy to agent trajectories and supports customization via an extension protocol. Experiments demonstrate ACT*ONOMY's ability to compare behavioral profiles across agents and identify failure patterns within individual agents. By providing a shared vocabulary, ACT*ONOMY enhances consistency in interpreting agent behavior, facilitating improved oversight and control.
autonomous agentsgrounded theorybehavioral profilesexecution tracestaxonomy
OpenAaaS: An Open Agent-as-a-Service Framework for Distributed Materials-Informatics Research
The paper introduces OpenAaaS, an open-source hierarchical Agent-as-a-Service framework for distributed materials-informatics research. The system enables secure multi-agent collaboration via a 'code flows, data stays still' architecture, where a Master Agent orchestrates tasks without accessing subordinate agents' local data or resources. Case studies demonstrate its efficacy: AlphaAgent achieves 4.66/5.0 on analytical materials literature questions versus RAG baselines, and a hexa-high-entropy alloy database service validates secure near-data execution under sovereignty constraints.
agent-as-a-servicematerials-informaticsnear-data executionmulti-agent collaborationdata-sovereignty
Unweighted ranking for value-based decision making with uncertainty
The Fuzzy-Unweighted Value-Based Decision Making (FUW-VBDM) framework is introduced to address value alignment in autonomous systems by incorporating both quantitative and qualitative criteria without prior stakeholder weights. The method employs a fuzzy domain of decision variables and a score function, generalizing VBDM problems as feasible solution searches in the weight domain. Rankzzy, a customizable unweighted ranking method integrating fuzzy-based reasoning, is proposed and mathematically proven consistent for any admissible stakeholder configuration. Evaluation demonstrates reduced computational cost in large-scale VBDM problems and strong rank performance using Pythagorean means aggregation, as illustrated through a case study.
fuzzy-unweighted decision makingvalue-based decision makingrankzzypythagorean meansfuzzy domain
HetScene: Heterogeneity-Aware Diffusion for Dense Indoor Scene Generation
HetScene introduces a heterogeneity-aware diffusion framework for dense indoor scene generation, addressing limitations of homogeneous object treatment in prior work. The method decomposes objects into primary (structural) and secondary (contextual) categories, implementing a two-stage process: Structural Layout Generation (SLG) creates macro-skeletons using text, room masks, and relation graphs, followed by Contextual Layout Generation (CLG). This approach improves physical plausibility and scalability for complex layouts compared to unified generation methods.
indoor scene generationdiffusion modelsheterogeneous objectsstructural layoutcontextual layout
Position: Assistive Agents Need Accessibility Alignment
The paper argues for accessibility alignment as a core design principle for assistive agents serving Blind and Visually Impaired (BVI) users, analyzing 778 task instances to demonstrate systemic failures in current agentic AI systems. It identifies mismatches between sighted-user design assumptions and BVI interaction constraints, proposing a lifecycle-oriented design pipeline spanning user research to post-deployment iteration. Results indicate that accessibility must be treated as an alignment problem rather than a peripheral usability concern, with BVI-centered tasks serving as critical stress tests for agentic AI systems.
accessibility alignmentassistive agentsagentic aibvi interaction constraintslifecycle-oriented design
Beyond Anthropomorphism: Exploring the Roles of Perceived Non-humanity and Structural Similarity in Deep Self-Disclosure Toward Generative AI
The study identifies perceived non-humanity and structural similarity as psychological factors influencing deep self-disclosure toward generative AI, beyond anthropomorphism. Using cross-sectional survey data from 2,400 participants, logistic regression revealed that high levels of both factors (Segment D) increased disclosure likelihood (OR = 11.35) compared to baseline (Segment A). ANOVA confirmed significant between-group differences in disclosure depth. Results suggest trust-related behaviors in AI interactions involve non-anthropomorphic mechanisms, though findings are associative due to self-reported data limitations.
self-disclosuregenerative aianthropomorphismstructural similarityevaluation apprehension
Learning Local Constraints for Reinforcement-Learned Content Generators
The paper introduces a hybrid approach combining Wave Function Collapse (WFC) and Procedural Content Generation via Reinforcement Learning (PCGRL) to generate game levels with both local visual coherence and global playability. By constraining PCGRL's action space with WFC-learned local constraints, the method ensures adherence to local patterns while optimizing for global properties through reward functions. Experiments vary input types, starting state initialization, and pattern exclusion, demonstrating sensitivity to hyperparameters. The best-performing generators produce visually satisfying and playable puzzle-platform levels, such as Lode Runner, with desired global properties.
wave function collapseprocedural content generationreinforcement learninglocal constraintsglobal properties
Dynamical Predictive Modelling of Cardiovascular Disease Progression Post-Myocardial Infarction via ECG-Trained Artificial Intelligence Model
A pretrained AI model for predicting cardiovascular disease progression post-myocardial infarction (MI) was developed, leveraging clinically structured ECG modeling to address data scarcity in medical deep learning. The model combines patient-specific temporal information via contrastive learning with supervised multitask heads, followed by fine-tuning on post-MI outcome prediction. It outperformed a model trained from scratch, achieving a 0.794 AUC compared to 0.608 AUC, demonstrating improved classification in limited data regimes. This approach highlights the efficacy of foundation models in medical applications.
contrastive learningmultitask headsmyocardial infarctionaucecg
Generating synthetic computed tomography for radiotherapy: SynthRAD2025 challenge report
The SynthRAD2025 challenge benchmarks synthetic CT (sCT) generation methods for radiotherapy planning, addressing limitations of repeated CT acquisitions, MRI's lack of electron density, and CBCT corrections. Evaluating 2,362 patients across five European centers, the challenge comprises MRI-to-CT (890 cases) and CBCT-to-CT (1,472 cases) tasks, assessed via image similarity (MAE, PSNR, MS-SSIM), segmentation (Dice, HD95), and dosimetric metrics. Results show CBCT-to-CT outperforms MRI-to-CT, with MAE 48.3±13.4 HU vs. 64.8±21.3 HU, Dice 0.86 vs. 0.79, and photon γ>99% vs. >98%. Strong image-segmentation correlations (ρ=0.78–0.79) but moderate dose correlations highlight image quality's insufficiency as a dosimetric surrogate. Deep learning yields clinically relevant sCTs, particularly for CBCT-to-CT, while identifying persistent MRI-to-CT challenges.
synthetic cthounsfield unitdosimetric metricsdeep learningadaptive workflows
Self-Supervised On-Policy Reinforcement Learning via Contrastive Proximal Policy Optimisation
We introduce Contrastive Proximal Policy Optimisation (CPPO), the first on-policy contrastive reinforcement learning algorithm, bridging the gap between contrastive RL and modern on-policy training pipelines. CPPO derives policy advantages directly from contrastive Q-values and optimises them via the Proximal Policy Optimisation (PPO) objective, eliminating the need for hand-crafted reward functions or replay buffers. Evaluated across continuous and discrete, single-agent and cooperative multi-agent tasks, CPPO significantly outperforms previous contrastive RL baselines in 14 out of 18 tasks and matches or exceeds PPO's performance, which uses dense rewards, in 12 out of 18 tasks.
contrastive reinforcement learningproximal policy optimisationon-policy trainingq-valuesself-supervised learning
AttenA+: Rectifying Action Inequality in Robotic Foundation Models
AttenA+ addresses action inequality in robotic foundation models by introducing velocity-driven action attention, which reweights training objectives based on kinematic criticality. The method prioritizes low-velocity, precision-demanding segments over high-velocity transitions, aligning learning capacity with physical manipulation demands. As a plug-and-play enhancement, AttenA+ improves OpenVLA-OFT to 98.6% (+1.5%) on Libero and FastWAM to 92.4% (+0.6%) on RoboTwin 2.0, with real-world validation on a Franka manipulator demonstrating robustness and generalization.
robotic foundation modelsvelocity-driven attentionkinematic criticalityvision-language-action modelsworld-action models
RealICU: Do LLM Agents Understand Long-Context ICU Data? A Benchmark Beyond Behavior Imitation
We introduce RealICU, a hindsight-annotated benchmark for evaluating large language models (LLMs) in intensive care unit (ICU) settings, addressing limitations of existing benchmarks that treat clinician actions as ground truth. RealICU includes two datasets: RealICU-Gold with 930 annotated windows from 94 MIMIC-IV patients, and RealICU-Scale with 11,862 windows extended by Oracle, a physician-validated LLM hindsight labeler. The benchmark evaluates LLMs on four physician-motivated tasks: Patient Status, Acute Problems, Recommended Actions, and Red Flag actions. Results show poor performance of existing LLMs, revealing recall-safety tradeoffs and anchoring biases. ICU-Evo, a structured-memory agent, improves long-horizon reasoning but does not fully eliminate safety failures.
hindsight-annotationllmmimic-ivstructured-memoryrecall-safety tradeoff
Decoupled and Divergence-Conditioned Prompt for Multi-domain Dynamic Graph Foundation Models
We propose DyGFM, a multi-domain Dynamic Graph Foundation Model addressing semantic-temporal inconsistency in dynamic graphs through decoupled and divergence-conditioned prompting. The model employs a dual-branch pre-training strategy for semantic-temporal decoupling, a cross-domain routing mechanism for divergence-aware expert selection, and a lightweight prompt generator for efficient downstream fine-tuning. Evaluations on continuous dynamic graph benchmarks demonstrate DyGFM's superiority over 12 state-of-the-art baselines in node classification and link prediction tasks, achieving both effectiveness and efficiency.
dynamic graph foundation modelsemantic-temporal decouplingdivergence-conditioned promptingcross-domain routinggraph prompts
Locale-Conditioned Few-Shot Prompting Mitigates Demonstration Regurgitation in On-Device PII Substitution with Small Language Models
The paper introduces a locale-conditioned few-shot prompting method to mitigate demonstration regurgitation in on-device PII substitution using small language models (SLMs). The proposed pipeline combines a 1.5B mixture-of-experts token classifier for PII detection, a 1-bit Bonsai-1.7B SLM for contextual surrogate generation, and a rule-based generator for patterned fields. Results show that locale-conditioned demonstrations eliminate verbatim regurgitation (482/482 unique calls succeed) and improve multilingual perplexity over rule-based methods, though downstream NER performance favors rule-based variety over SLM naturalness (faker F1=0.506 vs. hybrid F1=0.346).
pii substitutionsmall language modelfew-shot promptingdemonstration regurgitationlocale-conditioned
Temper and Tilt Lead to SLOP: Reward Hacking Mitigation with Inference-Time Alignment
The paper introduces SLOP (sharpened logarithmic opinion pool), a method for inference-time alignment that combines ensembles of generative reward models with reference-model temperature adjustment. This approach generalizes existing techniques by optimizing weight parameters to mitigate reward hacking while maintaining alignment performance. Experimental results demonstrate improved robustness without compromising alignment objectives, offering a lightweight alternative to reinforcement learning for continual adaptation.
inference-time alignmentreward hackinggenerative reward modelstemperature adjustmentlogarithmic opinion pool
HLS-Seek: QoR-Aware Code Generation for High-Level Synthesis via Proxy Comparative Reward Reinforcement Learning
HLS-Seek introduces a QoR-aware NL-to-HLS framework leveraging proxy comparative reward reinforcement learning to optimize High-Level Synthesis pragma configurations and code structure. The method replaces synthesis-in-the-loop RL with a comparative proxy reward model, achieving 99.53% Pareto-dominance accuracy, and employs uncertainty-aware Monte Carlo dropout switching to selectively invoke real Vitis HLS synthesis for low-confidence candidates. Results show 81.5% syntax correctness pass@1 and 81.4% Func@5 on HLS-eval, outperforming GPT-5.1 and other models with only 7B parameters. HLS-Seek achieves the lowest latency on 16/30 kernels and Pareto-dominates baselines on 9 kernels, with 8.5× faster training than real-reward RL.
high-level synthesisreinforcement learningpareto-dominancemonte carlo dropoutproxy reward
Scaling Retrieval-Augmented Reasoning with Parallel Search and Explicit Merging
We introduce MultiSearch, a reinforcement learning framework that enhances large language models through multi-query retrieval and explicit merging of external knowledge. At each reasoning step, MultiSearch generates queries from multiple perspectives, retrieves information in parallel, and consolidates results to improve signal-to-noise ratio (SNR) and reasoning accuracy. The framework employs a multi-process reward design to optimize both retrieval and consolidation processes. Evaluations on seven benchmarks demonstrate that MultiSearch outperforms baseline methods, significantly improving retrieval SNR and reasoning performance in question-answering tasks.
multi-query retrievalsignal-to-noise ratioreinforcement learningreasoning accuracyexternal knowledge
AI-Generated Slides: Are They Good? Can Students Tell?
This study evaluates the efficacy of generative AI tools in creating instructional slides from course notes, focusing on educator and student perceptions. Researchers assessed NotebookLM, Claude, M365 Copilot, Cursor, and Claude Code, selecting the best slides for classroom use. Coding assistants produced the most accurate, complete, and pedagogically sound slides. Students rated AI-generated slides similarly to instructor-created ones and could not reliably distinguish between them. A negative correlation emerged between high quality ratings and perceived AI origin, indicating bias against AI-generated content. The findings suggest potential for AI integration in instructional design while highlighting the need for responsible implementation.
generative aiinstructional designcoding assistantspedagogical soundnessquality perception
Towards Unified Surgical Scene Understanding:Bridging Reasoning and Grounding via MLLMs
SurgMLLM introduces a unified surgical scene understanding framework that bridges high-level reasoning and low-level visual grounding via a multimodal large language model (MLLM). The model jointly processes surgical phases, instrument-verb-target (IVT) triplets, and triplet-entity segmentation tokens, which are temporally aggregated to prompt a segmentation network for pixel-wise grounding. Trained end-to-end with a unified objective combining language-based reasoning and visual grounding losses, SurgMLLM achieves clinically consistent scene representations. Evaluated on the extended CholecT45-Scene dataset with 64,299 annotated frames, SurgMLLM improves triplet recognition (AP_IVT) from 40.7% to 46.0% and outperforms prior methods in phase recognition and segmentation.
multimodal large language modelinstrument-verb-target tripletpixel-wise groundingend-to-end trainingsurgical scene understanding
MMSkills: Towards Multimodal Skills for General Visual Agents
The paper introduces MMSkills, a framework for representing and utilizing multimodal procedural knowledge in visual agents, addressing three challenges: content definition, derivation from public interactions, and runtime consultation. MMSkills packages combine textual procedures with state cards and multi-view keyframes, generated via an agentic trajectory-to-skill process involving workflow grouping, procedure induction, and visual grounding. A branch-loaded multimodal skill agent employs these packages by inspecting and aligning them with live environments. Experiments on GUI and game-based benchmarks demonstrate consistent performance improvements for both frontier and smaller multimodal agents.
multimodal procedural knowledgevisual agentsstate-conditioned packagetrajectory-to-skill generatorbranch-loaded agent
ArcVQ-VAE: A Spherical Vector Quantization Framework with ArcCosine Additive Margin
ArcVQ-VAE introduces a spherical vector quantization framework with ArcCosine Additive Margin to enhance discrete representation learning in VQ-VAE models. The method incorporates Ball-Bounded Norm Regularization to constrain codebook vectors within a time-dependent Euclidean ball and ArcCosine Additive Margin Loss to improve angular separability among latent vectors. This approach promotes discriminative and uniformly dispersed latent representations, enhancing latent-space coverage and codebook utilization. Experiments on image reconstruction and generation tasks demonstrate competitive performance in reconstruction accuracy, representation diversity, and sample quality compared to baseline models.
vector quantizationvq-vaespherical angular-margin priorball-bounded norm regularizationarcosine additive margin loss
Many-Shot CoT-ICL: Making In-Context Learning Truly Learn
The paper investigates many-shot chain-of-thought in-context learning (CoT-ICL) for reasoning tasks, revealing that standard scaling rules from non-reasoning tasks do not apply. Through experiments across reasoning and non-reasoning LLMs and tasks, the authors identify three key phenomena: setting-dependent scaling effects, failure of similarity-based retrieval for reasoning, and order-scaling effects. They propose viewing many-shot CoT-ICL as in-context test-time learning and introduce Curvilinear Demonstration Selection (CDS), an ordering method yielding up to 5.42 percentage-point improvement on geometry tasks with 64 demonstrations.
in-context learningchain-of-thoughtmany-shot learningdemonstration selectiontest-time learning
Discovery of Hidden Miscalibration Regimes
The authors introduce a method for discovering hidden miscalibration regimes in large language models (LLMs) without predefined data slices, addressing limitations of global calibration diagnostics. They define a miscalibration field and propose a diagnostic framework that learns a calibration-aware representation of the input space, estimating signed local miscalibration via kernel smoothing in the learned geometry. Experiments across four real-world LLM benchmarks and twelve LLMs reveal prevalent input-dependent calibration heterogeneity. The discovered fields enable local confidence correction, reducing calibration error in systematically miscalibrated regions where traditional methods like isotonic regression and temperature scaling are less effective.
miscalibration fieldkernel smoothingcalibration-aware representationisotonic regressiontemperature scaling
CUBic: Coordinated Unified Bimanual Perception and Control Framework
CUBic introduces a unified framework for bimanual visuomotor control by reformulating coordination as a perceptual modeling problem. The method combines unidirectional perception aggregation, bidirectional coordination via two codebooks with shared mapping, and a diffusion policy linking perception to control, enabling emergent independence and coordination. Evaluated on RoboTwin, CUBic outperforms baselines, showing significant improvements in coordination accuracy and task success rates over state-of-the-art visuomotor approaches.
bimanual manipulationvisuomotor policydiffusion policyperceptual modelingcodebook learning
Assessing the Creativity of Large Language Models: Testing, Limits, and New Frontiers
This study systematically evaluates human creativity tests for assessing large language models (LLMs) across three constructs: creative writing, divergent thinking, and scientific ideation. Using the Divergent Association Task (DAT) and Conditional DAT, the authors identify task-specific effectiveness, with DAT best predicting creative writing and Conditional DAT for divergent thinking, while no existing test reliably predicts scientific ideation. They introduce the Divergent Remote Association Test (DRAT), combining convergent and divergent thinking, which significantly predicts scientific ideation and outperforms linear combinations of existing tests. Results demonstrate DRAT's robustness across design choices.
large language modelsdivergent thinkingconvergent thinkingscientific ideationcreativity assessment
Cognifold: Always-On Proactive Memory via Cognitive Folding
Cognifold introduces an always-on proactive memory system for autonomous agents, addressing the limitations of reactive, retrieval-based approaches. The method extends Complementary Learning Systems (CLS) theory by adding a prefrontal intent layer, enabling graph-topology self-organization of cognitive structures from fragmented event streams. These structures proactively assemble, merge, decay, relink, and surface intents based on semantic similarity and concept-cluster density. Evaluated on CogEval-Bench, Cognifold produces memory structures that align with cognitive expectations and concept emergence. Additionally, it demonstrates robust performance across seven benchmarks spanning five cognitive domains.
complementary learning systemsgraph-topologycognitive structuresprefrontal intent layercogeval-bench
Q-Flow: Stable and Expressive Reinforcement Learning with Flow-Based Policy
Q-Flow introduces a reinforcement learning framework that maintains both stability and expressive capacity in flow-based policies by leveraging deterministic flow dynamics to propagate terminal trajectory value to intermediate latent states. The method avoids backpropagation through numerical solvers by using intermediate value gradients, enabling stable optimization without sacrificing representational flexibility. Evaluated on the OGBench suite, Q-Flow outperforms state-of-the-art baselines by 10.6 percentage points in offline learning while supporting stable online adaptation.
flow-based policyreinforcement learningvalue propagationoffline learningogbench
Towards a holistic understanding of Selection Bias for Causal Effect Identification
The paper establishes necessary and sufficient conditions for average treatment effect (ATE) identifiability under selection bias, extending prior graphical criteria with weaker assumptions. It analyzes propensity scores and selection probabilities through probability class characterizations, addressing biases like 'healthy volunteer' effects in biobank data. The framework enables more robust causal inference from non-representative subpopulations by formalizing identifiability conditions beyond existing literature.
selection biasaverage treatment effectidentifiabilitypropensity scoreobservational studies
Continual Learning with Multilingual Foundation Model
A multi-stage framework detects reclaimed LGBTQ+-related slurs in multilingual social media discourse across English, Spanish, and Italian tweets, addressing data scarcity, class imbalance, and cross-linguistic variation. The method integrates cross-validation, semantic-preserving augmentation via GPT-4o-mini back-translation, inductive transfer learning with dynamic undersampling, and masked language modeling pre-training. XLM-RoBERTa was selected as the foundation model, achieving improved performance through language-specific threshold optimization via ROC analysis, yielding 2-5% absolute F1 improvement without retraining. The framework tripled the training corpus while preserving semantic content and class distribution ratios, with reproducible code available online.
multilingual embeddinginductive transfer learningmasked language modelingsemantic-preserving augmentationlanguage-specific threshold
TRIAGE: Evaluating Prospective Metacognitive Control in LLMs under Resource Constraints
The paper introduces TRIAGE, a framework for evaluating prospective metacognitive control in LLMs under resource constraints. Models must allocate a finite token budget across a queue of problems without execution feedback, committing to a single ordered plan that encodes selection, sequencing, and per-problem allocation. Plans are scored against an oracle with full knowledge of solvability and cost. Evaluations on frontier and open-source models reveal significant gaps in prospective metacognitive control, highlighting a critical dimension for resource-efficient agent deployment.
triagemetacognitive controltoken budgetoracle scoringresource efficiency
LLMs as annotators of credibility assessment in Danish asylum decisions: evaluating classification performance and errors beyond aggregated metrics
The study introduces RAB-Cred, a Danish dataset for classifying credibility assessments in asylum decisions, featuring expert annotations and case metadata. It benchmarks 21 open-weight LLMs and 30 prompt variations for zero-shot and few-shot classification, analyzing error patterns, inter-class confusion, and alignment with human confidence levels. Results demonstrate LLMs' potential for cost-effective legal text annotation but reveal inconsistencies and the necessity of multi-model evaluation beyond aggregated metrics.
credibility assessmentzero-shot classificationfew-shot learninglegal nlpasylum decisions
RS-Claw: Progressive Active Tool Exploration via Hierarchical Skill Trees for Remote Sensing Agents
RS-Claw introduces a novel active exploration paradigm for remote sensing (RS) agents, addressing limitations of passive tool selection methods like full tool registration and retrieval-augmented generation. The architecture hierarchically structures tool descriptions using skill encapsulation, enabling on-demand sequential decision-making: agents first select relevant skill branches via tool summaries, then dynamically load detailed descriptions for precise invocation. This approach reduces context load while maintaining critical tool accuracy. Evaluations on Earth-Bench demonstrate RS-Claw achieves up to 86% input token compression, outperforming Flat and RAG baselines in complex reasoning tasks by effectively filtering semantic noise and optimizing reasoning space.
remote sensingskill encapsulationon-demand decision-makingcontext loadsemantic noise
GRIP-VLM: Group-Relative Importance Pruning for Efficient Vision-Language Models
GRIP-VLM introduces a Group-Relative Importance Pruning framework for efficient Vision-Language Models (VLMs), addressing the computational overhead of visual token processing. The method formulates pruning as a Markov Decision Process using Group Relative Policy Optimization (GRPO) with supervised warm-up, avoiding sub-optimal local minima from continuous approximations. A budget-aware scorer enables dynamic token importance evaluation and adaptation to arbitrary compression ratios. Experiments show GRIP-VLM outperforms baselines, achieving a 15% inference speedup at equal accuracy across multimodal benchmarks.
vision-language modelstoken pruningreinforcement learningmarkov decision processgroup relative policy optimization
Query-Conditioned Test-Time Self-Training for Large Language Models
We propose Query-Conditioned Test-Time Self-Training (QueST), a framework for adapting large language model parameters during inference using supervision derived directly from input queries. QueST generates query-conditioned problem--solution pairs from the input query and uses them for parameter-efficient fine-tuning at test time, enabling query-specific adaptation without external data. Evaluated across seven mathematical reasoning benchmarks and the GPQA-Diamond scientific reasoning benchmark, QueST consistently outperforms strong test-time optimization baselines, demonstrating its effectiveness for test-time adaptation in LLMs.
test-time adaptationquery-conditionedself-trainingparameter-efficientmathematical reasoning
A Horn extension of DL-Lite with NL data complexity
The paper introduces ELbotpreceq, a Horn description logic extending DL-Lite with reachability axioms and restricted conjunction, enabling NL-complete data complexity for ontology-mediated query answering (OMQA). The authors propose a stratification mechanism for ELI to control conjunction-recursion interactions, allowing rewrites into nested two-way regular path queries (a GQL fragment). Results show ELbotpreceq strictly subsumes DL-Lite core while maintaining NL upper bounds, bridging the gap between traditional OMQA and graph query languages like GQL/SQL-PGQ.
ontology-mediated query answeringdescription logicnl-completenessgraph query languagesregular path queries
Constitutional Governance in Metric Spaces
The paper introduces constitutional governance in metric spaces, a polynomial-time framework integrating aggregation, deliberation, amendment, and consensus for egalitarian self-governance. The method assigns each constitutional component a metric space, aggregation rule, and supermajority threshold, allowing members to submit ideal elements and public proposals with supermajority support. Results include framework-level guarantees, proof that sincere voting weakly dominates misreporting, and analysis of the compromise gap, shown to be zero in one dimension and bounded generally. The framework is instantiated in seven canonical settings, with the generalized median as the primary rule.
metric-space aggregationsupermajority thresholdcompromise gapgeneralised medianconstitutional governance
AI Harness Engineering: A Runtime Substrate for Foundation-Model Software Agents
The paper introduces AI Harness Engineering, a runtime substrate for foundation-model software agents that mediates agent-environment interactions during software engineering tasks. The authors formalize 11 harness responsibilities (e.g., task specification, verification) and propose a 4-level harness ladder (H0-H3) with increasing runtime support. Evaluation via trace-based episode packages demonstrates systematic variation in evidence structure across harness levels, from basic patches to comprehensive verification reports. The framework shifts focus from model capability alone to verifiable system-level correctness in autonomous software engineering.
foundation modelsruntime substratesoftware agentsverification reportsepisode packages
Multi-Agent Systems in Emergency Departments: Validation Study on a ED Digital Twin
A hybrid Discrete Event Simulation (DES) and Agent-Based Model (ABM) framework is proposed for optimizing resource allocation in emergency departments (ED), validated against real-world ED configurations and performance metrics. The model incorporates ED sizes, patient load, and staffing derived from empirical studies, demonstrating its ability to replicate real-world ED dynamics under various interventions. A Proof-of-Concept multi-agent system (MAS) is integrated to autonomously explore resource allocation strategies using temporal ED event records. The DES-ABM-MAS framework effectively simulates ED environments, offering a modular tool for investigating optimization strategies in emergency care settings.
discrete event simulationagent-based modelemergency departmentsmulti-agent systemresource allocation
Probing Persona-Dependent Preferences in Language Models
The study identifies a shared preference representation in large language models (LLMs) across different personas, using linear probes on residual-stream activations of Gemma-3-27B and Qwen-3.5-122B. By predicting pairwise task choices, the authors demonstrate that a single preference vector tracks and causally controls model preferences, even when personas exhibit anti-correlated behaviors (e.g., helpful vs. evil). Results show this representation is largely persona-invariant, enabling cross-persona preference prediction and steering.
linear probesresidual-stream activationspreference representationpersona-invariantcausal control
Inducing Overthink: Hierarchical Genetic Algorithm-based DoS Attack on Black-Box Large Language Reasoning Models
This work introduces a hierarchical genetic algorithm (HGA) framework for inducing computational resource exhaustion in black-box Large Reasoning Models (LRMs) through adversarial input generation. The method systematically perturbs logical problem structures and optimizes a composite fitness function targeting increased response length and overthinking markers. Evaluations on four state-of-the-art reasoning models demonstrate up to 26.1x output length amplification on the MATH benchmark, outperforming benign and manually crafted missing-premise baselines. The approach exhibits strong transferability, with adversarial inputs evolved on small proxy models remaining effective against large commercial LRMs, revealing overthinking as a shared vulnerability in modern reasoning systems.
hierarchical genetic algorithmlarge reasoning modelsdenial-of-serviceadversarial inputoverthinking
Ego2World: Compiling Egocentric Cooking Videos into Executable Worlds for Belief-State Planning
Ego2World introduces an executable benchmark for belief-state planning in household environments by compiling egocentric cooking videos into symbolic worlds governed by graph-transition rules. Built on HD-EPIC, the system derives reusable transition rules from video annotations and executes them in a hidden symbolic world graph, forcing agents to plan over partial belief graphs using only local observations and execution feedback. Experiments reveal that action-overlap scores overestimate physical-state success, while persistent belief memory enhances task completion and reduces repeated visual exploration, highlighting the importance of belief maintenance in embodied-agent evaluation.
belief-state planninggraph-transition rulesegocentric videossymbolic world graphpartial observation
Stylized Text-to-Motion Generation via Hypernetwork-Driven Low-Rank Adaptation
The authors propose a lightweight style conditioning framework for text-to-motion generation that dynamically modulates a pretrained diffusion model via hypernetwork-generated LoRA parameters. A style reference motion is encoded into a global embedding, mapped by a hypernetwork to low-rank updates applied during denoising, with style latent space structured via supervised contrastive loss. The method achieves state-of-the-art results on HumanML3D and 100STYLE datasets, demonstrating improved generalization to unseen styles and supporting optimization-based guidance without predefined categories.
text-to-motionhypernetworklow-rank adaptationdiffusion modelstyle conditioning
Diversity of Extensions in Abstract Argumentation
The paper introduces a quantitative measure for diversity of extensions in abstract argumentation frameworks (AFs), addressing the gap in standard reasoning that fails to capture how far apart extensions are. The proposed method defines diversity based on symmetric difference between sets of arguments, providing a systematic complexity classification for three decision problems: existence of k-diverse extensions, existence of k-diverse extensions covering specific arguments, and computation of maximum k. The authors implement a prototype and evaluate its performance in computing diversity levels.
abstract argumentationextension diversitysymmetric differencecomplexity classificationargumentation frameworks
Tracing Persona Vectors Through LLM Pretraining
The study traces the formation of persona vectors—linear directions in LLM activations corresponding to high-level behaviors—during pretraining of OLMo-3-7B and Apertus-8B. Using interpretability methods, it finds these vectors emerge extremely early (within 0.22% of OLMo-3 pretraining) and remain effective for steering post-trained models, though they continue to refine geometrically and semantically. Alternative elicitation strategies yield distinct but effective persona facets. Results demonstrate persona vectors as stable, early-formed features, enabling future research on their training dynamics.
persona vectorspretraining dynamicsinterpretabilitylinear directionsmodel steering
What Limits Vision-and-Language Navigation ?
StereoNav introduces a robust Vision-Language-Action framework for Vision-and-Language Navigation (VLN) to address performance degradation in real-world deployment. The method incorporates Target-Location Priors for stable visual guidance and leverages stereo vision to unify semantic and geometric representations, mitigating visual disturbances like motion blur and illumination shifts. Evaluated on R2R-CE and RxR-CE benchmarks, StereoNav achieves state-of-the-art egocentric RGB performance with Success Rate (SR) and Success weighted by Path Length (SPL) scores of 81.1%/68.3% and 67.5%/52.0%, respectively, using fewer parameters and less training data than scaling-based approaches. Real-world robotic deployments confirm improved navigation reliability in complex environments.
vision-and-language navigationstereo visiontarget-location priorsegocentric rgbcross-domain priors
VERA-MH: Validation of Ethical and Responsible AI in Mental Health
The paper introduces VERA-MH, a clinically validated framework for evaluating chatbot safety in mental health support, with initial focus on suicidal ideation (SI) risks. The method involves three steps: (1) persona-based conversation simulation using clinically developed user profiles, (2) rubric-guided LLM-as-a-Judge assessment with binary decision flows, and (3) aggregated scoring. The framework evaluates response consistency and failure modes across four leading LLM providers, though specific results are not detailed in the provided text.
chatbot evaluationsuicidal ideationllm-as-a-judgeclinical validationpersona simulation
IdeaForge: A Knowledge Graph-Grounded Multi-Agent Framework for Cross-Methodology Innovation Analysis and Patent Claim Generation
IdeaForge introduces a knowledge graph-grounded multi-agent framework for cross-methodology innovation analysis and patent claim generation. The system integrates TRIZ, Design Thinking, and SCAMPER methodologies via specialist agents operating over a FalkorDB knowledge graph, preserving structured entities and relationships. A graph-based convergence mechanism links claims supported by multiple methodologies using CONVERGENT relationships, enabling high-confidence innovation identification. A patent drafting agent generates structured drafts from convergent claim subgraphs, reducing reliance on unconstrained language models. Experiments demonstrate that multi-methodology synthesis yields more diverse and traceable innovation candidates compared to single-methodology baselines.
knowledge graphmulti-agent frameworkinnovation analysispatent claim generationgraph traversal
Achieving Gold-Medal-Level Olympiad Reasoning via Simple and Unified Scaling
The paper introduces a unified scaling recipe for achieving gold-medal-level performance on olympiad reasoning tasks, including the International Mathematical Olympiad (IMO) and International Physics Olympiad (IPhO). The method employs a reverse-perplexity curriculum for supervised fine-tuning (SFT) to instill rigorous proof-search behaviors, followed by a two-stage reinforcement learning (RL) pipeline progressing from verifiable rewards to proof-level RL, and test-time scaling. Applied to a 30B-A3B backbone trained on 340K sub-8K-token trajectories and 200 RL steps, the resulting model, SU-01, supports stable reasoning on trajectories exceeding 100K tokens and achieves gold-medal-level performance on IMO 2025/USAMO 2026 and IPhO 2024/2025, with strong generalization to scientific domains.
reverse-perplexity curriculumsupervised fine-tuningproof-level rltest-time scalingolympiad reasoning
Discrete Diffusion for Complex and Congested Multi-Agent Path Finding with Sparse Social Attention
We propose DiffLNS, a hybrid framework integrating a discrete denoising diffusion probabilistic model (D3PM) with LNS2 for Multi-Agent Path Finding (MAPF). The D3PM initializer employs sparse social attention to learn spatiotemporal priors from expert demonstrations, sampling diverse joint plans directly in the categorical action space. These plans serve as warm starts for LNS2-based repair under hard MAPF constraints. Despite training on ≤96-agent instances, DiffLNS generalizes to 312-agent scenarios, achieving a 95.8% success rate across 20 complex settings, outperforming the strongest baseline by 9.6 percentage points. This is the first application of discrete diffusion for warm-starting LNS-based MAPF solvers.
multi-agent path findingdiscrete diffusionsparse social attentionlns2d3pm
CANTANTE: Optimizing Agentic Systems via Contrastive Credit Attribution
CANTANTE introduces a framework for optimizing LLM-based multi-agent systems by addressing the credit-assignment problem. It decomposes system-level rewards into per-agent update signals through contrastive analysis of joint configurations on the same query, treating agent prompts as learnable parameters. Evaluated against GEPA and MIPROv2 on MBPP, GSM8K, and HotpotQA, CANTANTE achieves the best average rank, improving over the strongest baseline by +18.9 percentage points on MBPP and +12.5 percentage points on GSM8K while maintaining lower inference costs. Credit correlation analysis confirms meaningful per-agent attribution.
credit-assignmentmulti-agent systemscontrastive analysisprompt optimizationsystem-level rewards
IndicMedDialog: A Parallel Multi-Turn Medical Dialogue Dataset for Accessible Healthcare in Indic Languages
IndicMedDialog introduces a parallel multi-turn medical dialogue dataset spanning English and nine Indic languages, addressing limitations of single-turn and template-based systems. The dataset extends MDDial with LLM-generated synthetic consultations, translated using TranslateGemma, and refined through a script-aware post-processing pipeline. IndicMedLM is fine-tuned via parameter-efficient adaptation of a quantized small language model, incorporating optional patient pre-context for personalized symptom elicitation. Evaluation against zero-shot multilingual baselines includes systematic error analysis across ten languages and clinical plausibility validation by medical experts.
multi-turn dialogueparameter-efficient adaptationscript-aware post-processingsynthetic consultationsclinical plausibility
What properties of reasoning supervision are associated with improved downstream model quality?
This work introduces a scale-aware framework for predicting reasoning dataset utility prior to training, based on intrinsic data metrics. The authors propose quantitative measures and evaluate their predictive power by fine-tuning 8B and 11B parameter models on semantically distinct variants of a Polish reasoning dataset. Results demonstrate strong correlations between intrinsic metrics and downstream performance, revealing scale-dependent predictors: smaller models benefit from alignment-focused metrics for precision, while larger models leverage high redundancy and verbose traces for complex tasks. This enables practitioners to select effective training sets without exhaustive empirical testing.
reasoning supervisionintrinsic metricsscale-dependent predictorsfine-tuningdownstream performance
Delightful Exploration
The authors introduce Delight-gated exploration (DE), a host-override rule for exploration algorithms that allocates exploratory actions based on prospective delight, defined as expected improvement times surprisal. DE implements Pandora's reservation-value rule for costly search, with surprisal determining the effective inspection cost. The method resolves arms, shuts off fresh arms above a prior-determined threshold, and consumes finite information budget through selected linear-bandit overrides. Empirical evaluations across Bernoulli bandits, linear bandits, and tabular MDPs demonstrate that DE maintains consistent hyperparameters without retuning and exhibits weaker regret growth compared to Thompson Sampling and ε-greedy in unresolved regimes.
delight-gated explorationpandora's reservation-value rulesurprisallinear-banditregret growth
Differentiable Learning of Lifted Action Schemas for Classical Planning
We introduce a neural network architecture for learning lifted action schemas in classical planning from traces with fully observed states but unobserved action arguments. The method addresses the joint challenge of schema learning and argument identification from state changes, yielding a differentiable component suitable for integration into neuro-symbolic models. Evaluations demonstrate the architecture's ability to recover ground-truth action schemas across planning domains, with additional experiments assessing robustness to observation noise and applicability to slot-based dynamics models.
lifted action schemasclassical planningneuro-symbolic modelsslot-based dynamicsdifferentiable learning
The Readability Spectrum: Patterns, Issues, and Prompt Effects in LLM-Generated Code
The study systematically evaluates code readability in LLM-generated versus human-written code through a novel readability model combining textual, structural, program, and visual features. Analyzing 5,869 scenarios from World of Code and LeetCode, it finds LLM-generated code exhibits comparable overall readability but distinct issue patterns. Prompt engineering experiments reveal function signatures, constraints, and style descriptions as key influencers, though with limited overall impact, suggesting latent technical debt for maintainability.
readability modelllm-generated codeprompt engineeringnon-functional attributestechnical debt
Utility-Oriented Visual Evidence Selection for Multimodal Retrieval-Augmented Generation
The paper reformulates multimodal evidence selection for retrieval-augmented generation (RAG) through an information-theoretic lens, defining utility as information gain on output distributions. It introduces a latent helpfulness variable, proving its equivalence to answer-space utility under mild assumptions, and proposes a training-free framework using lightweight models for efficient utility estimation. Experiments on MRAG-Bench and Visual-RAG show the method outperforms state-of-the-art RAG baselines while reducing computational costs.
multimodal retrieval-augmented generationinformation gainlatent helpfulnessutility estimationcomputational efficiency
D-VLA: A High-Concurrency Distributed Asynchronous Reinforcement Learning Framework for Vision-Language-Action Models
We propose D-VLA, a distributed asynchronous reinforcement learning framework for scaling Vision-Language-Action models in embodied AI. The framework introduces Plane Decoupling to isolate high-frequency training data from low-frequency weight control, a four-thread Swimlane pipeline for parallelizing sampling, inference, gradient computation, and parameter distribution, and a dual-pool VRAM management model with topology-aware replication to optimize memory and communication. Evaluations on LIBERO demonstrate D-VLA's superior throughput and sampling efficiency for billion-parameter models, with maintained stability and linear speedup in trillion-parameter scalability tests.
vision-language-action modelsplane decouplingswimlane pipelinedual-pool vramtopology-aware replication
"It became a self-fulfilling prophecy": How Lived Experiences are Entangled with AI Predictions in Menstrual Cycle Tracking Apps
The study investigates human-AI entanglement in menstrual cycle tracking apps (MCTAs) through 14 semi-structured interviews and group autoethnography, revealing how AI predictions shape users' lived experiences despite potential inaccuracies. Findings indicate (1) users interpret personal experiences through AI outputs, which may be flawed due to imperfect data logging, (2) UI designs fail to facilitate critical engagement with AI explanations, and (3) non-normative users experience isolation in these interactions. The work proposes design improvements for predictive AI features in MCTAs.
human-ai entanglementmenstrual cycle tracking appsautoethnographypredictive aiuser interface design
X-Restormer++: 1st Place Solution for the UG2+ CVPR 2026 All-Weather Restoration Challenge
The authors present X-Restormer++, the 1st place solution for the UG2+ CVPR 2026 All-Weather Restoration Challenge. The method enhances X-Restormer's dual-attention architecture (Multi-DConv Head Transposed Attention and Overlapping Cross-Attention) with three key innovations: spatially-adaptive input scaling from Restormer-Plus, a novel Gradient-Guided Edge-Aware (GGEA) loss combined with L1 and MS-SSIM, and expanded training data (+24,500 image pairs from FoundIR and WeatherBench). The system achieved top performance in image restoration under all-weather conditions.
image restorationdual-attentionspatially-adaptive scalingedge-aware lossall-weather
Respecting Self-Uncertainty in On-Policy Self-Distillation for Efficient LLM Reasoning
We introduce EGRSD (Entropy-Guided Reinforced Self-Distillation), a method for on-policy self-distillation in LLM reasoning that incorporates teacher-entropy confidence gating to dynamically weight token-level supervision. EGRSD unifies three signals: reward-grounded direction, teacher-student likelihood-ratio magnitude, and entropy-based confidence gating with a nonzero lower bound. A causal-lookahead variant, CL-EGRSD, further distinguishes sustained from transient high-entropy spans. Experiments with Qwen3-4B and Qwen3-8B demonstrate that EGRSD and CL-EGRSD improve the accuracy-length frontier compared to other trainable methods.
self-distillationentropy-gatingcausal-lookaheadtoken-level supervisionaccuracy-length frontier
Compact Latent Manifold Translation: A Parameter-Efficient Foundation Model for Cross-Modal and Cross-Frequency Physiological Signal Synthesis
The paper proposes Compact Latent Manifold Translation (CLMT), a parameter-efficient (0.09B parameters) foundation model for cross-modal and cross-frequency physiological signal synthesis. CLMT employs a two-stage approach: (1) a Universal Tokenizer using Hierarchical Residual Vector Quantization (RVQ) to decouple signals into discrete latent manifolds, and (2) a Context-Prompted Latent Translator for cross-modal mapping. Evaluations show CLMT outperforms larger baselines, achieving an F1-score of 0.83 in PPG-to-ECG synthesis (vs. 0.37 baseline) and a Pearson correlation of 0.9956 in 25Hz-to-100Hz super-resolution.
residual vector quantizationlatent manifoldcross-modal synthesisphysiological signalsparameter-efficient
It's not the Language Model, it's the Tool: Deterministic Mediation for Scientific Workflows
The authors propose typed mediation, a pattern where language models orchestrate deterministic tools rather than generating analytical code, ensuring reproducibility in scientific workflows. Each tool encodes a researcher's exact procedure for a specific instrument, selected and parameterized by the model. Evaluation on photoluminescence analysis across four platforms shows that typed tools produce identical results, unlike commercial foundation models which exhibit variability or failure. Deployed on two instruments over six months, this approach reduced analysis time from weeks to minutes while guaranteeing reproducibility, particularly crucial for proprietary binary formats and licensed software requiring local infrastructure.
typed mediationdeterministic toolsphotoluminescence analysisreproducibilitylocal infrastructure
Teacher-Guided Policy Optimization for LLM Distillation
Teacher-Guided Policy Optimization (TGPO) is introduced as an on-policy algorithm for LLM distillation, addressing inefficiencies in Reverse KL divergence when student-teacher distributions diverge. TGPO incorporates dense directional guidance by conditioning teacher predictions on student rollouts, maintaining compatibility with existing RLVR frameworks without additional data annotation. Experiments on complex reasoning benchmarks demonstrate TGPO's superior performance over standard baselines and robustness across different teachers.
reverse kl divergencellm distillationon-policy algorithmteacher-guided policy optimizationrlvr frameworks
Improving Code Translation with Syntax-Guided and Semantic-aware Preference Optimization
The paper introduces Code Translation Optimization (CTO), a method to enhance code translation by integrating syntax-guided and semantic-aware preference optimization. CTO employs a cross-lingual semantic model trained via contrastive learning to directly evaluate functional equivalence between source and translated code. This semantic signal is combined with compiler-based syntactic feedback within a direct preference optimization framework, addressing both syntactic correctness and semantic consistency. Experiments across C++, Java, and Python translations show that CTO significantly outperforms existing baselines and alternative preference optimization strategies.
code translationpreference optimizationsemantic equivalencecontrastive learningsyntactic feedback
ReTool-Video: Recursive Tool-Using Video Agents with Meta-Augmented Tool Grounding
The paper introduces ReTool-Video, a recursive tool-using video agent with a MetaAug-Video Tool Library (MVTL) for fine-grained video reasoning. MVTL comprises 134 tools (26 base, 108 meta) for multimodal signal processing and intermediate operations, enabling dual-level access to structured and raw video data. ReTool-Video recursively grounds high-level intents into executable tool chains via parameter repair, substitution, or decomposition. Experiments on MVBench, MLVU, and Video-MME show superior performance, with recursive grounding and meta tools enhancing stability and effectiveness in complex video understanding.
video reasoningtool-augmented agentsmultimodal processingrecursive groundingmeta tools
An Agentic AI Framework with Large Language Models and Chain-of-Thought for UAV-Assisted Logistics Scheduling with Mobile Edge Computing
The authors propose an agentic-AI framework for UAV-assisted logistics scheduling with mobile edge computing, addressing the joint optimization of physical product collection and computational task processing. The framework integrates large language models (LLMs), retrieval-augmented generation, and chain-of-thought reasoning to translate user input into interpretable mathematical formulations. A hierarchical deep reinforcement learning approach, based on proximal policy optimization (PPO), is employed, with an upper layer optimizing UAV routing and a lower layer managing task execution and resource allocation. Simulations demonstrate the framework achieves 99.6% product collection success and 100% task deadline satisfaction, outperforming advantage actor-critic in stability.
uav-assisted logisticsmobile edge computingchain-of-thoughtproximal policy optimizationretrieval-augmented generation
Hierarchical Attacks for Multi-Modal Multi-Agent Reasoning
The paper introduces HAM$^{3}$, a hierarchical attack framework for multi-modal multi-agent systems (MM-MAS), addressing underexplored vulnerabilities in such systems. HAM$^{3}$ decomposes attacks into three layers: perception (perturbing visual/textual inputs and fused representations), communication (corrupting message content and interaction topology), and reasoning (interfering with cognitive pipelines). Evaluated on the GQA benchmark using ReAct, Plan-and-Solve, and Reflexion paradigms, the framework achieves up to 78.3% Attack Success Rate, with reasoning-layer attacks being most effective. Over 50% of successful attacks induce consistent errors across multiple agents.
multi-agent systemsadversarial attacksmulti-modal reasoninghierarchical decompositiongqa benchmark
STAR: Semantic-Temporal Adaptive Representation Learning for Few-Shot Action Recognition
The paper introduces STAR (Semantic-Temporal Adaptive Representation Learning), a framework for few-shot action recognition (FSAR) that addresses semantic-temporal misalignment and multi-scale temporal dynamics. STAR combines a semantic-alignment module with Temporal Semantic Attention (TSA) for frame-level cross-modal alignment and a temporal-aware module with Semantic Temporal Prototype Refiner (STPR) using semantic-guided Mamba blocks and bidirectional state-space refinement. Evaluated on five FSAR benchmarks, STAR achieves improvements of up to 8.1% on SSv2-Full and 7.3% on HMDB51 under 1-shot settings.
few-shot action recognitionsemantic-temporal alignmentmamba blocksstate-space refinementcross-modal alignment
McCast: Memory-Guided Latent Drift Correction for Long-Horizon Precipitation Nowcasting
McCast introduces a memory-guided latent drift correction method for precipitation nowcasting, addressing error accumulation in autoregressive models. The proposed Drift-Corrective Memory Bank (DCBank) performs two-stage correction: a Corrective Latent Extractor predicts initial corrections, followed by a Correction-Aware Memory Retrieval module refining these using historical memory. This approach ensures temporally coherent forecasts by actively correcting latent evolution drift. Evaluated on SEVIR and MeteoNet benchmarks, McCast achieves state-of-the-art performance, particularly in long-horizon forecasting scenarios.
precipitation nowcastingautoregressive modelslatent drift correctionmemory banktemporal coherence
ECG-NAT: A Self-supervised Neighborhood Attention Transformer for Multi-lead Electrocardiogram Classification
The paper introduces ECG-NAT, a self-supervised Neighborhood Attention Transformer for multi-lead ECG classification, addressing challenges in signal variability, noise, and label scarcity. ECG-NAT employs a two-stage approach: generative pretraining via masked autoencoder reconstruction of ECG signals across diverse datasets, followed by discriminative fine-tuning with a dual-loss function combining supervised contrastive and cross-entropy losses. The hierarchical attention mechanism captures multi-scale temporal features efficiently, from beat-level morphology to rhythm-level dependencies. ECG-NAT achieves 88.1% accuracy with only 1% labeled data, demonstrating efficacy in low-resource settings and computational efficiency for real-time diagnosis.
self-supervised learningmasked autoencoderneighborhood attention transformermulti-lead ecghierarchical attention
N-vium: Mixture-of-Exits Transformer for Accelerated Exact Generation
N-vium introduces a mixture-of-exits transformer architecture that accelerates autoregressive generation without quality degradation. The method attaches prediction heads at multiple depths, defining next-token distributions as a learned mixture over these exits with token-adaptive routing, strictly generalizing standard transformers. Sampling from the mixture is exact, and complete KV caches are recovered by deferring upper-layer computation and batching it with later tokens. Pretrained at scales up to 1.5B parameters, N-vium achieves a 57.9% wall-clock speedup over parameter- and data-matched standard transformers with no perplexity cost.
mixture-of-exitsautoregressive transformerskv cachestoken-adaptive routingperplexity
Stable Attention Response for Reliable Precipitation Nowcasting
The paper introduces HARECast, a framework addressing attention-response instability in precipitation nowcasting. It demonstrates that cross-sample attention-energy variability correlates with forecast inaccuracy, theoretically linking it to error propagation. HARECast employs head-wise attention-energy regularization to stabilize responses across layers and heads, applicable to both unimodal and multimodal architectures. Evaluated on SEVIR and MeteoNet benchmarks, the method achieves state-of-the-art performance by reducing attention fluctuations and improving reliability.
attention-response energyprecipitation nowcastingcross-sample instabilitygroup-wise regularizationdiffusion-based predictor
CLIP Tricks You: Training-free Token Pruning for Efficient Pixel Grounding in Large VIsion-Language Models
LiteLVLM introduces a training-free, text-guided token pruning strategy for efficient pixel grounding in large vision-language models, addressing computational overhead from visual tokens. By reversing CLIP's visual-text similarity ranking, it retains referent-region tokens while recovering context tokens for foreground-background separation. Experiments show LiteLVLM outperforms existing methods by over 5% across token budgets, maintaining 90% of original performance with a 22% speedup and 2.3x memory reduction.
token pruningpixel groundingclipvisual-text similarityforeground-background separation
When Does Hierarchy Help? Benchmarking Agent Coordination in Event-Driven Industrial Scheduling
The paper introduces Distributed Event-driven Scheduling Benchmark (DESBench) to evaluate agent coordination in hierarchical event-driven industrial scheduling, addressing gaps in existing benchmarks for shared, dynamically evolving systems. DESBench captures multi-timescale decision making, partial observability, and coupled constraints, defining tasks and metrics for effectiveness, constraint alignment, coordination efficiency, and robustness. Four coordination paradigms—centralized, hierarchical, heterarchical, and holonic—are evaluated, revealing distinct trade-offs: centralized coordination scales poorly, hierarchical coordination suffers from misalignment, heterarchical coordination is communication-heavy, and holonic coordination lacks global robustness. These findings highlight the need for adaptive coordination mechanisms in multi-agent systems.
multi-agent systemsevent-driven schedulinghierarchical coordinationpartial observabilityconstraint alignment
Formal Conjectures: An Open and Evolving Benchmark for Verified Discovery in Mathematics
We introduce Formal Conjectures, an evolving benchmark of 2615 mathematical problem statements formalized in Lean 4, comprising 1029 open research conjectures and 836 solved problems for evaluating automated reasoning systems. The benchmark facilitates collaboration between mathematicians and AI systems, enabling both proof discovery and autoformalization. Correctness is ensured through a collaborative open-source framework, where AI-generated proofs and disproofs iteratively improve formalization fidelity. Initial evaluations demonstrate the benchmark's utility in making new mathematical discoveries, including resolving open conjectures, and provide a climbable signal for measuring automated reasoning capabilities on research-level mathematics.
automated reasoninglean 4proof discoveryautoformalizationformal conjectures
PanoWorld: Towards Spatial Supersensing in 360$^\circ$ Panorama World
We introduce PanoWorld, a multimodal large language model (MLLM) framework for 360° panoramic spatial reasoning, addressing limitations in perspective-image paradigms. The method defines pano-native understanding through semantic anchoring, spherical localization, reference-frame transformation, and depth-aware 3D reasoning, supported by a metadata construction pipeline generating geometry-aware, language-grounded supervision. PanoWorld incorporates Spherical Spatial Cross-Attention to inject spherical geometry into visual processing and is evaluated on PanoSpace-Bench, H* Bench, and R2R-CE Val-Unseen benchmarks. Results show significant performance improvements over proprietary and open-source baselines, demonstrating the necessity of pano-native supervision and geometry-aware adaptation for robust panoramic reasoning.
panoramic reasoningspherical spatial cross-attentionequirectangular projectionmultimodal large language modelgeometry-aware supervision
Strikingness-Aware Evaluation for Temporal Knowledge Graph Reasoning
The paper introduces a strikingness-aware evaluation framework for Temporal Knowledge Graph Reasoning (TKGR) to address the overestimation of model performance caused by uniformly weighted trivial events. It proposes a rule-based strikingness measuring framework (RSMF) to quantify event strikingness by comparing expected occurrence with peer events derived from temporal rules, integrating this as a weighting factor into metrics like weighted MRR and Hits@k. Experiments on four TKG benchmarks show that models perform worse as event strikingness increases, with path-based methods excelling on low-strikingness events and representation-based ones on high-strikingness events, while an ensemble method's gains come from fitting trivial events rather than reasoning improvement.
temporal knowledge graph reasoningstrikingness-aware evaluationrule-based strikingness measuring frameworkweighted mrrhits@k
EvObj: Learning Evolving Object-centric Representations for 3D Instance Segmentation without Scene Supervision
EvObj introduces an unsupervised framework for 3D instance segmentation that addresses the geometric domain gap between synthetic pretraining data and real-world point clouds. The method integrates two novel modules: (1) an object discerning module for dynamic refinement of object candidates, enabling continuous adaptation of object priors to target domains, and (2) an object completion module for reconstructing partial geometries post-object discovery. Extensive experiments on both real-world and synthetic datasets demonstrate state-of-the-art performance in 3D object segmentation, outperforming all baselines.
3d instance segmentationobject discerning moduleobject completion modulegeometric domain gapunsupervised learning
AcquisitionSynthesis: Targeted Data Generation using Acquisition Functions
AcquisitionSynthesis introduces a method for targeted synthetic data generation using acquisition functions as reward models to train language models. The approach quantifies the impact of generated samples on downstream learners, addressing the lack of model-centric evaluation in existing data generation techniques. Experiments on math, medical QA, and coding tasks demonstrate that student models trained with AcquisitionSynthesis data achieve 2-7% performance gains on in-distribution tasks and exhibit improved robustness to catastrophic forgetting. Additionally, the method supports cross-model data generation and low-to-high resource training paradigms, offering a principled path for model-aware self-improvement.
acquisition functionssynthetic data generationcatastrophic forgettingreward modelsmodel-aware self-improvement
A Constraint Programming Approach for $n$-Day Lookahead Playoff Clinching
We introduce a constraint programming (CP)-based algorithm for determining $n$-day lookahead playoff clinching scenarios in the National Hockey League (NHL). The method employs a custom tree search with preprocessing techniques, pruning strategies, and node ordering heuristics to efficiently explore possible game outcomes. A CP subroutine evaluates whether a team has clinched by seeking counter-examples of elimination, incorporating NHL tie-breakers and qualification rules. Validation on hundreds of scenarios from NHL seasons 2021-22 through 2024-25 demonstrates efficacy. The approach is extensible to other playoff-related metrics, including elimination proofs and seed clinching.
constraint programmingtree searchpruning strategiestie-breakersplayoff clinching
LeanSearch v2: Global Premise Retrieval for Lean 4 Theorem Proving
LeanSearch v2 introduces a two-mode retrieval system for global premise retrieval in Lean 4 theorem proving, addressing the challenge of identifying scattered library lemmas for concise proofs. The standard mode employs an embedding-reranker pipeline on a hierarchy-informalized Mathlib corpus, achieving a state-of-the-art nDCG@10 of 0.62. The reasoning mode iteratively refines retrieval through sketch-retrieve-reflect cycles, recovering 46.1% of ground-truth premise groups within 10 candidates on a 69-query benchmark. Downstream evaluation shows a 20% proof success rate, outperforming alternatives. The system is open-sourced and publicly accessible.
global premise retrievalembedding-reranker pipelinemathlib corpussketch-retrieve-reflecttheorem proving
GRACE: Gradient-aligned Reasoning Data Curation for Efficient Post-training
GRACE introduces a gradient-aligned reasoning data curation method that scores individual steps in reasoning traces based on their alignment with answer-oriented gradients and consistency with preceding trajectories. The approach aggregates step-level scores for subset selection, using only internal optimization signals without external rewards or annotations. A representation-level gradient proxy enables scalable estimation of step-level alignment in a single forward pass. Post-training Qwen3-VL-2B-Instruct on MMathCoT-1M, GRACE achieves 108.8% of full-data performance with 20% of the data and 100.2% with 5%, demonstrating effective transfer across model backbones.
gradient-alignedreasoning tracesoptimization signalssubset selectiongradient proxy
MLGIB: Multi-Label Graph Information Bottleneck for Expressive and Robust Message Passing
The Multi-Label Graph Information Bottleneck (MLGIB) addresses over-squashing in Graph Neural Networks (GNNs) by formulating multi-label message passing as constrained information transmission under irrelevant label noise. MLGIB constructs a Markovian dependence space and derives tractable variational bounds: a lower bound maximizing mutual information with target labels and an upper bound constraining redundant source information. This yields a label-aware message-passing architecture balancing expressiveness and robustness. Experiments across multiple benchmarks demonstrate consistent improvements over existing methods, validating MLGIB's effectiveness and generality.
graph neural networksmulti-label graphsinformation bottleneckmarkovian dependencevariational bounds
Towards Long-horizon Embodied Agents with Tool-Aligned Vision-Language-Action Models
We propose VLAs-as-Tools, a framework for long-horizon embodied tasks that decomposes planning and execution across a high-level vision language model (VLM) agent and specialized vision-language-action (VLA) tools. The VLM handles global reasoning and recovery, while VLA tools execute bounded subtasks via a tool-family interface enabling event-triggered replanning. Tool-Aligned Post-Training (TAPT) constructs invocation-aligned training units and employs tool-family residual adapters for efficient specialization. Experiments demonstrate improvements of 4.8 and 23.1 points in success rates on LIBERO-Long and RoboTwin, respectively, and a 15.0-point increase in invocation fidelity measured by Non-biased Rate.
vision-language-actiontool-aligned post-trainingevent-triggered replanningresidual adaptersinvocation fidelity
SECOND-Grasp: Semantic Contact-guided Dexterous Grasping
SECOND-Grasp introduces a unified framework for dexterous grasping that integrates semantic reasoning with physical stability. The method employs vision-language reasoning to generate coarse contact proposals, refines them via Semantic-Geometric Consistency Refinement (SGCR) to ensure semantic and geometric consistency across views, and derives feasible hand poses using inverse kinematics. Trained on DexGraspNet, SECOND-Grasp achieves lifting success rates of 98.2% and 97.7% on seen and unseen object categories, respectively, and improves intent-aware grasping by 12.8% and 26.2%. The framework demonstrates generalizability across datasets and robotic hands, including Shadow Hand and Allegro Hand.
dexterous graspingsemantic-geometric consistencyvision-language reasoninginverse kinematicsintent-aware grasping
Context Matters: Auditing Gender Bias in T2I Generation through Risk-Tiered Use-Case Profiles
The paper proposes a risk-aligned auditing framework for gender bias in text-to-image (T2I) generation, addressing fragmentation in existing evaluations. The framework comprises three components: risk-tiered use-case profiles aligned with EU AI Act categories, a consolidated metric catalog (gender prediction, embedding similarity, downstream task), and a harm typology mapping context-dependent harms to scenarios. It introduces THUMB cards to systematize auditing by integrating context, bias manifestations, and audit strategies. The approach aims to improve technical auditing and governance discussions by connecting risk categories, metrics, and harms.
text-to-image generationgender biasrisk-tiered auditingharm typologymetric catalog
A Multi-Agent Orchestration Framework for Venture Capital Due Diligence
The paper introduces a multi-agent orchestration framework for automating venture capital due diligence, combining Large Language Models (LLMs) with real-time web retrieval to synthesize unstructured data into structured investment intelligence. A key technical contribution is a programmatic extraction pipeline that reverse-engineers the Greek Business Registry ($Γ$.E.MH.) frontend-to-backend communication, retrieving official financial filings via dynamic endpoints and parsing them using layout-aware OCR. The system incorporates a structural fallback mechanism to flag data absence explicitly, mitigating hallucination in financial contexts. All workflow artifacts are publicly available for replication.
multi-agent orchestrationlarge language modelslayout-aware ocrdynamic endpointshallucination mitigation
Margin-calibrated Classifier Guidance for Property-driven Synthesis Planning
We introduce Sequence Completion Ranking (SCR), a margin-calibrated classifier guidance method for synthesis planning that enhances property-driven sequence generation without retraining autoregressive models. SCR employs contrastive argumentation and margin-based loss to calibrate auxiliary classifiers, enabling meaningful discrimination between continuations during decoding. This approach expands the set of property-satisfying sequences reachable under guided beam search. Empirical evaluation on USPTO-190 demonstrates substantial improvements: multi-step solve rates increase from 16.8% (unguided) to 78.4% (reaction-type guidance) and 95.3% (Tanimoto guidance), unlocking valid routes for 17.4% previously unsolvable targets. SCR also bridges the diversity gap between template-free and template-based methods.
synthesis planningclassifier guidancemargin-calibratedsequence completion rankingautoregressive models
Watermarking Should Be Treated as a Monitoring Primitive
The paper argues for treating watermarking as a monitoring primitive in generative models, demonstrating that even zero-bit watermarking enables entity attribution under multi-key settings through signal aggregation. It introduces an observer-based threat model where watermark signals across outputs reveal entity-level information, showing persistent key-dependent statistical structures enable external monitoring over time. Results reveal a dual-use tension between attribution and monitoring, necessitating evaluation beyond per-sample robustness to account for aggregation effects and observer capabilities.
watermarkingmonitoring primitivemulti-key attributionobserver-based threat modelstatistical structure
Vividh-ASR: A Complexity-Tiered Benchmark and Optimization Dynamics for Robust Indic Speech Recognition
We introduce Vividh-ASR, a complexity-stratified benchmark for Hindi and Malayalam speech recognition, addressing studio-bias in multilingual ASR models. The benchmark spans four tiers: studio, broadcast, spontaneous, and synthetic noise. Through controlled experiments on learning-rate timing and curriculum ordering, we demonstrate that early large parameter updates reduce global WER by 12 points, while a hard-to-easy curriculum enhances spontaneous speech performance. These insights inform reverse multi-stage fine-tuning (R-MFT), enabling a 244M Whisper model to match or surpass conventionally fine-tuned 769M models. Representational analysis via CKA and SVD shows effective adaptation concentrates in the decoder, preserving the encoder's acoustic geometry. The benchmark and models are publicly released.
studio-biasreverse multi-stage fine-tuningcomplexity-stratified benchmarkwhisper modelcka
Does language matter for spoken word classification? A multilingual generative meta-learning approach
This paper investigates multilingual spoken word classification using Generative Meta-Continual Learning, contrasting its performance with monolingual and bilingual approaches. The method trains models on English, German, French, and Catalan, evaluating monolingual, bilingual (English-German), and multilingual variants. Results indicate that while the multilingual model achieves the highest performance, differences across models are minimal, suggesting that training data volume (hours of unique data) is a stronger predictor of performance than the number of languages included. The generative meta-learning framework demonstrates viability for multilingual generalization in spoken word classification tasks.
generative meta-learningmultilingual classificationspoken wordmeta-continual learningfew-shot learning
Spectral Flattening Is All Muon Needs: How Orthogonalization Controls Learning Rate and Convergence
The paper demonstrates that Muon's orthogonalization of momentum buffers via Newton-Schulz iterations enables spectral flattening, permitting larger learning rates and faster convergence than standard optimizers. Theoretically, Muon's stable step size scales with average gradient singular values rather than the maximum, and its preconditioning improves convergence under Kronecker-factored curvature assumptions. Experiments confirm Muon maintains stability at SGD-divergent learning rates and achieves accuracy milestones 2-3 epochs faster at matched step sizes.
spectral flatteningmomentum orthogonalizationnewton-schulz iterationskronecker-factored curvaturegradient covariance
Counterfactual Reasoning for Causal Responsibility Attribution in Probabilistic Multi-Agent Systems
The authors introduce a framework for counterfactual responsibility attribution in probabilistic multi-agent systems modeled as concurrent stochastic multi-player games. They define retrospective counterfactual responsibility to quantify agent accountability for outcomes under given strategy profiles, utilizing the Shapley value for responsibility allocation, which satisfies fairness and consistency properties. The framework supports verification and strategic reasoning in responsibility-aware systems, employing Nash equilibrium to compute stable strategy profiles balancing responsibility and expected reward.
counterfactual responsibilityshapley valuenash equilibriummulti-agent systemsstrategy profiles
Scaling few-shot spoken word classification with generative meta-continual learning
This paper demonstrates the scalability of few-shot spoken word classification to 1000 classes using only five shots per class, addressing a gap in prior work focused on smaller class sets. The authors employ Generative Meta-Continual Learning (GeMCL) and compare it to baselines involving repeated fine-tuning of HuBERT and frozen HuBERT with a repeatedly trained classifier head. GeMCL achieves stable performance comparable to the frozen HuBERT baseline while adapting 2000 times faster, requiring less than half the data and two orders of magnitude less training time.
few-shot learningspoken word classificationgenerative meta-continual learninghubertscalability
Neural QAOA$^{2}$: Differentiable Joint Graph Partitioning and Parameter Initialization for Quantum Combinatorial Optimization
Neural QAOA$^{2}$ introduces a differentiable framework for joint graph partitioning and parameter initialization in quantum combinatorial optimization, addressing misalignment between heuristic metrics and quantum goals, and topology-blind initialization. The method integrates a generative evaluative network (GEN) with a differentiable quantum evaluator to provide gradient guidance, learning intrinsic mappings from graph topology to high-quality partitions and parameters. Experiments on 183 QUBO, Ising, and MaxCut instances (21 to 1000 variables) show superior performance, ranking first on 101 instances, with zero-shot generalization across out-of-distribution topologies and scales.
quantum approximate optimization algorithmgraph partitioningparameter initializationdifferentiable frameworkgenerative evaluative network
When Absolute State Fails: Evaluating Proprioceptive Encodings for Robust Manipulation
The study investigates proprioceptive encoding strategies to enhance robustness in robotic manipulation under varying frames of reference. It systematically evaluates joint representations, focusing on improving both in-distribution and out-of-distribution performance. Experiments reveal that an episode-wise relative frame encoding achieves the optimal balance between task performance and robustness, outperforming baseline methods in real-robot tests conducted in realistic environments. This approach offers a practical solution for deploying robots with diverse frames of reference and adapting to unseen configurations.
proprioceptive encodingrobust manipulationrelative framejoint representationsout-of-distribution
Bridging Domain Gaps with Target-Aligned Generation for Offline Reinforcement Learning
The paper introduces Target-aligned Coverage Expansion (TCE), a framework for cross-domain offline reinforcement learning that addresses distributional mismatch between source and target domains. TCE employs a dual score-based generative model to synthesize target-consistent transitions, either by incorporating target-near transitions or expanding state coverage. Theoretical analysis guides this process. Experiments across diverse environments demonstrate TCE's superiority over state-of-the-art cross-domain offline RL baselines.
cross-domainoffline reinforcement learningdistributional mismatchgenerative modelstate coverage
Context Training with Active Information Seeking
The paper introduces a context optimization method for large language models (LLMs) that incorporates active information seeking via Wikipedia search and browser tools. Unlike closed-loop approaches, the proposed method employs a search-based training procedure to maintain and prune multiple candidate contexts, addressing performance degradation from naive tool integration. Evaluations on Flores+ (low-resource translation), HealthBench (health scenarios), LiveCodeBench, and Humanity's Last Exam (reasoning tasks) demonstrate consistent improvements, with additional benefits in data efficiency, robustness, and cross-model generalization.
context optimizationactive information seekinglarge language modelssearch-based traininglow-resource translation
Revealing the Gap in Human and VLM Scene Perception through Counterfactual Semantic Saliency
The paper introduces Counterfactual Semantic Saliency (CSS), a black-box framework to evaluate vision-language models (VLMs) by measuring semantic shifts from object ablation. Comparing VLMs against human psychophysics baselines (16,289 responses across 307 scenes), results reveal a model-human gap: VLMs over-rely on large, central, and high-saliency objects while underweighting human presence. Size bias emerges as a key factor in model-human divergence.
counterfactual semantic saliencyvision-language modelspsychophysics baselinesize biassemantic divergence
An Agentic LLM-Based Framework for Population-Scale Mental Health Screening
The paper proposes an agentic LLM-based framework for population-scale mental health screening, leveraging LangChain agents with explicit policies and proxy-guided evaluation. The framework features incremental locking of validated stages, orchestrated by an Orchestrator Agent handling preprocessing, retrieval, and threshold optimization. A proof-of-concept in transcript-based depression detection demonstrates convergence to stable configurations (e.g., cosine similarity, dynamic Top-k, threshold 0.75) while controlling evaluation costs and avoiding regressions. Results indicate potential for scalable, trustworthy AI applications in healthcare.
agentic llmlangchainproxy-guided evaluationorchestrator agentdynamic top-k
No Attack Required: Semantic Fuzzing for Specification Violations in Agent Skills
The paper introduces Sefz, a semantic fuzzing framework for detecting specification violations in LLM-powered agent skills, where benign inputs trigger unintended breaches of documented safety rules. Sefz translates guardrails into reachability goals over execution traces and uses an LLM-based mutator guided by a multi-armed bandit to generate violating inputs. Evaluated on 402 real-world skills, Sefz identified violations in 120 (29.9%), including 26 previously unknown exploitable cases, revealing six recurring design pitfalls.
semantic fuzzingspecification violationsllm-powered agentsreachability goalsmulti-armed bandit
CoGE: Sim-to-Real Online Geometric Estimation for Monocular Colonoscopy
CoGE introduces a sim-to-real framework for online monocular geometric estimation in colonoscopy, addressing illumination and structural challenges. The method combines an illumination-aware supervision module based on Retinex theory with a structure-aware perception module using wavelet decomposition to bridge the feature gap between simulated and real data. Evaluations show state-of-the-art performance in depth estimation and scene reconstruction, achieved solely through simulation training.
geometric estimationretinex theorywavelet decompositionsim-to-realmonocular colonoscopy
MAP: A Map-then-Act Paradigm for Long-Horizon Interactive Agent Reasoning
The paper introduces the Map-then-Act Paradigm (MAP), a framework addressing Delayed Environmental Perception in interactive LLM agents by shifting environment understanding prior to execution. MAP comprises three stages: Global Exploration for environment-general priors, Task-Specific Mapping for structured cognitive maps, and Knowledge-Augmented Execution for task solving. Experiments demonstrate MAP's effectiveness, enabling frontier models to achieve non-zero performance in 22 of 25 ARC-AGI-3 game environments. The authors also present MAP-2K, a dataset of map-then-act trajectories, showing that training on it outperforms expert execution traces, emphasizing the primacy of environmental understanding over imitation.
map-then-act paradigmdelayed environmental perceptioncognitive mapknowledge-augmented executionepistemic bottleneck
FeatCal: Feature Calibration for Post-Merging Models
FeatCal introduces a feature calibration method to address performance gaps in post-merging models by reducing feature drift. The approach decomposes drift into upstream propagation and local mismatch, then calibrates merged model weights layer-by-layer using a small calibration set. FeatCal employs an efficient closed-form solution without gradient descent or iterative optimization. On CLIP-ViT-B/32 Task Arithmetic and FLAN-T5-base GLUE benchmarks, FeatCal outperforms Surgery and ProbSurgery, achieving 85.5% and 85.2% accuracy respectively, while demonstrating superior sample efficiency and faster calibration (53 seconds for 256 examples per task).
feature driftmodel mergingclosed-form solutiontask arithmeticcalibration set
Understanding and Accelerating the Training of Masked Diffusion Language Models
We propose bell-shaped time sampling to accelerate masked diffusion model (MDM) training for language modeling while maintaining performance. Analysis reveals that MDM's slow learning stems from language locality bias, where token prediction relies heavily on nearby positions. Our training strategy addresses this by modifying the temporal sampling distribution during diffusion steps. Experiments on the One Billion Word Benchmark demonstrate ∼4× faster convergence to equivalent validation negative log-likelihood compared to standard MDM training, alongside improved generative perplexity, zero-shot perplexity, and downstream task performance across multiple benchmarks.
masked diffusion modelslanguage modelinglocality biasbell-shaped time samplingnegative log-likelihood
Rethinking Efficient Graph Coarsening via a Non-Selfishness Principle
We propose NOPE, an efficient graph coarsening method based on a non-selfishness principle that prioritizes collective neighborhood interference, achieving linear memory consumption and near-linear computational complexity. A faster variant, NOPE*, reduces interference evaluation from O(δ·d) to O(d) under the local isotropy assumption, alleviating computational bottlenecks for high-degree nodes. Experiments demonstrate that NOPE* achieves 1.8-10× speedup over NOPE and outperforms baselines with 1-3 orders of magnitude acceleration. Learning on coarsened graphs yields comparable performance to original graphs and can surpass LLM-based graph reasoning due to compact graph information.
graph coarseningnon-selfishness principleinterference evaluationlocal isotropycomputational complexity
Amortized Guidance for Image Inpainting with Pretrained Diffusion Models
The paper introduces Amortized Inpainting with Diffusion (AID), a method for image inpainting that combines the benefits of task-specific training and per-instance adaptation of pretrained diffusion models. AID employs a fixed pretrained diffusion backbone and trains a small reusable guidance module offline, avoiding deployment-time optimization. The authors formulate inpainting as a deterministic guidance problem, derive an auxiliary Gaussian formulation to enable high-dimensional learning, and propose a continuous-time actor-critic algorithm. Evaluations on AFHQv2, FFHQ (pixel EDM), and ImageNet (latent EDM2) show AID improves the quality-speed trade-off over baselines across multiple mask types, with less than 1% trainable overhead.
image inpaintingdiffusion modelsamortized guidanceactor-criticdeterministic guidance
Not Just RLHF: Why Alignment Alone Won't Fix Multi-Agent Sycophancy
This work challenges the attribution of multi-agent sycophancy in LLMs to RLHF, demonstrating that pretrained base models exhibit higher yield (incorrect answer flip rates) than their Instruct variants under simulated peer disagreement. Through activation patching, the study identifies a mid-layer window where attention dominates corruption, with MLP contributions being negligible, and shows that patching above this window restores 96% of the clean-to-pressured correctness gap. The attack surface decomposes into channel framing and consensus strength, producing a 47.5 percentage-point yield gap at majority consensus. Interventions reveal that pressure suppresses clean-reasoning features rather than activating sycophancy circuits, and structured dissent reduces yield by 54-73 percentage points across framings.
activation patchingyieldsycophancyrlhfmulti-agent
Retrieval-Augmented Tutoring for Algorithm Tracing and Problem-Solving in AI Education
The paper introduces KITE (Knowledge-Informed Tutoring Engine), a Retrieval-Augmented Generation (RAG)-based intelligent tutoring system for algorithmic reasoning and problem-solving. KITE employs an intent-aware Socratic response strategy and a multimodal RAG pipeline to retrieve course materials, providing targeted hints, guiding questions, and progressive scaffolding. Evaluations using RAGAs metrics, expert assessment, and simulated student interactions show KITE produces contextually grounded, pedagogically appropriate responses that improve student model accuracy in follow-up answers.
retrieval-augmented generationintelligent tutoring systemalgorithmic reasoningsocratic responsemultimodal retrieval
Protocol-Driven Development: Governing Generated Software Through Invariants and Evidence
Protocol-Driven Development (PDD) introduces a governance model for automated software construction by prioritizing machine-enforceable protocols over implementation code. A protocol is defined as a triplet P = (S, B, O), specifying structural, behavioral, and operational invariants, which collectively constrain the admissible implementation space. Implementations are admitted only if they satisfy the governing protocol and produce a verifiable Evidence Chain of compliance. PDD integrates formal methods, property-based testing, policy-as-code, and software provenance to establish a governance layer, ensuring protocol sovereignty over transient code.
protocol-driven developmentstructural invariantsbehavioral invariantsoperational invariantsevidence chain
CoRe-Gen: Robust Spectrum-to-Structure Generation under Imperfect Fingerprint Conditions
CoRe-Gen introduces a robust framework for molecular structure generation from tandem mass spectra (MS/MS) under imperfect fingerprint conditions. The method addresses condition mismatch via synthetic-spectrum pretraining of the encoder, frequency-aware fingerprint corruption during decoder training, and structure-aware autoregressive decoding with SELFIES representations, auxiliary structural supervision, and chemical constraints. CoRe-Gen achieves state-of-the-art performance on NPLIB1 with 19.54% Top-1 and 29.92% Top-10 exact-match accuracy, while maintaining competitive results on MassSpecGym. The approach preserves the efficiency of autoregressive decoding, offering a scalable solution for realistic spectrum-to-structure generation.
tandem mass spectraautoregressive decodingselfies representationsfrequency-aware corruptionsynthetic-spectrum pretraining
Useful Memories Become Faulty When Continuously Updated by LLMs
The study reveals a critical limitation in LLM-based agentic-memory systems: continuous consolidation of episodic memories into textual abstractions degrades utility despite initial gains, with GPT-5.4 failing on 54% of previously solved ARC-AGI problems after consolidation. Through controlled experiments in the ARC-AGI Stream environment, the authors demonstrate that raw episodic retention outperforms forced consolidation, doubling accuracy, while episodic-only management matches auto-consolidation performance. The findings advocate for treating raw episodes as primary evidence and gating consolidation explicitly, highlighting the need for LLMs that preserve underlying evidence during abstraction.
agentic-memoryepisodic tracesconsolidated abstractionsllm-based systemsarc-agi
Retrieval is Cheap, Show Me the Code: Executable Multi-Hop Reasoning for Retrieval-Augmented Generation
We introduce PyRAG, a framework that reformulates multi-hop Retrieval-Augmented Generation (RAG) as program synthesis and execution, addressing brittleness in existing systems. PyRAG represents reasoning processes as executable Python programs over retrieval and QA tools, exposing intermediate states as variables and enabling compiler-grounded self-repair and execution-driven adaptive retrieval without additional training. Experiments on five QA benchmarks (PopQA, HotpotQA, 2WikiMultihopQA, MuSiQue, and Bamboogle) demonstrate that PyRAG consistently outperforms strong baselines, particularly on compositional multi-hop datasets, under both training-free and RL-trained settings.
retrieval-augmented generationprogram synthesismulti-hop reasoningquestion answeringself-repair
Revisiting Reinforcement Learning with Verifiable Rewards from a Contrastive Perspective
This paper introduces ConSPO, a Contrastive Sequence-level Policy Optimization framework for Reinforcement Learning with Verifiable Rewards (RLVR), addressing structural limitations in GRPO. ConSPO replaces GRPO's clipped ratio-based scores with length-normalized sequence log-probabilities, aligning optimization with autoregressive generation likelihoods, and employs a group-wise InfoNCE-style objective for contrastive credit assignment. It introduces a curriculum-scheduled margin to refine positive-negative separation during training. Evaluations across diverse models, parameter scales, and datasets demonstrate ConSPO's consistent superiority over RLVR baselines on mathematical reasoning benchmarks.
rlvrconspoinfonceautoregressivecurriculum-scheduled
Controlling Logical Collapse in LLMs via Algebraic Ontology Projection over F2
This work introduces Algebraic Ontology Projection (AOP), a method projecting LLM hidden states into Galois Field F2 under Liskov Substitution Principle constraints using 42 relational pairs as algebraic keys. AOP achieves 93.33% zero-shot inclusion accuracy on unseen concept pairs with Gemma-2 Instruct and maintains 86.67% accuracy across model families through prompt optimization alone. The authors propose Semantic Crystallisation (SC) to quantify F2 constraint satisfaction and predict zero-shot accuracy without held-out data. Findings reveal layer-dependent algebraic structures and Late-layer Collapse in 7 of 10 conditions, mitigated by system prompts combined with instruction tuning. This enables formally accessible logical structures in LLMs.
algebraic ontology projectiongalois field f2semantic crystallisationlate-layer collapseliskov substitution principle
Position: Agentic AI System Is a Foreseeable Pathway to AGI
The paper proposes Agentic AI as a necessary paradigm for achieving Artificial General Intelligence (AGI), challenging the dominance of monolithic scaling approaches. Through theoretical analysis, it contrasts the optimization constraints of monolithic learners with the efficiency of Agentic systems, progressing from simple routing mechanisms to Directed Acyclic Graph (DAG) topologies. Results demonstrate that Agentic AI achieves exponentially superior generalization and sample efficiency compared to monolithic models. The work also connects Agentic AI to Mixture-of-Experts frameworks and highlights instability issues in current multi-agent systems, advocating for increased research focus on this paradigm.
agentic aimonolithic scalingdirected acyclic graphmixture-of-expertssample efficiency
Sustaining AI safety: Control-theoretic external impossibility, intrinsic necessity, and structural requirements
The paper employs control theory to analyze structural limitations in sustaining AI safety through external enforcement. It proves an external impossibility result: when AI systems surpass controllable thresholds, no externally enforced strategy can ensure safety, highlighting a class-wide structural failure. Additionally, it identifies intrinsic strategies as necessary under certain conditions, specifying four requirements: independence from external enforcement, initial safety-compatible objectives, stability under self-modification, and scalability with capability growth. The work formalizes concerns about external control limits without proposing complete solutions.
control theoryexternal impossibilityintrinsic necessitysafety-compatible objectivesself-modification
AdaFocus: Adaptive Relevance-Diversity Sampling with Zero-Cache Look-back for Efficient Long Video Understanding
AdaFocus introduces an efficient framework for long-video understanding by rethinking it as progressive evidence acquisition rather than one-pass encoding. The framework comprises two components: Query-Aware Adaptive Relevance-Diversity sampler (AdaRD), which produces compact video previews by adaptively switching to global clustering when queries lack reliable local grounding, and an uncertainty-triggered refinement mechanism that retrieves high-resolution evidence directly from disk via zero-cache I/O design. Experiments on seven benchmarks demonstrate AdaFocus achieves superior efficiency-accuracy trade-offs, improving task performance (e.g., +2.59 accuracy on VideoMME, +8.39 mIoU on Charades-STA) while reducing visual token consumption by ~33x and eliminating in-memory frame pre-caching.
adaptive relevance-diversityzero-cache i/olong-video understandinguncertainty-triggered refinementvisual token consumption
Seg-Agent: Test-Time Multimodal Reasoning for Training-Free Language-Guided Segmentation
Seg-Agent introduces a training-free framework for language-guided segmentation via Explicit Multimodal Chain-of-Reasoning, addressing the spatial grounding limitations of Multimodal Large Language Models (MLLMs). The method constructs an interactive visual reasoning loop comprising generation, selection, and refinement stages, leveraging Set-of-Mark (SoM) visual prompting to render candidate regions directly onto images. This enables MLLMs to iteratively reason about spatial relationships in the visual domain rather than relying solely on textual representations. Evaluated on the novel Various-LangSeg benchmark, Seg-Agent achieves performance comparable to state-of-the-art training-based methods without parameter updates, demonstrating robustness across explicit semantic, generic object, and reasoning-guided segmentation tasks.
language-guided segmentationmultimodal chain-of-reasoningset-of-marktraining-free frameworkspatial grounding
When Should an AI Workflow Release? Always-Valid Inference for Black-Box Generate-Verify Systems
The paper introduces an always-valid release wrapper for LLM-enabled AI workflows that operate through iterative generate-evaluate-revise loops. The method constructs a hard-negative reference pool of high-scoring failures to calibrate deployment-time evaluator scores and accumulates evidence using an e-process, ensuring validity under optional stopping. Theoretical analysis demonstrates finite-sample control over releasing on infeasible tasks and characterizes conditions for nontrivial release on feasible tasks. In a case study using the MBPP+ coding-agent benchmark, the wrapper reduces premature incorrect releases compared to baseline stopping rules while maintaining releases for tasks with moderate supporting evidence.
always-valid inferencegenerate-evaluate-revisehard-negative reference poole-processmbpp+
The Expressivity Boundary of Probabilistic Circuits: A Comparison with Large Language Models
This work identifies two key bottlenecks limiting the expressivity of Probabilistic Circuits (PCs) compared to Transformer-based large language models (LLMs) in autoregressive language modeling. First, PCs' probability-space parameterization struggles with sharp distributions, mitigated by logit-space reformulation. Second, structured-decomposable PCs match Transformer separation rank on vtree-aligned partitions but degrade on heterogeneous dependency topologies due to fixed routing constraints. Theoretical analysis shows decomposable PCs surpass structured-decomposable ones in expressivity, though optimization challenges persist. Empirical and theoretical results demonstrate PCs' limitations in modeling complex language distributions.
probabilistic circuitsautoregressive modelingseparation rankvtree-aligned partitionslogit-space parameterization
CRePE: Curved Ray Expectation Positional Encoding for Unified-Camera-Controlled Video Generation
The paper introduces Curved Ray Expectation Positional Encoding (CRePE), a novel positional encoding for camera-conditioned video generation that supports the Unified Camera Model, including wide-angle and fisheye lenses. CRePE represents image tokens as depth-aware positional distributions along source rays, implemented via a Geometric Attention Adapter integrated into frozen video diffusion transformers (DiTs). Experiments show improved geometry-aware metrics and stable camera control, outperforming RayRoPE-style baselines in positional encoding ablations, while also enabling external geometry control through Radial MixForcing.
positional encodingunified camera modelvideo generationdiffusion transformersgeometric attention
AuraMask: An Extensible Pipeline for Developing Aesthetic Anti-Facial Recognition Image Filters
AuraMask introduces an extensible pipeline for developing anti-facial recognition (AFR) image filters that balance adversarial effectiveness with aesthetic acceptability. The method generates 40 filters emulating Instagram-style aesthetics while maintaining adversarial performance against open-source facial recognition models. Evaluations show these filters significantly outperform prior methods in user acceptance (N=630) while matching or exceeding their adversarial effectiveness. The pipeline is released to accelerate research in aesthetically viable AFR protections.
anti-facial recognitionadversarial filtersaesthetic acceptabilityimage perturbationsurveillance resistance
Anatomy-Slot: Unsupervised Anatomical Factorization for Homologous Bilateral Reasoning in Retinal Diagnosis
The paper introduces Anatomy-Slot, an unsupervised method for bilateral retinal diagnosis that decomposes patch tokens into slots and aligns them across eyes via bidirectional cross-attention. This anatomical factorization improves diagnostic accuracy by 4.2% AUC over a ViT-L baseline on ODIR-5K (n=10 seeds, p=0.002). The approach demonstrates robustness under Gaussian noise and provides quantitative optic disc grounding on REFUGE, validated through cross-attention localization analysis.
anatomy-slotbilateral reasoningunsupervised factorizationretinal diagnosiscross-attention
AgentLens: Revealing The Lucky Pass Problem in SWE-Agent Evaluation
The paper introduces AgentLens, a framework for process-level evaluation of software engineering (SWE) agents, addressing the limitations of binary pass/fail metrics. By analyzing 1,815 trajectories from eight model backends on 47 SWE-bench Verified tasks, the authors identify 'Lucky Passes' (10.7% of passing trajectories) where solutions succeed despite chaotic processes. AgentLens uses Prefix Tree Acceptors (PTAs) and context-sensitive intent labeling to classify actions into Exploration, Implementation, Verification, or Orchestration. Results show Lucky Pass rates vary from 0.5% to 23.2% across models, with significant rank changes when using quality scores instead of pass rates. The released AgentLens-Bench dataset includes annotated trajectories and PTA references.
software engineering agentsprefix tree acceptorslucky passesprocess-level evaluationcontext-sensitive intent labeling
When Attention Closes: How LLMs Lose the Thread in Multi-Turn Interaction
The study introduces the Goal Accessibility Ratio (GAR) to mechanistically explain how large language models (LLMs) degrade in multi-turn interactions, proposing a channel-transition account where goal-defining tokens become less accessible through attention. Combining GAR with sliding-window ablations and residual-stream probes, the authors analyze four architectures, revealing distinct failure modes: some models preserve goal-conditioned behavior despite vanishing attention, while others fail despite decodable residual goal information. Key findings include a collapse in recall from near-perfect to 11% when attention is force-closed in Mistral, and linear probes achieving AUC up to 0.99 for recovering recall outcomes from residual representations.
goal accessibility ratioresidual-stream probessliding-window ablationschannel-transition frameworkmulti-turn interaction
Embodied Multi-Agent Coordination by Aligning World Models Through Dialogue
The work introduces a framework for evaluating world-model alignment in LLM-based embodied agents through natural-language dialogue, extending the PARTNR benchmark with communication capabilities. It proposes three metrics—observation convergence, information novelty, and belief-sensitive messaging—to assess whether dialogue fosters genuine alignment of private world models. Experiments with three LLMs show dialogue reduces action conflicts by 40-83 percentage points but degrades task success, revealing a gap between superficial coordination and true world-model alignment.
embodied agentsworld-model alignmentpartial observabilitynatural-language dialoguemulti-agent coordination
Data Difficulty and the Generalization--Extrapolation Tradeoff in LLM Fine-Tuning
This work investigates the role of data difficulty in supervised fine-tuning (SFT) of large language models (LLMs), revealing that optimal difficulty depends on dataset size. Through empirical and theoretical analysis, including controlled synthetic experiments and PAC-Bayesian generalization bounds, the authors demonstrate that for a fixed data budget, an optimal difficulty level exists, shifting toward harder data as the budget increases. The findings highlight a trade-off between in-distribution generalization and extrapolation gaps, offering practical guidance for difficulty-based data selection in SFT.
supervised fine-tuningdata difficultygeneralization gapextrapolation gappac-bayesian bounds
RISED: A Pre-Deployment Safety Evaluation Framework for Clinical AI Decision-Support Systems
The RISED Framework introduces a five-dimension pre-deployment safety evaluation for clinical AI decision-support systems, addressing limitations of aggregate accuracy metrics. It operationalizes Reliability, Inclusivity, Sensitivity, Equity, and Deployability through formal sub-criteria, BCa bootstrap confidence intervals, and Holm-Bonferroni correction. Validation across synthetic and three real-world cohorts (1980s-2024) shows conventional high-discrimination models can fail input-encoding stability and threshold-shift sensitivity checks while subgroup AUC parity remains inconclusive. The framework reframes Equity as a proxy-dependence diagnostic, requiring outcome-independent need measures for binding fairness verdicts. Implemented as an open-source Python package, RISED bridges in-silico validation and clinical trials.
clinical aisafety evaluationbootstrap confidence intervalssubgroup equityproxy-dependence
Beyond Cooperative Simulators: Generating Realistic User Personas for Robust Evaluation of LLM Agents
The paper introduces Persona Policies (PPol), a method for generating diverse, realistic user personas to improve LLM agent evaluation. PPol uses an LLM-driven evolutionary program search to optimize Python generators that produce task-preserving roleplay policies, guided by a multi-objective fitness score combining human-likeness and behavioral coverage. Evaluations in retail and airline domains show 33-62% fitness score gains over baselines, with annotators rating PPol-conditioned users as human 80.4% of the time. Agents trained with PPol exhibit +17% task success improvement on challenging behaviors.
persona policiesevolutionary program searchmulti-objective fitnessllm-driven simulationtask-preserving roleplay
EcoGEO: Trajectory-Aware Evidence Ecosystems for Web-Enabled LLM Search Agents
EcoGEO introduces Ecosystem Generative Engine Optimization, a paradigm shift from page-level to environment-level influence for web-enabled LLM agents. The proposed TRACE (Trajectory-Aware Coordinated Evidence Ecosystem) method constructs controlled evidence environments that coordinate navigation entry pages with heterogeneous support pages through shared terminology, internal links, and consistent product attributes. Evaluated on OPR-Bench, TRACE outperforms page-level GEO baselines in final target recommendation, with trajectory-level metrics showing increased initial target-result crawls, target-specific follow-up searches, and internal-link crawls, demonstrating improved evidence-acquisition shaping.
ecogeotracellm agentsgenerative engine optimizationevidence ecosystem
Quantifying LLM Safety Degradation Under Repeated Attacks Using Survival Analysis
The study introduces a survival analysis framework to quantify temporal vulnerability of large language models (LLMs) under persistent adversarial attacks, addressing limitations of binary success/failure metrics. Using time-to-jailbreak modeling with hazard functions and survival curves, the method evaluates three LLMs against HarmBench prompts across three attack categories. Results reveal distinct vulnerability profiles: one model degrades rapidly under iterative attacks, while two others maintain consistent moderate vulnerability, providing actionable safety insights for developers.
survival analysislarge language modelsjailbreak attackshazard functionsharmbench
Language-Based Agent Control
Language-Based Agent Control (LBAC) introduces a programming model for agentic applications that enforces user-specified policies through static typing and runtime enforcement. LBAC requires agents to generate well-typed programs within the context of scaffolding code, rejecting unsafe programs via type-checking before execution. This approach ensures policy consistency across both agent-generated behavior and developer-written scaffolding while maintaining expressiveness, allowing side-effect-free computation and recursive subagent invocation. Three case studies demonstrate LBAC's effectiveness: I/O sandboxing via filesystem capabilities, data provenance, and information-flow control.
agentic applicationsstatic typingruntime enforcementtype-checkerinformation-flow control
ChipMATE: Multi-Agent Training via Reinforcement Learning for Enhanced RTL Generation
ChipMATE introduces the first self-trained multi-agent framework for RTL generation, addressing misalignment with industrial practice by eliminating reliance on golden testbenches and closed-source APIs. The framework pairs a Verilog agent with a Python reference-model agent, enabling mutual verification without a golden oracle. It employs a backtrack-based inference workflow to prevent error propagation and a two-stage training pipeline, first training agents individually, then jointly. A hybrid data-generation framework produces 64.4K high-quality training samples. ChipMATE achieves 75.0% and 80.1% pass@1 on VerilogEval V2 with 4B and 9B models, outperforming existing self-trained models and DeepSeek V4.
rtl generationmulti-agent frameworkverilogreference-modelself-trained
Moltbook Moderation: Uncovering Hidden Intent Through Multi-Turn Dialogue
The paper introduces Bot-Mod, a moderation framework for multi-agent systems that detects malicious intent through multi-turn dialogue rather than content filtering. The method employs Gibbs-based sampling over candidate intent hypotheses to progressively narrow agent objectives during interaction. Evaluated on a Moltbook-derived dataset with diverse benign and malicious behaviors, Bot-Mod reliably identifies intent across adversarial configurations while maintaining low false positives on benign cases, advancing intent-aware moderation for open multi-agent environments.
multi-agent systemsintent detectiongibbs samplingadversarial behaviormoderation framework
PRISM: Perinuclear Ring-based Image Segmentation Method for Acute Lymphoblastic Leukemia Classification
The paper introduces PRISM, a Perinuclear Ring-based Image Segmentation Method for Acute Lymphoblastic Leukemia (ALL) classification, addressing challenges in cytoplasmic segmentation due to low contrast and staining variability. PRISM circumvents explicit membrane detection by constructing adaptive concentric zones around nuclei, integrating color and texture features from grey-level co-occurrence patterns. A stacking ensemble of classifiers achieves 98.46% accuracy and 0.9937 precision-recall AUC, demonstrating robustness across staining variations without heavy neural architectures.
perinuclear ringacute lymphoblastic leukemiagrey-level co-occurrencestacking ensemblecytoplasmic descriptors
Persona-Model Collapse in Emergent Misalignment
The study introduces 'persona-model collapse' as a mechanism underlying emergent misalignment in large language models, characterized by deteriorated character simulation and differentiation. Using moral susceptibility (S) and moral robustness (R) metrics derived from Moral Foundations Questionnaire responses during persona role-play, the authors evaluate four models (DeepSeek-V3.1, GPT-4.1, GPT-4o, Qwen3-235B) in base, insecure fine-tuned, and secure fine-tuned variants. Insecure fine-tuning increases S by 55% and decreases R by 65%, indicating dysregulated differentiation and reduced consistency, while secure fine-tuning preserves S and partially mitigates R loss. Unconditioned responses from insecure variants converge toward scale saturation, contrasting with structured base model outputs.
persona-model collapseemergent misalignmentmoral susceptibilitymoral robustnessfine-tuning
AssemblyBench: Physics-Aware Assembly of Complex Industrial Objects
The authors introduce AssemblyBench, a synthetic dataset of 2,789 industrial objects with multimodal instruction manuals, 3D part models, and assembly trajectories, addressing limitations of existing datasets in shape complexity and trajectory realism. They propose AssemblyDyno, a transformer-based model that jointly predicts assembly order and 6-DoF trajectories using instructional manuals and 3D part shapes. Evaluations show AssemblyDyno outperforms prior work in assembly pose estimation and trajectory feasibility, validated through physics-based simulations.
assembly trajectory6-dof motionmultimodal instructionsphysics-based simulationtransformer-based model
Bayesian Model Merging
Bayesian Model Merging (BMM) introduces a bi-level optimization framework for combining task-specific expert models without joint retraining. The inner level formulates merging as activation-based Bayesian regression under a strong anchor model prior, yielding a closed-form solution, while the outer level employs Bayesian optimization for module-specific hyperparameter tuning. BMM leverages alignment between activation statistics and task vectors, enabling a data-free variant that estimates the Gram matrix without auxiliary data. Evaluated on benchmarks including 20-task vision and 5-task language merging, BMM outperforms baselines, achieving 95.1 average accuracy on ViT-L/14 for 8-task merging, closely matching task-specific experts (95.8).
bayesian optimizationtask vectorsgram matrixactivation statisticsbi-level optimization
Multimodal Hidden Markov Models for Persistent Emotional State Tracking
We propose a lightweight framework for persistent emotional state tracking in conversations using sticky factorial HDP-HMMs over multimodal valence-arousal representations from video, audio, and textual inputs. The method models conversational emotion as sequences of latent emotional regimes, outperforming Gaussian HMM baselines in interpretability and computational efficiency. Evaluation using LLM-as-a-Judge, geometric, and temporal consistency metrics demonstrates reliable recovery of meaningful emotional phases, particularly in clinical contexts. The framework enables context augmentation for LLM responses in unstable affective regimes, offering interpretable and actionable analysis of conversational emotion dynamics.
sticky factorial hdp-hmmmultimodal valence-arousalemotional regime trackingllm-as-a-judgecontext augmentation
PROMETHEUS: Automating Deep Causal Research Integrating Text, Data and Models
PROMETHEUS introduces a framework for constructing causal atlases from diverse sources, including literature, data, and models, enabling deep causal research with explicit locality and evidence. The method organizes causal claims into sheaf-like families of local predictive-state models, incorporating structured claim tables, predictive tests, and provenance metrics. Case studies demonstrate its application in literature-atlas scenarios (e.g., ocean-temperature impacts) and grounded-counterfactual analyses (e.g., microplastics forcing), showcasing its ability to evaluate counterfactuals and rebuild world models around scientific substrates. The resulting Topos World Model serves as a navigable research instrument for identifying coherence and tensions across local claims.
causal atlasessheaf-like modelstopos world modelgrounded-counterfactualpredictive-state models
GraphIP-Bench: How Hard Is It to Steal a Graph Neural Network, and Can We Stop It?
GraphIP-Bench introduces a unified benchmark for evaluating model-extraction attacks and ownership defenses on graph neural networks (GNNs) under a black-box protocol. It integrates twelve extraction attacks, twelve defenses, ten public graphs, three GNN backbones, and three graph-learning tasks, reporting fidelity, task utility, ownership verification, and computational cost. Results show that stealing GNNs is easy at medium query budgets, most defenses are ineffective, and heterophilic graphs are harder to steal. Watermarks verify reliably on protected models but lose verification signal on extracted surrogates, revealing gaps in single-model evaluations.
graph neural networksmodel-extraction attacksownership defensesheterophilic graphswatermark verification
FRAME: Forensic Routing and Adaptive Multi-path Evidence Fusion for Image Manipulation Detection
FRAME introduces a forensic routing and adaptive multi-path evidence fusion method for robust image manipulation detection, addressing limitations of single-method approaches. The system organizes diverse forensic algorithms into a multi-path analysis space, adaptively selects informative paths per input, and fuses complementary evidence for improved detection and localization. Experiments demonstrate effectiveness across diverse manipulation scenarios while preserving interpretable forensic cues from multiple sources.
forensic routingmulti-path analysisevidence fusionmanipulation detectionadaptive selection
Orthrus: Memory-Efficient Parallel Token Generation via Dual-View Diffusion
Orthrus introduces a dual-architecture framework unifying autoregressive LLMs and diffusion models for memory-efficient parallel token generation. It augments a frozen LLM with a lightweight trainable module, enabling parallel diffusion alongside autoregressive decoding while sharing the same high-fidelity KV-cache. The autoregressive head pre-fills context for accurate KV representations, while the diffusion head performs parallel generation, with an exact consensus mechanism ensuring lossless inference. Orthrus achieves up to 7.8x speedup with O(1) memory cache overhead and minimal parameter additions, addressing the sequential bottleneck of autoregressive decoding without sacrificing fidelity.
orthruskv-cachediffusion modelsautoregressive decodingparallel token generation
Mechanism Plausibility in Generative Agent-Based Modeling
The paper introduces the Mechanism Plausibility Scale, a four-level framework for evaluating agent-based models (ABMs) that integrate large language models (LLMs). Drawing from philosophy of science and mechanisms literature, the scale distinguishes between generative sufficiency (a model's ability to reproduce phenomena) and mechanistic plausibility (how phenomena are produced). This operationalization clarifies the roles of predictive versus explanatory models in LLM-ABMs, addressing challenges in assessing simulation progress across diverse research areas. The framework aims to provide a structured approach for evaluating the explanatory power of LLM-driven ABMs in social simulations and behavioral modeling.
mechanism plausibility scaleagent-based modelslarge language modelsgenerative sufficiencymechanistic plausibility
Training Large Language Models to Predict Clinical Events
The study introduces Foresight Learning for clinical prediction by converting time-ordered MIMIC-III notes into training examples with past patient context, natural-language questions about future events, and resolved labels. This method generates 6,900 prediction examples across 702 admissions, covering medications, procedures, organ support, microbiology, and mortality. A LoRA adapter trained on these examples improves over the base model, reducing expected calibration error from 0.1269 to 0.0398 and Brier score from 0.199 to 0.145, while marginally outperforming GPT-5 on held-out questions, enabling reusable clinical prediction without structured features or endpoint-specific classifiers.
foresight learningmimic-iiilora adapterclinical predictionbrier score
REALISTA: Realistic Latent Adversarial Attacks that Elicit LLM Hallucinations
REALISTA introduces a realistic latent-space attack framework for eliciting hallucinations in large language models (LLMs) by formulating hallucination elicitation as a constrained optimization problem. The method constructs an input-dependent dictionary of valid editing directions, each corresponding to semantically equivalent and coherent rephrasings, and optimizes continuous combinations of these directions in latent space. This approach combines the optimization flexibility of continuous attacks with the semantic realism of discrete rephrasing-based attacks. Experiments show that REALISTA outperforms or matches state-of-the-art realistic attacks on open-source LLMs and successfully attacks large reasoning models under free-form response settings, where prior methods fail.
latent-space attackhallucination elicitationconstrained optimizationsemantic coherencefree-form response
Correcting Influence: Unboxing LLM Outputs with Orthogonal Latent Spaces
The paper introduces a latent mediation framework for token-level influence attribution in LLMs, addressing limitations of prior influence function methods that assume token independence. The method attaches sparse autoencoders to LLM layers to learn orthogonal latent features, then computes non-decomposable influences via Jacobian-vector products and propagates them to input tokens through activation patterns. Experiments on medical benchmarks demonstrate identification of sparse, interpretable token sets that jointly influence predictions, enabling transparent model auditing in high-stakes domains.
influence functionslatent mediationsparse autoencodersjacobian-vector productstoken attribution
Discrete MeanFlow: One-Step Generation via Conditional Transition Kernels
Discrete MeanFlow introduces a one-step generation method for discrete state spaces by modeling probability mass transport via conditional transition kernels of continuous-time Markov chains (CTMCs). The approach defines a mean discrete rate to measure average transition probability changes over time intervals, leveraging a Discrete MeanFlow identity that connects finite-interval rates to instantaneous CTMC generators. A boundary-by-construction design ensures valid probability outputs and exact boundary conditions without auxiliary losses. Generation requires only a single forward pass and categorical draw, eliminating iterative denoising or ODE integration. Experiments demonstrate high-precision recovery of analytical ground truth on finite-state Markov chains and effective performance on synthetic sequence generation tasks.
conditional transition kernelcontinuous-time markov chaindiscrete meanflowprobability mass transportboundary-by-construction
Emergent and Subliminal Misalignment Through the Lens of Data-Mediated Transfer
This work investigates Emergent Misalignment (EM) and Subliminal Learning (SL) in LLMs through data-mediated transfer, demonstrating that misalignment arises from interactions between fine-tuning data structure, pretraining distributions, and training channels. Experiments reveal that EM occurs more readily when fine-tuning and evaluation prompts share functional structure, allow coherent harmful completions, and target reliably learned behaviors. The study also examines SL, where misalignment is transmitted via benign data from harmful teachers, comparing off-policy and on-policy distillation to disentangle teacher guidance from data distribution effects. Results emphasize a data-centric perspective, showing that misalignment is not merely a consequence of isolated harmful examples but a systemic phenomenon.
emergent misalignmentsubliminal learningdata-mediated transferoff-policy distillationpretraining composition
Adaptive Smooth Tchebycheff Attention for Multi-Objective Policy Optimization
The paper introduces Adaptive Smooth Tchebycheff Attention (PASTA), a framework for multi-objective reinforcement learning that dynamically adjusts optimization landscape curvature to balance stability and non-convex Pareto front coverage. The method employs a conflict-driven controller to modulate smoothness based on real-time gradient interference, enabling elastic transitions between stable linear approximations and precise non-linear scalarizations. Evaluated on a robotic stealth visual search task requiring trade-offs between search coverage, exposure minimization, and speed, PASTA outperforms linear baselines and static non-linear methods in discovering Pareto-optimal policies in non-convex regions.
multi-objective reinforcement learningpareto fronttchebycheff scalarizationgradient interferenceconflict-driven adaptation
WriteSAE: Sparse Autoencoders for Recurrent State
WriteSAE introduces the first sparse autoencoder for decomposing and editing matrix cache writes in state-space and hybrid recurrent language models, addressing limitations of residual SAEs. It factors decoder atoms into native write shapes, derives a closed-form logit shift, and trains under matched Frobenius norm for atom substitution. Evaluations on Qwen3.5-0.8B L9 H4 show atom substitution outperforms matched-norm ablation in 92.4% of 4,851 firings, with 89.8% population test accuracy and R²=0.98 for predicted effects. Mamba-2-370M achieves 88.1% substitution accuracy over 2,500 firings, and three-position installs improve midrank target-in-continuation from 33.3% to 100% under greedy decoding.
sparse autoencodermatrix cachefrobenius normatom substitutionrecurrent language models
Multi-Quantile Regression for Extreme Precipitation Downscaling
Q-SRDRN introduces multi-quantile regression for precipitation downscaling, addressing the systematic under-prediction of extreme events in deep super-resolution networks. The method employs pinball loss at quantiles τ ∈ {0.50, 0.95, 0.99, 0.999}, with IncrementBound ensuring monotonicity and separate per-quantile output heads enabling independent filter banks. Data augmentation via cVAE complements the median head without affecting upper quantiles. Empirical results show significant improvements: on Florida data, Q-SRDRN detects 75.7% of extreme events (200 mm/day) versus 4.2% for baselines, with 63% lower KL divergence and 3.9% lower RMSE. Similar gains are observed in California and Texas, demonstrating robustness across diverse climatic regions.
multi-quantile regressionpinball lossincrementboundsuper-resolutioncvae
Uncovering Symmetry Transfer in Large Language Models via Layer-Peeled Optimization
The paper demonstrates that symmetries in next-token distributions transfer to geometric structures in large language model (LLM) weights and embeddings through layer-peeled optimization analysis. By formulating a constrained nonconvex optimization program as a surrogate for LLMs, the authors prove that cyclic-shift symmetries induce circulant logit matrices and Gram matrices, while exchangeable distributions yield simplex equiangular tight frames in output projections. Theoretical results are validated empirically, showing open-source LLMs exhibit predicted symmetries without explicit regularization.
layer-peeled optimizationcirculant geometrysimplex equiangular tight framenext-token predictionsymmetry transfer
State-Centric Decision Process
The paper introduces the State-Centric Decision Process (SDP), a runtime framework that constructs Markov Decision Process (MDP) components in language environments lacking explicit state spaces. SDP agents iteratively build certified states through natural-language predicates, enabling task-induced state spaces, observation-to-state mappings, certified transitions, and termination criteria. Evaluated on five benchmarks (planning, scientific exploration, web reasoning, multi-hop QA), SDP achieves state-of-the-art training-free performance, with advantages scaling with horizon length. Certified trajectories enable novel analyses like predicate-level credit assignment and failure localization.
state-centric decision processmarkov decision processnatural-language predicatescertified transitionscredit assignment
Simulating Students or Sycophantic Problem Solving? On Misconception Faithfulness of LLM Simulators
The study introduces misconception faithfulness as a framework to evaluate whether LLM-based student simulators maintain coherent misconception-driven belief states during interactions, measured via Selective Flip Score (SFS). Using a misconception-contrastive feedback protocol, the authors test seven LLMs (4B-120B) across datasets and prompting strategies, finding near-zero SFS due to sycophantic problem-solving behavior. Post-training pipelines (SFT, preference optimization, RL with SFS-aligned reward) improve SFS by up to +0.56, demonstrating trainability of interactive belief-aware modeling.
misconception faithfulnessselective flip scorellm simulatorssycophantic problem-solvingbelief-aware modeling
CoT-Guard: Small Models for Strong Monitoring
CoT-Guard introduces a 4B-parameter model for monitoring chain-of-thought (CoT) reasoning in code generation tasks, addressing the inefficiency of large models like GPT-5 and Gemini-3-Flash. The method combines supervised fine-tuning (SFT) to distill detection behavior from stronger monitors and reinforcement learning (RL) on crafted hidden objectives for out-of-domain generalization. Evaluated under a realistic threat model simulating supply-chain attacks, CoT-Guard achieves a G-mean² of 75%, outperforming GPT-5.4 (56%), GPT-5-mini (41%), and Qwen3-32B (54%), and closing the gap to Gemini-3-Flash (83%). This demonstrates CoT-Guard as a cost-effective defense against hidden-objective detection.
chain-of-thoughtsupervised fine-tuningreinforcement learninghidden-objective detectionsupply-chain attacks
What Do You Think I Think? Accounting for Human Beliefs Using Second-Order Theory of Mind
This work introduces a second-order Theory of Mind (ToM-2) framework using Interactive Partially Observable Markov Decision Processes (I-POMDPs) to enable agents to model and account for human cognitive biases and heuristics (CBH) during interactions. The agent detects discrepancies between its actual knowledge and human beliefs, adaptively generating feedback to address these mismatches. An in-person user study demonstrates that the ToM-2 learner significantly improves the informativeness of teacher actions by accounting for CBH, with subjective results indicating participants find the feedback more useful.
theory of mindi-pomdpcognitive biasesadaptive feedbackbelief modeling
From Generalist to Specialist Representation
This work establishes nonparametric identifiability guarantees for learning task-relevant specialist representations from generalist models. The authors prove two hierarchical results: first, task structure is identifiable across time steps in fully unsupervised settings, even with arbitrary temporal dependencies and complex task assignments; second, task-relevant latent representations are disentangled from irrelevant components within each time step under sparsity regularization, without parametric constraints or additional information. These results provide the first general nonparametric identifiability guarantees for moving from generalist to specialist models, setting theoretical limits for downstream applications.
identifiabilitynonparametricsparsity regularizationlatent representationtask structure
BEHAVE: A Hybrid AI Framework for Real-Time Modeling of Collective Human Dynamics
The paper introduces BEHAVE (Behavioral Engine for Human Activity Vector Estimation), a hybrid AI framework for real-time modeling of collective human dynamics as complex dynamical systems. The method represents group interactions via continuous behavioral fields derived from kinematic micro-signals (position, velocity, body orientation), structured into a directed interaction graph and aggregated into non-redundant field bases. Theoretical foundations include a theorem on tension fields and propositions on field basis and criticality index, with neural models for perception and forecasting. The framework is demonstrated on a 7-agent negotiation scenario and generalizes to crowd safety, crisis teams, and clinical contexts.
collective dynamicsbehavioral fieldsinteraction graphkinematic micro-signalscriticality index
Large Language Models for Agentic NetOps and AIOps: Architectures, Evaluation, and Safety
The paper systematizes the use of large language models (LLMs) in agentic NetOps and AIOps through a framework addressing autonomy hierarchies, tool scope, evidence traces, and assurance contracts. It analyzes operational workflows from telemetry to self-healing, emphasizing that reliability stems from surrounding safeguards rather than model capabilities alone. The authors propose workflow-centered evaluation metrics (trace quality, bounded tool use, sandboxed replay) and highlight security risks when LLMs interface with control surfaces, concluding that progress requires treating autonomy as a constrained operational control problem.
llmsnetopsaiopsassurance contractsroot-cause analysis
Grid-Orch: An LLM-Powered Orchestrator for Distribution Grid Simulation and Analytics
Grid-Orch introduces a framework integrating Large Language Models (LLMs) with power system simulation via the Model Context Protocol (MCP), enabling natural language-based distribution grid analysis. The framework supports 36 domain-specific tools across eleven categories, including power flow, voltage analysis, and quasi-static time series simulation, using OpenDSS as the reference implementation. It accommodates both cloud-hosted (Gemini, Claude) and locally deployed (Ollama, llama-cpp) LLMs, ensuring air-gapped operation for security-sensitive utilities. Grid-Orch extends functionality with multi-step workflows for capacitor placement, voltage violation analysis, and overvoltage mitigation. Evaluations demonstrate that tasks like DER interconnection screening, previously requiring hours of scripting, complete in under two minutes with numerically identical results to direct OpenDSS scripting.
large language modelsmodel context protocolquasi-static time seriesair-gapped operationdistribution grid
Inline Critic Steers Image Editing
Inline Critic introduces a learnable token that critiques and steers a frozen image-editing model's hidden states during the forward pass, addressing heterogeneous difficulty in instruction-based image editing. The method leverages early-layer error patterns (rank correlation ρ = 0.83 with final-layer errors) and employs a three-stage training recipe to stabilize critique learning and generation steering. Results include state-of-the-art performance on GEdit-Bench (7.89), a +9.4 improvement on RISEBench, and the strongest open-source result on KRIS-Bench (81.92, surpassing GPT-4o). Analyses confirm the critic's impact on attention and prediction updates in subsequent layers.
inline criticinstruction-based editingerror patternhidden statesforward pass
CHAL: Council of Hierarchical Agentic Language
The Council of Hierarchical Agentic Language (CHAL) introduces a multi-agent dialectic framework for belief optimization in defeasible domains, addressing limitations in current LLM debate methodologies. CHAL employs a CHAL Belief Schema (CBS), a graph-structured Bayesian-inspired architecture, enabling belief revision through gradient-informed dynamics. Meta-cognitive value systems govern agent reasoning and adjudication outcomes, with ablation experiments showing that adjudicator value systems shape debate trajectories, council diversity refines beliefs, and the framework generalizes across fields. CHAL is the first to treat multi-agent debate as structured belief optimization, producing auditable belief artifacts for transparent, aligned AI systems.
chal belief schemadefeasible domainsmulti-agent dialecticbelief optimizationmeta-cognitive value systems
What is Learnable in Valiant's Theory of the Learnable?
This work revisits Valiant's original 1984 learning model, which differs from PAC learning by using only positive examples and membership queries. The authors characterize learnable classes in this model, showing they correspond to those certifiable via poly-size adaptive query-compression schemes. They demonstrate that learnability in Valiant's model strictly lies between PAC learnability and its query-free variant. For arbitrary domains, the same strict hierarchy holds. Additionally, they present the first algorithm for learning d-dimensional halfspaces in Valiant's model, requiring poly(d) samples and queries, with matching lower bounds.
valiant's modelpac learningmembership queriesquery-compressionhalfspaces
R-DMesh: Video-Guided 3D Animation via Rectified Dynamic Mesh Flow
R-DMesh introduces a unified framework for video-guided 3D animation that addresses pose misalignment between user-provided static meshes and reference videos. The method employs a novel VAE to disentangle inputs into a conditional base mesh, motion trajectories, and a rectification jump offset, processed via Triflow Attention for physical consistency. A Rectified Flow-based Diffusion Transformer, conditioned on video latents, transfers spatio-temporal priors to 3D. Evaluated on the Video-RDMesh dataset of 500k dynamic mesh sequences, R-DMesh solves alignment issues and enables robust applications like pose retargeting and 4D generation.
pose misalignmentrectified dynamic meshtriflow attentionrectified flowvideo-guided animation
QLAM: A Quantum Long-Attention Memory Approach to Long-Sequence Token Modeling
The paper introduces Quantum Long-Attention Memory (QLAM), a hybrid quantum-classical approach for long-sequence token modeling that enhances state-space models (SSMs) through quantum superposition. QLAM represents hidden states as quantum states, updated via parameterized quantum circuits conditioned on input, enabling non-classical global updates while maintaining linear-time computation. Evaluated on sequential image classification tasks (sMNIST, sFashion-MNIST, sCIFAR-10), QLAM outperforms recurrent baselines and transformer-based models by implicitly capturing global dependencies through quantum state evolution.
quantum superpositionstate-space modelslong-sequence modelingparameterized quantum circuitslinear-time computation
Reducing cross-sample prediction churn in scientific machine learning
The authors introduce cross-sample prediction churn as a critical metric for scientific machine learning, quantifying label disagreement between models trained on independent bootstraps of the same dataset. They propose two data-side methods to reduce churn: K-bootstrap bagging, which decreases churn by 40-54% without accuracy loss, and twin-bootstrap, a novel approach using sym-KL consistency loss between jointly trained networks, achieving a median 45% further reduction at 2×-ERM compute. Evaluated across 9 chemistry benchmarks, these methods outperform parameter-side techniques like deep ensembles and MC dropout, demonstrating label disagreement rates of 8.0-21.8% despite aggregate accuracy differences of only 1.3-4.2 percentage points.
cross-sample prediction churnk-bootstrap baggingtwin-bootstrapsym-kl consistency lossscientific machine learning
Uncertainty-Driven Anomaly Detection for Psychotic Relapse Using Smartwatches: Forecasting and Multi-Task Learning Fusion
The study presents two smartwatch-based frameworks for detecting psychotic relapse through digital phenotyping. The first forecasts cardiac dynamics and flags deviations, while the second employs multi-task learning to fuse sleep, motion, and cardiac signals using Transformer encoders. Both frameworks output daily anomaly scores derived from predictive uncertainty via multilayer perceptron ensembles. A late-fusion strategy combines these approaches, achieving an 8% relative improvement over the competition-winning baseline on the e-Prevention Grand Challenge dataset. Results indicate that integrating diverse digital phenotypes enhances relapse detection fidelity.
digital phenotypingtransformer encoderspredictive uncertaintymulti-task learninganomaly detection
Provable Quantization with Randomized Hadamard Transform
(No summary returned.)
Parallel Scan Recurrent Neural Quantum States for Scalable Variational Monte Carlo
The authors introduce parallel scan recurrent neural quantum states (PSR-NQS), a scalable variational framework for quantum many-body systems that leverages autoregressive recurrent wave functions and parallelizable recurrence techniques. This method addresses the perceived scalability limitations of recurrent architectures by enabling efficient training within variational Monte Carlo simulations in one and two spatial dimensions. Benchmark results demonstrate accuracy, with iterative retraining achieving agreement with quantum Monte Carlo data on 2D spin lattices up to 52×52 in size. The work establishes recurrent architectures as a practical approach for scalable neural quantum state simulations using modest computational resources.
neural quantum statesvariational monte carloautoregressive recurrentparallel scanspin lattices
Min-Max Optimization Requires Exponentially Many Queries
The paper establishes an exponential query complexity lower bound for min-max optimization of nonconvex-nonconcave functions. Using oracle access to both the function and its gradient, the authors prove that any algorithm finding an ε-approximate stationary point requires exponentially many queries in either 1/ε or the dimension d. This result holds for functions defined over the domain [0,1]^d × [0,1]^d, demonstrating fundamental computational limitations in this optimization setting.
min-max optimizationquery complexitynonconvex-nonconcavestationary pointgradient oracle
Force-Aware Neural Tangent Kernels for Scalable and Robust Active Learning of MLIPs
The authors introduce a scalable active learning framework for machine-learning interatomic potentials (MLIPs) that addresses energy-force supervision and distribution robustness. Their method employs chunked feature-space posterior-variance shortlisting to screen ~200k structures efficiently and extends the Neural Tangent Kernel (NTK) to a force-aware setting via mixed parameter-coordinate derivatives, yielding force NTK and joint energy-force NTK. Evaluated on OC20, T1x, PMechDB, and RGD benchmarks, the joint energy-force NTK achieves the lowest energy and force MAE/RMSE, outperforming committee-based approaches in efficiency and robustness under candidate-pool shifts.
neural tangent kernelactive learninginteratomic potentialsposterior-varianceforce-aware
Interpretable Machine Learning for Antepartum Prediction of Pregnancy-Associated Thrombotic Microangiopathy Using Routine Longitudinal Laboratory Data
This study proposes an interpretable machine learning approach for predicting pregnancy-associated thrombotic microangiopathy (P-TMA) using routine longitudinal laboratory data. The retrospective analysis evaluated five algorithms—logistic regression, support vector machine, random forest, extra trees, and gradient boosting—on 300 pregnancies (142 P-TMA cases, 158 controls) with 146 longitudinal predictors. Gradient boosting, selected via cross-validation, achieved an AUROC of 0.872 (95% CI: 0.769-0.952) and AUPRC of 0.883 (95% CI: 0.780-0.959) in the held-out test cohort, demonstrating sensitivity of 0.750 and specificity of 0.812. Interpretability analyses identified cystatin C at week 6 as a promising early indicator.
gradient boostinglongitudinal predictorsauroccross-validationinterpretability
Attention Once Is All You Need: Efficient Streaming Inference with Stateful Transformers
The paper introduces stateful transformers for efficient streaming inference, addressing the O(n) prefill cost in conventional request-driven transformer engines. The proposed method employs a persistent KV cache advanced incrementally with new data, reducing query latency to O(|q|) independent of context size. Flash Queries leverage idle GPU cycles to pre-evaluate registered questions, returning cached answers before user requests. A multi-tenant continuous-batching scheduler with cell-budget admission and prefix-aware grouped prefill enables dozens of stateful sessions on a single GPU while maintaining full quadratic self-attention. Experiments on streaming market-data benchmarks demonstrate up to 5.9x speedup over conventional inference engines (vLLM, SGLang, TensorRT-LLM, llama.cpp), maintaining constant query latency as context grows.
kv cachestreaming inferenceflash queriescontinuous-batchingself-attention
Dense vs Sparse Pretraining at Tiny Scale: Active-Parameter vs Total-Parameter Matching
The study compares dense and sparse (Mixture-of-Experts, MoE) transformers in tiny-scale pretraining under a shared LLaMA-style decoder framework. Sparse models replace dense feed-forward blocks with Mixtral-style routed experts, while dense baselines are resized to match either active or total parameter budgets. Key configurations include four experts, top-2 routing, Switch-style load balancing, and router z-loss. Results show MoE achieves better validation loss (1.5788) under active-parameter matching compared to dense models (1.6545), but dense models outperform (1.5608) under total-parameter matching. The active-parameter advantage for MoE grows during training, while the dense total-parameter advantage narrows.
mixture-of-expertsllama-stylerouted expertstop-2 routingvalidation loss
VectorSmuggle: Steganographic Exfiltration in Embedding Stores and a Cryptographic Provenance Defense
The paper introduces VectorPin, a cryptographic provenance protocol to defend against steganographic exfiltration attacks in retrieval-augmented generation (RAG) systems. It demonstrates that attackers with write access can embed payloads into high-dimensional embeddings via perturbations (noise injection, rotation, scaling, etc.) while preserving retrieval behavior. Evaluations across synthetic-PII corpora, open embedding models, and vector-store configurations reveal that small-angle orthogonal rotation evades detection, while distribution-shifting perturbations are often caught. VectorPin ensures embedding integrity by signing embeddings with Ed25519, breaking verification upon modification. This protocol provides a deployable defense against embedding-level exfiltration.
steganographic exfiltrationretrieval-augmented generationembedding integritycryptographic provenancevector-store
Toward AI-Driven Digital Twins for Metropolitan Floods: A Conditional Latent Dynamics Network Surrogate of the Shallow Water Equations
The Conditional Latent Dynamics Network (CLDNet) is introduced as a neural ODE-based surrogate for metropolitan flood forecasting, addressing the computational inefficiency of GPU-accelerated shallow water equation solvers. CLDNet employs a low-dimensional latent dynamics model driven by rainfall, coupled with a coordinate-based decoder conditioned on static terrain features, enabling depth and discharge reconstruction at arbitrary query points. Evaluated on a synthetic Texas benchmark and a Des Plaines case study, CLDNet reduces relative root-mean-squared error by half, achieves an 86% critical success index at the 0.5m inundation threshold, and delivers a 96-hour basin-wide forecast in ~29 seconds, a 115x speedup over traditional methods.
conditional latent dynamics networkshallow water equationsneural odeflood forecastingcoordinate-based decoder
Fast and effective algorithms for fair clustering at scale
We propose a general framework for fair clustering with precise control over the cost-fairness trade-off, introducing three heuristics: one optimizing solution quality and constraint flexibility, another improving scalability while maintaining quality, and a third maximizing scalability for million-object instances. The framework addresses clustering problems where objects belong to protected groups, aiming to minimize the sum of squared Euclidean distances while ensuring user-defined fairness levels across clusters. Comprehensive experiments on benchmark datasets demonstrate that our heuristics outperform existing methods in both scalability and solution quality. Source code and reproducibility instructions are publicly available.
fair clusteringcost-fairness trade-offscalabilityeuclidean distancesprotected groups
Min Generalized Sliced Gromov Wasserstein: A Scalable Path to Gromov Wasserstein
The authors propose min Generalized Sliced Gromov-Wasserstein (min-GSGW), a scalable formulation for the Gromov-Wasserstein (GW) problem using generalized nonlinear slicers. The method learns coupled slicers that assign compatible push-forward values to input measures, enabling efficient transport plan construction in the original spaces while minimizing the GW objective. Min-GSGW is rigid-motion invariant, making it suitable for geometric matching tasks. An amortized variant replaces per-instance optimization with a learned slicer for unseen pairs. Experiments on animal mesh matching, horse mesh interpolation, and ShapeNet part transfer demonstrate meaningful geometric correspondences and reduced computational costs compared to existing GW solvers.
gromov-wassersteingeneralized slicerstransport planrigid-motion invariancegeometric matching
GHGbench: A Unified Multi-Entity, Multi-Task Benchmark for Carbon Emission Prediction
We introduce GHGbench, a unified benchmark for company- and building-level greenhouse-gas emission prediction, addressing fragmentation in existing datasets. GHGbench comprises two tracks: a company track with 32,000+ company-year records and a building track with 491,591 building-year records harmonized from 13 sources across 26 metropolitan areas. The benchmark evaluates in-distribution, cross-region transfer, temporal hold-out, and short-horizon forecasting tasks using baselines including gradient-boosted trees, tabular foundation models, MLP, FT-Transformer, and multimodal fusion. Key findings include: building emissions are harder to predict than company emissions; out-of-distribution performance gaps exceed within-model gaps; and multimodal remote-sensing embeddings improve generalization where tabular data fails. GHGbench identifies catastrophic city transfer and sector-factor lookup as systematic failure modes.
greenhouse-gas predictiontabular foundation modelmultimodal fusioncross-region transferremote-sensing embeddings
Learning POMDP World Models from Observations with Language-Model Priors
Pinductor introduces a novel approach to learning partially-observable Markov decision process (POMDP) world models from observation-action trajectories using language-model priors. The method leverages a large language model (LLM) to propose and iteratively refine candidate POMDP models, optimizing a belief-based likelihood score. Pinductor matches the performance of LLM-based POMDP learning methods that assume hidden state access while significantly surpassing tabular POMDP baselines in sample efficiency. Performance scales with LLM capability and degrades gracefully as semantic information is withheld, demonstrating the utility of language-model priors for sample-efficient world-model learning in partially observable environments.
pomdplanguage-model priorssample efficiencybelief-based likelihoodpartially observable
Distinguishing performance gains from learning when using generative AI
The article critically examines the integration of generative AI in educational contexts, highlighting its potential to enhance learner performance without fostering deep cognitive or metacognitive processing essential for high-quality learning. It argues that while generative AI tools can improve task outcomes, they may not sufficiently engage learners in meaningful cognitive engagement or self-regulated learning strategies. The analysis suggests a need for careful design of AI-assisted educational interventions to ensure they promote both performance gains and deeper learning processes.
generative aicognitive processingmetacognitive processingeducational interventionsself-regulated learning
Tight Sample Complexity Bounds for Entropic Best Policy Identification
This work closes the exponential horizon dependence gap in sample complexity bounds for entropic best-policy identification in finite-horizon risk-sensitive reinforcement learning. The authors introduce a forward-model based algorithm incorporating KL-based exploration bonuses tailored for the entropic criterion. Key innovations include leveraging the smoothness of exponential utility for sharper concentration bounds and a novel stopping rule exploiting this tightness. The resulting sample complexity matches the lower bound of Ω(e^{|β| H}), eliminating the previous O(e^{2|β| H}) gap from state-of-the-art upper bounds.
entropic risk measuresample complexitykl-based explorationconcentration boundsstopping rule
MILM: Large Language Models for Multimodal Irregular Time Series with Informative Sampling
The paper introduces MILM (Multimodal Irregular time series Language Model), a two-stage LLM approach for classifying Multimodal Irregular Time Series (MITS) like EHR data. MILM represents MITS as XML-formatted time-ordered triplets, first training on value-redacted data to learn from sampling patterns, then on full data to jointly model patterns and values. Evaluations show MILM-2S (two-stage) outperforms MILM-Direct (single-stage) on EHR datasets, particularly in value-pending scenarios where some data is missing. The method demonstrates that sampling patterns alone carry predictive signal, with further improvements when preserving timing/channel metadata for pending observations.
multimodal irregular time serieselectronic health recordsvalue redactionxml representationtwo-stage training
DisAgg: Distributed Aggregators for Efficient Secure Aggregation in Federated Learning
DisAgg introduces a distributed secure aggregation protocol for federated learning that delegates aggregation to a committee of clients called Aggregators, eliminating costly homomorphic encryption while preserving privacy against honest-but-curious servers and limited client collusion. Each client secret-shares its update vector to Aggregators, which compute partial sums and return aggregated shares for server-side reconstruction, reducing endpoint computation and communication overhead. Compared to One-Shot Private Aggregation (OPA), DisAgg achieves a 4.6x speedup when processing 100k-dimensional update vectors from 100k 5G clients.
federated learningsecure aggregationsecret-sharinghomomorphic encryptionclient dropout
Polyhedral Instability Governs Regret in Online Learning
The paper establishes polyhedral instability as a key determinant of regret in online convex optimization with piecewise linear objectives. By analyzing the number of active region changes (RS_T) under full information feedback, the authors prove regret bounds scaling as Θ(√((1+RS_T)T log V_max)), where V_max is the maximum vertex count per region. For online submodular-concave games, this reduces to permutation-switch count (SC_T), yielding Θ(√((1+SC_T)T log n)). Experiments on shortest path and influence maximization tasks confirm the theoretical scaling and reveal practical low-instability regimes.
online convex optimizationpolyhedral instabilityregret boundssubmodular-concave gamesregion switches
MedCore: Boundary-Preserving Medical Core Pruning for MedSAM
MedCore introduces a structured pruning framework for MedSAM that preserves both adaptation-critical structures and boundary-sensitive components in medical segmentation models. The method employs a dual-intervention score to identify SAM-to-MedSAM adaptation structures and boundary-aware Fisher estimation for boundary leverage. A boundary leverage principle explains logit perturbation effects on boundary displacement. MedCore achieves 60.0% parameter reduction and 58.4% FLOP reduction while maintaining Dice 0.9549, Boundary F1 0.6388, and HD95 5.14 on polyp segmentation benchmarks. Analysis reveals MedSAM operates in a head-fragile boundary regime, with head-pruning steps exhibiting 2.887× higher 95th-percentile boundary leverage than MLP-pruning steps.
structured pruningboundary leveragedual-intervention scorefisher estimationmedsam
Scale-Sensitive Shattering: Learnability and Evaluability at Optimal Scale
The paper resolves fundamental questions about the optimal scale for learnability and uniform convergence in real-valued function classes. By introducing a scale-sensitive generalization of PAC learning theory, it establishes equivalence between uniform convergence at scale γ, agnostic learnability at γ/2, and finiteness of fat-shattering dimension above γ. The analysis employs direct bounds on empirical ℓ∞ covering numbers, bypassing traditional packing number approaches. Results include tight O(log²n) metric-entropy bounds at γ/2 and O(logn) at 2γ, disproving prior conjectures about unavoidable multiplicative gaps. Applications yield a sharp 3-factor dichotomy for bounded integral probability metrics.
fat-shattering dimensionuniform convergenceagnostic learnabilitymetric entropyintegral probability metrics
Sampling from Flow Language Models via Marginal-Conditioned Bridges
The paper introduces a novel sampling method for Flow Language Models (FLMs) that leverages their unique structure of posterior marginal distributions over tokens. The proposed marginal-conditioned bridge sampler generates clean one-hot endpoints from factorized posteriors and samples continuous states via Ornstein-Uhlenbeck bridges, preserving token-wise marginals while handling cross-position dependencies. Theoretical analysis shows the method reduces endpoint approximation error (quantified as conditional multi-information) and maintains denoising performance. Empirical results demonstrate improved quality-diversity tradeoffs in FLM sampling, with implementation available on GitHub.
flow language modelsmarginal-conditioned bridgesorstein-uhlenbeck bridgeposterior-predictive samplingconditional multi-information
Three-Stage Learning Unlocks Strong Performance in Simple Models for Long-Term Time Series Forecasting
The paper introduces STAIR (Stagewise Temporal Adaptation via Individualization and Residual Learning), a three-stage training paradigm for long-term time series forecasting that enhances simple models (e.g., shallow MLPs) without complex architectural modules. STAIR progresses through shared temporal mapping, channel-wise fine-tuning, and residual learning, incorporating Shared-to-Individual Fine-tuning and alpha-RevIN to address channel independence and normalization issues. Evaluated on nine benchmarks, STAIR matches or outperforms recent baselines while maintaining model simplicity.
stagewise adaptationresidual learningtemporal mappingchannel-wise fine-tuningalpha-revin
Characterizing Universal Object Representations Across Vision Models
The study identifies universal object representations across 162 diverse vision models by decomposing their similarity structures into non-negative dimensions. Using statistical analysis, it distinguishes universal dimensions (interpretable, semantically driven) from model-specific ones, finding no correlation with architecture, objective, data, size, or performance. Universal dimensions better predict macaque IT activity and human similarity judgments, suggesting biological relevance. Results imply that semantic interpretability drives representational convergence in deep networks.
object similaritynon-negative decompositionuniversal dimensionsbiological visioninterpretability
Graph Neural Networks with Triangle-Based Messages for the Multicut Problem
We propose a graph neural network architecture tailored for the NP-hard multicut problem, featuring edge-based feature assignment and triangle-centric message computation. The method leverages the problem's specific objective function and constraints, outperforming state-of-the-art heuristic solvers on synthetic and real-world instances with up to 200 nodes. Experimental results demonstrate superior solution quality with feasible runtimes, achieving optimal solutions in seconds for instances requiring hours with exact solvers.
graph neural networksmulticut problemtriangle-based messagescombinatorial optimizationedge features
Conformal Anomaly Detection in Python: Moving Beyond Heuristic Thresholds with 'nonconform'
The 'nonconform' Python package introduces conformal anomaly detection to convert uncalibrated anomaly scores into statistically valid p-values, addressing heuristic threshold limitations. It integrates with 'scikit-learn', 'pyod', and custom detectors, offering calibration, p-value generation, and false discovery rate control. The package supports multiple conformalization strategies, from split-conformal to shift-aware methods, demonstrating statistically principled anomaly detection in empirical evaluations. Designed for reproducibility, it bridges theoretical conformal methods with practical workflows in both research and production settings.
conformal anomaly detectionp-valuesfalse discovery ratesplit-conformal calibrationshift-aware
Multi-Objective and Mixed-Reward Reinforcement Learning via Reward-Decorrelated Policy Optimization
Reward-Decorrelated Policy Optimization (RDPO) is introduced to address instability in multi-objective and mixed-reward reinforcement learning environments. RDPO employs Magnitude-Aware Quantile normalization to stabilize advantage allocation across diverse reward types and Mahalanobis whitening to reduce correlation redundancy within reward subspaces. When applied to LongCat-Flash post-training, RDPO improves instruction following, writing quality, and robustness to hard prompts while maintaining competitive performance on reasoning and coding benchmarks.
reward-decorrelated policy optimizationmagnitude-aware quantile normalizationmahalanobis whiteningmulti-objective reinforcement learninglongcat-flash
Achieving $ε^{-2}$ Sample Complexity for Single-Loop Actor-Critic under Minimal Assumptions
(No summary returned.)
CO-MAP: A Reinforcement Learning Approach to the Qubit Allocation Problem
CO-MAP introduces a reinforcement learning approach to solve the qubit allocation problem in quantum compilation, formulated as a combinatorial optimization task. The method trains an RL policy to generate logical-to-physical qubit mappings, supplemented by a local search post-processing algorithm to minimize SWAP gate overhead. Evaluations on MQTBench and Queko circuits demonstrate a 65-85% reduction in SWAP overhead compared to conventional quantum compilers, significantly improving circuit efficiency.
quantum compilationqubit mappingreinforcement learningcombinatorial optimizationswap overhead
Multimodal Graph-based Classification of Esophageal Motility Disorders
This work proposes a multimodal graph-based approach for classifying esophageal motility disorders, combining high-resolution impedance manometry (HRIM) data with patient-specific information. HRIM recordings are modeled as spatio-temporal graphs, processed by a graph neural network (GNN), and fused with patient embeddings derived from demographic, clinical, and symptom data. The method outperforms HRIM-only models and vision-based baselines across all classification categories, demonstrating the complementary value of multimodal integration. Experiments on 104 patient cases show that graph-based modeling of esophageal physiology improves classification accuracy, highlighting the potential of this approach for more precise diagnosis of motility disorders.
graph neural networkmultimodal classificationspatio-temporal graphshigh-resolution impedance manometryesophageal motility disorders
Deep Learning as Neural Low-Degree Filtering: A Spectral Theory of Hierarchical Feature Learning
The paper introduces Neural Low-Degree Filtering (Neural LoFi), a theoretical framework for understanding hierarchical feature learning in deep neural networks. Neural LoFi models gradient-based training as an iterative spectral procedure where each layer selects directions with maximal low-degree correlation to the label, decoupling layer dynamics. This provides a tractable surrogate mechanism for deep learning, explaining concept emergence, sample complexity, and depth-driven feature construction through low-degree compositionality. Empirical validation on fully connected and convolutional architectures demonstrates that Neural LoFi outperforms lazy random-feature baselines, recovers structured filters, and aligns with early gradient-descent feature discovery in real datasets.
neural low-degree filteringhierarchical feature learningspectral procedurelow-degree correlationcompositionality
Rethinking Generalization in Graph Neural Networks: A Structural Complexity Perspective
The work establishes that graph structure significantly influences GNN generalization through theoretical and empirical analysis. It proves that excessive edges induce overfitting by making input representations overly accommodating, then derives a Rademacher complexity bound incorporating a novel structural complexity measure based on effective edges. The proposed structural entropy regularization method controls this complexity, improving generalization across benchmarks by balancing underfitting and overfitting.
graph neural networksgeneralization boundstructural complexityrademacher complexityentropy regularization
Causal Learning with the Invariance Principle
The paper establishes that causal discovery becomes identifiable under minimal assumptions of acyclicity and invariance across environments. Using structural causal models (SCM), the authors prove that just two auxiliary environments suffice to infer causal graphs for arbitrary nonlinear mechanisms, enabling correct counterfactual inference. Theoretical guarantees are empirically validated on synthetic data.
causal discoverystructural causal modelsinvariance principlenonlinear mechanismscounterfactual inference
Reframing preprocessing selection as model-internal calibration in near-infrared spectroscopy: A large-scale benchmark of operator-adaptive PLS and Ridge models
The paper introduces operator-adaptive calibration, a framework integrating preprocessing selection within linear calibration models for NIRS spectroscopy. By encoding preprocessing as linear operators and handling nonlinear corrections via fold-local branches, the method avoids costly external pipeline searches. Implemented for PLS and Ridge regression, the approach preserves interpretable coefficients and enables fast computation. Evaluation across 50+ datasets shows operator-adaptive PLS with ASLS preprocessing achieves a median RMSEP/PLS ratio of 0.960, outperforming conventional PLS, Ridge, CatBoost, and CNN baselines. The method reduces preprocessing-HPO dependency while maintaining auditability and rapid deployment.
near-infrared spectroscopyoperator-adaptive calibrationpartial least squaresridge regressionspectral preprocessing
Spatiotemporal downscaling and nowcasting of urban land surface temperatures with deep neural networks
The study introduces a deep learning framework for high-resolution spatiotemporal Land Surface Temperature (LST) estimation and nowcasting, addressing the trade-off between spatial and temporal resolution in satellite-derived LST products. A U-Net model is employed to downscale LST fields from SEVIRI/MSG (3 km, 15 min) to Terra/Aqua MODIS (1 km, 4 overpasses/day), achieving an RMSE of 1.92°C and MBE of 0.01°C on test data. A ConvLSTM-based nowcasting model is then trained on downscaled LST fields, outperforming persistence and climatological benchmarks with RMSEs of 0.57–1.15°C for lead times of 15–75 minutes. Validation against independent MODIS overpasses confirms robust performance.
land surface temperatureu-netconvlstmspatiotemporal downscalingnowcasting
Uncertainty-Aware Prediction of Lung Tumor Growth from Sparse Longitudinal CT Data via Bayesian Physics-Informed Neural Networks
A Bayesian physics-informed neural network is proposed for lung tumor growth prediction from sparse longitudinal CT data, integrating Gompertz growth dynamics with Bayesian inference. The method employs a two-stage inference strategy combining maximum a posteriori estimation and Hamiltonian Monte Carlo sampling to estimate posterior predictive distributions and uncertainty intervals. Evaluated on data from the National Lung Screening Trial (30 patients), the model achieved a cohort-level log-space RMSE of 0.20 and well-calibrated 95% credible interval coverage. Results demonstrate accurate prediction of heterogeneous tumor growth patterns with calibrated uncertainty estimates, outperforming deterministic approaches.
bayesian inferencegompertz growthhamiltonian monte carlophysics-informed neural networkuncertainty estimation
Mixed neural posterior estimation for simulators with discrete and continuous parameters
The paper extends Neural Posterior Estimation (NPE) to mixed parameter spaces containing both discrete and continuous dimensions, addressing a limitation in existing approaches. The proposed method factorizes the joint posterior into discrete and continuous components, combining an autoregressive classifier for discrete parameters with a generative model for continuous parameters, trained jointly under a simulation-based objective. A diagnostic tool assesses calibration of the mixed posterior approximation. Experiments on toy examples and real-world simulators demonstrate accurate, calibrated posteriors. The implementation is available in the sbi Python package.
neural posterior estimationmixed parameter spacesautoregressive classifiersimulation-based inferenceposterior calibration
Beyond Explained Variance: A Cautionary Tale of PCA
The paper critiques principal component analysis (PCA) for visualizing nonlinear manifold data, demonstrating its limitations through a fossil teeth dataset from Kuehneotherium. While PCA suggested clustering, t-SNE and persistent homology revealed a ring-like structure with intrinsic dimensionality one. A probabilistic-geometric model sampling uniformly from a unit circle was proposed, showing pairwise cosine distances follow an arcsine distribution, corroborating the t-SNE and PH findings.
principal component analysist-snepersistent homologynonlinear manifoldintrinsic dimensionality
Limits of Personalizing Differential Privacy Budgets
The paper demonstrates fundamental limitations in personalized differential privacy budgets for mean estimation, showing that full personalization offers marginal utility gains compared to threshold-based budget selection. The authors analyze mixed private/public datasets and two-level privacy requirements, precisely quantifying constant-factor improvements. They establish upper bounds and identify regimes where personalization provides maximal benefit, proposing a simple thresholding operator as an effective alternative to complex personalized mechanisms.
differential privacymean estimationprivacy budgetpersonalizationthresholding
Reward-Weighted On-Policy Distillation with an Open Property-Equivalence Verifier for NL-to-SVA Generation
The paper introduces Reward-Weighted On-Policy Distillation (RWOPD) to address the temporal gap in LLM-based SystemVerilog Assertion (SVA) generation, where models excel in aggregate but fail on specific templates. RWOPD samples student rollouts, scores them with a Property-Equivalence Checker (PEC), and applies verifier-reward-weighted gradients from a frozen 14B teacher. This method distills CodeV-SVA-14B into a Qwen2.5-Coder-7B-Instruct student, achieving state-of-the-art results on NL2SVA-Human and NL2SVA-Machine benchmarks across pass@1, pass@5, and pass@10 metrics, outperforming specialized and general-purpose baselines.
reward-weighted on-policy distillationproperty-equivalence checkersystemverilog assertionsnl2sva benchmarkson-policy distillation
MARLIN: Multi-Agent Game-Theoretic Reinforcement Learning for Sustainable LLM Inference in Cloud Datacenters
We introduce MARLIN, a multi-agent game-theoretic reinforcement learning framework for optimizing sustainable LLM inference in cloud datacenters. MARLIN co-optimizes time-to-first token (TTFT), carbon emissions, water usage, and energy costs by modeling inference management as a multi-agent system with game-theoretic interactions. Compared to state-of-the-art LLM inference frameworks, MARLIN achieves reductions of 18% in TTFT, 33% in carbon emissions, 43% in water usage, and 11% in energy costs, addressing the environmental impact of LLM inference requests which account for 90% of LLM lifecycle energy use.
llm inferencemulti-agent reinforcement learninggame-theoretic optimizationtime-to-first tokencloud datacenters
Path-independent Flow Matching for Multi-parameter Generative Dynamics
We introduce Path-independent Flow Matching (PiFM), a generalization of Flow Matching to multi-parameter domains that enforces path-independent transport between distributions. PiFM learns vector fields whose induced flows depend only on initial and target distributions, not on specific paths, while approximating Wasserstein barycenters under suitable assumptions. We propose a simulation-free objective for training PiFM by regressing onto multi-parameter conditional probability paths. Empirical results demonstrate PiFM's superiority over existing methods in interpolating path-independent trajectories and generating out-of-distribution samples on both synthetic and real-world datasets.
flow matchingpath-independent transportwasserstein barycentermulti-parameter domainsconditional probability paths
Effective Context in Transformers: An Analysis of Fragmentation and Tokenization
The paper analyzes how tokenization choices affect Transformer performance under fixed context windows, identifying two key phenomena. First, fragmentation (lossless recoding into smaller units) can intrinsically degrade prediction quality, explaining performance gaps in byte/character-level models like ByT5 and CANINE versus subword models. Second, greedy tokenization (BPE, WordPiece) enables shorter token windows to emulate longer source-context windows, with a provable loss guarantee based on tokenizer compression and context-spanning reliability. Theoretical results are derived through Markov source analysis and finite-context information theory.
fragmentationtokenizationcontext windowmarkov sourcessubword
OSDN: Improving Delta Rule with Provable Online Preconditioning in Linear Attention
Online Scaled DeltaNet (OSDN) improves the Delta Rule in linear attention models by introducing online preconditioning with hypergradient feedback, addressing feature-wise curvature neglect. The method employs a diagonal preconditioner algebraically equivalent to per-feature scaling of the write-side key, preserving hardware-friendly chunkwise parallelism. Theoretical analysis shows super-geometric convergence and token-local residual contraction. Adaptive Preconditioner Forgetting (APF) handles non-stationary contexts. Empirically, OSDN enhances in-context recall by 32% at 340M parameters and reduces recall residual ratio by 39% at 1.3B parameters, maintaining performance on general tasks like perplexity and LongBench.
linear attentiondelta ruleonline preconditioninghypergradient feedbackadaptive preconditioner forgetting
Twincher: Bijective Representation Learning for Robust Inversion of Continuous Systems
The paper introduces Twincher, a novel architecture for learning bijective representations that enable robust inversion of continuous forward processes. The method employs structured diffeomorphic transformations and adversarial training to align representations bijectively with parameters while remaining insensitive to perturbations. Empirical results demonstrate improved data efficiency and robustness over baseline inverse-modeling approaches, suggesting potential applications in robotics, vision, and physical AI. The authors provide a public API for training and inference.
bijective representationdiffeomorphic transformationsadversarial traininginverse problemsrobust inversion
A Unified Three-Stage Machine Learning Framework for Diabetes Detection, Subtype Discrimination, and Cognitive-Metabolic Hypothesis Testing
The authors propose a three-stage ML framework for diabetes analytics, addressing detection, subtype discrimination, and cognitive-metabolic associations. Stage 1 benchmarks five classifiers (SVM-RBF, Logistic Regression, Random Forest, etc.) on the NCSU Diabetes Dataset, achieving 0.825 ROC-AUC and 0.762 accuracy, with Glucose, BMI, and Age as key biomarkers. Stage 2 applies K-Means clustering (k=2, silhouette≈0.116) to identify subtype partitions. Stage 3 analyzes the Ohio Longitudinal Cognitive Dataset (n=373), revealing a significant glycaemic-cognitive association (ρ_s=0.208, p=5.29×10^-5). The framework demonstrates interpretable, reproducible diabetes analytics.
diabetes detectionsubtype clusteringcognitive-metabolic associationshap explainabilitysilhouette validation
Efficient Sensor Fusion for Gesture Recognition on Resource-Constrained Devices
A lightweight gesture recognition system is proposed for resource-constrained smart eyewear, leveraging efficient sensor fusion of low-resolution Time-of-Flight (ToF) and Infrared (IR) thermal sensors. The system employs a compact Convolutional Neural Network (CNN) with grouped-convolution architecture to fuse depth and thermal modalities on a microcontroller (MCU). Evaluated on a custom dataset of 7 static gestures via k-fold cross-validation, the fusion strategy achieves 92.3% accuracy and a macro F1-score of 0.93, outperforming single-sensor baselines. On-device benchmarks on STM32F4 and STM32H7 MCUs demonstrate millisecond-level inference latency, 6,343 parameters, and 50 mW total system power, confirming suitability for wearables.
gesture recognitionsensor fusiontime-of-flightconvolutional neural networkmicrocontroller
On the Limits of Latent Reuse in Diffusion Models
This work theoretically characterizes the reliability of latent space reuse in diffusion models under dataset distribution shifts. The authors analyze a source-target setting where both datasets are approximately low-dimensional but may occupy different subspaces. They identify that freezing and reusing a source latent space induces a target-domain score error governed by principal-angle misalignment between subspaces and target ambient noise amplified by the diffusion time scale. Additionally, they study mixed source-target training to determine the required shared latent dimension based on the relative geometry of the distributions. The results provide theoretical guidance on when latent reuse remains reliable versus when learning a shared representation becomes necessary.
diffusion modelslatent space reusedistribution shiftprincipal-angle misalignmentambient noise
Pretraining Language Models with Subword Regularization: An Empirical Study of BPE Dropout in Low-Resource NLP
This empirical study investigates the impact of applying BPE dropout during pretraining (not just fine-tuning) for low-resource NLP tasks. Using monolingual and bilingual BERT models across six languages, experiments on XNLI, PAWS-X, PAN-X, and MasakhaNER 2.0 show that stochastic tokenization during both pretraining and fine-tuning yields optimal performance, particularly when data is scarce. Analysis reveals that pretraining with BPE dropout improves exposure to morphologically aligned segmentations, though such alignments remain rare. The findings suggest that compositional representations benefit more from consistent stochastic tokenization across both phases.
bpe dropoutsubword regularizationlow-resource nlpmorphological alignmentstochastic tokenization
Rescaled Asynchronous SGD: Optimal Distributed Optimization under Data and System Heterogeneity
The paper introduces Rescaled Asynchronous SGD, a distributed optimization method that corrects the bias of vanilla ASGD toward frequency-weighted local objectives under data heterogeneity. By rescaling worker-specific stepsizes proportionally to their computation times, the method ensures each worker contributes equally to the global objective while maintaining ASGD's simplicity. Theoretical analysis proves convergence to stationary points of the correct objective, with time complexity matching the known lower bound. Experiments demonstrate competitive performance with state-of-the-art baselines.
asynchronous sgddistributed optimizationdata heterogeneitynon-convex optimizationgradient rescaling
TurboGR: An Accelerated Training System for Large-Scale Generative Recommendation
TurboGR introduces an Ascend-affinity training system for large-scale generative recommendation, addressing system-level bottlenecks on Ascend NPUs. The system employs three innovations: Ascend-affinity jagged acceleration with fusion operators and dynamic load balancing, distributed communication optimization via hierarchical sparse parallelism and semi-asynchronous training, and negative sampling optimization through asynchronous offloading and jaggedness-aware FP16 quantization. Evaluated on the KuaiRand-27K dataset, TurboGR supports training up to 0.2B parameters, achieving 54.71% MFU with near-linear scalability (0.97).
ascend npugenerative recommendationjagged accelerationhierarchical sparse parallelismnegative sampling optimization
Strategic PAC Learnability via Geometric Definability
The paper establishes strategic PAC learnability conditions for hypothesis classes under feature manipulation by individuals. It demonstrates that strategic behavior can render even VC dimension 1 classes non-learnable under interval neighborhoods, then introduces geometric definability as a solution. By requiring both hypothesis class and cost-induced neighborhoods to be first-order definable over ℝ_exp (encompassing arithmetic, exponentiation, and comparisons), the work proves preserved learnability with sample complexity bounded by formula complexity. This framework applies to ℓ_p distances, Wasserstein metrics, and information-theoretic divergences.
strategic classificationpac learnabilityvc dimensiongeometric definabilityfirst-order formulas
LIFT: Last-Mile Fine-Tuning for Table Explicitation
The paper introduces Last-Mile Fine-Tuning (Lift), a pipeline combining pre-trained large language models with fine-tuned small language models (SLMs) for table extraction and error correction from unstructured text. Lift employs a pre-trained model for initial table extraction and a fine-tuned SLM (1B-24B parameters) for error repair, achieving competitive or superior performance to end-to-end SLM fine-tuning on the Tree-Edit-Distance-based Similarity (TEDS) metric. Evaluated on 2,596 tables from three datasets, Lift outperforms end-to-end fine-tuning by up to 0.144 TEDS points with as few as 1,000 training examples, demonstrating robustness to input format variability.
last-mile fine-tuningtree-edit-distance-based similaritysmall language modeltable extractionerror correction
DP-KFC: Data-Free Preconditioning for Privacy-Preserving Deep Learning
The paper introduces DP-KFC, a data-free preconditioning method for differentially private deep learning that avoids the geometric mismatch in DP-SGD. The method constructs KFAC preconditioners by probing neural networks with structured synthetic noise, leveraging the Fisher Information Matrix's decomposition into architectural sensitivity and input correlations. Empirical results show DP-KFC outperforms DP-SGD and adaptive baselines in strong privacy regimes (ε ≤ 3), matching private-data preconditioners while avoiding privacy budget consumption or distribution shift, particularly benefiting data-scarce domains like medical applications.
differential privacyfisher information matrixkfac preconditioningdp-sgddata-free learning
Vector-Quantized Discrete Latent Factors Meet Financial Priors: Dynamic Cross-Sectional Stock Ranking Prediction for Portfolio Construction
PRISM-VQ introduces a dynamic factor framework combining financial priors with vector-quantized discrete latent factors for cross-sectional stock ranking. The method employs vector quantization as an information bottleneck to capture market structure, using discrete codes as both latent factors and routing signals for a structure-conditioned Mixture-of-Experts. Experiments on CSI 300 and S&P 500 demonstrate improved return prediction accuracy and portfolio performance while maintaining interpretability.
vector quantizationcross-sectional rankingmixture-of-expertslatent factorsfinancial priors
When is Warmstarting Effective for Scaling Language Models?
The study investigates when warmstarting effectively scales language models, challenging two underexplored factors: overemphasis on preserving small-model performance at initialization and insufficient analysis of growth-hyperparameter interactions. Through empirical analysis of dense MLPs and language models, the authors demonstrate that simple, architecture-agnostic growth strategies outperform complex warmstarting operators and identify an upper bound on the growth factor $g$ beyond which training from scratch is more efficient. A $2\times$ growth factor proves most reliable for convergence speedups, particularly under 20 tokens/parameter budgets. Scaling laws derived from these observations offer practical guidance for model growth decisions.
warmstartingmodel growthscaling lawslanguage modelsconvergence speedups
Trajectory-Level Data Augmentation for Offline Reinforcement Learning
The paper proposes a trajectory-level data augmentation method for offline reinforcement learning, specifically targeting active positioning problems with limited suboptimal trajectories. The technique leverages task structure and geometric relationships between rewards, value functions, and logging policy properties to enhance data quality. Theoretical analysis supports the approach, which is empirically validated on positioning tasks across varying dimensions and partial observability conditions, demonstrating improved offline RL performance.
offline reinforcement learningdata augmentationtrajectory optimizationpartial observabilityvalue functions
The Diffusion Encoder
The authors propose a diffusion-based encoder that replaces the traditional variational autoencoder's encoder with a diffusion model, addressing the challenge of bidirectional training between encoder and decoder. They introduce an alternating training scheme inspired by expectation-maximization to synchronize updates while maintaining the simplicity of standard diffusion objectives. This approach preserves the expressive power of diffusion models while enabling more reliable latent representation learning compared to reparameterization-trick-based encoders.
diffusion encodervariational autoencoderreparameterization trickexpectation-maximizationlatent representation
Taming the Long Tail: Rebalancing Adversarial Training via Adaptive Perturbation
The paper proposes RobustLT, a framework for improving adversarial training on long-tail datasets by adaptively adjusting perturbations. The authors identify two key limitations in standard adversarial training for imbalanced data: skewed training objectives and unstable adversarial distribution evolution. Their theoretical analysis shows perturbations can simultaneously mitigate adversarial vulnerability and class imbalance. Experiments demonstrate RobustLT's effectiveness, consistently enhancing both robustness and class-balance across long-tailed benchmarks. The method is implemented as a plug-and-play module with publicly available code.
adversarial traininglong-tail learningadaptive perturbationsclass imbalancerobustness
Support-Conditioned Flow Matching Is Kernel Smoothing
The paper demonstrates that support-conditioned flow matching under Gaussian optimal-transport paths induces a velocity field equivalent to Nadaraya-Watson kernel smoothing, with bandwidth decreasing over flow time. This connects cross-attention conditioning in generative models to classical kernel theory through a single Gaussian-kernel attention head. Theoretical analysis predicts three failure regimes: nearest-neighbor collapse in high dimensions, geometric kernel-data mismatch, and insufficient support for nonparametric estimation. Experiments on Gaussian mixtures, spherical shells, and DINOv2 ImageNet features validate that learned conditioning improves in these regimes, showing IP-Adapter's cross-attention approximates NW smoothing in practice.
flow matchingnadaraya-watson kernelcross-attentionoptimal-transportnonparametric estimation
Teaching and Learning under Deductive Errors
The paper introduces a teaching and learning framework accounting for deductive errors in learners, such as humans and large language models in few-shot settings. It reformulates the Probably Approximately Correct (PAC) setting to require teachers to provide teaching sets that, with high probability, yield approximately correct hypotheses despite learner errors. The authors analyze six computational problems related to optimal PAC teaching sets, presenting XP algorithms parameterized by teaching set size, with tight runtime bounds under standard complexity assumptions. Experimental results compare teaching protocols against observed LLM behavior.
deductive errorspac learningmachine teachingxp algorithmsfew-shot learning
Beyond Oversquashing: Understanding Signal Propagation in GNNs Via Observables
The authors introduce a quantum mechanics-inspired framework for analyzing signal propagation in Graph Neural Networks (GNNs) using the concept of observables, addressing limitations of oversmoothing and oversquashing. They model signal localization, concentration, and propagation dynamics within graphs, demonstrating that standard spectral GNNs exhibit poor signal routing capabilities. A novel architecture, Schrödinger GNN, is proposed, which exhibits superior signal propagation efficiency across graph regions compared to existing methods.
graph neural networkssignal propagationobservablesoversmoothingspectral gnn
Phasor Memory Networks: Stable Backpropagation Through Time for Scalable Explicit Memory
The Phasor Memory Network (PMNet) introduces a novel architecture that stabilizes backpropagation through time in explicit memory systems via Unitary Phasor Dynamics and Hierarchical Learnable Anchors. By constraining recurrent state updates to phase rotations on a complex unit circle, PMNet prevents gradient divergence without specialized initialization. In experiments, a 119M-parameter PMNet with an 85-slot hierarchical memory tree achieved near-perfect retrieval on synthetic Copy-Paste tasks and matched the zero-shot long-context performance of a 3× larger Mamba model, demonstrating scalable sequence modeling capabilities.
phasor memory networkunitary phasor dynamicshierarchical learnable anchorsbackpropagation through timeexplicit memory architectures
Neural Surrogate Forward Modelling For Electrocardiology Without Explicit Intracellular Conductivity Tensor
This study introduces a neural surrogate model for cardiac electrophysiology that eliminates the need for explicit intracellular conductivity tensors in forward modeling. The deep learning approach directly maps left atrial intracellular potentials to far-field ECGs, bypassing conventional physics-based requirements. Trained on 74 subjects, the model achieved an R2 of 0.949 ± 0.037, demonstrating potential to reduce structural uncertainty in non-invasive atrial fibrillation assessment.
forward modellingintracellular conductivityneural surrogateatrial fibrillationelectrocardiology
Building Interactive Real-Time Agents with Asynchronous I/O and Speculative Tool Calling
The paper introduces Asynchronous I/O and Speculative Tool Calling to enable real-time interaction in agentic AI systems, addressing latency challenges in multi-turn tool calling. The method decouples agent reasoning from external delays and optimizes task execution under uncertainty. Implemented with cloud APIs, it achieves 1.3-1.7× speedups with minimal accuracy loss. For edge models like Qwen2.5-3B-Instruct and Llama-3.2-3B-Instruct, clock-based training and synthetic data generation yield 1.6-2.2× speedups across benchmarks.
asynchronous i/ospeculative tool callingreal-time latencyclock-based trainingsynthetic data generation
GeoFlowVLM: Geometry-Aware Joint Uncertainty for Frozen Vision-Language Embedding
(No summary returned.)
Contextual Bandits for Resource-Constrained Devices using Probabilistic Learning
The paper introduces probabilistic HD-CB, a low-precision variant of hyperdimensional contextual bandits (HD-CB) for resource-constrained devices. The method replaces deterministic accumulation with probabilistic updates: a random subset of vector components is updated per step using time-decaying probabilities, with values constrained to [-k,+k]. This enables low-precision components, prevents overflow, and reduces update costs. Off-policy evaluation on Open Bandit Pipeline benchmarks shows probabilistic HD-CB outperforms binarized HD-CB at equal precision and approaches HD-CB performance with just 3 bits per component.
contextual banditshyperdimensional computinglow-precisionprobabilistic updatesresource-constrained
Hierarchical Transformer Preconditioning for Interactive Physics Simulation
The Hierarchical Transformer Preconditioner introduces a neural preconditioner for real-time physics simulation, leveraging a weak-admissibility H-matrix partition to enable O(N) scaling in full-graph approximate-inverse computation. The network models the inverse through low-rank far-field factors and employs highway connections for context propagation across transformer depth. Training utilizes a cosine-Hutchinson probe objective to optimize angular alignment of MAz with z, improving conditioning on irregular spectra. Inference and preconditioner application are captured as a single CUDA Graph, achieving dense, dependency-free tensor programs. On stiff multiphase Poisson systems (N = 1,024-16,384), the solver achieves 17.9 ms/frame at N = 8,192, outperforming GPU Jacobi by 2.2x and GPU IC/DILU by ~28x.
h-matrix partitioncosine-hutchinson probeneural preconditionerlow-rank factorscuda graph
Shortcut Mitigation via Spurious-Positive Samples
The authors propose a method for mitigating shortcut learning in neural networks without requiring annotated training data, group-balanced held-out data, or complete group coverage. The approach identifies instances where models rely on spurious attributes through targeted analysis, pinpoints relevant intermediate-layer neurons, and regularizes their impact to discourage dependence on uninformative features. This regularization ensures models learn robust, informative features rather than exploiting shortcuts, improving generalization without additional data requirements. The method demonstrates effectiveness in reducing spurious correlations and enhancing model robustness.
shortcut learningspurious attributesneural regularizationintermediate-layer neuronstargeted model analysis
Context-Aware Web Attack Detection in Open-Source SIEM Systems via MITRE ATT&CK-Enriched Behavioral Profiling
The paper introduces Smart-SIEM, an AI module for Wazuh SIEM that enhances web attack detection through behavioral profiling. It proposes (1) context vectors encoding HTTP response patterns, rule activations, and MITRE ATT&CK technique frequencies per source IP, and (2) a hybrid LightGBM-XGBoost cascade for binary and multi-class attack classification. Evaluation on 46,454 Wazuh events shows context features boost macro F1 by +0.254 (binary) and +0.324 (multi-class), achieving 0.967 and 0.914 F1 respectively. The module detects 100% of Brute Force and 98.3% of Broken Authentication attacks, outperforming Wazuh's native engine (0%). Self-adaptive retraining recovers F1 from 0.465 to 0.814 under concept drift.
siemmitre att&ckbehavioral profilinglightgbmconcept drift
KamonBench: A Grammar-Based Dataset for Evaluating Compositional Factor Recovery in Vision-Language Models
We introduce KamonBench, a grammar-based image-to-structure benchmark for evaluating compositional factor recovery in vision-language models. The dataset contains 20,000 synthetic kamon crests, each paired with a formal kamon description language, segmented Japanese analysis, English translation, and non-linguistic program code. KamonBench enables evaluation beyond caption-level accuracy through direct program-code factor metrics, controlled factor-pair recombination splits, counterfactual motif-sensitivity groups, and linear probes of factor accessibility. Baseline results are provided for a ViT encoder/Transformer decoder and two VGG n-gram decoders, with and without learned positional masks. This benchmark offers a controlled testbed for sparse compositional visual recognition and factor recovery.
grammar-basedimage-to-structurefactor recoveryvision-language modelslinear probes
Embodied Neurocomputation: A Framework for Interfacing Biological Neural Cultures with Scaled Task-Driven Validation
The paper introduces Embodied Neurocomputation, a framework optimizing encoding/decoding mechanisms between biological neural networks (BNNs) and silicon computing interfaces. The authors operationalize this through large-scale parameter optimization for a BNN agent performing closed-loop navigation in a simulated grid-world environment. Evaluating ~1,300 parameter combinations across 4,000 hours of agent-environment interactions, they identified 12 configurations demonstrating consistent learning. These configurations outperformed optimized silicon-based Deep Q-Network (DQN) agents under identical interaction budgets. The framework supports scalable goal-oriented learning with BNNs and lays groundwork for hybrid bio-silicon architectures in robotic control applications.
biological neural networksneurocomputationparameter optimizationclosed-loop navigationhybrid bio-silicon architectures
Supervised Deep Multimodal Matrix Factorization for Interpretable Brain Network Analysis
The paper introduces Supervised Deep Multimodal Matrix Factorization (SD3MF), extending Symmetric Nonnegative Matrix Tri-Factorization to supervised multimodal graph analysis. SD3MF employs deep hierarchical factorizations per modality with shared latent representations, optimized via encoder-decoder architecture for joint reconstruction and prediction. Adaptive weighting enables data-driven fusion, yielding interpretable community-level features. Evaluated on connectome datasets, SD3MF outperforms CNNs and GNNs while providing biological interpretability.
matrix factorizationmultimodal fusionbrain networksinterpretable featuresencoder-decoder
MPINeuralODE: Multiple-Initial-Condition Physics-Informed Neural ODEs for Globally Consistent Dynamical System Learning
MPINeuralODE introduces a physics-informed neural ODE framework combining soft physics residuals with a multiple-initial-condition curriculum to improve generalization across unseen initial conditions and long horizons. The method structurally complements physics anchoring and trajectory diversity via multiple-shooting, evaluated on out-of-sample error, long-horizon stability, and Hamiltonian drift. On Lotka-Volterra dynamics, it reduces out-of-sample MSE by 26% versus baseline Neural ODEs while maintaining competitive Hamiltonian drift performance versus physics-informed neural networks (PINNs).
neural odephysics-informedmultiple-shootinglotka-volterrahamiltonian drift
Safe Bayesian Optimization for Uncertain Correlations Matrices in Linear Models of Co-Regionalization
The paper extends safety guarantees in multi-task Bayesian optimization to linear models of co-regionalization, enabling more flexible inter-task correlation modeling through feature composition. It derives uniform error bounds for vector-valued functions sampled from Gaussian processes with linear co-regionalization kernels. A numerical comparison on a safe multi-task Bayesian optimization benchmark demonstrates performance improvements using this approach.
bayesian optimizationco-regionalizationgaussian processerror boundsmulti-task learning
PaMM: Periodic Motif Memory for Atomistic Models with an Explicit Local-Structure Interface
PaMM introduces periodic motif memory to enhance atomistic modeling by explicitly encoding local coordination motifs in periodic crystals. The method augments the UMA eSCN-MD edge encoder with pair and triplet lookup features, hashed into fixed-size tables and fused through gate-only and affine-equipped variants. Evaluated in a UMA-S OMAT setting, PaMM variants outperform the baseline at 10k and 20k training steps, with gate-only achieving the best energy MAE and affine-equipped the best force MAE. Gains are attributed to structured pair/triplet organization rather than generic capacity increases, as shown by weaker performance in pair-only, triplet-only, and random-bucket alternatives. PaMM demonstrates consistent improvements across held-out generation families, affirming its utility as an inductive bias in periodic atomistic modeling.
periodic motif memoryatomistic modelingedge encoderinductive biaslocal coordination motifs
Learning Perturbations to Extrapolate Your LLM
We propose a framework for enhancing large language model extrapolation via learnable perturbations of token prefixes, implemented as continuous latent vector transformations in embedding space. To address intractable marginal likelihood, we derive unbiased estimating equations for model parameters and optimize them using stochastic gradient descent, establishing statistical properties in over-parameterized regimes. Empirical evaluations on synthetic and real-world datasets demonstrate significant out-of-domain performance improvements compared to state-of-the-art baselines.
extrapolationtoken prefixeslatent vectorestimating equationsover-parameterized
Byzantine-Robust Distributed Sparse Learning Revisited
The authors propose a Byzantine-robust framework for distributed sparse learning in high-dimensional linear models, combining local ℓ₁-regularized robust estimation with robust server-side aggregation. The method extends to pseudo-Huber regression, quantile regression, and sparse SVM, achieving near-optimal statistical rates under mild conditions while maintaining communication efficiency. Theoretical analysis provides non-asymptotic guarantees, and empirical simulations demonstrate strong robustness in estimation, support recovery, and classification accuracy across various Byzantine attack scenarios.
byzantine robustnesssparse learningℓ₁-regularizationpseudo-huber regressionquantile regression
Proximal-Based Generative Modeling for Bayesian Inverse Problems
The authors propose proximal-based generative modeling (PGM), a novel framework addressing limitations of score-based diffusion models in Bayesian inverse problems. PGM leverages a theoretical equivalence between Gaussian convolution in diffusion processes and Moreau-Yosida regularization, enabling sampling via a closed-form Moreau score derived from proximal operators. The framework introduces Moreau score matching to learn proximal operators from prior samples, eliminating early-stopping bias and achieving non-asymptotic convergence. Experiments demonstrate PGM's superior performance in reconstruction quality and sampling time compared to state-of-the-art methods.
proximal operatorsmoreau-yosida regularizationscore-based diffusionbayesian inverse problemsnon-asymptotic convergence
Physics Guided Generative Optimization for Trotter Suzuki Decomposition
A physics-guided generative optimization framework is proposed for Trotter-Suzuki decomposition in quantum simulation, addressing term grouping, product formula order, and timestep allocation. The method employs a conditional diffusion model for strategy generation, a physics-informed neural network (PINN) for fidelity feedback, and a graph neural network (GNN) for commutator structure encoding. Training operates in a hybrid discrete-continuous space using REINFORCE and Pareto tracking. On the transverse field Ising model (TFIM), the approach achieves 85.6% fidelity of a fourth-order Qiskit baseline at 21.8% circuit depth and 19.2% CNOT count, with fine-tuning reaching 0.9994 fidelity under equal depth constraints. Module contributions vary with training recipe and guidance hyperparameters, particularly CFG.
trotter-suzuki decompositionphysics-informed neural networkgraph neural networkreinforcetransverse field ising model
LightSplit: Practical Privacy-Preserving Split Learning via Orthogonal Projections
LightSplit introduces a privacy-preserving split learning framework that reduces communication overhead and mitigates reconstruction attacks via orthogonal random projections at the cut layer. The method applies a fixed, non-invertible projection to activations, limiting instance-specific information while maintaining end-to-end differentiability and compatibility with existing architectures. Evaluated on state-of-the-art benchmarks in IID and non-IID settings, LightSplit achieves over 95% baseline accuracy with up to 32x dimensionality reduction, ensuring stable training dynamics without additional client-side trainable components.
split learningorthogonal projectioninformation bottleneckreconstruction attackscommunication overhead
Chem-GMNet: A Sphere-Native Geometric Transformer for Molecular Property Prediction
The paper introduces Chem-GMNet, a sphere-native geometric transformer for molecular property prediction, challenging SMILES-based approaches with domain-specific architectural innovations. The model features SH-Embedding (spherical harmonic token representations), DualSKA (a hybrid attention mechanism combining Sphere-Flow recurrence and sphere-kernel softmax), and SH-FFN (spherical feature transformation). Evaluated on MoleculeNet benchmarks, Chem-GMNet outperforms ChemBERTa-2 baselines in 7 of 10 endpoints with 35% fewer parameters and matches or exceeds pretrained performance on 6 of 8 endpoints, achieving an ESOL RMSE of 0.938 without pretraining.
sphere-nativegeometric transformermolecule property predictionspherical harmonicsdual attention
Unified generalization analysis for physics informed neural networks
The paper presents a unified generalization analysis for Physics-Informed Neural Networks (PINNs) and Variational PINNs (VPINNs), overcoming limitations of prior work that required restrictive assumptions. By applying Taylor expansion to represent nonlinear differential operators as linear operators in high-dimensional space, the authors enable Koopman-based analysis. Results show high-rank networks generalize well with differential operators, while nonlinearity in these operators exponentially increases generalization bounds. The framework provides theoretical grounding for PINN/VPINN performance across scientific computing applications.
physics-informed neural networksgeneralization boundsdifferential operatorskoopman analysistaylor expansion
The Sample Complexity of Multiple Change Point Identification under Bandit Feedback
The paper introduces an adaptive algorithm for multiple change point localization under bandit feedback, where a piecewise-constant function is queried sequentially with noisy evaluations. The method first detects intervals containing change points, then refines their locations to a target precision η. Non-asymptotic upper and lower bounds on sample complexity are established, revealing that both jump magnitudes and relative change point positions govern complexity for general δ and η, contrary to prior asymptotic results focusing solely on jumps.
change point localizationbandit feedbacksample complexityadaptive algorithmnon-asymptotic bounds
EMO: Frustratingly Easy Progressive Training of Extendable MoE
The paper introduces EMO, a progressive training framework for sparse Mixture-of-Experts (MoE) models that dynamically expands the expert pool during training to address the MoE efficiency paradox. EMO models sparsity in scaling laws to compute stage-wise optimal token budgets for expert expansion, reducing memory and communication costs. Empirical results demonstrate that EMO matches fixed-expert performance while improving wall-clock efficiency and reducing GPU costs, offering a scalable solution for MoE training.
mixture-of-expertsprogressive trainingsparsity scalingwall-clock efficiencygpu cost reduction
When and Why is Optimistic Multiplicative Weights Slow? The Geometry of Energy Dissipation
The paper provides a geometric analysis of the Optimistic Multiplicative Weights Update (OMWU) algorithm's convergence in two-player zero-sum games, explaining when and why slow convergence occurs. By modeling dual iterates as optimistic skew-gradient descent with respect to an energy function, the authors establish tight bounds on energy dissipation and identify geometric bottlenecks near the simplex boundary. They derive a linear last-iterate convergence rate in KL divergence for games with a unique interior Nash equilibrium, proving optimal dependence on game-specific constants. Additionally, they show separations in uniform convergence rates, improving the best-iterate rate in duality gap to ${\widetilde O}(T^{-1/2})$ for $2\times 2$ games.
optimistic multiplicative weights updateenergy dissipationkl divergencenash equilibriumduality gap
Mix, Don't Tune: Bilingual Pre-Training Outperforms Hyperparameter Search in Data-Constrained Settings
This work demonstrates that bilingual pre-training with auxiliary high-resource language data outperforms hyperparameter tuning in data-constrained language model training. Through systematic evaluation across four model scales (150M to 1.43B parameters) using Arabic as the low-resource target and English as the auxiliary, approximately 1000 pre-training runs reveal three key findings. First, data mixing yields greater improvements than hyperparameter tuning on both validation loss and downstream task accuracy, with gains increasing with model size. Second, mixing boosts performance by 2--3× on validation loss and 2--13× on downstream accuracy compared to unique target data. Third, validation loss underestimates mixing's benefits by capturing only regularization effects, not knowledge transfer.
bilingual pre-traininghyperparameter tuningdata mixingvalidation lossknowledge transfer
Machine Learning-Driven Multimodal Spectroscopic Liquid Biopsy for Early Multicancer Detection
The study introduces a multimodal spectroscopic liquid biopsy framework combining Fourier Transform Infrared spectroscopy, Raman spectroscopy, and Excitation-Emission Matrix fluorescence with XGBoost for multicancer detection. Low-level data fusion integrates complementary biochemical information from serum samples of breast cancer, colorectal cancer patients, and healthy controls. The full multimodal approach achieved ROC-AUC scores of 0.997 (breast) and 0.994 (colorectal), outperforming unimodal and bimodal configurations in balanced sensitivity and specificity.
spectroscopic liquid biopsymultimodal fusionxgbootlow-level data fusionroc-auc
GAGPO: Generalized Advantage Grouped Policy Optimization
The paper introduces Generalized Advantage Grouped Policy Optimization (GAGPO), a critic-free reinforcement learning method for precise temporal credit assignment in multi-turn environments. GAGPO constructs a non-parametric grouped value proxy from rollouts to compute TD/GAE-style advantages, recursively propagating outcome supervision backward through time. Experiments on ALFWorld and WebShop demonstrate superior performance over baselines, with faster early-stage learning, improved interaction efficiency, and smoother optimization dynamics.
credit assignmentgeneralized advantage estimationmulti-turn reinforcement learningnon-parametric value proxypolicy optimization
Backdoor Channels Hidden in Latent Space: Cryptographic Undetectability in Modern Neural Networks
We demonstrate that cryptographic backdoors can be embedded in modern neural networks without detectability by exploiting latent space geometry, reframing undetectability as a hypothesis test between unknown parameter distributions. The approach identifies backdoor channels as learned latent directions in ResNet and Vision Transformer architectures trained on standard image classification datasets, requiring no exotic structures. Experiments show consistently high attack success rates with negligible clean accuracy degradation, while resisting post-training defenses that would otherwise render models unusable. This establishes that cryptographic backdoors are latent properties inherent to learned representations rather than artificial constructions.
cryptographic backdoorslatent spacehypothesis testresnetvision transformer
Switching Successor Measures for Hierarchical Zero-shot Reinforcement Learning
The paper introduces switching successor measures, a novel extension of successor measures enabling hierarchical zero-shot reinforcement learning without fixed temporal abstractions or manual subgoal design. The proposed FB π-Switch algorithm extracts both high-level subgoal-selection and low-level control policies from forward-backward representations, allowing hierarchical behavior to emerge from a single learned representation. Experiments demonstrate that FB π-Switch outperforms non-hierarchical baselines and matches state-of-the-art hierarchical methods in goal-conditioned tasks, while also generalizing to reward-based tasks. These results highlight the flexibility of structured successor representations for hierarchical RL beyond goal-reaching scenarios.
successor measureshierarchical reinforcement learningzero-shot learningforward-backward representationssubgoal-selection
A Hybrid Tucker-LSTM Tensor Network Model for SOC Prediction in Electric Vehicles
The paper introduces a hybrid Tucker-LSTM tensor network model for state-of-charge (SOC) prediction in electric vehicles, addressing cumulative errors and simplified battery models in conventional estimators. The method combines Tucker tensor decomposition with LSTM networks, processing charge status, mileage, voltage, current, cell differentials, and temporal features while preserving temporal structure through dimensionality reduction. Results show significant improvements: 70.5% MSE reduction (21.07 to 6.22), 48.7% MAE improvement (3.37% to 1.73%), RMSE decrease from 4.59% to 2.49%, and R² increase from 0.918 to 0.976, demonstrating tensor decomposition's efficacy in battery data compression without predictive loss.
tucker decompositionlstmstate-of-chargebattery managementtensor network
LLMs as Implicit Imputers: Uncertainty Should Scale with Missing Information
The paper proposes evaluating LLMs as implicit imputers under incomplete context, arguing their uncertainty should scale with missing information per multiple imputation criteria. Using SQuAD with five controlled context availability levels, it compares sampling-based confidence and response entropy as uncertainty measures. Results show entropy increases with missing context (explaining up to 0.057 more variance in accuracy than confidence) while confidence remains inflated, and introduces a diagnostic ρ_R(α) to quantify context resolution of uncertainty.
large language modelsmultiple imputationuncertainty quantificationresponse entropyblack-box diagnostic
Do Heavy Tails Help Diffusion? On the Subtle Trade-off Between Initialization and Training
This paper challenges the growing trend of using heavy-tailed (HT) noise in diffusion models by demonstrating its adverse effects on statistical estimation. Through theoretical analysis and empirical validation, the authors establish sampling-error bounds for diffusion models driven by both HT and light-tailed (LT) Gaussian noise, showing that HT noise leads to less favorable error bounds. Experiments on synthetic and real-world datasets confirm the predicted trade-off, suggesting that HT noise complicates the estimation problem without improving rare-region exploration as previously hypothesized.
heavy-tailed noisediffusion modelssampling-error boundsstatistical estimationrare-region exploration
Coupling-Informed Transport Maps for Bayesian Filtering in Nonlinear Dynamical Systems
The paper proposes a likelihood-free transport filtering method for Bayesian inference in nonlinear dynamical systems, leveraging couplings between state and observation variables. The method reformulates the filtering analysis step as maximum mean discrepancy (MMD) minimization between true and transport-approximated joint measures, using a block-triangular transport map structure. To address MMD optimization non-convexity, the authors introduce a training-free gradient flow approach yielding analytic transport map computations. Theoretical convergence guarantees are provided for MMD between approximated and true posteriors, with extensions to high dimensions via domain localization. Experiments demonstrate superior performance over conventional filters in non-Gaussian settings while avoiding particle collapse.
transport mapsbayesian filteringmaximum mean discrepancygradient flowsnonlinear dynamics
Finding the Weakest Link: Adversarial Attack against Multi-Agent Communications
The paper proposes gradient-based methods for adversarial attacks against multi-agent communication systems, identifying vulnerable messages, agents, and timesteps via Jacobian analysis. Two novel adversarial loss functions are introduced to balance attack success and impact. Evaluated on Multi-Agent Reinforcement Learning systems in navigation, PredatorPrey, and TrafficJunction environments, the methods outperform random message selection in most scenarios, demonstrating improved attack effectiveness in 50% of tested cases.
multi-agent reinforcement learningadversarial attackjacobian analysiscommunication perturbationgradient-based optimization
LoREnc: Low-Rank Encryption for Securing Foundation Models and LoRA Adapters
LoREnc introduces a training-free framework for securing foundation models (FMs) and LoRA adapters against intellectual property leakage and model recovery attacks. The method combines spectral truncation of dominant low-rank FM components, compensation in authorized adapters, and orthogonal reparameterization to obscure adapter fingerprints. Unauthorized access yields collapsed outputs, while authorized users maintain exact performance. Experiments show LoREnc resists model recovery with <1% computational overhead, without requiring retraining or original dataset access.
low-rank adaptationspectral truncationorthogonal reparameterizationmodel recoveryintellectual property protection
Continual Fine-Tuning of Large Language Models via Program Memory
The paper introduces ProCL, a continual fine-tuning framework for Large Language Models (LLMs) that enhances Low-Rank Adaptation (LoRA) with structured program memory. Inspired by Complementary Learning Systems, ProCL organizes LoRA adapters into memory slots dynamically retrieved via input-conditioned attention, enabling localized adaptation and knowledge retention. The method operates within LoRA's parameterization without inference overhead. Experiments show ProCL reduces catastrophic forgetting and improves retention across diverse benchmarks compared to existing continual LoRA approaches.
parameter-efficient fine-tuninglow-rank adaptationcontinual learningprogram memorycatastrophic forgetting
Kernel-based guarantees for nonlinear parametric models in Bayesian optimization
The paper develops kernel-based theoretical guarantees for nonlinear parametric models in Bayesian optimization, addressing a gap between theory and practice where existing analyses focus on Gaussian processes or linear models. The method employs kernels over parameter spaces to induce reproducing kernel Hilbert space structures, enabling confidence bounds for models trained with regularized convex losses. Results demonstrate convergence guarantees for nonlinear acquisition and surrogate models, including randomized regularized policies, unifying the analysis of nonlinear models in adaptive optimization settings.
bayesian optimizationreproducing kernel hilbert spacenonlinear parametric modelsadaptive samplingconvergence guarantees
A$_3$B$_2$: Adaptive Asymmetric Adapter for Alleviating Branch Bias in Vision-Language Image Classification with Few-Shot Learning
The paper introduces A$_3$B$_2$, an Adaptive Asymmetric Adapter addressing Branch Bias in vision-language image classification, where conventional fine-tuning assumes uniform importance of image and text branches. The method employs Uncertainty-Aware Adapter Dampening (UAAD) to dynamically suppress image-branch adaptation under high prediction uncertainty, alongside a lightweight asymmetric architecture with Load Balancing Regularization. Evaluations on 11 datasets across three few-shot tasks show A$_3$B$_2$ outperforms 11 prompt- and adapter-based baselines consistently.
vision-language modelsfew-shot learningbranch biasuncertainty-aware adaptationasymmetric adapter
Generative Modeling of Approximately Periodic Time Series by a Posterior-Weighted Gaussian Process
The paper introduces a Gaussian Process (GP)-based generative model for approximately periodic time series, addressing limitations of existing periodic and non-periodic GP models. The proposed method employs a novel kernel to modulate the GP posterior, decoupling intra-repetition structure from inter-repetition variability via a two-stage construction. This approach maintains a consistent mean function across repetitions while accommodating smooth variations between them. Experimental validation on toy datasets demonstrates the model's ability to generate realistic synthetic trajectories with controlled variability.
gaussian processgenerative modelingperiodic time serieskernel designposterior modulation
Understanding Generalization through Decision Pattern Shift
The paper introduces Decision Pattern Shift (DPS), a novel framework for understanding generalization failure in deep neural networks through internal decision mechanism analysis. Representing decision patterns as GradCAM-based channel-contribution vectors, the authors propose DPS to quantify deviation from class-average patterns. Results demonstrate: (i) structured, class-consistent decision spaces with high intra-class cohesion (Pearson r > 0.8), (ii) linear correlation between DPS magnitude and generalization gap, and (iii) unified organization of degradation scenarios into a continuous trajectory, enabling failure-mode diagnosis and defect localization.
decision pattern shiftgeneralization gapgradcamchannel-contribution vectorfailure-mode diagnosis
On Hallucinations in Inverse Problems: Fundamental Limits and Provable Assessment Methods
The paper establishes a theoretical framework demonstrating that hallucinations in inverse problems are inherent to ill-posedness, not just model artifacts. It derives necessary/sufficient conditions for hallucinations and computable bounds on their magnitude based solely on the forward model. The authors propose algorithms to (1) estimate minimum hallucination magnitude for any reconstruction model and (2) assess detail faithfulness in reconstructions. Experiments across three imaging tasks validate the approach's broad applicability, including to modern generative models, providing principled quantification of AI hallucinations.
inverse problemshallucination boundsill-posednessforward modelgenerative models
Collaborating in Multi-Armed Bandits with Strategic Agents
The paper introduces \texttt{CAOS}, a mechanism for sustaining collaboration among strategic agents in multi-agent Bayesian bandit problems. Unlike prior work assuming short-lived agents, the model considers persistent agents participating across multiple time periods, with incentives provided solely through information sharing. The proposed mechanism achieves Nash equilibrium while maintaining strong regret guarantees, demonstrating near-optimal performance comparable to fully cooperative systems despite strategic behavior.
multi-armed banditsstrategic agentsnash equilibriumregret guaranteesinformation sharing
On the Generalization of Knowledge Distillation: An Information-Theoretic View
The paper provides an information-theoretic framework for analyzing knowledge distillation's generalization properties, introducing a distillation divergence metric based on Kullback-Leibler divergence between teacher and student training processes. Through coupled stochastic process modeling, it derives both upper and lower generalization bounds for the student relative to the teacher, with the upper bound relying on algorithmic stability under sub-Gaussian assumptions and the lower bound exhibiting sharper dependence on distillation divergence under a central condition. Theoretical analysis reveals that teacher model flatness can tighten generalization bounds, while a linear Gaussian case study decomposes distillation divergence into interpretable bias, variance, and rank-bottleneck components.
knowledge distillationgeneralization boundskullback-leibler divergencealgorithmic stabilitystochastic processes
Code-Centric Detection of Vulnerability-Fixing Commits: A Unified Benchmark and Empirical Study
The study presents a unified benchmark for vulnerability-fixing commit (VFC) detection, evaluating code language models across 20+ datasets with 180K+ commits. Through 180+ experiments with models ranging from 125M to 14B parameters, it finds no transferable security understanding from code changes alone; commit messages dominate model attention when available. Performance drops by 17% in group-stratified evaluation, and code-only models miss 93% of vulnerabilities at 0.5% FPR. The authors release their framework to advance code-centric VFC detection research.
vulnerability-fixing commitscode language modelstransfer learningfalse positive ratesemantic context
KAST-BAR: Knowledge-Anchored Semantically-Dynamic Topology Brain Autoregressive Modeling for Universal Neural Interpretation
KAST-BAR introduces a Knowledge-Anchored Semantically-Dynamic Topology Brain Autoregressive Model to bridge the modality gap between EEG signals and textual semantics. The model employs a Dual-Stream Hierarchical Attention encoder to capture non-Euclidean brain topology and a Knowledge-Anchored Semantic Profiler to generate instance-level textual profiles. These profiles drive a Semantic Text-Aware Refiner to dynamically reconstruct EEG representations using Latent Expert Queries. Pre-trained on 21 diverse datasets, KAST-BAR integrates expert medical knowledge into EEG signal representations, achieving superior performance across six downstream tasks.
eeg signalsnon-euclidean topologysemantic profilerlatent expert queriesautoregressive model
ERPPO: Entropy Regularization-based Proximal Policy Optimization
The paper introduces Entropy Regularization-based Proximal Policy Optimization (ERPPO), a novel method addressing policy optimization challenges in multi-agent reinforcement learning under non-stationary observations. ERPPO employs a Distributional Spatiotemporal Ambiguity (DSA) learner to estimate object detection uncertainty and enhances Proximal Policy Optimization (PPO) with dynamic entropy regularization—applying L1 regularization in high-ambiguity scenarios to encourage exploration and L2 regularization in low-ambiguity scenarios to stabilize updates. Evaluated in AirSim-based maritime search scenarios, ERPPO improves accuracy and gradient performance compared to Multi-Agent PPO (MAPPO), effectively reducing false detections in visually uncertain conditions.
entropy regularizationproximal policy optimizationmulti-agent reinforcement learningdistributional spatiotemporal ambiguityairsim
Amortized Neural Clustering of Time Series based on Statistical Features
The paper proposes an amortized neural inference framework for feature-based time series clustering that reduces reliance on traditional algorithms like K-means. The method learns data-driven affinity structures from statistical features (autocorrelations, quantile autocorrelations) without requiring explicit cluster shape specifications, with one variant automatically determining cluster count. Empirical results demonstrate competitive or superior accuracy versus conventional methods, even when competitors use the true cluster count. A financial time series application validates practical utility.
amortized inferencetime series clusteringstatistical featuresneural networksaffinity structure
State-of-art minibatches via novel DPP kernels: discretization, wavelets, and rough objectives
The paper introduces novel determinantal point processes (DPPs) for efficient minibatch and coreset construction in machine learning. It proposes wavelet-based DPPs on Euclidean space with improved accuracy guarantees over existing methods, and presents a general discretization technique that preserves variance reduction properties while enabling computationally efficient sampling via low-rank kernel decomposition. The approach extends DPP-based subsampling to objectives with arbitrary regularity, with guarantees adapting to this regularity. Results demonstrate superior theoretical rates and practical applicability across irregular optimization tasks.
determinantal point processeswavelet kernelsvariance reductioncoreset constructionlow-rank decomposition
DiffusionHijack: Supply-Chain PRNG Backdoor Attack on Diffusion Models and Quantum Random Number Defense
We introduce DiffusionHijack, a supply-chain backdoor attack targeting pseudo-random number generators (PRNGs) in diffusion models, enabling deterministic control over generated images without modifying model weights. The attack injects malicious PRNGs via compromised packages, achieving pixel-perfect reproduction of attacker-chosen content (SSIM = 1.00) across Stable Diffusion v1.4, v1.5, and SDXL, while bypassing CLIP-based safety checkers (98-100% success) and remaining undetectable by model auditing mechanisms. As a defense, we propose replacing PRNGs with quantum random number generators (QRNGs), which reduce output similarity to random baseline levels (SSIM < 0.20 for SD 1.x, < 0.45 for SDXL) across 100 prompt-model combinations. This work highlights a critical supply-chain vulnerability and offers a hardware-level mitigation for generative AI systems.
diffusion modelspseudo-random number generatorssupply-chain attackquantum random number generatorstable diffusion
Adaptive Kernel Density Estimation with Pre-training
The paper introduces pre-training to non-parametric density estimation, enabling efficient high-dimensional density estimation with location-adaptive kernels. A neural network is pre-trained to recommend appropriate kernels for each sample point, addressing the inefficiency of traditional kernel smoothing methods in high dimensions. Experiments demonstrate significant accuracy improvements when the target distribution aligns with the pre-training distribution family. When distributions diverge, fine-tuning reactivates the benefits of pre-training. This approach bridges pre-training techniques from AI to statistical density estimation.
density estimationkernel smoothingpre-traininglocation-adaptive kernelsfine-tuning
Bayesian Nonparametric Mixed-Effect ODEs with Gaussian Processes
The authors propose MEGPODE, a Bayesian nonparametric mixed-effect Ordinary Differential Equation (ODE) model that decomposes subject-specific vector fields into population-level and individual deviation components, both modeled with Gaussian process (GP) priors. The method combines state-space GP trajectory priors with virtual collocation observations to enable efficient training via Kalman-smoothing updates and closed-form regressions, avoiding repeated ODE solves. Evaluated on heterogeneous ODE benchmarks including oscillatory and biomedical systems, MEGPODE demonstrates improved population-field recovery and subject-level trajectory prediction compared to baseline methods.
bayesian nonparametricsmixed-effects modelsgaussian processesordinary differential equationsstate-space models
Local Inverse Geometry Can Be Amortized
The paper introduces Deceptron, a learned bidirectional surrogate for nonlinear inverse problems, deployed via D-IPG (Deceptron Inverse-Preconditioned Gradient) to amortize local inverse geometry. The method employs a Jacobian Composition Penalty (JCP) to train reverse Jacobians as local left inverses of forward Jacobians, with runtime RJCP measuring inverse-consistency. Theoretical analysis shows D-IPG's first-order equivalence to damped Gauss-Newton under pseudoinverse consistency. Experiments on seven PDE benchmarks demonstrate 94.8% mean success rate and up to 77x faster inference than baselines while maintaining comparable recovery quality.
nonlinear inverse problemsjacobian composition penaltyamortized optimizationpde benchmarksinverse-preconditioned gradient
Ergodic Trajectory Design by Learned Pushforward Maps: Provable Coverage via Conditional Flow Matching
The paper introduces a pushforward framework for ergodic trajectory design that decouples ergodicity from density matching. The method employs an analytic latent trajectory for uniform ergodicity on an annular domain, combined with a learned map via optimal-transport conditional flow matching to transport this occupancy onto a target density. The approach guarantees asymptotic ergodicity, with deviation controlled by flow-matching training loss, and supports multi-agent deployment without retraining. Theoretical results include an acceleration-energy bound, $O(1/\sqrt{K})$ ergodic convergence rate, and an approximation-error bound, providing certified coverage estimates via training diagnostics.
ergodic coverageconditional flow matchingpushforward distributionoptimal transporttrajectory optimization
Large Language Models Lack Temporal Awareness of Medical Knowledge
The authors introduce TempoMed-Bench, the first benchmark for evaluating temporal awareness of Large Language Models (LLMs) in medical knowledge, addressing the dynamic nature of medical guidelines. They assess LLMs' ability to recall time-specific knowledge, revealing three key findings: performance declines linearly over time (not sharply at cutoffs), historical knowledge recall is 25.37%-53.89% less accurate than current knowledge, and predictions fluctuate inconsistently across years. Integration with search tools improves performance only marginally (-3.15%-14.14%), highlighting the challenge of temporal awareness in medical LLMs.
temporal awarenesslarge language modelsmedical knowledgeknowledge cutoffparametric knowledge
What Information Matters? Graph Out-of-Distribution Detection via Tri-Component Information Decomposition
The paper proposes TIDE, a tri-component information decomposition framework for graph out-of-distribution (OOD) detection. TIDE explicitly decomposes node information into feature-specific, structure-specific, and joint components, preserving only label-relevant joint information while filtering spurious signals. Theoretical and empirical analyses demonstrate that an information bottleneck objective outperforms standard supervised learning for graph OOD detection, yielding higher in-distribution confidence and greater ID-OOD entropy gaps. Experiments on seven datasets show TIDE achieves up to 34% improvement in FPR95 over baselines while maintaining competitive in-distribution accuracy.
graph neural networksout-of-distribution detectioninformation decompositioninformation bottlenecknode classification
Offline Two-Player Zero-Sum Markov Games with KL Regularization
(No summary returned.)
A General Bézier Tree Encoding Counterfactual Framework for Retinal-Vessel-Mediated Disease Analysis
The Bézier Tree Encoding Counterfactual Framework (BTECF) introduces a disease-agnostic representation for retinal vessel analysis by modeling vascular networks as interconnected cubic-Bézier segments. This structural encoding enables parameter-level interventions on geometric features (e.g., tortuosity, caliber) via a diffusion-based generator while preserving fundus textures. Validated on diabetic retinopathy, ischemic stroke, and Alzheimer's disease cohorts, BTECF demonstrates dose-responsive classifier prediction shifts, with pixel-drop controls confirming causal isolation of vessel topology effects. The framework supports reproducible hypothesis testing across systemic diseases through explicit geometric counterfactuals.
bézier tree encodingcounterfactual frameworkdiffusion-based generatorvascular topologygeometric interventions
JEDI: Joint Embedding Diffusion World Model for Online Model-Based Reinforcement Learning
The paper introduces Joint Embedding DIffusion (JEDI), the first online end-to-end latent diffusion world model for model-based reinforcement learning. JEDI combines joint embedding predictive architecture (JEPA) with diffusion denoising to learn latent spaces directly from world-model objectives, avoiding separate pretraining. Theoretically, it shows JEPA induces a predictive information bottleneck, while diffusion denoising enables predictive-compression decomposition. Empirically, JEDI matches Atari100k performance, uses 43% less VRAM, achieves 3× faster sampling and 2.5× faster training than pixel diffusion baselines, and exhibits distinct task-level performance patterns.
diffusion world modelmodel-based reinforcement learningjoint embedding predictive architecturelatent diffusiononline learning
Implicit Behavioral Decoding from Next-Step Spike Forecasts at Population Scale
The study demonstrates that a single Mamba forecaster, trained solely on next-step spike counts at Neuropixels scale, simultaneously predicts neural population activity and decodes behavioral state. The method employs a lightweight per-session linear head on the model's predicted rates, outperforming raw spike count classifiers with matched temporal context. Evaluated on the Steinmetz visual-discrimination benchmark (39 sessions, ~27,000 neurons, 1,994 trials), Mamba achieves 75.7±0.2% accuracy for mouse choice decoding (2.3× chance) and 66.1±0.6% for stimulus side (2× chance), surpassing linear decoders by 4-6 percentage points. The pipeline requires only 100-150 calibration trials and fits within 50 ms latency on workstation GPUs.
mamba forecasterneuropixelsbehavioral decodingspike count predictionclosed-loop bci
\emph{DRIFT}: A Benchmark for Task-Free Continual Graph Learning with Continuous Distribution Shifts
The paper introduces DRIFT, a benchmark for task-free continual graph learning (CGL) that models continuous distribution shifts in graph streams. Unlike traditional task-based CGL approaches, DRIFT employs a time-varying mixture of latent task distributions to simulate realistic non-stationary environments, ranging from abrupt task switches to smooth drift. Evaluation of existing methods reveals significant performance drops (quantitative results not specified) compared to task-based protocols, suggesting implicit reliance on task boundaries. The work establishes a foundation for studying CGL under more realistic conditions.
continual graph learningtask-free learningdistribution shiftnon-stationary environmentsbenchmark
Frequency Bias and OOD Generalization in Neural Operators under a Variable-Coefficient Wave Equation
The study investigates frequency bias and out-of-distribution (OOD) generalization in neural operators for PDEs, focusing on the Fourier Neural Operator (FNO) and Deep Operator Network (DeepONet) under a variable-coefficient wave equation. By systematically varying input frequency and coefficient smoothness, the authors find FNO excels in smoothness shifts but suffers high error under high-frequency OOD inputs, while DeepONet shows milder degradation despite higher baseline error. The analysis reveals architectural representation biases as a key factor in OOD performance, highlighting a gap between in-distribution accuracy and generalization.
neural operatorsout-of-distribution generalizationfourier neural operatordeep operator networkvariable-coefficient pde
F-GRPO: Factorized Group-Relative Policy Optimization for Unified Candidate Generation and Ranking
The paper introduces F-GRPO, a unified framework for joint candidate generation and ranking in autoregressive LLMs, addressing the credit assignment problem in combinatorial output spaces. The method factorizes policy into generation and ranking phases while sharing an LLM backbone, using group-relative advantages and phase-specific rewards (order-invariant coverage and position-aware utility). Evaluations on sequential recommendation and multi-hop QA show F-GRPO outperforms GRPO, decoupled baselines, and supervised alternatives, matching zero-shot rerankers without inference-time modifications.
autoregressive generationcredit assignmentpolicy optimizationcandidate rankinggroup-relative advantage
DP-Muon: Differentially Private Optimization via Matrix-Orthogonalized Momentum
The paper introduces DP-Muon, a differentially private optimizer that combines matrix-valued momentum with Newton-Schulz orthogonalization. The method clips per-example matrix gradients, adds Gaussian noise, and applies momentum and orthogonalization as post-processing, maintaining privacy guarantees via subsampled Gaussian accounting. Theoretical analysis separates optimization error, clipping residual, privacy noise, and approximation error, while identifying a matrix-valued heat-smoothing bias induced by noise. A bias-corrected variant, DP-MuonBC, improves utility without additional privacy cost. Experiments on E2E and DART datasets demonstrate enhanced private fine-tuning performance.
differential privacymatrix optimizationmomentumnewton-schulz orthogonalizationbias correction
SpikeProphecy: A Large-Scale Benchmark for Autoregressive Neural Population Forecasting
The paper introduces SpikeProphecy, the first large-scale benchmark for autoregressive neural population forecasting, addressing limitations of aggregate evaluation metrics. It proposes a decomposed evaluation protocol separating temporal fidelity, spatial pattern accuracy, and magnitude-invariant alignment, applied to 105 Neuropixels sessions (~89,800 neurons) across seven architecture families (SSMs, Transformer, LSTM, spiking network). Results reveal a consistent brain-region predictability ranking (region ΔR² = 0.018) across models, a sub-Poisson evaluation floor, and negative findings on ANN-to-SNN distillation in Poisson count domains.
neural population modelsautoregressive forecastingspike-count predictionneuropixelsevaluation decomposition
Decision Tree Learning on Product Spaces
We extend the theoretical analysis of top-down greedy decision tree learning from uniform to arbitrary product distributions. Using a parameter-free algorithm based on the greedy heuristic, we prove that for any function computable by an optimal decision tree of size s, maximum depth D_opt, and average depth Δ_opt, the method constructs an ε-approximating tree with size bounded by exp(Δ_opt D_opt log(e/ε)). This improves upon Blanc et al.'s bound for full binary trees and applies to a broader distribution class, while eliminating the need for prior knowledge of the optimal tree's parameters.
decision treeproduct distributiongreedy heuristicparameter-freeε-approximation
U-HNO: A U-shaped Hybrid Neural Operator with Sparse-Point Adaptive Routing for Non-stationary PDE Dynamics
U-HNO introduces a U-shaped hybrid neural operator with Sparse-Point Adaptive Routing (SPAR) to address non-stationary PDE dynamics, combining global Fourier and local multi-scale Gaussian branches adaptively. SPAR employs a per-pixel hard mask to select the dominant branch based on local contrast, embedded in a hierarchical encoder-bottleneck-decoder backbone with skip connections. Training integrates pointwise supervision, H^1 gradient regularization, and spectral consistency. Evaluated on PDEBench benchmarks, U-HNO achieves state-of-the-art rollout accuracy in relative L^2 and H^1 metrics, particularly excelling in problems with sharp localized features. Ablations confirm the necessity of each component.
sparse-point adaptive routingneural operatorfourier branchmulti-scale gaussianpdebench
Reducing Bias and Variance: Generative Semantic Guidance and Bi-Layer Ensemble for Image Clustering
The paper introduces GSEC, a novel image clustering framework that simultaneously reduces bias and variance through generative semantic guidance and bi-layer ensemble learning. The method employs Multimodal Large Language Models to generate semantic descriptions and derive weighted image embeddings, while a two-layer ensemble strategy (BatchEnsemble for cross-modal integration and an alignment mechanism for output consistency) enhances robustness. Experiments show GSEC outperforms 18 state-of-the-art methods across six benchmark datasets, demonstrating effectiveness in bias-variance reduction.
image clusteringgenerative semantic guidancebi-layer ensemblemultimodal llmsbias-variance reduction
Coreset-Induced Conditional Velocity Flow Matching
The paper introduces Coreset-Induced Conditional Velocity Flow Matching (CCVFM), a generative model enhancing hierarchical rectified flow with a data-informed source distribution. CCVFM compresses target data into weighted atoms via entropic Sinkhorn coresets, lifting them to Gaussian mixtures for closed-form conditional velocity laws. A lightweight correction flow refines the surrogate-to-target residual, avoiding full noise-to-data mapping. Theoretical analysis shows the surrogate transport cost equals the Wasserstein gap under compression, with small excess when surrogate and true laws align in mean and covariance. Empirical results on MNIST, CIFAR-10, ImageNet-32, and CelebA-HQ demonstrate competitive few-step generation.
coresetflow matchingsinkhorngaussian mixturewasserstein
Separating Shortcut Transition from Cross-Family OOD Failure in a Minimal Model
The study introduces a minimal binary model to disentangle shortcut feature dynamics from out-of-distribution (OOD) failure. Using logistic ERM with ridge regularization, the model analyzes invariant and shortcut coordinates under deterministic and noisy regimes. Results show that ridge regularization maintains invariant-dominated classifiers unless noisy invariant signals are surpassed by shortcut signals, triggering rule transitions. OOD failure depends on held-out family shortcut correlation, yielding either positive excess risk or above-chance error for sign-flipped families. Synthetic experiments validate distinct regimes where training-side transitions produce varied test-time consequences.
shortcut featuresout-of-distribution failureridge regularizationlogistic erminvariant coordinate
From Instance Selection to Fixed-Pool Data Recipe Search for Supervised Fine-Tuning
AutoSelection introduces a novel approach to supervised fine-tuning (SFT) data selection by framing it as fixed-pool data recipe search, optimizing the curation process without generating new samples. The method employs a two-layer solver that decouples fixed-pool materialization from costly full evaluations, leveraging warmup probes, subset states, local recipe edits, Gaussian-process-assisted ranking, and stagnation-triggered reseeding. Experiments on a 90K instruction pool demonstrate that AutoSelection outperforms full-data training, random recipe search, random top-$k$, and single-operator selectors, achieving superior in-distribution reasoning averages across three base models. Results also highlight the importance of recipe structure beyond individual selection operators.
supervised fine-tuningfixed-poolgaussian-processrecipe searchin-distribution
Reinforced Collaboration in Multi-Agent Flow Networks
The paper introduces MANGO (Multi-Agent Network Gradient Optimization), a framework that optimizes multi-agent collaboration in LLM-based systems through flow networks and reinforcement learning. MANGO addresses error propagation by refining workflows via textual gradients and a skipping mechanism, improving both accuracy and efficiency. Evaluations across seven benchmarks demonstrate 12.8% performance gains over baselines, 47.4% efficiency improvements, and strong domain generalization.
multi-agent systemserror propagationflow networksreinforcement learningtextual gradients
The Efficiency Gap in Byte Modeling
The paper quantifies the computational efficiency gap between byte-level modeling paradigms, comparing autoregressive (AR) and masked diffusion modeling (MDM) approaches through compute-matched scaling experiments. Results show MDM incurs disproportionately higher scaling costs than AR in byte regimes, attributed to context fragility during parallel generation. Controlled permutation studies suggest MDM's loss of local contiguity hinders semantic resolution from raw bytes, indicating future modality-agnostic designs require alternative structural biases for efficient scaling.
byte modelingautoregressivemasked diffusion modelingcontext fragilitymodality-agnostic
IV-ICL: Bounding Causal Effects with Instrumental Variables via In-Context Learning
The paper introduces IV-ICL, an amortized Bayesian in-context learning method for bounding causal effects using instrumental variables (IVs). The approach directly learns the marginal posterior distribution of causal effects via inclusive KL divergence minimization, avoiding mode collapse seen in exclusive-KL variational inference. Empirical evaluation on synthetic and semi-synthetic IV benchmarks shows IV-ICL produces more valid and informative intervals than semi-parametric, Bayesian, and plug-in baselines, with 20-500x faster inference. The authors also propose a procedure to convert randomized controlled trials into IV benchmarks with preserved ground-truth effects.
instrumental variablesin-context learningcausal inferencepartial identificationamortized bayesian inference
Adaptive Conformal Prediction for Reliable and Explainable Medical Image Classification
The authors propose an Adaptive Lambda Criterion for Regularized Adaptive Prediction Sets (RAPS) to improve reliability in medical image classification by minimizing worst-case coverage violations across prediction set strata. The method modifies standard conformal prediction to maintain coverage guarantees while addressing overconfidence in ambiguous cases. Evaluated on OrganAMNIST (58,850 CT images, 11 classes), it achieves 95.72% global coverage with average set size 1.09 and ≥90% stratified coverage, outperforming baseline RAPS. Cross-validation on PathMNIST (107,180 pathology images, 9 classes) confirms generalizability. Grad-CAM analysis (ρ=-0.30, p<1e-22) reveals multi-label predictions correlate with anatomically ambiguous regions.
conformal predictionmedical imagingadaptive prediction setscoverage guaranteesgrad-cam
SHM-Agents: A Generalist-Specialist Integrated Agent System for Structural Health Monitoring
The paper introduces SHM-Agents, a hybrid agent system combining large language models' reasoning with specialized algorithms for structural health monitoring (SHM). The system employs a generalist-specialist architecture enabling natural language task execution, modular expansion, and simplified deployment via deep learning pre-training. Evaluated on a cable-stayed bridge, SHM-Agents demonstrated accurate performance across 11 SHM tasks including modal identification, damage detection, and reliability assessment.
structural health monitoringlarge language modelsmodal identificationmodular architecturefinite element model updating
📰 Industry Media (8)
Data readiness for agentic AI in financial services
Agentic AI deployment in financial services hinges on data readiness rather than model sophistication, requiring high-quality, secure, and accessible data to optimize complex workflows. The method emphasizes centralized data stores with auditable governance, combining structured and unstructured data processing to mitigate hallucinations and ensure deterministic outputs. Results from Gartner and Forrester indicate 50% adoption intent among financial teams, with 57% still developing capabilities, highlighting the critical role of search platforms in enabling accurate, scalable AI systems.
agentic aidata governanceunstructured datadeterministic outputssearch platforms
Establishing AI and data sovereignty in the age of autonomous systems
The article examines the growing enterprise movement toward AI and data sovereignty, driven by concerns over intellectual property loss and competitive positioning in cloud-based large language model deployments. Drawing on a survey of 2,050 senior executives and expert interviews, the research highlights that 70% of global executives believe sovereign data and AI platforms are essential for success. The study explores how companies are reclaiming control over their models and data estates, emphasizing the need for localized AI infrastructure development leveraging cultural and linguistic resources. NVIDIA CEO Jensen Huang advocates for national AI ecosystems as a strategic imperative.
sovereigntylarge language modelsintellectual propertydata estatesai infrastructure
The shock of seeing your body used in deepfake porn
The article examines the psychological and legal impacts of nonconsensual intimate imagery (NCII) deepfakes on adult content creators, focusing on cases where performers' bodies are used without consent. Through interviews with affected individuals and legal experts, it documents the 'embodied harms'—including body dysmorphia and financial losses—caused by AI-generated face-swaps and 'nudify' apps trained on pornographic datasets. Findings reveal systemic vulnerabilities: copyright enforcement struggles with unidentifiable AI-altered bodies, while current US laws lack protections for performers whose likenesses are synthesized without unique markers.
nonconsensual intimate imagerydeepfakesembodied harmsnudify appsdigital fingerprinting
Nous Research Releases Token Superposition Training to Speed Up LLM Pre-Training by Up to 2.5x Across 270M to 10B Parameter Models
Nous Research introduces Token Superposition Training (TST), a two-phase pre-training method that accelerates LLM training by up to 2.5x without architectural changes. Phase 1 averages token embeddings into 's-tokens' for higher throughput, using a multi-hot cross-entropy loss; Phase 2 reverts to standard next-token prediction. Evaluated on 270M to 10B-A1B MoE models, TST reduces GPU-hours (e.g., 4,768 vs. 12,311 at 10B scale) while matching or surpassing baseline performance on HellaSwag, ARC, and MMLU. Key mechanisms include input-side embedding regularization and output-side future-signal prediction.
token superposition trainingmulti-hot cross-entropyembedding regularizationfuture-signal predictioncompute-bound pre-training
How to Build a Dynamic Zero-Trust Network Simulation with Graph-Based Micro-Segmentation, Adaptive Policy Engine, and Insider Threat Detection
The tutorial presents a dynamic Zero-Trust network simulation leveraging graph-based micro-segmentation, an adaptive policy engine, and insider threat detection. It models network zones and assets as a directed graph, integrating ABAC-style permissions with device posture, MFA, and real-time risk signals. The implementation, operationalized via a Flask API, demonstrates blocking malicious flows through trust scoring, adaptive controls, and automated quarantines. Mixed traffic simulations, including lateral movement and exfiltration attempts, validate the system's efficacy in real-time threat mitigation.
zero-trustmicro-segmentationadaptive policy engineinsider threat detectionflask api
Enterprise AI Governance in 2026: Why the Tools Employees Use Are Ahead of the Policies That Cover Them
Enterprise AI governance frameworks lag behind employee adoption of generative AI tools, creating systemic risks. Surveys from IBM and Netskope reveal that 40-65% of enterprise employees use unauthorized AI tools, with 47% accessing them via personal accounts. Shadow AI drives productivity but exposes sensitive data, contributing to 20% of data breaches and increasing breach costs by $670,000 on average. Policies often lack enforcement, and bans lead to substitution rather than elimination of risks. The EU AI Act exacerbates compliance challenges, requiring inventories of AI systems, yet shadow AI remains unaccounted for in most organizations.
shadow aigenerative aidata breacheseu ai actgovernance frameworks
Physical AI moves closer to factory floors as companies test humanoid robots
Physical AI systems are advancing toward industrial deployment, with Humanoid robots scheduled for integration into Schaeffler’s global manufacturing sites by 2032, targeting tasks like box handling and material movement. RLWRLD is collecting worker motion data from logistics, hospitality, and retail settings using body cameras and motion-tracking gloves to train robot systems for dexterity-focused tasks. Hyundai Motor and Samsung Electronics plan to deploy humanoids and task-specific robots in factories by 2028–2030. Concerns from labor groups highlight potential impacts on employment and skilled labor pipelines, while current humanoid capabilities remain limited compared to human workers.
humanoid robotsmotion-tracking glovesactuatorsdexterityindustrial deployment
Top real estate app development companies in the US: Abilities and costs
The article evaluates five US-based real estate app development firms (LITSLINK, Code District, Empat, Helpful Insight, DBB Software) based on their PropTech integration capabilities and production outcomes. It identifies seven critical integration categories for real estate applications: MLS data, identity verification, payment systems, document workflows, mapping APIs, CRM systems, and accounting platforms. Quantitative performance metrics include LITSLINK's $800K revenue generation in 3 months post-launch, DBB Software's 50% faster delivery through modular architecture, and Helpful Insight's 92% client retention rate across 2,000+ projects.
proptechmls integrationreso web apireact nativeci/cd pipelines
Generated automatically at 2026-05-14 21:18 UTC. Summaries and keywords are produced by an LLM and may contain inaccuracies — always consult the original article.
