All digests
ResearchersENArtificial Intelligencedaily

[Artificial Intelligence] Daily digest — 276 papers, 0 strong connections (2026-04-17)

DeepScience — Artificial Intelligence
DeepScience
Artificial Intelligence · Daily Digest
April 17, 2026
276
Papers
11/11
Roadblocks Active
3
Connections
⚡ Signal of the Day
• A cluster of papers today converges on a single uncomfortable finding: current AI models confidently commit to wrong answers and rarely know when to stop talking.
• From VLMs that exhibit 'answer inertia' (doubling down on early wrong predictions rather than revising them) to frontier models almost never abstaining on unanswerable questions, the evidence is building that overconfidence is a structural failure mode, not an edge case.
• Watch the abstention and calibration space closely — the MM-AQA benchmark and the Reasoning Dynamics study together suggest this is a measurable, reproducible problem that benchmark designers and model trainers will need to explicitly target.
📄 Top 10 Papers
RadAgent: A tool-using AI agent for stepwise interpretation of chest computed tomography
RadAgent breaks chest CT interpretation into a sequence of specialized tool calls — organ segmentation, disease classification, slice extraction — guided by a clinician-reviewed checklist and trained with reinforcement learning. It outperforms a direct 3D vision-language model by 6 macro-F1 points on accuracy and 24.7 points on robustness to adversarial inputs, and adds faithful reasoning traces the baseline cannot produce at all. This matters because it shows modular, tool-using agents can genuinely improve clinical reliability over end-to-end models, not just benchmark scores.
█████████ 0.9 reasoning-reliability Preprint
Reasoning Dynamics and the Limits of Monitoring Modality Reliance in Vision-Language Models
Across 18 vision-language models and three benchmarks, this study tracks how model confidence evolves during chain-of-thought reasoning and finds that models lock in early predictions and reinforce them rather than revising when evidence contradicts the initial guess — a pattern the authors call 'answer inertia'. Models are also consistently swayed by misleading text even when the visual evidence alone would be enough to answer correctly. This is a structural problem: reasoning steps that look deliberate are often just elaborate justifications for snap judgments.
█████████ 0.9 reasoning-reliability Preprint
Knowing When Not to Answer: Evaluating Abstention in Multimodal Reasoning Systems
The authors construct MM-AQA, a 2,079-instance benchmark of unanswerable multimodal questions built by systematically removing visual evidence or modality relevance from MMMU and MMLongBench-Doc. Frontier vision-language models almost never abstain under standard prompting — and simple confidence baselines beat standard prompting for knowing when to stay silent. Multi-agent debate systems help, but introduce a direct accuracy-abstention trade-off, revealing that current architectures have no principled mechanism for recognizing the limits of their own knowledge.
█████████ 0.9 hallucination-grounding Preprint
SWE-TRACE: Optimizing Long-Horizon SWE Agents Through Rubric Process Reward Models and Heuristic Test-Time Scaling
SWE-TRACE addresses a core failure mode in AI software engineering agents: sparse, end-of-task rewards make training unstable for long multi-step tasks. It combines dense intermediate feedback via a Rubric-Based Process Reward Model with a memory-augmented RL loop, and distills a 60K-instance training corpus biased toward the shortest correct solution paths. The heuristic test-time scaling component reuses the learned reward signals at inference to prune bad actions without adding latency. Reproducibility is low as no code or data is released, so results should be treated as preliminary.
██████████ 0.8 reasoning-reliability Preprint
UniDoc-RL: Coarse-to-Fine Visual RAG with Hierarchical Actions and Dense Rewards
UniDoc-RL trains a vision-language model agent to retrieve and read document images through a coarse-to-fine hierarchy: first locating the right page region, then cropping and zooming for precision, using GRPO reinforcement learning with separate rewards at each stage. This staged reward design avoids the need for a separate value network while achieving up to 17.7% gains over prior visual RAG methods. Code and data are publicly released, making this one of the more reproducible results in today's batch.
██████████ 0.8 multimodal-understanding Preprint
OpenMobile: Building Open Mobile Agents with Task and Trajectory Synthesis
OpenMobile generates synthetic training data for mobile phone agents by first building a memory of what each app can do through exploration, then synthesizing multi-step task instructions grounded in that memory. A policy-switching rollout alternates between a learner model and an expert to collect error-recovery trajectories that standard imitation learning misses. Fine-tuned on this data, Qwen3-VL-8B reaches 64.7% success on AndroidWorld — competitive results on an open model without proprietary fine-tuning data. Both code and dataset are publicly released.
██████████ 0.8 agent-tool-use Preprint
Feedback-Driven Execution for LLM-Based Binary Analysis
FORGE applies a forest of LLM agents to binary vulnerability analysis, where each agent runs a reasoning-action-observation loop that incrementally builds evidence rather than attempting full analysis in a single pass. Across 591 binaries, the system identifies 1,274 vulnerabilities at 72.3% precision — a domain where conventional tools struggle with the open-ended reasoning required. The dynamic tree structure bounds per-agent context while allowing parallel exploration, which is a practical architectural solution to the long-context problem in agentic systems.
██████████ 0.8 reasoning-reliability Preprint
How Do LLMs and VLMs Understand Viewpoint Rotation Without Vision? An Interpretability Study
Using a synthetic text-only benchmark (VRUBench) and mechanistic analysis via linear probing and causal head interventions, this study finds that LLMs and VLMs encode viewpoint orientation information in their hidden states but consistently fail to bind that information to the correct object when answering questions — a binding failure that manifests as hallucination in the final transformer layers. Humans solve these tasks at 100%; models fall far short. Targeted fine-tuning of only the identified key attention heads partially repairs the deficit without broad performance degradation.
██████████ 0.8 interpretability Preprint
SGA-MCTS: Decoupling Planning from Execution via Training-Free Atomic Experience Retrieval
SGA-MCTS uses Monte Carlo Tree Search offline to explore solution spaces and distills successful trajectories into abstract State-Goal-Action atoms — reusable reasoning templates stripped of domain-specific details. At inference, a hybrid symbolic-semantic retrieval system re-grounds these atoms into the current problem context, allowing frozen open-weights models to match reported SOTA performance without any task-specific fine-tuning. The claim that this rivals GPT-5-class systems is striking but unverified pending full paper and code release.
██████████ 0.8 reasoning-reliability Preprint
Autonomous Evolution of EDA Tools: Multi-Agent Self-Evolved ABC
This paper demonstrates a multi-agent LLM system that iteratively rewrites portions of ABC, a 1.2-million-line C codebase for logic synthesis used in chip design, compiling and formally verifying each change before accumulating improvements across generations. The system discovers optimizations beyond human-designed heuristics on standard benchmarks, which is a meaningful result because logic synthesis quality directly affects chip performance and power. Reproducibility is uncertain — the authors use a significantly expanded private fork of ABC and have not confirmed artifact release.
██████████ 0.8 agent-tool-use Preprint
🔬 Roadblock Activity
Roadblock Papers Status Signal
Efficiency & Scaling 116 Active The largest roadblock by volume today, with SWE-TRACE's heuristic test-time scaling and SGA-MCTS both proposing inference-time efficiency gains that avoid the latency costs of standard parallel sampling.
Reasoning Reliability 111 Active Multiple independent papers today document the same failure mode — models committing early to wrong answers and rationalizing rather than revising — suggesting answer inertia is a systemic and measurable property of current architectures.
Agent Tool Use 75 Active RadAgent, OpenMobile, FORGE, and the EDA evolution system all demonstrate that structured tool orchestration with intermediate verification consistently outperforms end-to-end model calls on complex multi-step tasks.
Interpretability 74 Active The viewpoint rotation study uses causal head interventions to pinpoint where binding failures occur in transformer layers, offering a replicable methodology for localizing specific reasoning deficits rather than just measuring them.
Multimodal Understanding 70 Active UniDoc-RL and ProVoice-Bench both highlight that multimodal models still struggle when evidence must be actively retrieved or when audio context requires proactive rather than reactive reasoning.
Hallucination & Grounding 62 Active The abstention benchmark MM-AQA makes the hallucination problem concrete and quantifiable: models almost never refuse to answer unanswerable questions, and simple confidence baselines beat sophisticated prompting for knowing when to stay silent.
Alignment & Safety 56 Active Activity today is largely theoretical or speculative in this roadblock, with no high-confidence empirical papers; the two Machine-Native Intelligence submissions have severe methodological limitations and should not be treated as evidence.
Data Quality & Curation 30 Active OpenMobile's policy-switching trajectory synthesis — deliberately collecting error-recovery examples that pure imitation learning misses — is the most concrete data curation advance in today's papers.
Embodied AI 20 Active The World-Value-Action model proposes latent-space planning for robotic vision-language-action systems but is truncated and unverifiable; no strong empirical signal from this roadblock today.
Long Context 19 Active FORGE's forest-of-agents architecture provides a practical engineering answer to long-context limits by bounding each agent's context window while enabling parallel exploration of large binary codebases.
Encoding & Memory 1 Low Minimal activity today; only one paper touches this roadblock, suggesting it is not a focus area in the current publication cycle.
View Full Analysis
DeepScience — Cross-domain scientific intelligence
Sources: arXiv · OpenAlex · Unpaywall
deepsci.io