All digests
ResearchersENArtificial Intelligencedaily

[Artificial Intelligence] Daily digest — 237 papers, 0 strong connections (2026-04-14)

DeepScience — Artificial Intelligence
DeepScience
Artificial Intelligence · Daily Digest
April 14, 2026
237
Papers
10/10
Roadblocks Active
2
Connections
⚡ Signal of the Day
• Today's AI research is dominated by a split diagnosis: multiple benchmarks confirm that current AI agents fail badly on real-world tasks, while a parallel wave of inference-time and context-management engineering papers offers partial remedies without new training.
• PaperScope and BankerToolBench independently show that even the best frontier systems (GPT-5.4, OpenAI Deep Research) fail nearly half of rigorous evaluation criteria and produce zero client-ready outputs — the gap between demo and deployment remains large across both scientific reasoning and professional workflow domains.
• Watch the convergence of active context curation (ContextCurator), dynamic reasoning context (SWE-AGILE), and role-orchestrated inference (Three Roles): these three papers independently arrive at the same insight — that smarter use of fixed model weights at inference time may close more of the performance gap than adding parameters.
📄 Top 10 Papers
Escaping the Context Bottleneck: Active Context Curation for LLM Agents via Reinforcement Learning
ContextCurator trains a small 7B policy model via reinforcement learning to decide what information an agent needs to keep, prune, or summarize — decoupling memory management from task execution. On real web navigation tasks, this raises success rates from 36.4% to 41.2% while cutting token use by 8.8%, and on multi-hop research tasks it cuts token consumption by a factor of 8 with improved accuracy. This matters because context overflow is one of the hardest practical blockers for deploying long-running agents: this is the first RL-trained system that handles it as a learned skill rather than a hard rule.
█████████ 0.9 long-context Preprint
Three Roles, One Model: Role Orchestration at Inference Time to Close the Performance Gap Between Small and Large Agents
By deploying the same frozen 8B model in three distinct roles — summarizer, agent, and corrector — within a single inference pipeline, this work roughly doubles task completion rates on multi-step tool-use benchmarks without any additional training. A scaffolded Qwen3-8B outperforms the much larger DeepSeek-Coder 33B on structured tasks, suggesting that architecture of execution matters as much as raw model size. This is a practical result: organizations constrained to small, locally-deployable models can recover substantial capability purely through inference design.
█████████ 0.9 agent-tool-use Preprint
Agentic Aggregation for Parallel Scaling of Long-Horizon Agentic Tasks
Instead of running one agent and hoping for the best, AggAgent runs multiple parallel trajectories and deploys a lightweight aggregation agent to synthesize them — inspecting and searching across solutions rather than just voting on final answers. This yields up to 5.3% absolute improvement on agentic benchmarks and 10.3% on deep research tasks, with overhead bounded by a single extra rollout. The key insight is that parallel test-time compute is cheap to scale, and aggregation is a learnable skill that extracts more signal from diversity than majority vote ever can.
█████████ 0.9 agent-tool-use Preprint
PaperScope: A Multi-Modal Multi-Document Benchmark for Agentic Deep Research Across Massive Scientific Papers
PaperScope constructs 2,400 QA pairs across 25,495 AI papers and 11 sub-tasks to stress-test whether AI systems can genuinely reason across large bodies of scientific literature. Top systems including OpenAI Deep Research and Tongyi Deep Research achieve limited scores, with multi-hop reasoning over multi-modal evidence proving especially difficult. This benchmark matters because it exposes the gap between systems that retrieve well in isolation and systems that can synthesize evidence across documents the way a skilled researcher actually does.
█████████ 0.9 reasoning-reliability Preprint
Retrieval as Generation: A Unified Framework with Self-Triggered Information Planning
GRIP embeds retrieval decisions directly into the autoregressive token stream: a model emits special control tokens to trigger retrieval, reformulate queries, or terminate search — all within a single decoding pass and without any external retrieval controller. This end-to-end coordination means the model learns when it actually needs outside information rather than always retrieving or never retrieving, which reduces both hallucination and wasted computation. The mechanism is important because it eliminates the architectural seam between generation and retrieval that currently forces RAG pipelines into brittle two-stage designs.
█████████ 0.9 hallucination-grounding Preprint
SWE-AGILE: A Software Agent Framework for Efficiently Managing Dynamic Reasoning Context
SWE-AGILE addresses the tension in software engineering agents between deep chain-of-thought reasoning (which bloats context) and shallow reactive agents (which miss edge cases) by maintaining only a sliding window of detailed reasoning and compressing older steps into concise digests. Trained on just 2,200 trajectories with a token-compression reward signal, it sets a new state-of-the-art for 7B–8B models on SWE-Bench-Verified. This shows that explicit context lifecycle management — not just bigger models — is the key lever for sustained reasoning in multi-turn coding tasks.
█████████ 0.9 reasoning-reliability Preprint
BankerToolBench: Evaluating AI Agents in End-to-End Investment Banking Workflows
Developed in collaboration with 502 investment bankers from firms including Goldman Sachs and JPMorgan, this benchmark evaluates nine frontier LLMs on realistic end-to-end workflows producing Excel, PowerPoint, and PDF deliverables. The best model (GPT-5.4) fails nearly half of rubric criteria, and 0% of outputs were rated client-ready by practitioners. This hard number matters: it quantifies exactly how far current agents are from professional deployment thresholds in a high-stakes domain where errors have direct financial consequences.
█████████ 0.9 agent-tool-use Preprint
Synthius-Mem: Brain-Inspired Hallucination-Resistant Persona Memory Achieving 94.4% Memory Accuracy and 99.6% Adversarial Robustness on LoCoMo
Synthius-Mem organizes conversational memory into six cognitive domains inspired by human memory architecture, achieving 94.37% accuracy on the LoCoMo benchmark — surpassing both the previous best system (MemMachine at 91.69%) and human performance (87.9 F1). Adversarial robustness reaches 99.55%, a metric no competing system reports. This matters for deployed conversational agents, where persistent persona consistency and resistance to contradictory inputs are the difference between a useful assistant and an unreliable one.
█████████ 0.9 hallucination-grounding Preprint
Relax: An Asynchronous Reinforcement Learning Engine for Omni-Modal Post-Training at Scale
Relax is a reinforcement learning training infrastructure that decouples data generation, model updates, and reward computation into independent asynchronous services, enabling fully async training to run 2x faster than co-located baselines on 30B multimodal models without sacrificing convergence. The architecture is built to handle text, image, video, and audio natively in a single stack, addressing the fragmented tooling problem that currently forces teams to maintain separate pipelines per modality. For organizations doing post-training at scale, this is a direct reduction in compute cost for RLHF and GRPO runs.
██████████ 0.8 efficiency-scaling Preprint
Minimal Embodiment Enables Efficient Learning of Number Concepts in Robot
A neural network trained to count through physical robotic interaction reaches 96.8% accuracy using only 10% of the training data that a vision-only baseline requires — and the advantage persists even when visual-motor correspondences are randomized, indicating embodiment acts as a structural learning prior rather than a direct information source. The trained model spontaneously develops logarithmically-tuned number-selective units and a mental number line organization matching biological findings. This suggests that embodied interaction shapes representation in ways that transfer to abstract numerical reasoning, with implications for how much data efficient AI systems might need if grounded in physical experience.
██████████ 0.8 embodied-ai Preprint
🔬 Roadblock Activity
Roadblock Papers Status Signal
Efficiency & Scaling 89 Active Async RL infrastructure (Relax) and inference-time role orchestration both show 2x compute efficiency gains without retraining, suggesting the field is shifting from parameter scaling toward execution architecture as the primary efficiency lever.
Reasoning Reliability 88 Active PaperScope and BankerToolBench independently confirm that multi-hop, multi-source reasoning remains a hard failure mode for frontier systems, while SWE-AGILE and Three Roles demonstrate inference-time mitigations that partially recover lost performance.
Multimodal Understanding 87 Active Activity is high but today's top papers focused more on text-heavy agent reasoning; multimodal robotic grounding (CLASP) and vision-language murder mystery games address perception-reasoning coupling but remain early-stage.
Agent Tool Use 64 Active Three independent papers (Three Roles, AggAgent, ContextCurator) converge on inference-time orchestration as a scalable path to better agent tool use, while BankerToolBench provides a sobering professional-grade failure baseline.
Hallucination & Grounding 55 Active GRIP's retrieval-as-generation paradigm and Synthius-Mem's structured persona memory both attack hallucination through architectural changes — tighter retrieval integration and domain-structured memory — rather than post-hoc filtering.
Interpretability 52 Active Interpretability-focused papers are present in volume today but none reached the top tier; the VaCoAl hyperdimensional computing paper claims novel semantic selectivity mechanisms but has low confidence and withheld source code.
Alignment & Safety 51 Active OOM-RL proposes financial loss as an alignment signal but has near-zero reproducibility; RedShell demonstrates that fine-tuned LLMs can generate functional offensive code with high syntactic validity, raising dual-use concerns.
Data Quality & Curation 34 Active The physics simulator paper shows synthetic simulation data can substitute for human-labeled data with zero-shot sim-to-real transfer, offering a scalable curation path for domains where ground truth is expensive to collect.
Embodied AI 30 Active The minimal embodiment paper's finding that physical interaction acts as a structural learning prior — not just an information source — is the most theoretically interesting embodied-AI result today, with implications for data-efficient grounding.
Long Context 27 Active ContextCurator (8x token reduction with maintained accuracy) and SWE-AGILE (sliding-window reasoning digests) both demonstrate that learned compression policies outperform naive context truncation for long-horizon agents.
View Full Analysis
DeepScience — Cross-domain scientific intelligence
Sources: arXiv · OpenAlex · Unpaywall
deepsci.io