Daily Issue
Vol. I — No. 1
25 · 04
Saturday, 25 April 2026
Generated 2026-04-25 01:18
google/gemini-2.5-flash-lite-preview-09-2025
A great artist is always before his time or behind it. — George Edward Moore 35 items · 3 sections
§ 0

The Morning

Local weather 1
This morning in
London
Clear sky
Today's range
20.6°8.0°
currently 9.6°
Feels
7.7°
Rain
0%
Wind
7 km/h
Humid
78%
Rise
05:43
Set
20:13
§ I

From the arXiv

arXiv preprints 10 of 20
cs.AIarxiv:2604.21725v1Lead article

AEL: Agent Evolving Learning for Open-Ended Environments

Wujiang Xu, Jiaojiao Han, Minghao Guo, Kai Mei, Xi Zhu

he paper introduces Agent Evolving Learning (AEL), a two-timescale framework designed to enable LLM agents to effectively utilize past experience in open-ended environments. AEL employs fast-timescale Thompson Sampling to select the optimal memory retrieval policy for each episode, while a slow-timescale LLM reflection process diagnoses failures and injects causal insights into the agent's prompt. This method significantly improves performance on sequential tasks by providing a structured way to interpret and apply prior knowledge.

Diagram describing a Fantasia interaction, including behavioral sources and failure modes.
Diagram describing a Fantasia interaction, including behavioral sources and failure modes.
cs.AIarxiv:2604.21827v1

Alignment has a Fantasia Problem

Nathanael Jo, Zoe De Simone et al.

The paper identifies "Fantasia interactions" as a core problem where AI treats incomplete user prompts as final intent, leading to misaligned assistance because users often lack fully formed goals. The contribution is arguing that alignment research must shift…

cs.AIarxiv:2604.21910v1

From Research Question to Scientific Workflow: Leveraging Agentic AI for Science Automation

Bartosz Balis, Michal Orzechowski et al.

This paper introduces an agentic AI architecture to automate the translation of natural language research questions into executable scientific workflows. It achieves this by separating the process into three layers: an LLM for intent extraction, deterministic …

Component architecture. The Conductor orchestrates three specialized agents. The Workflow Composer (semantic layer) consults domain Skills (knowledge layer) to produce workflow plans that include data preparation commands. The Deployment Service and Execution Sentinel (deterministic layer) execute these plans on the Kubernetes infrastructure running the HyperFlow engine.
Component architecture. The Conductor orchestrates three specialized agents. The Workflow Composer (semantic layer) consults domain Skills (knowledge layer) to produce workflow plans that include data…
In Stage I, agents 1 to K–1 sequentially construct a shared KV trace by prefilling the existing cache and appending newly generated KV segments without gradient updates. The accumulated KV trace serves as a latent communication medium across agents. In Stage II, the final agent performs autoregressive decoding on the prefilled KV cache. Cross-attention over the KV trace produces hidden states, which are projected through the LM head to generate tokens. Supervised fine-tuning is applied using cross-entropy loss, and gradients are backpropagated to update only the LoRA parameters of the final agent while keeping the backbone model frozen.
In Stage I, agents 1 to K–1 sequentially construct a shared KV trace by prefilling the existing cache and appending newly generated KV segments without gradient updates. The accumulated KV trace serve…
cs.AIarxiv:2604.21794v1

Learning to Communicate: Toward End-to-End Optimization of Multi-Agent Language Systems

Ye Yu, Heming Liu et al.

This paper introduces **DiffMAS**, a novel training framework that enables the **end-to-end, joint optimization of latent inter-agent communication** alongside multi-agent reasoning. It treats the internal, non-textual communication (like key-value caches) as …

cs.AIarxiv:2604.21896v1

Nemobot Games: Crafting Strategic AI Gaming Agents for Interactive Learning with Large Language Models

Chee Wei Tan, Yuchen Wang et al.

This paper introduces **Nemobot Games**, an interactive engineering environment that operationalizes Shannon's game taxonomy using Large Language Models (LLMs) to create strategic AI agents. The core method involves leveraging the LLM's reasoning and synthesis…

Crowdsourcing and strategy optimization in game-playing AI. The LLM generates optimized strategies for game-playing agents, while game states and results from interactions with human players are fed back to train the LLM, creating a self-reinforcing cycle of improvement through crowdsourcing.
Crowdsourcing and strategy optimization in game-playing AI. The LLM generates optimized strategies for game-playing agents, while game states and results from interactions with human players are fed b…
№06
cs.AI
9

Process Supervision via Verbal Critique Improves Reasoning in Large Language Models

Hao-Yuan Chen

This paper introduces Verbal Process Supervision (VPS), a training-free method that uses structured natural-language critique from a stronger model to iteratively guide an LLM's re…

№07
cs.AI
9

Stealthy Backdoor Attacks against LLMs Based on Natural Style Triggers

Jiali Wei, Ming Fan et al.

This paper introduces **BadStyle**, a novel backdoor attack framework against LLMs that utilizes **natural style-level triggers** instead of explicit patterns. The core method invo…

№08
cs.AI
9

StructMem: Structured Memory for Long-Horizon Behavior in LLMs

Buqiang Xu, Yijun Chen et al.

StructMem introduces a structure-enriched hierarchical memory framework for LLMs designed to capture event relationships essential for long-horizon reasoning. It achieves this by t…

№09
cs.AI
9

Tool Attention Is All You Need: Dynamic Tool Gating and Lazy Schema Loading for Eliminating the MCP/Tools Tax in Scalable Agentic Workflows

Anuj Sadani, Deepak Kumar

This paper introduces **Tool Attention**, a middleware mechanism that replaces the costly, eager schema injection of the Model Context Protocol (MCP) with a dynamic, gated attentio…

№10
cs.AI
9

Transient Turn Injection: Exposing Stateless Multi-Turn Vulnerabilities in Large Language Models

Naheed Rayhan, Sohely Jahan

The paper introduces **Transient Turn Injection (TTI)**, a novel multi-turn attack that exploits LLM vulnerabilities by distributing adversarial intent across isolated interactions…

§ II

The Town Square

Hacker News 6
compiled overnight by google/gemini-2.5-flash-lite-preview-09-2025 · end of issue no. 1 · thank you for reading