Ad
Skip to content

Deepmind suggests AI should occasionally assign humans busywork so we do not forget how to do our jobs

Image description
Nano Banana Pro prompted by THE DECODER

Key Points

  • Researchers from Google Deepmind have proposed a framework for "intelligent AI delegation" covering how authority, responsibility, and accountability are transferred, whether between humans and AI or among AI agents themselves.
  • A central ethical recommendation is to deliberately build in inefficiencies: the system should occasionally assign tasks to humans that AI could handle on its own, so that people retain the experience needed to intervene competently in critical situations.
  • This is supposed to address the "paradox of automation," where over-reliance on AI leaves human supervisors unable to act when it matters most.

AI agents might soon autonomously distribute tasks to other agents and humans. Google Deepmind researchers argue in a new paper that current methods aren't up to the job and propose a framework to fix that.

The team's concept of "intelligent AI delegation" covers transferring authority, responsibility, and accountability, defining clear roles, and establishing mechanisms for trust. The framework applies to every combination: humans delegating to AI, AI agents delegating to each other, and AI systems handing tasks back to humans.

What organizational theory reveals about AI agent networks

The researchers borrow heavily from how human organizations work. They apply the principal-agent problem to AI systems: a principal delegates to an agent whose goals don't necessarily align with their own. With AI agents, this shows up less as deliberate deception and more as alignment issues like reward hacking, where a system exploits loopholes in its objective function.

How many agents an orchestrator or human supervisor can reliably monitor matters, too. The "authority gradient" concept from aviation shows that large competence gaps between superiors and subordinates stifle communication. For AI agents, sycophancy—the tendency to tell the user what they want to hear—can mean an agent won't even push back on tasks it shouldn't take on.

Ad
DEC_D_Incontent-1

Verification as the core principle of AI delegation

The framework rests on five pillars: continuous agent evaluation, dynamic task redistribution when conditions change, traceable documentation of all decisions, reputation systems that coordinate open marketplaces, and safeguards that prevent individual errors from cascading through the entire network.

Flussdiagramm des Delegationsprozesses. Ausgehend von einem Ziel werden Aufgaben nach Kritikalität, Komplexität und Ressourcen bewertet und in parallele oder sequenzielle Teilschritte zerlegt. Eine Prüfung fragt, ob das Ergebnis verifizierbar ist. Falls nein, wird die Aufgabe rekursiv weiter zerlegt. Falls ja, wird entschieden, ob ein Mensch oder KI-Agent die Aufgabe übernimmt. Es folgen Vorschlagsgenerierung mit Metriken zu Erfolgsrate, Kosten und Dauer, die Auswahl des besten Vorschlags sowie die Zuweisung über ein zentrales Register oder einen dezentralen Marktplatz mit Reputationsprüfung. Am Ende steht ein Smart Contract, der Monitoring, Datenschutz, Autonomie und Verifikation regelt.
From target setting to smart contracts: how intelligent delegation works under the proposed framework. | Image: Deepmind

The key idea is "contract-first decomposition:" a task can only be delegated if its outcome can be verified. If a subtask is too subjective, too expensive, or too complex, it needs to be broken down further. Instead of centralized directories, the researchers argue for decentralized marketplaces with smart contracts that protect both sides.

Security is a major concern. The paper flags malicious agents that steal data or deliver falsified results, "agentic viruses"—self-propagating prompts—and "cognitive monoculture." If too many agents run on the same handful of foundation models, a single vulnerability could take down large parts of the network.

Diagramm des adaptiven Koordinationszyklus. Links sind externe Auslöser wie Aufgabenänderungen, Ressourcenwechsel, Prioritätsänderungen und Sicherheitswarnungen sowie interne Auslöser wie Leistungsabfall, Budgetüberschreitung, Verifikationsfehler und nicht reagierende Agenten dargestellt. Diese fließen in einen vierstufigen Zyklus aus Umgebungsüberwachung, Problemerkennung, Ursachendiagnose und Bewertung von Reaktionsszenarien. Anschließend wird geprüft, ob die Aktion reversibel ist, wie dringend sie ist und ob eine Eskalation nötig ist. Bei irreversiblen Aktionen erfolgt sofortige Terminierung oder menschliche Eskalation. Andernfalls wird zwischen dezentraler Marktauktion und zentralem Agenten als Orchestrierungsmethode gewählt.

Ad
DEC_D_Incontent-2

The adaptive coordination cycle: external and internal triggers kick off a multi-stage response chain. | Image: Deepmind

Deliberate inefficiency as a safeguard against human skill loss

One recommendation stands out: the researchers suggest deliberately building in inefficiencies. The system should occasionally assign tasks to people that it could handle on its own, specifically so humans don't lose their skills. The idea comes from the "Automation Paradox:" if AI takes over all routine tasks, human supervisors lose the hands-on experience they need to step in when something goes wrong. You end up with a fragile system where people are technically responsible but no longer understand what's actually happening.

The researchers also flag a "moral crumple zone," a setup where people have no real control over outcomes but sit in delegation chains specifically to absorb liability when things go wrong.

The team also looked at how well existing agent protocols like Anthropic's MCP, Google's A2A, and the Agent Payments Protocol hold up for this kind of delegation. According to the paper, none of them fully meet the requirements. MCP only offers binary access without finer-grained authorization levels, while A2A lacks support for cryptographic result verification.

AI News Without the Hype – Curated by Humans

As a THE DECODER subscriber, you get ad-free reading, our weekly AI newsletter, the exclusive "AI Radar" Frontier Report 6× per year, access to comments, and our complete archive.

Source: Arxiv