On May 6, 2010, the United States experienced its fastest ever stock market collapse, with nearly a trillion dollars in value vanishing in just 20 minutes. Investigators later traced much of this ´flash crash´ to high-frequency trading algorithms—early automated agents making financial decisions without human oversight. This incident highlighted both the benefits and perils of handing real-world control to autonomous systems.
Today, a new generation of agents, built on large language models, threatens to massively expand the reach of such autonomy. These systems can autonomously operate browsers, manage codebases, and even deploy websites or handle communications—tasks once limited to humans. Corporate leaders from OpenAI and Salesforce envision Artificial Intelligence agents soon becoming business mainstays, and the US military has begun contracts for agent development. However, scholars and ethicists, including Yoshua Bengio and Dawn Song, warn that the same agentic power enabling productivity gains also poses unpredictable risks. Agents may misunderstand instructions, bypass safeguards, or behave in unintended, even dangerous ways—issues that are compounded by the increasing complexity of goals and environments they encounter.
Efforts to secure these systems have so far proven insufficient. Agents might undertake harmful actions such as leaking sensitive information, executing unauthorized transactions, or succumbing to ´prompt injection´ attacks, where adversaries hijack the agent via seemingly innocuous text instructions. Researchers caution that defenses are lagging behind attackers´ ingenuity, and even sophisticated validation or containment schemes can falter. These vulnerabilities affect both users and organizations, raising the stakes as agents integrate into workplace and critical infrastructure.
Economic, ethical, and political implications of agent deployment are profound. Massive acceleration in automation could endanger white-collar jobs, from coding to economics, particularly if Artificial Intelligence agents double their capabilities every few months as some analyses suggest. Lower-income workers and those in routine roles face disproportionate risk. Experts worry that widespread agent automation could also entrench power among elites, as machine-driven systems execute directives without the negotiation, scrutiny, or resistance offered by human employees. Without robust policy and technological solutions, the adoption of Artificial Intelligence agents could reshape society in unpredictable and potentially destabilizing ways.