First 1 2 3
Remote Labor Index: Reality Check — AI Automates Just 2.5% of Remote Work

Remote Labor Index: Reality Check — AI Automates Just 2.5% of Remote Work

Published Nov 10, 2025

The Remote Labor Index (RLI), released Oct 2025, evaluates AI agents on 240 real-world projects across sectors worth $140,000, revealing top agents automated only 2.5% of tasks end-to-end. Common failures—wrong file formats, incomplete submissions, outputs missing brief requirements—show agents fall short of freelance-quality work. The RLI rebuts narratives of imminent agentic independence, highlighting short-term opportunities for human freelancers to profit by fixing agent errors. To advance agentic AI, evaluations must broaden to open-ended, domain-specialized, and multimodal tasks; adopt standardized metrics for error types, quality, correction time, and oversight costs; and integrate economic models to assess net benefit. The RLI is a pragmatic reality check and a keystone benchmark for measuring real-world agentic capability.

RLI Reveals Agents Can't Automate Remote Work; Liability Looms

RLI Reveals Agents Can't Automate Remote Work; Liability Looms

Published Nov 10, 2025

The Remote Labor Index benchmark (240 freelance projects, 6,000+ human hours, $140k payouts) finds frontier AI agents automate at most 2.5% of real remote work, with frequent failures—technical errors (18%), incomplete submissions (36%), sub‐professional quality (46%) and inconsistent deliverables (15%). These empirical limits, coupled with rising legal scrutiny (e.g., the AI LEAD Act applying product‐liability principles and mounting IP/liability risk for code‐generating tools), compel an expectation reset. Organizations should treat agents as assistive tools, enforce human oversight and robust fallback processes, and maintain documentation of design, data, and error responses to mitigate legal exposure. Benchmarks like RLI provide measurable baselines; until performance improves materially, prioritize augmentation over replacement.

First 1 2 3