RLI Reveals Agents Can't Automate Remote Work; Liability Looms

RLI Reveals Agents Can't Automate Remote Work; Liability Looms

Published Nov 10, 2025

The Remote Labor Index benchmark (240 freelance projects, 6,000+ human hours, $140k payouts) finds frontier AI agents automate at most 2.5% of real remote work, with frequent failures—technical errors (18%), incomplete submissions (36%), sub‐professional quality (46%) and inconsistent deliverables (15%). These empirical limits, coupled with rising legal scrutiny (e.g., the AI LEAD Act applying product‐liability principles and mounting IP/liability risk for code‐generating tools), compel an expectation reset. Organizations should treat agents as assistive tools, enforce human oversight and robust fallback processes, and maintain documentation of design, data, and error responses to mitigate legal exposure. Benchmarks like RLI provide measurable baselines; until performance improves materially, prioritize augmentation over replacement.