TimeWarp: Evaluating Web Agents by Revisiting the Past

Md Farhan Ishmam and Kenneth Marino

TimeWarp is a benchmark for evaluating the robustness of agents to temporal changes in web UI. TimeWarp consists of three web environments: Wiki, News, and Shop, each with six UI versions across different eras of the internet. The benchmark also includes TimeTraj, a method for scalably collecting trajectories via human-refined plans, and TimeWarp-BC, a variant of Behavior Cloning (BC) to train agents better via knowledge distillation on complex tasks that require memory and planning.

TimeWarp Environment

TimeWarp Dataset

Four categories of tasks: Wiki, News, Shop, and Multi-Environment.

231 unique goals × 6 environments = 1386 tasks.

TimeWarp environment and dataset overview

Dataset Viewer

View and explore the TimeWarp dataset on Hugging Face .

TimeTraj and TimeWarp-BC

TimeTraj is a scalable method for collecting trajectories from a single human-refined plan per task, which is used to automatically generate trajectories across versions.

TimeWarp-BC extends Behavior Cloning by training the web agent on the teacher's full response, including action, thinking, planning, and memory tokens.

Demo of TimeWarp-BC Agent

What is the population difference between the two special administrative regions of China? 1 / 3

Findings

Untrained web agents, especially visual ones, are vulnerable to temporal changes in the web UI. TimeWarp-BC improves both performance and robustness over vanilla behavior cloning.
Main TimeWarp results table

Success Rate (%) of LLM and VLM models on TimeWarp tasks using the observation setting: HTML, Accessibility Tree (AXT), UI Screenshot (SS), Set of Marks (SoM), or the AXT training setting: Behavior Clong (BC) on versions 1 to 6, version 6 only, and TimeWarp-BC (TW).

Training on multiple versions generalize better.
Generation results

Training on more versions generally improve performance in held-out versions (in pink).

Continually training on versions cause catastrophic forgetting and cross-dataset training doesn't help.
Continual evaluation results

(a) Continually training on version 1 from version 6, degrades the performance on version 6. (b) Training on WebArena-Lite then TimeWarp also degrades performance.

Training on thinking, memory, and planning tokens help.
ATMP analysis

Behavior cloning performance peaks when trained with (T)hinking, (M)emory, and (P)lanning tokens.

Agents fail slowly.
Trajectory visualization

Performance of web agents across versions and environments is correlated with the trajectory length.

LoRA doesn't improve performance but improves trajectory length.
LoRA analysis

Full Fine-tuning (FFT) outperforms LoRA at varying adapter ranks. However, the average trajectory length of LoRA is lower than that of FFT.

More training data and training context length improves performance.
Ablation study results

(a) Success rate of TimeWarp-BC (TW) agent grows almost linearly with more training data but not the Behavior Cloning (BC) on version 6 agent. (b) Highest success rate is achieved when trained with 64k context for 3 epochs.

Agents fail in unseen challenges due to temporal changes in the web UI.
Error analysis

Web agent getting stuck due to a popup ad in shop version 5.

Citation

@misc{timewarp2026,
      title={TimeWarp: Evaluating Web Agents by Revisiting the Past}, 
      author={Md Farhan Ishmam and Kenneth Marino},
      year={2026},
      eprint={2603.04949},
      archivePrefix={arXiv},
      primaryClass={cs.AI},
      url={https://arxiv.org/abs/2603.04949}, 
  }

Acknowledgements

We would like to thank Nejd Khadija for helping with the task and plan annotation.