Run a single experiment iteration. Edit the target file, evaluate, keep or discard.
Run exactly ONE experiment iteration: review history, decide a change, edit, commit, evaluate.
/ar:run engineering/api-speed # Run one iteration
/ar:run # List experiments, let user pick
If no experiment specified, run python {skill_path}/scripts/setup_experiment.py --list and ask the user to pick.
# Read experiment config
cat .autoresearch/{domain}/{name}/config.cfg
# Read strategy and constraints
cat .autoresearch/{domain}/{name}/program.md
# Read experiment history
cat .autoresearch/{domain}/{name}/results.tsv
# Checkout the experiment branch
git checkout autoresearch/{domain}/{name}
Review results.tsv:
Strategy escalation:
Edit only the target file specified in config.cfg. Change one thing. Keep it simple.
git add {target}
git commit -m "experiment: {short description of what changed}"
python {skill_path}/scripts/run_experiment.py \
--experiment {domain}/{name} --single
Read the script output. Tell the user:
After every 10th experiment (check results.tsv line count), update the Strategy section of program.md with patterns learned.