We gave the agent access to our K8s cluster with H100s and H200s and let it provision its own GPUs. Over 8 hours: ~910 experiments instead of ~96 sequentially Discovered that scaling model width mattered more than all hparam tuning Taught itself to exploit heterogeneous hardware: use H200s for validation, screen ideas on H100s Blog: https://blog.skypilot.co/scaling-autoresearch/ submitted by /u/Shot-Patience-9874
Originally posted by u/Shot-Patience-9874 on r/ClaudeCode
You must log in or # to comment.
