Love this. Going from "interpret the UI and guess what to click" to "here are the functions you can call" is how it should have always worked. Since the tools are just JS on the page, could the site surface what the agent is doing to the user in real time? Would be cool for human-in-the-loop stuff where you want to approve actions before they execute.