> The instinct should be to tweak the agent to do it right.
I'm extremely doubtful of this. It doesn't save time to tell it "you have an error on line 19", because that's (often) just as much work as fixing the error. Likewise, saying "be careful and don't make mistakes" is not going to achieve anything. So how can you possibly tweak the agent to "do it right" reliably without human intervention? That's not even a solved problem for working with _humans_ who don't have the context window limitations, let alone an LLM that deletes everything past 30k tokens.
I'm not touching code. I'm trying out the feature, and there's any number of things to tweak (because I missed some detail during planning, or agent made bad assumption, etc).
Improving the agent means improving the code base such that the agent can effectively work on it.
It can not Com as a surprise that an agent is better at working on a well documented code base with clear architecture.
On the other hand, if you expect that an agent can add the right amount of ketchup to your undocumented speghatti code, then you will continue to have a bad time.
Musicians are already experiencing this. The likes of Suno are churning out high quality songs with only a minimal amount of prompting material.
One can roughly prototype a song, giving it the structure, melody, harmony, rhythm, lyrics that a finished song might have, upload it and request a cover in a particular style. The output will often resemble a highly competent human performance.
This is one contract that mutates.
Other websites have EU terms, us terms, etc and declare which one you are covered under.
reply