Multi turn review of code written by cc reviewed by codex works pretty well. Been one of the only ways to be able to deliver larger scoped features without constant bugs. I've seen them do 10-15 rounds of fix and review until complete.
Also implemented this as a gh action, works well for sentry to gh to auto triage to fix pr.
Yes I’ve had a lot of success with this too. I found with prompt tightening I seldom do more than 5 rounds now, but it also does an explicit plan step with plan review.
Currently I’m authoring with codex and reviewing with opus.
> yes it does, and it's annoying as all hell. Dirt, sun, etc all pop an alert about degraded performance
As with all things FSD, it does sometimes and not others. I've driving my parents' Tesla with FSD engaged and it did complain when the windshield got dirty but didn't say anything when it drove into fog. (I took over manually.)
Out of curiosity, was the camera view compromised? I would probably take over too, but like the poster above, I get the warning in all kinds of conditions.
Has it improved lately? I used it for a year and until I switched to cursor, I felt the AI coding stuff was a complete sham, because copilot was essentially useless.
No, it lets good engineers parallelize work. I can be adding a route to the backend while Cline with Sonnet 3.7 adds a button to the frontend. Boilerplate work that would take 20-30 minutes is handled by a coding agent. With Claude writing some of the backend routes with supervision, you've got a very efficient workflow. I do something like this daily in a 80k loc codebase.
I look forward to good standard integrations to assign a ticket to an agent and let it go through ci and up for a preview deploy & pr. I think there's lots of smaller issues that could be raised and sorted without much intervention.
The research posted demonstrates the opposite of that within the scope of sequence lengths they studied. The model has future tokens strongly represented well in advance.
reply