Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I don't get you guys that are getting such bad results.

Are you guys just trying to one shot stuff? Are you not using agents to iterate on things? Are you not putting agents against each other (have one code, one critique/test the code, and put them in a loop)?

I still look at the code that's produced, I'm not THAT far down the "vibe coding" path that I'm trusting everything being produced, but I get phenomenal results and I don't actually write any code any more.

So like, yeah, first pass the llm will create my feature and there's definitely some poorly written code or duplicate code or other code smells, but then I tell another agent to review and find all these problems. Then that review gets fed back in to the agent that created the feature. Wham, bam, clean code.

I'm not using gastown or ralph wiggum ($$$) but reading the docs, looking over how things work, I can see how it all comes together and should work. They've been built out to automatically do the review + iteration loop that I do.



My feeling has been that 'serious' software engineers aren't particularly suited to use these tools. Most don't have an interest in managing people or are attracted to the deterministic nature of computing. There's a whole psychology you have to learn when managing people, and a lot of those skills transfer to wrangling AI agents from my experience.

You can't be too prescriptive or verbose when interacting with them, you have to interact with them a bit to start understanding how they think and go from there to determine what information or context to provide. Same for understanding their programming styles, they will typically do what they're told but sometimes they go on a tangent.

You need to know how to communicate your expectations. Especially around testing and interaction with existing systems, performance standards, technology, the list goes on.


All our best performing devs/engineers are using the tools the most.

I think this is something a lot of people are telling themselves though, sure.


Best performing by what metric? There aren't meaningful ways to measure engineer "performance" that makes them comparable as far as I know.


Your org doesn't track engineering impact?

What about git stats?

I can tell you the guys that are consistently pushing code AND having the biggest impact are using LLM tools.


Are we measuring productivity by lines of code again? This was treated as unserious for decades.


Why ignore where I mention engineering impact??? Come on, be real here


What git stats do you have that show “impact”?

The OP was right to assume it was lines of code. Another assumption could be number of commits, which also doesn’t measure impact.


Track engineering impact and git stats were two separate suggestions in that comment. Every org tracks impact through performance reviews.


Probably because you mentioned "git stats".

What you meant by that?


High number of days with commits, merging and shipping code consistently (some people/project will ship multiple times a day/week, some projects move a little slower).

That plus the completion of high impact projects makes good strong engineers.

Those are the people I see using LLMs


So quantity of code?


It lets 0.05X developers be 0.2X developers and 1X developers be 0.9-1.1X developers.

The problem is some 0.05X developers thought they were 0.5X and now they think they're 2X.


Nah, our best devs/engineers use the tools the most.

In my real life experience it's been the middling devs that always talk about "ai slop" and how the tools can't do their jobs.


On our team there's a very clear distinction between three groups:

- those who have embraced AI and learned to use it well

- those who have embraced AI but treat it as a silver bullet

- those who reject AI

First group is by far the most productive and adds the most value to the team.


Yeah, it's similar where I'm at.

If anything the silver bullet people are mostly managers and C levels... some of which don't even use the tools themselves.

Of the devs that rejected it at first, the ones with the same sentiment I'm seeing online in threads like these, we forced one to give it a try. He now treats totters between using it well and treating it as a silver bullet. I still hear him incredulous about the things claude does at meetings, "I had to do <thing> and I thought I'd let claude get a crack at it... did it in one shot"


I mean, that fits with what I said.


I mean, not all workplaces hire the best.


I have some success but by the time I'm done I'm often not sure if I saved any time.


My (former) coworker who’s heavy into this stuff produced a lot of unmaintainable slop on his way out while singing agents praises to hire-ups. He also felt he was getting a lot of value and had no issues.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: