Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Yes, this is it. The idea that LLMs somehow write this deceptive code that magically looks right but isn't is just silly. Why would that be the case? If someone finds they are good at writing code (hard to define of course but take a "measure" like long term maintainability for example) but they fail to catch bad code in review it is just an issue with their skill. Reviewing code can be trained just as writing code can be. A good first step might be to ask oneself: "how would I have approached this".


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: