Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

The approach in the paper specifically addresses the case where an LLM can usually solve a task when it requires few steps, but fails for the same kind of task with more steps because it randomly gets a step in the middle wrong and then derails. It can't do anything for tasks that the LLM can't solve even when there's just a few steps.

In other words, it compensates for random error, not systematic error.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: