Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I think OP is suggesting that Apple / AMD / Intel do the work of integrating their NPUs into popular libraries like `llama.cpp`. Which might make sense. My impression is that by the time the vendors support a certain model with their NPUs the model is too old and nobody cares anyway. Whereas llama.cpp keeps up with the latest and greatest.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: