Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

This makes me think it would be nice to see some kinda child of modern transformer architecture and neural ODEs. There was such interesting work a few years ago on how neural ode/pdes could be seen as a sort of continuous limit of layer depth. Maybe models could learn cool stuff if the embeddings were somehow dynamical model solutions or something.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: