Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Yes. The path dependence for current attention based LLMs is enormous.


At the same time, there is now a ton of data for training models to act as useful assistants, and benchmarks to compare different assistant models. The wide availability and ease of obtaining new RLHF training data will make it more feasible to build models on new architectures I think.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: