Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

This particular model has a very low barrier; the model size is smaller than Stable Diffusion which is running easily on consumer hardware for inference, though training is more resource intensive (but not out of reach of consumers, whether through high-end consumer hardware or affordable cloud resources.)

For competitive LLMs targeting text generation, especially for training, a compute-based barrier is more significant.



Yeah that’s fair. I intended my comment to be more of a reflection on the culture in general, but the motivations in this instance are probably different.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: