Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

> Conversely, if a model starts generating text so good that it can be used to train new models, then that should give us confidence in the quality of that text.

It's seems the best one could hope for is that recycling generating text into new training data would be not detrimental. But it's really difficult for me to imagine how this would ever be useful. It seems this would imply that the LLM had somehow managed to expand the dimension of vector space spanned by the original training data. Which sounds either impossible or like the model became sentient.



> It seems this would imply that the LLM had somehow managed to expand the dimension of vector space spanned by the original training data.

The number of dimensions? Well, not by itself I guess. But the span of output compared to training data? Sure, why not?

I think it's also worth pointing out there's a difference between text produced by an LLM looped on itself, which arguably may not contain any new information and would be like repeatedly recompressing the same JPG, and text produced by LLM/human interaction. The latter is indirectly recording new knowledge simply because people's prompts are not random. Even with human part of the conversation discarded, feeding such LLM output back into training data would end up selectively emphasizing associations, which is a good signal too (even if noisier than new human-created text).




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: