IT can be, though. High quality synthetic data for training is a thing, see Phi-2 [0] (and many more examples, people having been using GPT-4 outputs for finetuning a LOT to great avail).
> AI retraining on its own output is not a great thing.
For most forms of GenAI, including image generators, using existing models to generate synthetic data (and then doing human labelling, etc., as appropriate) for fine tuning is quite common and important. Why do you think its not a great thing?
ChatGPT trains on its own output by exploring its own sampling space using reinforcement learning, Dalle 3 was trained on a dataset of synthetic captions, they are both SOTA, you can use AI to train AI models.
tomrod|2 years ago
AI retraining on its own output is not a great thing.
danielbln|2 years ago
[0] https://www.microsoft.com/en-us/research/blog/phi-2-the-surp...
dragonwriter|2 years ago
For most forms of GenAI, including image generators, using existing models to generate synthetic data (and then doing human labelling, etc., as appropriate) for fine tuning is quite common and important. Why do you think its not a great thing?
GaggiX|2 years ago