I am referring to this question/scenario Train neural network with unlimited training data but unfortunately I can not comment.
As I am not seeing any training batch multiple times I would guess that my model has less chance to overfit and generalize better. But since I am able to generate infinite data, do I still need to generate test data at some point? I could train my model unitl I reach a very low loss value (if possible) and then, since I generate the data in the same way as before (same distribution), I would already know that my model will perform well on the new unseen data as it was trained on data it only saw once.
What additionally confuses me regarding this topic is that most literature deals with optimization algorithms where one assumes one data set due to practical reasons. But generating always new data on the spot makes the mathematical setting actually more cleaner.