Reinforcement learning makes the model better over time, so why should there be fewer and fewer good results?
If you’re talking about the rate of improvement going down, then yes, of course. That’s bound to happen (unless you have an actual intelligence explosion, but in that case you won’t know what “good results” even mean anyway).
It will still, over time, give fewer and fewer good results to be fed back into it.
Reinforcement learning makes the model better over time, so why should there be fewer and fewer good results?
If you’re talking about the rate of improvement going down, then yes, of course. That’s bound to happen (unless you have an actual intelligence explosion, but in that case you won’t know what “good results” even mean anyway).