"Google said in a statement: “Quality raters are employed by our suppliers and are temporarily assigned to provide external feedback on our products. Their ratings are one of many aggregated data points that help us measure how well our systems are working, but do not directly impact our algorithms or models.” GlobalLogic declined to comment for this story." (emphasis mine)
How is this not a straight up lie? For this to be true they would have to throw away labeled training data.
Is there a reason not to use validation data in your next round of training data? Or is it more efficient to reuse validation and instead get more training data?
I'd have thought that if you kept the same validation you'd risk over fitting.
Clearly that does make it hard to measure. I'd think you'd want "equivalent" validation (like changing the SATs every year), though I imagine that's not really a meaningful concept.
More recent models actually use "reinforcement learning from AI feedback", where the task of assigning a reward is essentially fed back into the model itself. Human feedback is then only used to ground the training, on selected examples (potentially even entirely artificial ones) where the AI is most highly uncertain about what feedback should be given.
How is this not a straight up lie? For this to be true they would have to throw away labeled training data.