←back to thread

283 points Brajeshwar | 4 comments | | HN request time: 0.001s | source
Show context
iandanforth ◴[] No.45231600[source]
"Google said in a statement: “Quality raters are employed by our suppliers and are temporarily assigned to provide external feedback on our products. Their ratings are one of many aggregated data points that help us measure how well our systems are working, but do not directly impact our algorithms or models.” GlobalLogic declined to comment for this story." (emphasis mine)

How is this not a straight up lie? For this to be true they would have to throw away labeled training data.

replies(4): >>45231651 #>>45231697 #>>45231758 #>>45232359 #
1. yobbo ◴[] No.45232359[source]
> For this to be true they would have to throw away labeled training data.

That's how validation works.

replies(1): >>45233162 #
2. jfengel ◴[] No.45233162[source]
Is there a reason not to use validation data in your next round of training data? Or is it more efficient to reuse validation and instead get more training data?
replies(1): >>45233504 #
3. parineum ◴[] No.45233504[source]
You'd have to recreate your validation if you trained your model on it every iteration and then they wouldn't be consistent enough to show any trends
replies(1): >>45240383 #
4. jfengel ◴[] No.45240383{3}[source]
I'd have thought that if you kept the same validation you'd risk over fitting.

Clearly that does make it hard to measure. I'd think you'd want "equivalent" validation (like changing the SATs every year), though I imagine that's not really a meaningful concept.