←back to thread

1503 points participant3 | 1 comments | | HN request time: 0s | source
Show context
MgB2 ◴[] No.43574927[source]
Idk, the models generating what are basically 1:1 copies of the training data from pretty generic descriptions feels like a severe case of overfitting to me. What use is a generational model that just regurgitates the input?

I feel like the less advanced generations, maybe even because of their limitations in terms of size, were better at coming up with something that at least feels new.

In the end, other than for copyright-washing, why wouldn't I just use the original movie still/photo in the first place?

replies(13): >>43575052 #>>43575080 #>>43575231 #>>43576085 #>>43576153 #>>43577026 #>>43577350 #>>43578381 #>>43578512 #>>43578581 #>>43579012 #>>43579408 #>>43582494 #
ramraj07 ◴[] No.43578381[source]
So I train a model to say y=2, and then I ask the model to guess the value of y and it says 2, and you call that overfitting?

Overfitting is if you didn't exactly describe Indiana Jones and then it still gave Indiana Jones.

replies(2): >>43578447 #>>43579929 #
1. whywhywhywhy ◴[] No.43579929[source]
What would most humans draw when you describe such a well known character by their iconic elements. Think if you deviated and acted a pedant about it people would think you're just trying to prove a point or being obnoxious.