←back to thread

688 points crescit_eundo | 1 comments | | HN request time: 0s | source
Show context
lukev ◴[] No.42143161[source]
I don't necessarily believe this for a second but I'm going to suggest it because I'm feeling spicy.

OpenAI clearly downgrades some of their APIs from their maximal theoretic capability, for the purposes of response time/alignment/efficiency/whatever.

Multiple comments in this thread also say they couldn't reproduce the results for gpt3.5-turbo-instruct.

So what if the OP just happened to test at a time, or be IP bound to an instance, where the model was not nerfed? What if 3.5 and all subsequent OpenAI models can perform at this level but it's not strategic or cost effective for OpenAI to expose that consistently?

For the record, I don't actually believe this. But given the data it's a logical possibility.

replies(3): >>42143229 #>>42143264 #>>42144445 #
1. TZubiri ◴[] No.42143229[source]
Stallman may have its flaws, but this is why serious research occurs with source code (or at least with binaries)