←back to thread

688 points crescit_eundo | 10 comments | | HN request time: 1.283s | source | bottom
1. lukev ◴[] No.42143161[source]
I don't necessarily believe this for a second but I'm going to suggest it because I'm feeling spicy.

OpenAI clearly downgrades some of their APIs from their maximal theoretic capability, for the purposes of response time/alignment/efficiency/whatever.

Multiple comments in this thread also say they couldn't reproduce the results for gpt3.5-turbo-instruct.

So what if the OP just happened to test at a time, or be IP bound to an instance, where the model was not nerfed? What if 3.5 and all subsequent OpenAI models can perform at this level but it's not strategic or cost effective for OpenAI to expose that consistently?

For the record, I don't actually believe this. But given the data it's a logical possibility.

replies(3): >>42143229 #>>42143264 #>>42144445 #
2. TZubiri ◴[] No.42143229[source]
Stallman may have its flaws, but this is why serious research occurs with source code (or at least with binaries)
3. zeven7 ◴[] No.42143264[source]
Why do you doubt it? I thought it was well known that Chat GPT has degraded over time for the same model, mostly for cost saving reasons.
replies(1): >>42143324 #
4. permo-w ◴[] No.42143324[source]
ChatGPT is - understandably - blatantly different in the browser compared to the app, or it was until I deleted it anyway
replies(1): >>42143446 #
5. lukan ◴[] No.42143446{3}[source]
I do not understand that. The app does not do any processing, just a UI to send text to and from the server.
replies(1): >>42144820 #
6. com2kid ◴[] No.42144445[source]
> OpenAI clearly downgrades some of their APIs from their maximal theoretic capability, for the purposes of response time/alignment/efficiency/whatever.

When ChatGPT3.5 first came out, people were using it to simulate entire Linux system installs, and even browsing a simulated Internet.

Cool use cases like that aren't even discussed anymore.

I still wonder what sort of magic OpenAI had and then locked up away from the world in the name of cost savings.

Same thing with GPT 4 vs 4o, 4o is obviously worse in some ways, but after the initial release (when a bunch of people mentioned this), the issue has just been collectively ignored.

replies(2): >>42144529 #>>42146045 #
7. golol ◴[] No.42144529[source]
You can still do this. People just lost interest in this stuff because it became clear to ehich degree the simulation is really being done (shallow).

Yet I do wish we had access to less finetuned/distilled/RLHF'd models.

8. isaacfrond ◴[] No.42144820{4}[source]
There is a small difference between the app and the browser. before each session, the llm is started with a systems prompt. these are different for the app and the browser. You can find them online somewhere, but iirc the app is instructed to give shorter answers
replies(1): >>42149617 #
9. ipsum2 ◴[] No.42146045[source]
People are doing this all the time with Claude 3.5.
10. bongodongobob ◴[] No.42149617{5}[source]
Correct, it's different in a mobile browser too, the system prompt tells it to be brief/succinct. I always switch to desktop mode when using it on my phone.