←back to thread

507 points martinald | 1 comments | | HN request time: 0s | source
Show context
WhitneyLand ◴[] No.45055133[source]
Model context limits are not “artificial” as claimed.

The largest context window a model can offer at a given quality level depends on the context size the model was pretrained with as well as specific fine tuning techniques.

It’s not simply a matter of considering increased costs.

replies(1): >>45055226 #
Der_Einzige ◴[] No.45055226[source]
Context extension methods exist and work. Please educate yourself about these rather than confidentially saying wrong things.
replies(1): >>45074750 #
1. WhitneyLand ◴[] No.45074750[source]
Not sure what you’re disagreeing with? Context window size limits are not artificial. It takes real time/money/resources to increase them.

There are a few ways to approach the problem. Pre-training on longer context lengths I’ve already mentioned. Fine-tuning techniques (like LongRoPE) I’ve already mentioned.

Inference time context extension tricks I didn’t mention because the papers I’ve seen seem to suggest there’s often problems with quality or unfavorable tradeoffs.

There’s no magic way around these limits, it’s a real engineering problem.