I use AI a lot in cases where I need a bit more than 16k input length (GPT3.5's context window limit). GPT3.5's performance is normally fine for me, but I have to use GPT4 to get a longer context window, at a much increased inference price for the many queries I end up racking up over a long session.
The Claude 3 family of models are the first ones that seem to have very respectable performance and have longer (200k) context windows across the entire family (Opus + Sonnet + Haiku). So I'm very excited about the 'Sonnet' model (the middle quality model).
TLDR: It's exciting to see the benchmark results of Opus, but I think Sonnet might enable more new real world use cases than Opus, when considering the context window and the relatively low cost.
Claude 4 and still 200k context size
Perplexity limits the Claude 3 Opus Context window to 30k tokens : perplexity_ai
Perplexity limits the Claude 3 Opus Context window to 30k tokens
Claude 3 context window is a big deal
Does TextCortex offer Text Generation API?
Which languages does TextCortex support?
What's the difference between TextCortex, Microsoft Copilot & ChatGPT?
Videos
I like Claude 3.7 a lot, but context size was the only downsize. Well, looks like we need to wait one more year for 1M context model.
Even 400K will be a massive improvement! Why 200k?