These LLMs cost millions per day to run, and even with super popular paid API services, the revenue is nowhere near enough to cover mind-boggling costs. What happens next?
The big industry AI leaders raise prices to cover OpEx, and companies realize a low-paid human is just less of a headache compared to an LLM?
Do they just keep absorbing billions in losses like Uber did until cab companies were destroyed, then enjoy no competition?
Are they holding out until a model is capable of displacing enough people that it actually is a good value for business customers?
Question in the title... With Gemini already at a 90% cheaper than Anthropic, and now Deepseek following suit. How will users be able to justify the Claude API price?
I built out all my AI features of my software to use Anthropic, but now given the context size I'm producing and what I expect users to produce in terms of context size, it becomes harder and harder for me to justify the Anthropic price.
Have they released any news recently on breakthrough? potentially making Sonnet more affordable?
It seems quite easy to make a wrapper around the OpenAI API and Anthropic API to get to use GPT4/Claude3 paying per-call, instead of paying for the subscriptions (which are really overpriced, especially if – like me – you like to compare results across models). If I had to be subscribed to everything, it would be 60$ per month or so, while I'm guessing if I paid per call it would be a lot less.
However, the only wrappers I've seen are quite bad (this only supports OpenAI, doesn't support editing or nice visualization of past conversations. This seemed more promising, but is out of date and doesn't support Claude3 or OpenAI). There are some others, but none really seemed to be that good in UI, or allow to plug in to most models (e.g. through LangChain).
Some features that are nice to have:
- GUI
- Supports OpenAI/Anthropic/Gemini
- Editing (even for previous messages, like the OpenAI frontend)
- Saving conversations
- Conversation titles
Does anyone have any recommendations?
I'm excited to share my recent side-by-side comparison of Anthropic's Claude 3.5 Sonnet and OpenAI's GPT-4o models. Using my AI-powered trading platform NexusTrade as a testing ground, I put these models through their paces on complex financial tasks.
Some key findings:
✅ Claude excels at reasoning and human-like responses, creating a more natural chat experience
✅ GPT-4o is significantly faster, especially when chaining multiple prompts
✅ Claude performed better on complex portfolio configuration tasks
✅ GPT-4o handled certain database queries more effectively
✅ Claude is nearly 2x cheaper for input tokens and has a 50% larger context window
While there's no clear winner across all scenarios, I found Claude 3.5 Sonnet to be slightly better overall for my specific use case. Its ability to handle complex reasoning tasks and generate more natural responses gives it an edge, despite being slower.
Does this align with your experience? Have you tried out the new Claude 3.5 Sonnet model? What did you think?
Also, if you want to read a full comparison, check out the detailed analysis here