I was looking to buy the open air API for my simple NLP classification problem.
Given the current price for chat gpt 4o 2.5$/1 M input tokens I have calculated that it would cost me less than 2$ a month to use the API?
My output is 3 class classification so the output cost is nearly next to nothing.
I feel like something is off..
Does anybody have any real life experience using their API?
EDIT: The answer appears to be "sign up to platform.openai.com instead of ChatGPT, because then you only get charged for the tokens you use, and not the $20/month ChatGPT charge"
Thanks to everyone who answered, I'm up and running, I'll feedback if it starts costing too much!
EDIT 2: Apparently google is too hard for a lot of people, so here's a FAQ for all of those who hijacked this for something else:
-
Just read the docs on the OpenAI integration, it's all there, no hardware required unless you want to talk to it in which case you'll need one of the hardware voice assistants.
-
I'm using it to make my smart home more intelligent - there are loads of examples on Youtube of what people are doing, I want to use OpenAI to do the same thing, so I followed the tutorials on there and got it working
Yes, this is blunt, yes, I think people should share knowledge, but I'm also not going to do your homework for you.
================
I don't have the money or the interest to spend on running a local LLM, so I want to run hosted.
I've noticed the OpenAI API is billed "per million tokens" rather than ChatGPT which is billed at $20USD/month, so I'm starting to work out how much it will cost me to run OpenAI as the backend for my HA setup.
Please note that I am only interested in hearing from people who are already running OpenAI with HA - if you're not doing this, I'm sure your project is awesome and if this doesn't work then I'll definitely be interested in what I should use instead, but right now I need this specific question answered.
Thanks in advance for your time!
I’ve just paid $5 for 20k tokens is it for real?
Has anyone worked out the average monthly cost that you could be paying, if you build an app with openAI's ChatGPT API?
What's the rough monthly cost per user? And how much fee you have to be collecting from the user, to break even? Or how much ad you have to be showing?
Is it financially feasible to actually use OpenAI's API to build something?
Let's say we build a Replika's clone, a chat bot that you can chat with.
Assuming we use the chat-gpt3.5-turbo API, which costs:
USD0.002/1000 tokens
Regardless of what the bot is doing, telling stories, summarising PDF, whatever, we have to be inevitably stuffing a lot of past conversations or the "context" of the conversation into the prompt, and effectively using up all 4000 tokens in every interaction.
So for every question and answer from AI, we use:
full 4000 tokens.
That will be:
USD0.008 per interaction
And assuming we built this app and shipped, user started using. Assume an active user ask a question to a bot once every 5 minute, and they interact with your app for about 2 hours per day:
That will be:
12 interactions per hour or
24 interactions per day or
720 interactions per month
Based on the cost of 0.008 per interaction, the cost for 1 active user will be:
720x0.008 = USD5.76 for chat-gpt3.5-turbo
(And i am not even talking about GPT4's pricing, which is roughly 20 times more expensive).
My understanding from my past apps is that, there is no way, that Google Admobs banner, interstitial ad, etc. can contribute USD5.76 for each active user. (Or can it?)
And therefore, the app can't be an ad-sponsored free app. It has to be a paid app. It has to be an app that is collecting substantially more than USD5.76 per month from each user to be profitable.
Or imagine, we don't sell to end user directly, we build a "chat bot plugin" for organisations for their employees, or for their customers. So if this organisation has 1000 monthly active users, we have to be collecting way more than USD5760 per month?
I hope I was wrong somewhere in the calculation here. What do you think?
TLDR If I build a Replika clone and I have users as sticky as Replika users, monthly fee per user to OpenAI is $5.76 and my user monthly subscription is $8 (Replika).
Let’s say someone implements GPT-4 in their app. Every time a user of the app does something within the app that calls on GPT, the creator of the app has to pay for it, correct? That’s because there’s a limited amount of “tokens”. So let’s say there’s 10,000 users of the app. There’s no way that isn’t super expensive right? Or am I missing something?
Hey there! I wanted to share a tool I built, that I think might be useful to the people in this subreddit.
I find that ChatGPT Plus is amazing but $20/month is quite a lot - especially compared to what I would spend by accessing GPT from the API instead. I know a lot of people are not aware of this, so I built a calculator that asks you questions about your normal usage of chatGPT, and tells you how much you would spend if you switched to the API.
I thought I'd share the calculator I built - let me know what you think!
Here is the link: https://www.quartzite.ai/tools/openai-api-cost-calculator
Thanks!
Hello, i have been developing a side project that utilizes openai gpt4o latest api for its vision capabilities.
I am trying to make a cost analysis, my api requests pretty consistent with around 34k input and 2k output however the charges i am having varies very different.
I should be paying about 10.3 cents per request however it changes between 13-20 cents per request.
What am i doing wrong here ? Thanks.
So i crunched some numbers today.
Im trying to make a chat gpt driven app, and i looked at what would happen if i scaled up. Im currently using $.02 daily, which is a fair estimate. Now, running those numbers up,
Hundreds (e.g., 100 users): Daily cost: 100 users * $0.02/user = $2
Monthly cost: $2/day * 30 days = $60
Annual cost: $60/month * 12 months = $720
Thousands (e.g., 1,000 users):
Daily cost: 1,000 users * $0.02/user = $20
Monthly cost: $20/day * 30 days = $600
Annual cost: $600/month * 12 months = $7,200
Tens of Thousands (e.g., 10,000 users):
Daily cost: 10,000 users * $0.02/user = $200
Monthly cost: $200/day * 30 days = $6,000
Annual cost: $6,000/month * 12 months = $72,000
How the hell can any startup afford this?? These prices feel exorbitant. And trust me, im trying to minmax my token usage as much as i can here, but it hurts when you get charged for tokens sent, returned, in chat history and system prompt.
Idk, whats yall’s opinion? Has anyone made a gpt app that didnt break the bank?
Edit: just woke up, ouch my karma
Edit 2: seeing alot of comments asking my business plan, im not trying to out myself here but generally speaking i expect the service to be something like the following:
-users would pay a one time fee to access the app for a period of time, typically a few months. Chats are also rate limited to 15/3 hours
There was one pretty helpful comment out there pointing out that simply charging users the equivalent of $.04 a day would solve alot of issues, and honestly I agree so shoutout to that guy wherever he is.
Apparently 70k is considered normal for VC funding, which is nuts to me. I ran a firebase app for a year with about 100 active users and spent $.12 on bandwidth, so the jump is jarring.
Im still standing by my statement. Lower level startups will get gate kept by this pricing, leaving only the giants to monopolize it. Our only hope is for PALM to have better pricing or wizardLM to catch up.
For people that didn't know already, you can sign up for an OpenAI API key and pay for your usage by the token, without even having a ChatGPT account. The screenshots you see are from my usage dashboard after sending one Codex prompt, everything was reset to zero before that. Here are more details about the request, it's intentionally non-scientific, because I wanted it to represent a typical moderately complex task in a real workflow:
- Using VSCode Codex extension, I sent the prompt "Go through this codebase and prophylactically fix any issues that may arise when it is tested on MacOS"
- No concurrent requests or subagents or anything fancy like that
- The codebase is about 7000 lines
- Codex worked for about 7 minutes
- Context was mostly full at the time of my sending my prompt, and Codex automatically compacted my context in the middle of responding
- GPT-5.4, xHigh, not using speed mode
- Note: if you're wondering why the screenshot shows "34 requests", it's because any time Codex executes a command and goes back to thinking, that is technically a separate API request.
My thoughts: you could look at this in one of two ways. Either their API is extremely profitable, or users that burn through usage limits every day are losing OpenAI a ton of money. Even if you assume OpenAI takes a generous 90% profit margin on API tokens, this single prompt would've incurred a cost of $0.35. Considering that, the $20/mo tier usage limits are pretty generous IMO.
I'm trying to build a voice agent for a B2C and I never realized how expensive it is. I can get it's easy to be profitable for B2B agents since you reduce payroll(s), but I don't get how this could be profitable for B2C.
Do you charge per usage or just price it very expensive?
Basically the post. I'm surprised at the number of people paying for plus and not using any of the features beyond basic LLM stuff. It's so easy to get an API key and $20 of heavy usage would probably still go far.
That said, I use it from the shell, and was wondering if anyone knows of particularly good web interfaces where you can just plug your API key in. Or alternatively, if there is a local app for Mac or Windows that would be more secure.
Hi all,
Would appreciate some clarity on my use case, as I'm not confident that it will be cost effective for me to use the API.
My plan is that perhaps somewhere around 30-50 times a day, I would be asking the API "Does this blurb of text contain this string, or something that looks like this string?" And I would only need to get back a "Yes" or "No".
The blurb of text could be anywhere from 20 to 500 characters or so.
How would I calculate this based on the pricing guide?
Anybody have a general cost per hour they're seeing with the 4o and 4o mini realtime audio API since the price decrease and improved caching?
I know that before, people were saying they were hitting $60+ per hour.
New GPT-4o and GPT-4o mini realtime snapshots at lower cost
We’re releasing gpt-4o-realtime-preview-2024-12-17 as part of the Realtime API beta with improved voice quality, more reliable input (especially for dictated numbers), and reduced costs. Due to our efficiency improvements, we’re dropping the audio token price by 60% to $40/1M input tokens and $80/1M output tokens. Cached audio input costs are reduced by 87.5% to $2.50/1M input tokens.
We’re also bringing GPT-4o mini to the Realtime API beta as gpt-4o-mini-realtime-preview-2024-12-17. GPT-4o mini is our most cost-efficient small model and brings the same rich voice experiences to the Realtime API as GPT-4o. GPT-4o mini audio price is $10/1M input tokens and $20/1M output tokens. Text tokens are priced at $0.60/1M input tokens and $2.40/1M output tokens. Cached audio and text both cost $0.30/1M tokens.
These snapshots are available in the Realtime API(opens in a new window) and also in the Chat Completions API(opens in a new window) as gpt-4o-audio-preview-2024-12-17 and gpt-4o-mini-audio-preview-2024-12-17.New GPT-4o and GPT-4o mini realtime snapshots at lower costWe’re releasing gpt-4o-realtime-preview-2024-12-17
as part of the Realtime API beta with improved voice quality, more
reliable input (especially for dictated numbers), and reduced costs. Due
to our efficiency improvements, we’re dropping the audio token price by
60% to $40/1M input tokens and $80/1M output tokens. Cached audio input
costs are reduced by 87.5% to $2.50/1M input tokens.
According to this post by The Verge, which quotes the New York Times:
Roughly 10 million ChatGPT users pay the company a $20 monthly fee, according to the documents. OpenAI expects to raise that price by two dollars by the end of the year, and will aggressively raise it to $44 over the next five years, the documents said.
That could be a strong motivator for pushing people to the "LocalLlama Lifestyle".
I’m conceptualizing a SaaS that uses OpenAI’s models on the backend but am having trouble calculating potential cost of use. I understand there is a token model, but how many tokens are typically used in each request? I have ChatGPT Plus, for example - is there a way to see the value in tokens for my searches?