larrytheliquid
larrytheliquid
CCConvex Community
Created by codebird on 11/21/2024 in #support-community
Pricing
got it, so basically if you have e.g. an LLM call to o1 that takes for 2-3 minutes to respond, then you are charged for those 2-3 minutes of waiting?
20 replies
CCConvex Community
Created by codebird on 11/21/2024 in #support-community
Pricing
@Jamie sorry to dig up an older thread, but i searched for pricing/billing and this came up. besides the changes you've alluded to for the "scale" plan for cheaper pricing on cache hits, ive got a separate question. the current pricing mentions charging per gb/hr. since a lot of js/ts apps are mostly awaiting IO, i was wondering if the this aspect of the pricing refers to being charged per wall-clock gb/hr, or only for actual cpu-bound computation (i.e., not charging when waiting for IO from a long-running API request, like an LLM call)?
20 replies