DeepSeek revealed the new rates on Wednesday, and they will take effect on Thursday after midnight.
From the hours of 12.30am to 8.30am China time, API access to the V3 model will be available at a 50 per cent discount. This comes to US$0.035 per million tokens for cache hits, US$0.135 per million tokens for cache misses, and US$0.55 per million tokens for output.
Access to the start-up’s R1 reasoning model will be available during the same hours at a 75 per cent discount. The two models are now priced identically during off-peak hours.
Context length – the maximum number of tokens a model can process at one time – is 64,000 tokens for both models. A token in AI refers to a fundamental unit of data processed by the algorithm, which can be a word, number, or even a punctuation mark. DeepSeek bills users based on the total number of input and output tokens processed by its models.
By comparison, OpenAI’s o1 reasoning model is priced at US$15 per million input tokens, US$7.5 per million cached input tokens, and US$60 per million output tokens, with a context length of 200,000 tokens.