Product Pricing
Explanation: The prices listed below are all inclusive of tax. Batch API inference costs are 60% of the standard model price, ideal for large-scale tasks with low real-time requirements. Here, 1M = 1,000,000. The prices in the table represent the cost per 1M tokens consumed.Notes
- Batch API currently only supports the
kimi-k2.5model - Batch API is not subject to real-time concurrency limits, ideal for bulk tasks
- Tasks must complete within the specified
completion_window, otherwise they expire - See the Batch API Guide for detailed usage instructions