Skip to main content

Product Pricing

Explanation: The prices listed below are all inclusive of tax. Batch API inference costs are 60% of the standard model price, ideal for large-scale tasks with low real-time requirements. Here, 1M = 1,000,000. The prices in the table represent the cost per 1M tokens consumed.

Notes

  • Batch API currently only supports the kimi-k2.5 model
  • Batch API is not subject to real-time concurrency limits, ideal for bulk tasks
  • Tasks must complete within the specified completion_window, otherwise they expire
  • See the Batch API Guide for detailed usage instructions