Updated on 2025-06-16 GMT+08:00

Viewing a Public Inference Service

During the trial period of an inference endpoint, you can directly use a public inference service for inference. The current public inference service is deployed based on the open-source large model. The service list is as follows (subject to the actual inference service).

Table 1 Public inference service

Name

Description

Free Quota

Maximum Context Length

Prompt Template Length

Maximum Output Tokens

QWEN_2_72B

With 72 billion parameters, Qwen2 outperforms most previous open-weight models in multiple benchmark tests in terms of language understanding, generation, multilingual capabilities, coding, mathematics, and inference. It is also competitive with proprietary models.

During the OBT, a quota of 1 million tokens is provided for free. After the quota is used up, the service is unavailable and the tokens cannot be purchased again. The validity period is 90 days after the service is enabled. If the validity period expires, the service becomes invalid.

16,000

23

16,360