Viewing a Public Inference Service
During the trial period of an inference endpoint, you can directly use a public inference service for inference. The current public inference service is deployed based on the open-source large model. The service list is as follows (subject to the actual inference service).
Name |
Description |
Free Quota |
Maximum Context Length |
Prompt Template Length |
Maximum Output Tokens |
---|---|---|---|---|---|
QWEN_2_72B |
With 72 billion parameters, Qwen2 outperforms most previous open-weight models in multiple benchmark tests in terms of language understanding, generation, multilingual capabilities, coding, mathematics, and inference. It is also competitive with proprietary models. |
During the OBT, a quota of 1 million tokens is provided for free. After the quota is used up, the service is unavailable and the tokens cannot be purchased again. The validity period is 90 days after the service is enabled. If the validity period expires, the service becomes invalid. |
16,000 |
23 |
16,360 |
Feedback
Was this page helpful?
Provide feedbackThank you very much for your feedback. We will continue working to improve the documentation.See the reply and handling status in My Cloud VOC.
For any further questions, feel free to contact us through the chatbot.
Chatbot