Introduction to Inference
After an AI model is developed, you can use it to create an AI application and quickly deploy the application as an inference service. The AI inference capabilities can be integrated into your IT platform by calling APIs.
- Develop a model: Models can be developed in ModelArts or your local development environment. A locally developed model must be uploaded to Huawei Cloud OBS.
- Create an AI application: Import the model file and inference file to the ModelArts model repository and manage them by version. Use these files to build an executable AI application.
- Deploy as a service: Deploy the AI application as a container instance in the resource pool and register inference APIs that can be accessed externally.
- Perform inference: Add the function of calling the inference APIs to your application to integrate AI inference into the service process.
Deploying an AI Application as a Service
- Real-time service
Deploy an AI application as a web service with real-time test UI and monitoring supported.
- Batch service
Deploy an AI application as a batch service that performs inference on batch data and automatically stops after data processing is complete.
- Edge service
Deploy an AI application as a web service on an edge node through Intelligent EdgeFabric (IEF).
Next Article: Managing AI Applications
Did this article solve your problem?
Thank you for your score!Your feedback would help us improve the website.