Help Center/ ModelArts/ FAQs/ ModelArts Standard Inference Deployment/ How Do I Speed Up Real-Time Service Prediction in ModelArts?
Updated on 2025-08-28 GMT+08:00

How Do I Speed Up Real-Time Service Prediction in ModelArts?

  • When deploying a real-time service, select instance specifications with better performance for faster prediction. For example, use GPs instead of CPUs.
  • When deploying a real-time service, add the number of instances.

    If you set the number of instances to 1, the standalone computing mode is used. If you set the number of instances to a value greater than 1, the distributed computing mode is used. Configure this parameter based on site requirements.

  • The inference speed is closely related to the model complexity. Try to optimize the model for faster prediction.

    ModelArts provides model version management to facilitate source tracing and repeated model tuning.