Suspension During Training
Symptom 1
According to the logs of the nodes on which a training job runs, an error occurred on a node but the job did not exit, leading to the job suspension.
Solution 1
Check the error cause and rectify the fault.
Symptom 2
The job is stuck in sync-batch-norm or the training speed is lowered down. If sync-batch-norm is enabled for PyTorch, the training speed is lowered down because all node data must be synchronized on each batch normalization layer in every iteration, which leads to heavy communication traffic.
Solution 2
Disable sync-batch-norm, or upgrade the PyTorch version to 1.10.
Symptom 3
The job is stuck in TensorBoard.
Solution 3
Set a local path for storage, for example, cache/tensorboard. Do not store data in OBS.
Symptom 4
When PyTorch dataloader is used to read data, the job is stuck in data reading, and logs stop to update.
Solution 4
When using dataloader to read data, set num_work to a small value.
Feedback
Was this page helpful?
Provide feedbackThank you very much for your feedback. We will continue working to improve the documentation.See the reply and handling status in My Cloud VOC.
For any further questions, feel free to contact us through the chatbot.
Chatbot