MindYOLO ModelArts Training Quick Start¶
This article mainly introduces the training method of MindYOLO using the ModelArts platform. ModelArts related tutorial reference Help Center
Prepare data and code¶
Use the OBS service to upload data sets. For related operation tutorials, see OBS User Guide to obtain the [AK] of this account (https:// docs.xckpjs.com/zh-cn/browsertg/obs/obs_03_1007.html), please consult the corresponding platform administrator or account person in charge for the server address. If the AK is not in the location specified in the user guide, please also consult the platform administrator or account person in charge. .
operate:
- Log in to obs browser+
- Create a bucket -> create a new folder (eg: coco)
- To upload data files, please place the data files in a separate folder (that is, coco in the use case). The code will copy the data in the obs bucket, and the copied content will be all files in this folder (such as coco). document. Without creating a new folder, you cannot select the complete data set.
Prepare code¶
Also use the OBS service to upload the training code.
Operation: Create a bucket -> Create a new folder (such as: mindyolo) -> Upload the code file, create an output folder at the same level of mindyolo to store training records, and create a log folder to store logs.
Create new algorithm¶
- Select Algorithm Management->Create in the tab.
- Customize the algorithm name, select Ascend-Powered-Engine for the prefabricated framework, select the MindSpore-2.0 version image for the master branch, and select the MindSpore-1.8.1 version image for the r0.1 branch. Set the code directory, startup file, input, and output. and superparameters.
- If you need to load pre-trained weights, you can select the uploaded model file in the model selection and add the ckpt_dir parameter in the running parameters.
- The startup file is train.py
- To run super parameters, enable_modelarts needs to be added, and the value is True.
- The running super parameter config path refers to the directory of the running environment preview in the training job, such as /home/ma-user/modelarts/user-job-dir/mindyolo/configs/yolov5/yolov5n.yaml
- If distributed training scenarios are involved, the hyperparameter is_parallel needs to be added and set to True when running in distributed mode and False when running on a single card.
Create new job¶
- Select in the ModelArts service: Training Management -> Training Jobs -> Create a training job, set the job name, and choose not to include it in the experiment; Create Method -> My Algorithm, select the newly created algorithm;
- Training input -> Data storage location, select the obs data bucket just created (coco in the example), select the output folder when preparing the code for training output, and set the config hyperparameter value according to the running environment preview;
- Select the resource pool, specifications, computing nodes, and select the log folder when creating the code for the job log path.
- Submit training and it will be running after queuing.
Modify job¶
Select Rebuild on the training job page to modify the selected job configuration.