Understanding the workspace interior
User Roles
- Reader- Although this role allows you to look at everything, but, it cannot change any data, action, or anything that could change the state of the resource.
- Contributor- This one lets you to look as well as change everything except the user roles and rights on the resource.
- Owner- This role permits you to use any action on a specific resource.
- AzureML Data Scientist- This one can only create or delete compute resources or modify the workspace settings.
- AzureML Metrics Writer- It can only write metrics to the workspace.
Experiments
Hence, to keep track of the iterations of our model training, we define them as runs and align them to a construct called an experiment, that can collect all the information concerning a specific model we want to train. To achieve this, we have to connect every training script run to a specific experiment.
Datasets & Datastores
Compute Targets
- Compute instance- it's a single virtual machine generally used for development, as a notebook server, or as a target for training and inference.
- Compute cluster- it's a multi-node cluster of machines typically used for complex training and production environments for inference.
Environments
- Curated environments uses predefined environments having typical runtimes as well as ML frameworks.
- System-managed environments (use default behavior) builds environments starting from a base image with dependency management via Conda.
- User-managed environments builds environments by either starting from a base image through Docker steps (while still allowing you to handle all libraries as well as dependencies by yourself), or by creating a complete custom Docker image.
Runs
- A training script- It performs the actual ML training which basically takes your source folder with all source files, zips it, and sends it to the compute target.
- An environment- The ML environment described previously.
- A compute target- The target compute instance or cluster that the run will be executed in.
While during and after the execution, the run tracks as well as collects the following information:
- Log files- contains the log files generated during the execution and any statement we add to the logging.
- Metrics- includes standard run metrics and any type of object (values, images, and tables) that we want to track specifically during the run.
- Snapshots- consists of a copy of the source directory containing our training scripts (using the ZIP file that we already require for the run configuration).
- Output files- have the files generated by the algorithm (the model) and any file we additionally want to attach to the run.
Registered Models
Hence, the model registry allows you to keep track of the different results you achieved through training and also helps you to deploy different versions of the model production, development, and test environments.
Deployments & Deployment Endpoints
However, if you want to abstract multiple model deployments behind a common endpoint, you can define an endpoint service which is a separate service in Azure ML that can offer a common domain for multiple model deployments, perform Secure Socket Layer (SSL)/Transport Layer Security (TLS) termination. and also permits traffic allocation between deployments.
Pipelines
Pipelines are used to facilitate workflows and bring automation to every step of the ML chain.
Comments
Post a Comment