From 4c102ff2760600cd7fc709db363aea66b41943d5 Mon Sep 17 00:00:00 2001 From: Amagash Date: Sat, 11 Sep 2021 20:23:26 +0200 Subject: [PATCH 1/7] add devpmm pet id --- .../17-Introduction/README.md | 12 +++++------ 5-Data-Science-In-Cloud/18-Low-Code/README.md | 8 ++++---- .../18-Low-Code/assignment.md | 2 +- 5-Data-Science-In-Cloud/19-Azure/README.md | 20 +++++++++---------- .../19-Azure/assignment.md | 2 +- 5 files changed, 22 insertions(+), 22 deletions(-) diff --git a/5-Data-Science-In-Cloud/17-Introduction/README.md b/5-Data-Science-In-Cloud/17-Introduction/README.md index 920e323..93d9ad2 100644 --- a/5-Data-Science-In-Cloud/17-Introduction/README.md +++ b/5-Data-Science-In-Cloud/17-Introduction/README.md @@ -66,7 +66,7 @@ The steps necessary to create this projects are the following: * Create an output sink and specify the job output * Start the job -To view the full process, check out the [documentation](https://docs.microsoft.com/en-us/azure/stream-analytics/stream-analytics-twitter-sentiment-analysis-trends). +To view the full process, check out the [documentation](https://docs.microsoft.com/en-us/azure/stream-analytics/stream-analytics-twitter-sentiment-analysis-trends?ocid=AID3041109). ### Scientific papers analysis @@ -75,9 +75,9 @@ Let’s take another example, of a project created by [Dmitry Soshnikov](http:// Dmitry created a tool that analyses COVID papers. By reviewing this project, you will see how you can create a tool that extracts knowledge from scientific papers, gains insights, and get a tool that helps researchers navigate large collections of papers in a meaningful way. Let's see the different steps used for this: -* Extracting and pre-processing information with [Text Analytics for Health](https://docs.microsoft.com/en-us/azure/cognitive-services/text-analytics/how-tos/text-analytics-for-health?WT.mc_id=academic-40229-cxa) -* Using [Azure ML](https://azure.microsoft.com/services/machine-learning/?WT.mc_id=academic-40229-cxa) to parallelize the processing -* Storing and querying information with [Cosmos DB](https://azure.microsoft.com/services/cosmos-db/?WT.mc_id=academic-40229-cxa) +* Extracting and pre-processing information with [Text Analytics for Health](https://docs.microsoft.com/en-us/azure/cognitive-services/text-analytics/how-tos/text-analytics-for-health?ocid=AID3041109) +* Using [Azure ML](https://azure.microsoft.com/services/machine-learning/?ocid=AID3041109) to parallelize the processing +* Storing and querying information with [Cosmos DB](https://azure.microsoft.com/services/cosmos-db/?ocid=AID3041109) * Create an interactive dashboard for data exploration and visualization using Power BI To see the full process, visit [Dmitry’s blog](https://soshnikov.com/science/analyzing-medical-papers-with-azure-and-text-analytics-for-health/). @@ -88,8 +88,8 @@ As you can see, we can leverage Cloud services in many ways to perform Data Scie ## Footnote Sources: -* https://azure.microsoft.com/overview/what-is-cloud-computing -* https://docs.microsoft.com/azure/stream-analytics/stream-analytics-twitter-sentiment-analysis-trends +* https://azure.microsoft.com/overview/what-is-cloud-computing?ocid=AID3041109 +* https://docs.microsoft.com/azure/stream-analytics/stream-analytics-twitter-sentiment-analysis-trends?ocid=AID3041109 * https://soshnikov.com/science/analyzing-medical-papers-with-azure-and-text-analytics-for-health/ ## Post-Lecture Quiz diff --git a/5-Data-Science-In-Cloud/18-Low-Code/README.md b/5-Data-Science-In-Cloud/18-Low-Code/README.md index f80ef70..5c82627 100644 --- a/5-Data-Science-In-Cloud/18-Low-Code/README.md +++ b/5-Data-Science-In-Cloud/18-Low-Code/README.md @@ -43,7 +43,7 @@ Table of contents: ### 1.1 What is Azure Machine Learning? Data scientists expend a lot of effort exploring and pre-processing data, and trying various types of model-training algorithms to produce accurate models, which is time consuming, and often makes inefficient use of expensive compute hardware. -[Azure ML](https://docs.microsoft.com/EN-US/azure/machine-learning/overview-what-is-azure-machine-learning) is a cloud-based platform for building and operating machine learning solutions in Azure. It includes a wide range of features and capabilities that help data scientists prepare data, train models, publish predictive services, and monitor their usage. Most importantly, it helps data scientists increase their efficiency by automating many of the time-consuming tasks associated with training models; and it enables them to use cloud-based compute resources that scale effectively to handle large volumes of data while incurring costs only when actually used. +[Azure ML](https://docs.microsoft.com/EN-US/azure/machine-learning/overview-what-is-azure-machine-learning?ocid=AID3041109) is a cloud-based platform for building and operating machine learning solutions in Azure. It includes a wide range of features and capabilities that help data scientists prepare data, train models, publish predictive services, and monitor their usage. Most importantly, it helps data scientists increase their efficiency by automating many of the time-consuming tasks associated with training models; and it enables them to use cloud-based compute resources that scale effectively to handle large volumes of data while incurring costs only when actually used. Azure ML provides all the tools developers and data scientists need for their machine learning workflows, including: @@ -98,7 +98,7 @@ Once you have the dataset, we can start the project in Azure. ## 2. Low code/No code training of a model in Azure ML Studio ### 2.1 Create an Azure ML workspace -To train a model in Azure ML you first need to create an Azure ML workspace. The workspace is the top-level resource for Azure Machine Learning, providing a centralized place to work with all the artifacts you create when you use Azure Machine Learning. The workspace keeps a history of all training runs, including logs, metrics, output, and a snapshot of your scripts. You use this information to determine which training run produces the best model. [Learn more](https://docs.microsoft.com/en-us/azure/machine-learning/concept-workspace) +To train a model in Azure ML you first need to create an Azure ML workspace. The workspace is the top-level resource for Azure Machine Learning, providing a centralized place to work with all the artifacts you create when you use Azure Machine Learning. The workspace keeps a history of all training runs, including logs, metrics, output, and a snapshot of your scripts. You use this information to determine which training run produces the best model. [Learn more](https://docs.microsoft.com/en-us/azure/machine-learning/concept-workspace?ocid=AID3041109) It is recommended to use the most up-to-date browser that's compatible with your operating system. The following browsers are supported: @@ -226,7 +226,7 @@ Great now that the dataset is in place and the compute cluster is created, we ca ### 2.4 Low code/No Code training with AutoML Traditional machine learning model development is resource-intensive, requiring significant domain knowledge and time to produce and compare dozens of models. -Automated machine learning (AutoML), is the process of automating the time-consuming, iterative tasks of machine learning model development. It allows data scientists, analysts, and developers to build ML models with high scale, efficiency, and productivity all while sustaining model quality. It greatly accelerates the time it takes to get production-ready ML models with great ease and efficiency. [Learn more](https://docs.microsoft.com/en-us/azure/machine-learning/concept-automated-ml) +Automated machine learning (AutoML), is the process of automating the time-consuming, iterative tasks of machine learning model development. It allows data scientists, analysts, and developers to build ML models with high scale, efficiency, and productivity all while sustaining model quality. It greatly accelerates the time it takes to get production-ready ML models with great ease and efficiency. [Learn more](https://docs.microsoft.com/en-us/azure/machine-learning/concept-automated-ml?ocid=AID3041109) 1. In the [Azure ML workspace](https://ml.azure.com/) that we created earlier click on "Automated ML" in the left menu and select the dataset you just uploaded. Click Next. @@ -355,7 +355,7 @@ Look more closely at the model explanations and details that AutoML generated fo In this lesson, you learned how to train, deploy and consume a model to predict heart failure risk in a Low code/No code fashion in the cloud. If you have not done it yet, look more closely at the model explanations that AutoML generated for the top models and try to understand why the best model is better than the other ones. -You can go further into Low code/No code AutoML by reading this [documentation](https://docs.microsoft.com/en-us/azure/machine-learning/tutorial-first-experiment-automated-ml). +You can go further into Low code/No code AutoML by reading this [documentation](https://docs.microsoft.com/en-us/azure/machine-learning/tutorial-first-experiment-automated-ml?ocid=AID3041109). ## Assignment diff --git a/5-Data-Science-In-Cloud/18-Low-Code/assignment.md b/5-Data-Science-In-Cloud/18-Low-Code/assignment.md index d4963bc..c7ddf90 100644 --- a/5-Data-Science-In-Cloud/18-Low-Code/assignment.md +++ b/5-Data-Science-In-Cloud/18-Low-Code/assignment.md @@ -2,7 +2,7 @@ ## Instructions -We saw how to use the Azure ML platform to train, deploy and consume a model in a Low code/No code fashion. Now look around for some data that you could use to train an other model, deploy it and consume it. You can look for datasets on [Kaggle](https://kaggle.com) and [Azure Open Datasets](https://azure.microsoft.com/en-us/services/open-datasets/catalog/?WT.mc_id=academic-15963-cxa). +We saw how to use the Azure ML platform to train, deploy and consume a model in a Low code/No code fashion. Now look around for some data that you could use to train an other model, deploy it and consume it. You can look for datasets on [Kaggle](https://kaggle.com) and [Azure Open Datasets](https://azure.microsoft.com/en-us/services/open-datasets/catalog?ocid=AID3041109). ## Rubric diff --git a/5-Data-Science-In-Cloud/19-Azure/README.md b/5-Data-Science-In-Cloud/19-Azure/README.md index 51dd4ba..9dca6cc 100644 --- a/5-Data-Science-In-Cloud/19-Azure/README.md +++ b/5-Data-Science-In-Cloud/19-Azure/README.md @@ -54,7 +54,7 @@ Key areas of the SDK include: - Use automated machine learning, which accepts configuration parameters and training data. It automatically iterates through algorithms and hyperparameter settings to find the best model for running predictions. - Deploy web services to convert your trained models into RESTful services that can be consumed in any application. -[Learn more about the Azure Machine Learning SDK](https://docs.microsoft.com/en-us/python/api/overview/azure/ml/?view=azure-ml-py) +[Learn more about the Azure Machine Learning SDK](https://docs.microsoft.com/en-us/python/api/overview/azure/ml?ocid=AID3041109) In the [previous lesson](../18-tbd/README.md), we saw how to train, deploy and consume a model in a Low code/No code fashion. We used the Heart Failure dataset to generate and Heart failure prediction model. In this lesson, we are going to do the exact same thing but using the Azure Machine Learning SDK. @@ -107,7 +107,7 @@ Now that we have a Notebook, we can start training the model with Azure ML SDK. ### 2.5 Training a model -First of all, if you ever have a doubt, refer to the [Azure ML SDK documentation](https://docs.microsoft.com/en-us/python/api/overview/azure/ml/?view=azure-ml-py). In contains all the necessary information to understand the modules we are going to see in this lesson. +First of all, if you ever have a doubt, refer to the [Azure ML SDK documentation](https://docs.microsoft.com/en-us/python/api/overview/azure/ml?ocid=AID3041109). In contains all the necessary information to understand the modules we are going to see in this lesson. #### 2.5.1 Setup Workspace, experiment, compute cluster and dataset @@ -155,7 +155,7 @@ df.describe() ``` #### 2.5.2 AutoML Configuration and training -To set the AutoML configuration, use the [AutoMLConfig class](https://docs.microsoft.com/en-us/python/api/azureml-train-automl-client/azureml.train.automl.automlconfig(class)?view=azure-ml-py). +To set the AutoML configuration, use the [AutoMLConfig class](https://docs.microsoft.com/en-us/python/api/azureml-train-automl-client/azureml.train.automl.automlconfig(class)?ocid=AID3041109). As described in the doc, there are a lot of parameters with which you can play with. For this project, we will use the following parameters: @@ -207,18 +207,18 @@ RunDetails(remote_run).show() ### 3.1 Saving the best model -The `remote_run` an object of type [AutoMLRun](https://docs.microsoft.com/en-us/python/api/azureml-train-automl-client/azureml.train.automl.run.automlrun?view=azure-ml-py). This object contains the method `get_output()` which returns the best run and the corresponding fitted model. +The `remote_run` an object of type [AutoMLRun](https://docs.microsoft.com/en-us/python/api/azureml-train-automl-client/azureml.train.automl.run.automlrun?ocid=AID3041109). This object contains the method `get_output()` which returns the best run and the corresponding fitted model. ```python best_run, fitted_model = remote_run.get_output() ``` -You can see the parameters used for the best model by just printing the fitted_model and see the properties of the best model by using the [get_properties()](https://docs.microsoft.com/en-us/python/api/azureml-core/azureml.core.run(class)?view=azure-ml-py#azureml_core_Run_get_properties) method. +You can see the parameters used for the best model by just printing the fitted_model and see the properties of the best model by using the [get_properties()](https://docs.microsoft.com/en-us/python/api/azureml-core/azureml.core.run(class)?view=azure-ml-py#azureml_core_Run_get_properties?ocid=AID3041109) method. ```python best_run.get_properties() ``` -Now register the model with the [register_model](https://docs.microsoft.com/en-us/python/api/azureml-train-automl-client/azureml.train.automl.run.automlrun?view=azure-ml-py#register-model-model-name-none--description-none--tags-none--iteration-none--metric-none-) method. +Now register the model with the [register_model](https://docs.microsoft.com/en-us/python/api/azureml-train-automl-client/azureml.train.automl.run.automlrun?view=azure-ml-py#register-model-model-name-none--description-none--tags-none--iteration-none--metric-none-?ocid=AID3041109) method. ```python model_name = best_run.properties['model_name'] script_file_name = 'inference/score.py' @@ -231,9 +231,9 @@ model = best_run.register_model(model_name = model_name, ``` ### 3.2 Model Deployment -Once the best model is saved, we can deploy it with the [InferenceConfig](https://docs.microsoft.com/en-us/python/api/azureml-core/azureml.core.model.inferenceconfig?view=azure-ml-py) class. InferenceConfig represents the configuration settings for a custom environment used for deployment. The [AciWebservice](https://docs.microsoft.com/en-us/python/api/azureml-core/azureml.core.webservice.aciwebservice?view=azure-ml-py) class represents a machine learning model deployed as a web service endpoint on Azure Container Instances. A deployed service is created from a model, script, and associated files. The resulting web service is a load-balanced, HTTP endpoint with a REST API. You can send data to this API and receive the prediction returned by the model. +Once the best model is saved, we can deploy it with the [InferenceConfig](https://docs.microsoft.com/en-us/python/api/azureml-core/azureml.core.model.inferenceconfig?view=azure-ml-py?ocid=AID3041109) class. InferenceConfig represents the configuration settings for a custom environment used for deployment. The [AciWebservice](https://docs.microsoft.com/en-us/python/api/azureml-core/azureml.core.webservice.aciwebservice?view=azure-ml-py) class represents a machine learning model deployed as a web service endpoint on Azure Container Instances. A deployed service is created from a model, script, and associated files. The resulting web service is a load-balanced, HTTP endpoint with a REST API. You can send data to this API and receive the prediction returned by the model. -The model is deployed using the [deploy](https://docs.microsoft.com/en-us/python/api/azureml-core/azureml.core.model(class)?view=azure-ml-py#deploy-workspace--name--models--inference-config-none--deployment-config-none--deployment-target-none--overwrite-false--show-output-false-) method. +The model is deployed using the [deploy](https://docs.microsoft.com/en-us/python/api/azureml-core/azureml.core.model(class)?view=azure-ml-py#deploy-workspace--name--models--inference-config-none--deployment-config-none--deployment-target-none--overwrite-false--show-output-false-?ocid=AID3041109) method. ```python from azureml.core.model import InferenceConfig, Model @@ -296,7 +296,7 @@ Congratulations! You just consumed the model deployed and trained on Azure ML wi There are many other things you can do through the SDK, unfortunately, we can not view them all in this lesson. But good news, learning how to skim through the SDK documentation can take you a long way on your own. Have a look at the Azure ML SDK documentation and find the `Pipeline` class that allows you to create pipelines. A Pipeline is a collection of steps which can be executed as a workflow. -**HINT:** Go to the [SDK documentation](https://docs.microsoft.com/en-us/python/api/overview/azure/ml/?view=azure-ml-py) and type keywords in the search bar like "Pipeline". You should have the `azureml.pipeline.core.Pipeline` class in the search results. +**HINT:** Go to the [SDK documentation](https://docs.microsoft.com/en-us/python/api/overview/azure/ml/?view=azure-ml-py?ocid=AID3041109) and type keywords in the search bar like "Pipeline". You should have the `azureml.pipeline.core.Pipeline` class in the search results. ## Post-Lecture Quiz @@ -316,7 +316,7 @@ Congratulations! You just consumed the model deployed and trained on Azure ML wi 3. It can be used throught a Graphical User Interface ## Review & Self Study -In this lesson, you learned how to train, deploy and consume a model to predict heart failure risk with the Azure ML SDK in the cloud. Check this [documentation](https://docs.microsoft.com/en-us/python/api/overview/azure/ml/?view=azure-ml-py) for further information about the Azure ML SDK. Try to create your own model with the Azure ML SDK. +In this lesson, you learned how to train, deploy and consume a model to predict heart failure risk with the Azure ML SDK in the cloud. Check this [documentation](https://docs.microsoft.com/en-us/python/api/overview/azure/ml/?view=azure-ml-py?ocid=AID3041109) for further information about the Azure ML SDK. Try to create your own model with the Azure ML SDK. ## Assignment diff --git a/5-Data-Science-In-Cloud/19-Azure/assignment.md b/5-Data-Science-In-Cloud/19-Azure/assignment.md index cc9b42d..ad1850f 100644 --- a/5-Data-Science-In-Cloud/19-Azure/assignment.md +++ b/5-Data-Science-In-Cloud/19-Azure/assignment.md @@ -2,7 +2,7 @@ ## Instructions -We saw how to use the Azure ML platform to train, deploy and consume a model with the Azure ML SDK. Now look around for some data that you could use to train an other model, deploy it and consume it. You can look for datasets on [Kaggle](https://kaggle.com) and [Azure Open Datasets](https://azure.microsoft.com/en-us/services/open-datasets/catalog/?WT.mc_id=academic-15963-cxa). +We saw how to use the Azure ML platform to train, deploy and consume a model with the Azure ML SDK. Now look around for some data that you could use to train an other model, deploy it and consume it. You can look for datasets on [Kaggle](https://kaggle.com) and [Azure Open Datasets](https://azure.microsoft.com/en-us/services/open-datasets/catalog/?ocid=AID3041109). ## Rubric From 2f0f79239f2dade9b30e49996b4abbb7be9d384b Mon Sep 17 00:00:00 2001 From: Amagash Date: Mon, 13 Sep 2021 11:01:17 +0200 Subject: [PATCH 2/7] update PET ID --- .../17-Introduction/README.md | 8 ++++---- 5-Data-Science-In-Cloud/18-Low-Code/README.md | 8 ++++---- .../18-Low-Code/assignment.md | 2 +- 5-Data-Science-In-Cloud/19-Azure/README.md | 18 +++++++++--------- 5-Data-Science-In-Cloud/19-Azure/assignment.md | 2 +- 5 files changed, 19 insertions(+), 19 deletions(-) diff --git a/5-Data-Science-In-Cloud/17-Introduction/README.md b/5-Data-Science-In-Cloud/17-Introduction/README.md index 93d9ad2..ea50f8f 100644 --- a/5-Data-Science-In-Cloud/17-Introduction/README.md +++ b/5-Data-Science-In-Cloud/17-Introduction/README.md @@ -66,7 +66,7 @@ The steps necessary to create this projects are the following: * Create an output sink and specify the job output * Start the job -To view the full process, check out the [documentation](https://docs.microsoft.com/en-us/azure/stream-analytics/stream-analytics-twitter-sentiment-analysis-trends?ocid=AID3041109). +To view the full process, check out the [documentation](https://docs.microsoft.com/en-us/azure/stream-analytics/stream-analytics-twitter-sentiment-analysis-trends?WT.mc_id=academic-40229-cxa&ocid=AID30411099). ### Scientific papers analysis @@ -75,9 +75,9 @@ Let’s take another example, of a project created by [Dmitry Soshnikov](http:// Dmitry created a tool that analyses COVID papers. By reviewing this project, you will see how you can create a tool that extracts knowledge from scientific papers, gains insights, and get a tool that helps researchers navigate large collections of papers in a meaningful way. Let's see the different steps used for this: -* Extracting and pre-processing information with [Text Analytics for Health](https://docs.microsoft.com/en-us/azure/cognitive-services/text-analytics/how-tos/text-analytics-for-health?ocid=AID3041109) -* Using [Azure ML](https://azure.microsoft.com/services/machine-learning/?ocid=AID3041109) to parallelize the processing -* Storing and querying information with [Cosmos DB](https://azure.microsoft.com/services/cosmos-db/?ocid=AID3041109) +* Extracting and pre-processing information with [Text Analytics for Health](https://docs.microsoft.com/en-us/azure/cognitive-services/text-analytics/how-tos/text-analytics-for-health?WT.mc_id=academic-40229-cxa&ocid=AID3041109) +* Using [Azure ML](https://azure.microsoft.com/services/machine-learning?WT.mc_id=academic-40229-cxa&ocid=AID3041109) to parallelize the processing +* Storing and querying information with [Cosmos DB](https://azure.microsoft.com/services/cosmos-db?WT.mc_id=academic-40229-cxa&ocid=AID3041109) * Create an interactive dashboard for data exploration and visualization using Power BI To see the full process, visit [Dmitry’s blog](https://soshnikov.com/science/analyzing-medical-papers-with-azure-and-text-analytics-for-health/). diff --git a/5-Data-Science-In-Cloud/18-Low-Code/README.md b/5-Data-Science-In-Cloud/18-Low-Code/README.md index 5c82627..46b5c42 100644 --- a/5-Data-Science-In-Cloud/18-Low-Code/README.md +++ b/5-Data-Science-In-Cloud/18-Low-Code/README.md @@ -43,7 +43,7 @@ Table of contents: ### 1.1 What is Azure Machine Learning? Data scientists expend a lot of effort exploring and pre-processing data, and trying various types of model-training algorithms to produce accurate models, which is time consuming, and often makes inefficient use of expensive compute hardware. -[Azure ML](https://docs.microsoft.com/EN-US/azure/machine-learning/overview-what-is-azure-machine-learning?ocid=AID3041109) is a cloud-based platform for building and operating machine learning solutions in Azure. It includes a wide range of features and capabilities that help data scientists prepare data, train models, publish predictive services, and monitor their usage. Most importantly, it helps data scientists increase their efficiency by automating many of the time-consuming tasks associated with training models; and it enables them to use cloud-based compute resources that scale effectively to handle large volumes of data while incurring costs only when actually used. +[Azure ML](https://docs.microsoft.com/EN-US/azure/machine-learning/overview-what-is-azure-machine-learning?WT.mc_id=academic-40229-cxa&ocid=AID3041109) is a cloud-based platform for building and operating machine learning solutions in Azure. It includes a wide range of features and capabilities that help data scientists prepare data, train models, publish predictive services, and monitor their usage. Most importantly, it helps data scientists increase their efficiency by automating many of the time-consuming tasks associated with training models; and it enables them to use cloud-based compute resources that scale effectively to handle large volumes of data while incurring costs only when actually used. Azure ML provides all the tools developers and data scientists need for their machine learning workflows, including: @@ -98,7 +98,7 @@ Once you have the dataset, we can start the project in Azure. ## 2. Low code/No code training of a model in Azure ML Studio ### 2.1 Create an Azure ML workspace -To train a model in Azure ML you first need to create an Azure ML workspace. The workspace is the top-level resource for Azure Machine Learning, providing a centralized place to work with all the artifacts you create when you use Azure Machine Learning. The workspace keeps a history of all training runs, including logs, metrics, output, and a snapshot of your scripts. You use this information to determine which training run produces the best model. [Learn more](https://docs.microsoft.com/en-us/azure/machine-learning/concept-workspace?ocid=AID3041109) +To train a model in Azure ML you first need to create an Azure ML workspace. The workspace is the top-level resource for Azure Machine Learning, providing a centralized place to work with all the artifacts you create when you use Azure Machine Learning. The workspace keeps a history of all training runs, including logs, metrics, output, and a snapshot of your scripts. You use this information to determine which training run produces the best model. [Learn more](https://docs.microsoft.com/en-us/azure/machine-learning/concept-workspace?WT.mc_id=academic-40229-cxa&ocid=AID3041109) It is recommended to use the most up-to-date browser that's compatible with your operating system. The following browsers are supported: @@ -226,7 +226,7 @@ Great now that the dataset is in place and the compute cluster is created, we ca ### 2.4 Low code/No Code training with AutoML Traditional machine learning model development is resource-intensive, requiring significant domain knowledge and time to produce and compare dozens of models. -Automated machine learning (AutoML), is the process of automating the time-consuming, iterative tasks of machine learning model development. It allows data scientists, analysts, and developers to build ML models with high scale, efficiency, and productivity all while sustaining model quality. It greatly accelerates the time it takes to get production-ready ML models with great ease and efficiency. [Learn more](https://docs.microsoft.com/en-us/azure/machine-learning/concept-automated-ml?ocid=AID3041109) +Automated machine learning (AutoML), is the process of automating the time-consuming, iterative tasks of machine learning model development. It allows data scientists, analysts, and developers to build ML models with high scale, efficiency, and productivity all while sustaining model quality. It greatly accelerates the time it takes to get production-ready ML models with great ease and efficiency. [Learn more](https://docs.microsoft.com/en-us/azure/machine-learning/concept-automated-ml?WT.mc_id=academic-40229-cxa&ocid=AID3041109) 1. In the [Azure ML workspace](https://ml.azure.com/) that we created earlier click on "Automated ML" in the left menu and select the dataset you just uploaded. Click Next. @@ -355,7 +355,7 @@ Look more closely at the model explanations and details that AutoML generated fo In this lesson, you learned how to train, deploy and consume a model to predict heart failure risk in a Low code/No code fashion in the cloud. If you have not done it yet, look more closely at the model explanations that AutoML generated for the top models and try to understand why the best model is better than the other ones. -You can go further into Low code/No code AutoML by reading this [documentation](https://docs.microsoft.com/en-us/azure/machine-learning/tutorial-first-experiment-automated-ml?ocid=AID3041109). +You can go further into Low code/No code AutoML by reading this [documentation](https://docs.microsoft.com/en-us/azure/machine-learning/tutorial-first-experiment-automated-ml?WT.mc_id=academic-40229-cxa&ocid=AID3041109). ## Assignment diff --git a/5-Data-Science-In-Cloud/18-Low-Code/assignment.md b/5-Data-Science-In-Cloud/18-Low-Code/assignment.md index c7ddf90..fb359bf 100644 --- a/5-Data-Science-In-Cloud/18-Low-Code/assignment.md +++ b/5-Data-Science-In-Cloud/18-Low-Code/assignment.md @@ -2,7 +2,7 @@ ## Instructions -We saw how to use the Azure ML platform to train, deploy and consume a model in a Low code/No code fashion. Now look around for some data that you could use to train an other model, deploy it and consume it. You can look for datasets on [Kaggle](https://kaggle.com) and [Azure Open Datasets](https://azure.microsoft.com/en-us/services/open-datasets/catalog?ocid=AID3041109). +We saw how to use the Azure ML platform to train, deploy and consume a model in a Low code/No code fashion. Now look around for some data that you could use to train an other model, deploy it and consume it. You can look for datasets on [Kaggle](https://kaggle.com) and [Azure Open Datasets](https://azure.microsoft.com/en-us/services/open-datasets/catalog?WT.mc_id=academic-40229-cxa&ocid=AID3041109). ## Rubric diff --git a/5-Data-Science-In-Cloud/19-Azure/README.md b/5-Data-Science-In-Cloud/19-Azure/README.md index 9dca6cc..13d2096 100644 --- a/5-Data-Science-In-Cloud/19-Azure/README.md +++ b/5-Data-Science-In-Cloud/19-Azure/README.md @@ -54,7 +54,7 @@ Key areas of the SDK include: - Use automated machine learning, which accepts configuration parameters and training data. It automatically iterates through algorithms and hyperparameter settings to find the best model for running predictions. - Deploy web services to convert your trained models into RESTful services that can be consumed in any application. -[Learn more about the Azure Machine Learning SDK](https://docs.microsoft.com/en-us/python/api/overview/azure/ml?ocid=AID3041109) +[Learn more about the Azure Machine Learning SDK](https://docs.microsoft.com/en-us/python/api/overview/azure/ml?WT.mc_id=academic-40229-cxa&ocid=AID3041109) In the [previous lesson](../18-tbd/README.md), we saw how to train, deploy and consume a model in a Low code/No code fashion. We used the Heart Failure dataset to generate and Heart failure prediction model. In this lesson, we are going to do the exact same thing but using the Azure Machine Learning SDK. @@ -107,7 +107,7 @@ Now that we have a Notebook, we can start training the model with Azure ML SDK. ### 2.5 Training a model -First of all, if you ever have a doubt, refer to the [Azure ML SDK documentation](https://docs.microsoft.com/en-us/python/api/overview/azure/ml?ocid=AID3041109). In contains all the necessary information to understand the modules we are going to see in this lesson. +First of all, if you ever have a doubt, refer to the [Azure ML SDK documentation](https://docs.microsoft.com/en-us/python/api/overview/azure/ml?WT.mc_id=academic-40229-cxa&ocid=AID3041109). In contains all the necessary information to understand the modules we are going to see in this lesson. #### 2.5.1 Setup Workspace, experiment, compute cluster and dataset @@ -155,7 +155,7 @@ df.describe() ``` #### 2.5.2 AutoML Configuration and training -To set the AutoML configuration, use the [AutoMLConfig class](https://docs.microsoft.com/en-us/python/api/azureml-train-automl-client/azureml.train.automl.automlconfig(class)?ocid=AID3041109). +To set the AutoML configuration, use the [AutoMLConfig class](https://docs.microsoft.com/en-us/python/api/azureml-train-automl-client/azureml.train.automl.automlconfig(class)?WT.mc_id=academic-40229-cxa&ocid=AID3041109). As described in the doc, there are a lot of parameters with which you can play with. For this project, we will use the following parameters: @@ -207,18 +207,18 @@ RunDetails(remote_run).show() ### 3.1 Saving the best model -The `remote_run` an object of type [AutoMLRun](https://docs.microsoft.com/en-us/python/api/azureml-train-automl-client/azureml.train.automl.run.automlrun?ocid=AID3041109). This object contains the method `get_output()` which returns the best run and the corresponding fitted model. +The `remote_run` an object of type [AutoMLRun](https://docs.microsoft.com/en-us/python/api/azureml-train-automl-client/azureml.train.automl.run.automlrun?WT.mc_id=academic-40229-cxa&ocid=AID3041109). This object contains the method `get_output()` which returns the best run and the corresponding fitted model. ```python best_run, fitted_model = remote_run.get_output() ``` -You can see the parameters used for the best model by just printing the fitted_model and see the properties of the best model by using the [get_properties()](https://docs.microsoft.com/en-us/python/api/azureml-core/azureml.core.run(class)?view=azure-ml-py#azureml_core_Run_get_properties?ocid=AID3041109) method. +You can see the parameters used for the best model by just printing the fitted_model and see the properties of the best model by using the [get_properties()](https://docs.microsoft.com/en-us/python/api/azureml-core/azureml.core.run(class)?view=azure-ml-py#azureml_core_Run_get_properties?WT.mc_id=academic-40229-cxa&ocid=AID3041109) method. ```python best_run.get_properties() ``` -Now register the model with the [register_model](https://docs.microsoft.com/en-us/python/api/azureml-train-automl-client/azureml.train.automl.run.automlrun?view=azure-ml-py#register-model-model-name-none--description-none--tags-none--iteration-none--metric-none-?ocid=AID3041109) method. +Now register the model with the [register_model](https://docs.microsoft.com/en-us/python/api/azureml-train-automl-client/azureml.train.automl.run.automlrun?view=azure-ml-py#register-model-model-name-none--description-none--tags-none--iteration-none--metric-none-?WT.mc_id=academic-40229-cxa&ocid=AID3041109) method. ```python model_name = best_run.properties['model_name'] script_file_name = 'inference/score.py' @@ -233,7 +233,7 @@ model = best_run.register_model(model_name = model_name, Once the best model is saved, we can deploy it with the [InferenceConfig](https://docs.microsoft.com/en-us/python/api/azureml-core/azureml.core.model.inferenceconfig?view=azure-ml-py?ocid=AID3041109) class. InferenceConfig represents the configuration settings for a custom environment used for deployment. The [AciWebservice](https://docs.microsoft.com/en-us/python/api/azureml-core/azureml.core.webservice.aciwebservice?view=azure-ml-py) class represents a machine learning model deployed as a web service endpoint on Azure Container Instances. A deployed service is created from a model, script, and associated files. The resulting web service is a load-balanced, HTTP endpoint with a REST API. You can send data to this API and receive the prediction returned by the model. -The model is deployed using the [deploy](https://docs.microsoft.com/en-us/python/api/azureml-core/azureml.core.model(class)?view=azure-ml-py#deploy-workspace--name--models--inference-config-none--deployment-config-none--deployment-target-none--overwrite-false--show-output-false-?ocid=AID3041109) method. +The model is deployed using the [deploy](https://docs.microsoft.com/en-us/python/api/azureml-core/azureml.core.model(class)?view=azure-ml-py#deploy-workspace--name--models--inference-config-none--deployment-config-none--deployment-target-none--overwrite-false--show-output-false-?WT.mc_id=academic-40229-cxa&ocid=AID3041109) method. ```python from azureml.core.model import InferenceConfig, Model @@ -296,7 +296,7 @@ Congratulations! You just consumed the model deployed and trained on Azure ML wi There are many other things you can do through the SDK, unfortunately, we can not view them all in this lesson. But good news, learning how to skim through the SDK documentation can take you a long way on your own. Have a look at the Azure ML SDK documentation and find the `Pipeline` class that allows you to create pipelines. A Pipeline is a collection of steps which can be executed as a workflow. -**HINT:** Go to the [SDK documentation](https://docs.microsoft.com/en-us/python/api/overview/azure/ml/?view=azure-ml-py?ocid=AID3041109) and type keywords in the search bar like "Pipeline". You should have the `azureml.pipeline.core.Pipeline` class in the search results. +**HINT:** Go to the [SDK documentation](https://docs.microsoft.com/en-us/python/api/overview/azure/ml/?view=azure-ml-py?WT.mc_id=academic-40229-cxa&ocid=AID3041109) and type keywords in the search bar like "Pipeline". You should have the `azureml.pipeline.core.Pipeline` class in the search results. ## Post-Lecture Quiz @@ -316,7 +316,7 @@ Congratulations! You just consumed the model deployed and trained on Azure ML wi 3. It can be used throught a Graphical User Interface ## Review & Self Study -In this lesson, you learned how to train, deploy and consume a model to predict heart failure risk with the Azure ML SDK in the cloud. Check this [documentation](https://docs.microsoft.com/en-us/python/api/overview/azure/ml/?view=azure-ml-py?ocid=AID3041109) for further information about the Azure ML SDK. Try to create your own model with the Azure ML SDK. +In this lesson, you learned how to train, deploy and consume a model to predict heart failure risk with the Azure ML SDK in the cloud. Check this [documentation](https://docs.microsoft.com/en-us/python/api/overview/azure/ml/?view=azure-ml-py?WT.mc_id=academic-40229-cxa&ocid=AID3041109) for further information about the Azure ML SDK. Try to create your own model with the Azure ML SDK. ## Assignment diff --git a/5-Data-Science-In-Cloud/19-Azure/assignment.md b/5-Data-Science-In-Cloud/19-Azure/assignment.md index ad1850f..3985315 100644 --- a/5-Data-Science-In-Cloud/19-Azure/assignment.md +++ b/5-Data-Science-In-Cloud/19-Azure/assignment.md @@ -2,7 +2,7 @@ ## Instructions -We saw how to use the Azure ML platform to train, deploy and consume a model with the Azure ML SDK. Now look around for some data that you could use to train an other model, deploy it and consume it. You can look for datasets on [Kaggle](https://kaggle.com) and [Azure Open Datasets](https://azure.microsoft.com/en-us/services/open-datasets/catalog/?ocid=AID3041109). +We saw how to use the Azure ML platform to train, deploy and consume a model with the Azure ML SDK. Now look around for some data that you could use to train an other model, deploy it and consume it. You can look for datasets on [Kaggle](https://kaggle.com) and [Azure Open Datasets](https://azure.microsoft.com/en-us/services/open-datasets/catalog?WT.mc_id=academic-40229-cxa&ocid=AID3041109). ## Rubric From c43c64d36ef94858c266014320e73395520c21eb Mon Sep 17 00:00:00 2001 From: Ghost911-s Date: Tue, 14 Sep 2021 13:44:05 +0200 Subject: [PATCH 3/7] minor typos --- 4-Data-Science-Lifecycle/14-Introduction/README.md | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/4-Data-Science-Lifecycle/14-Introduction/README.md b/4-Data-Science-Lifecycle/14-Introduction/README.md index 24d8626..92cb75b 100644 --- a/4-Data-Science-Lifecycle/14-Introduction/README.md +++ b/4-Data-Science-Lifecycle/14-Introduction/README.md @@ -1,6 +1,6 @@ # Introduction to the Data Science Lifecycle -At this point you've probably come to the realization that that data science is a process. This process can be broken down into 5 stages: +At this point you've probably come to the realization that data science is a process. This process can be broken down into 5 stages: - Capturing - Processing @@ -17,7 +17,7 @@ This lesson focuses on 3 parts of the life cycle: capturing, processing and main ## Capturing The first stage of the lifecycle is very important as the next stages are dependent on it. It’s practically two stages combined into one: acquiring the data and defining the purpose and problems that need to be addressed. -Defining the goals of the project will require deeper context into the problem or question. First, we need to identify and acquire those who need their problem solved. These may be stakeholders in a business or sponsors of the project who can help identify who or what will benefit from this project as well as what, and why they need it. A well-defined goal should be measurable and quantifiable to define an acceptable result. +Defining the goals of the project will require deeper context into the problem or question. First, we need to identify and acquire those who need their problem solved. These may be stakeholders in a business or sponsors of the project, who can help identify who or what will benefit from this project as well as what, and why they need it. A well-defined goal should be measurable and quantifiable to define an acceptable result. Questions a data scientist may ask: - Has this problem been approached before? What was discovered? @@ -39,7 +39,7 @@ Questions a data scientist may ask about the data: ## Processing -The processing stage of the lifecycle focuses on discovering patterns in the data as well as modeling. Some techniques used to in the processing stage requires statistical methods to uncover the patterns. Typically, this would be a tedious task for a human to do with a large data set and will rely on computers to do the heavy lifting to speed up the process. This stage is also where data science and machine learning will intersect. As you learned in the first lesson, machine learning is the process of building models to understand the data. Models are a representation of the relationship between variables in the data that help predict outcomes. +The processing stage of the lifecycle focuses on discovering patterns in the data as well as modeling. Some techniques used in the processing stage require statistical methods to uncover the patterns. Typically, this would be a tedious task for a human to do with a large data set and will rely on computers to do the heavy lifting to speed up the process. This stage is also where data science and machine learning will intersect. As you learned in the first lesson, machine learning is the process of building models to understand the data. Models are a representation of the relationship between variables in the data that help predict outcomes. Common techniques used in this stage are covered in the ML for Beginners curriculum. Follow the links to learn more about them: From 7d8bb5f72ceebece4e2e76cd0d14580272c11b5c Mon Sep 17 00:00:00 2001 From: Rohit Yadav Date: Wed, 15 Sep 2021 02:14:52 +0530 Subject: [PATCH 4/7] Update case for tables --- .../05-relational-databases/README.md | 12 ++++++------ 1 file changed, 6 insertions(+), 6 deletions(-) diff --git a/2-Working-With-Data/05-relational-databases/README.md b/2-Working-With-Data/05-relational-databases/README.md index b72d7e1..9dbb81a 100644 --- a/2-Working-With-Data/05-relational-databases/README.md +++ b/2-Working-With-Data/05-relational-databases/README.md @@ -10,7 +10,7 @@ A relational database has at its core tables. Just as with the spreadsheet, a ta Let's begin our exploration by starting a table to store information about cities. We might start with their name and country. You could store this in a table as follows: -| city | country | +| City | Country | | -------- | ------------- | | Tokyo | Japan | | Atlanta | United States | @@ -22,7 +22,7 @@ Notice the column names of **city**, **country** and **population** to describe Chances are, the table above seems relatively familiar to you. Let's start to add some additional data to our burgeoning database - annual rainfall (in millimeters). We'll focus on the years 2018, 2019 and 2020. If we were to add it for Tokyo, it might look something like this: -| city | country | year | amount | +| City | Country | Year | Amount | | ----- | ------- | ---- | ------ | | Tokyo | Japan | 2020 | 1690 | | Tokyo | Japan | 2019 | 1874 | @@ -32,7 +32,7 @@ What do you notice about our table? You might notice we're duplicating the name OK, let's try something else. Let's add new columns for each year: -| city | country | 2018 | 2019 | 2020 | +| City | Country | 2018 | 2019 | 2020 | | -------- | ------------- | ---- | ---- | ---- | | Tokyo | Japan | 1445 | 1874 | 1690 | | Atlanta | United States | 1779 | 1111 | 1683 | @@ -46,7 +46,7 @@ This is why we need multiple tables and relationships. By breaking apart our dat Let's return to our data and determine how we want to split things up. We know we want to store the name and country for our cities, so this will probably work best in one table. -| city | country | +| City | Country | | -------- | ------------- | | Tokyo | Japan | | Atlanta | United States | @@ -56,7 +56,7 @@ But before we create the next table, we need to figure out how to reference each ### cities -| city_id | city | country | +| city_id | City | Country | | ------- | -------- | ------------- | | 1 | Tokyo | Japan | | 2 | Atlanta | United States | @@ -68,7 +68,7 @@ With our cities table created, let's store the rainfall. Rather than duplicating ### rainfall -| rainfall_id | city_id | year | amount | +| rainfall_id | city_id | Year | Amount | | ----------- | ------- | ---- | ------ | | 1 | 1 | 2018 | 1445 | | 2 | 1 | 2019 | 1874 | From 98914b7a27c90847db188f90c4382bae23c3b616 Mon Sep 17 00:00:00 2001 From: Amagash Date: Wed, 15 Sep 2021 01:48:28 +0200 Subject: [PATCH 5/7] removed localisation string --- 5-Data-Science-In-Cloud/17-Introduction/README.md | 4 ++-- 5-Data-Science-In-Cloud/18-Low-Code/README.md | 8 ++++---- 5-Data-Science-In-Cloud/18-Low-Code/assignment.md | 2 +- 5-Data-Science-In-Cloud/19-Azure/assignment.md | 2 +- 5-Data-Science-In-Cloud/19-Azure/notebook.ipynb | 2 +- 5 files changed, 9 insertions(+), 9 deletions(-) diff --git a/5-Data-Science-In-Cloud/17-Introduction/README.md b/5-Data-Science-In-Cloud/17-Introduction/README.md index ea50f8f..9fc4900 100644 --- a/5-Data-Science-In-Cloud/17-Introduction/README.md +++ b/5-Data-Science-In-Cloud/17-Introduction/README.md @@ -66,7 +66,7 @@ The steps necessary to create this projects are the following: * Create an output sink and specify the job output * Start the job -To view the full process, check out the [documentation](https://docs.microsoft.com/en-us/azure/stream-analytics/stream-analytics-twitter-sentiment-analysis-trends?WT.mc_id=academic-40229-cxa&ocid=AID30411099). +To view the full process, check out the [documentation](https://docs.microsoft.com/azure/stream-analytics/stream-analytics-twitter-sentiment-analysis-trends?WT.mc_id=academic-40229-cxa&ocid=AID30411099). ### Scientific papers analysis @@ -75,7 +75,7 @@ Let’s take another example, of a project created by [Dmitry Soshnikov](http:// Dmitry created a tool that analyses COVID papers. By reviewing this project, you will see how you can create a tool that extracts knowledge from scientific papers, gains insights, and get a tool that helps researchers navigate large collections of papers in a meaningful way. Let's see the different steps used for this: -* Extracting and pre-processing information with [Text Analytics for Health](https://docs.microsoft.com/en-us/azure/cognitive-services/text-analytics/how-tos/text-analytics-for-health?WT.mc_id=academic-40229-cxa&ocid=AID3041109) +* Extracting and pre-processing information with [Text Analytics for Health](https://docs.microsoft.com/azure/cognitive-services/text-analytics/how-tos/text-analytics-for-health?WT.mc_id=academic-40229-cxa&ocid=AID3041109) * Using [Azure ML](https://azure.microsoft.com/services/machine-learning?WT.mc_id=academic-40229-cxa&ocid=AID3041109) to parallelize the processing * Storing and querying information with [Cosmos DB](https://azure.microsoft.com/services/cosmos-db?WT.mc_id=academic-40229-cxa&ocid=AID3041109) * Create an interactive dashboard for data exploration and visualization using Power BI diff --git a/5-Data-Science-In-Cloud/18-Low-Code/README.md b/5-Data-Science-In-Cloud/18-Low-Code/README.md index 46b5c42..be95ef2 100644 --- a/5-Data-Science-In-Cloud/18-Low-Code/README.md +++ b/5-Data-Science-In-Cloud/18-Low-Code/README.md @@ -43,7 +43,7 @@ Table of contents: ### 1.1 What is Azure Machine Learning? Data scientists expend a lot of effort exploring and pre-processing data, and trying various types of model-training algorithms to produce accurate models, which is time consuming, and often makes inefficient use of expensive compute hardware. -[Azure ML](https://docs.microsoft.com/EN-US/azure/machine-learning/overview-what-is-azure-machine-learning?WT.mc_id=academic-40229-cxa&ocid=AID3041109) is a cloud-based platform for building and operating machine learning solutions in Azure. It includes a wide range of features and capabilities that help data scientists prepare data, train models, publish predictive services, and monitor their usage. Most importantly, it helps data scientists increase their efficiency by automating many of the time-consuming tasks associated with training models; and it enables them to use cloud-based compute resources that scale effectively to handle large volumes of data while incurring costs only when actually used. +[Azure ML](https://docs.microsoft.com/azure/machine-learning/overview-what-is-azure-machine-learning?WT.mc_id=academic-40229-cxa&ocid=AID3041109) is a cloud-based platform for building and operating machine learning solutions in Azure. It includes a wide range of features and capabilities that help data scientists prepare data, train models, publish predictive services, and monitor their usage. Most importantly, it helps data scientists increase their efficiency by automating many of the time-consuming tasks associated with training models; and it enables them to use cloud-based compute resources that scale effectively to handle large volumes of data while incurring costs only when actually used. Azure ML provides all the tools developers and data scientists need for their machine learning workflows, including: @@ -98,7 +98,7 @@ Once you have the dataset, we can start the project in Azure. ## 2. Low code/No code training of a model in Azure ML Studio ### 2.1 Create an Azure ML workspace -To train a model in Azure ML you first need to create an Azure ML workspace. The workspace is the top-level resource for Azure Machine Learning, providing a centralized place to work with all the artifacts you create when you use Azure Machine Learning. The workspace keeps a history of all training runs, including logs, metrics, output, and a snapshot of your scripts. You use this information to determine which training run produces the best model. [Learn more](https://docs.microsoft.com/en-us/azure/machine-learning/concept-workspace?WT.mc_id=academic-40229-cxa&ocid=AID3041109) +To train a model in Azure ML you first need to create an Azure ML workspace. The workspace is the top-level resource for Azure Machine Learning, providing a centralized place to work with all the artifacts you create when you use Azure Machine Learning. The workspace keeps a history of all training runs, including logs, metrics, output, and a snapshot of your scripts. You use this information to determine which training run produces the best model. [Learn more](https://docs.microsoft.com/azure/machine-learning/concept-workspace?WT.mc_id=academic-40229-cxa&ocid=AID3041109) It is recommended to use the most up-to-date browser that's compatible with your operating system. The following browsers are supported: @@ -226,7 +226,7 @@ Great now that the dataset is in place and the compute cluster is created, we ca ### 2.4 Low code/No Code training with AutoML Traditional machine learning model development is resource-intensive, requiring significant domain knowledge and time to produce and compare dozens of models. -Automated machine learning (AutoML), is the process of automating the time-consuming, iterative tasks of machine learning model development. It allows data scientists, analysts, and developers to build ML models with high scale, efficiency, and productivity all while sustaining model quality. It greatly accelerates the time it takes to get production-ready ML models with great ease and efficiency. [Learn more](https://docs.microsoft.com/en-us/azure/machine-learning/concept-automated-ml?WT.mc_id=academic-40229-cxa&ocid=AID3041109) +Automated machine learning (AutoML), is the process of automating the time-consuming, iterative tasks of machine learning model development. It allows data scientists, analysts, and developers to build ML models with high scale, efficiency, and productivity all while sustaining model quality. It greatly accelerates the time it takes to get production-ready ML models with great ease and efficiency. [Learn more](https://docs.microsoft.com/azure/machine-learning/concept-automated-ml?WT.mc_id=academic-40229-cxa&ocid=AID3041109) 1. In the [Azure ML workspace](https://ml.azure.com/) that we created earlier click on "Automated ML" in the left menu and select the dataset you just uploaded. Click Next. @@ -355,7 +355,7 @@ Look more closely at the model explanations and details that AutoML generated fo In this lesson, you learned how to train, deploy and consume a model to predict heart failure risk in a Low code/No code fashion in the cloud. If you have not done it yet, look more closely at the model explanations that AutoML generated for the top models and try to understand why the best model is better than the other ones. -You can go further into Low code/No code AutoML by reading this [documentation](https://docs.microsoft.com/en-us/azure/machine-learning/tutorial-first-experiment-automated-ml?WT.mc_id=academic-40229-cxa&ocid=AID3041109). +You can go further into Low code/No code AutoML by reading this [documentation](https://docs.microsoft.com/azure/machine-learning/tutorial-first-experiment-automated-ml?WT.mc_id=academic-40229-cxa&ocid=AID3041109). ## Assignment diff --git a/5-Data-Science-In-Cloud/18-Low-Code/assignment.md b/5-Data-Science-In-Cloud/18-Low-Code/assignment.md index fb359bf..afaba4d 100644 --- a/5-Data-Science-In-Cloud/18-Low-Code/assignment.md +++ b/5-Data-Science-In-Cloud/18-Low-Code/assignment.md @@ -2,7 +2,7 @@ ## Instructions -We saw how to use the Azure ML platform to train, deploy and consume a model in a Low code/No code fashion. Now look around for some data that you could use to train an other model, deploy it and consume it. You can look for datasets on [Kaggle](https://kaggle.com) and [Azure Open Datasets](https://azure.microsoft.com/en-us/services/open-datasets/catalog?WT.mc_id=academic-40229-cxa&ocid=AID3041109). +We saw how to use the Azure ML platform to train, deploy and consume a model in a Low code/No code fashion. Now look around for some data that you could use to train an other model, deploy it and consume it. You can look for datasets on [Kaggle](https://kaggle.com) and [Azure Open Datasets](https://azure.microsoft.com/services/open-datasets/catalog?WT.mc_id=academic-40229-cxa&ocid=AID3041109). ## Rubric diff --git a/5-Data-Science-In-Cloud/19-Azure/assignment.md b/5-Data-Science-In-Cloud/19-Azure/assignment.md index 3985315..bdd1864 100644 --- a/5-Data-Science-In-Cloud/19-Azure/assignment.md +++ b/5-Data-Science-In-Cloud/19-Azure/assignment.md @@ -2,7 +2,7 @@ ## Instructions -We saw how to use the Azure ML platform to train, deploy and consume a model with the Azure ML SDK. Now look around for some data that you could use to train an other model, deploy it and consume it. You can look for datasets on [Kaggle](https://kaggle.com) and [Azure Open Datasets](https://azure.microsoft.com/en-us/services/open-datasets/catalog?WT.mc_id=academic-40229-cxa&ocid=AID3041109). +We saw how to use the Azure ML platform to train, deploy and consume a model with the Azure ML SDK. Now look around for some data that you could use to train an other model, deploy it and consume it. You can look for datasets on [Kaggle](https://kaggle.com) and [Azure Open Datasets](https://azure.microsoft.com/services/open-datasets/catalog?WT.mc_id=academic-40229-cxa&ocid=AID3041109). ## Rubric diff --git a/5-Data-Science-In-Cloud/19-Azure/notebook.ipynb b/5-Data-Science-In-Cloud/19-Azure/notebook.ipynb index 8a068aa..03eb893 100644 --- a/5-Data-Science-In-Cloud/19-Azure/notebook.ipynb +++ b/5-Data-Science-In-Cloud/19-Azure/notebook.ipynb @@ -86,7 +86,7 @@ "cell_type": "markdown", "source": [ "## Create a Compute Cluster\n", - "You will need to create a [compute target](https://docs.microsoft.com/en-us/azure/machine-learning/concept-azure-machine-learning-architecture#compute-target) for your AutoML run." + "You will need to create a [compute target](https://docs.microsoft.com/azure/machine-learning/concept-azure-machine-learning-architecture#compute-target) for your AutoML run." ], "metadata": {} }, From e33c1847edabbfa1ce4dd08a3131214a51ed5d0c Mon Sep 17 00:00:00 2001 From: Anupam Mishra <66557767+anupamishra333@users.noreply.github.com> Date: Wed, 15 Sep 2021 19:11:15 +0530 Subject: [PATCH 6/7] Update README.md --- README.md | 20 ++++++++++---------- 1 file changed, 10 insertions(+), 10 deletions(-) diff --git a/README.md b/README.md index 84a6e8c..ba8bd13 100644 --- a/README.md +++ b/README.md @@ -25,16 +25,16 @@ In addition, a low-stakes quiz before a class sets the intention of the student ## Each lesson includes: -- optional sketchnote -- optional supplemental video -- pre-lesson warmup quiz -- written lesson -- for project-based lessons, step-by-step guides on how to build the project -- knowledge checks -- a challenge -- supplemental reading -- assignment -- post-lesson quiz +- Optional sketchnote +- Optional supplemental video +- Pre-lesson warmup quiz +- Written lesson +- For project-based lessons, step-by-step guides on how to build the project +- Knowledge checks +- A challenge +- Supplemental reading +- Assignment +- Post-lesson quiz > **A note about quizzes**: All quizzes are contained [in this app](https://red-water-0103e7a0f.azurestaticapps.net/), for 40 total quizzes of three questions each. They are linked from within the lessons but the quiz app can be run locally; follow the instruction in the `quiz-app` folder. They are gradually being localized. From 45a29177a3a89bf7e316e9c460d5ad5cd2eeb0f7 Mon Sep 17 00:00:00 2001 From: Jalen McGee <87670464+JalenMcG@users.noreply.github.com> Date: Wed, 15 Sep 2021 14:23:27 -0400 Subject: [PATCH 7/7] Inserted the link to the communication lesson --- 4-Data-Science-Lifecycle/README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/4-Data-Science-Lifecycle/README.md b/4-Data-Science-Lifecycle/README.md index 384fadf..c17bc4e 100644 --- a/4-Data-Science-Lifecycle/README.md +++ b/4-Data-Science-Lifecycle/README.md @@ -8,7 +8,7 @@ In these lessons, you'll explore some of the aspects of the Data Science lifeycl 1. [Introduction](14-Introduction/README.md) 2. [Analyzing](15-Analyzing/README.md) -3. [Communication](16-Communication/README.md) +3. [Communication](https://github.com/microsoft/Data-Science-For-Beginners/tree/main/4-Data-Science-Lifecycle/16-communication) ### Credits