You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
* updated sample.json and Teacherkit.ipynb
* remove ANSWERS
* update sample.json w pip install -r requirements.txt
* fixes to bracnh bc-update: fix broken links, consolodate data to common data folder, answers in pw protected zip
* fix broken links lower level READMEs
* Update sample.json
replace sample.json with Piotr's updates last month
* Update sample.json
Co-authored-by: Jimmy Wei <[email protected]>
Copy file name to clipboardExpand all lines: AI-and-Analytics/Jupyter/Introduction_to_Machine_Learning/01_Introduction_to_Machine_Learning_and_Tools/Introduction_to_Machine_Learning_and_Toolkit-ANSWERS.ipynb
Copy file name to clipboardExpand all lines: AI-and-Analytics/Jupyter/Introduction_to_Machine_Learning/01_Introduction_to_Machine_Learning_and_Tools/Introduction_to_Machine_Learning_and_Toolkit.ipynb
Copy file name to clipboardExpand all lines: AI-and-Analytics/Jupyter/Introduction_to_Machine_Learning/01_Introduction_to_Machine_Learning_and_Tools/README.md
+7-45Lines changed: 7 additions & 45 deletions
Original file line number
Diff line number
Diff line change
@@ -19,8 +19,13 @@ At the end of this course you will be able to:
19
19
- Demonstrate supervised learning algorithms
20
20
- Explain key concepts like under- and over-fitting, regularization, and cross-validation
21
21
- Classify the type of problem to be solved, choose the right algorithm, tune parameters, and validate a model
22
-
- Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware
23
-
- Download and Install Intel® Extension for Scikit-learn* powered by oneAPI
22
+
- Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware using two lines of code
23
+
```python
24
+
from sklearnex import patch_sklearn
25
+
patch_sklearn()
26
+
```
27
+
- Reference describing: Download and Install Intel® Extension for Scikit-learn* powered by oneAPI
28
+
24
29
25
30
26
31
## License
@@ -38,46 +43,3 @@ Third party program Licenses can be found here: [third-party-programs.txt](https
38
43
- Statistics
39
44
40
45
41
-
## Syllabus
42
-
43
-
- 11 Modules (18 hours)
44
-
- 11 Lab Exercises
45
-
46
-
-----------------------
47
-
| Modules | Description | Duration |
48
-
| :--- | :------ | :------ |
49
-
|[01_Introduction_to_Machine_Learning_and_Tools](01_Introduction_to_Machine_Learning_and_Tools/Introduction_to_Machine_Learning_and_Toolkit.ipynb)| + Demonstrate supervised learning algorithms.<br> + Explain key concepts like under- and over-fitting, regularization, and cross-validation<br>: + Classify the type of problem to be solved <br> + Choose an algorithm, tune parameters, and validate a model <br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 60 min |
50
-
|[02-Introduction_to_Supervised_Learning_and_K_Nearest_Neighbors](02-Introduction_to_Supervised_Learning_and_K_Nearest_Neighbors/Supervised_Learning_and_K_Nearest_Neighbors_Exercises.ipynb)| + Explain supervised learning and how it can be applied to regression and classification problems.<br> + Apply K-Nearest Neighbor (KNN) algorithm for classification <br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 120 min |
51
-
|[03-Train_Test_Splits_Validation_Linear_Regression](03-Train_Test_Splits_Validation_Linear_Regression/Train_Test_Splits_Validation_Linear_Regression.ipynb)| + Explain the difference between over-fitting and under-fitting a model<br> + Describe Bias-variance tradeoffs<br> + Find the optimal training and test data set splits, cross-validation, and model complexity versus error<br> + Apply a linear regression model for supervised learning.<br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 120 min |
52
-
|[04-Regularization_and_Gradient_Descent](04-Regularization_and_Gradient_Descent/Regularization_and_Gradient_Descent_Exercises.ipynb)| + Describe Logistic regression and how it differs from linear regression<br> + Identify metrics for classification errors and scenarios in which they can be used<br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 120 min |
53
-
|[05-Logistic_Regression_and_Classification_Error_Metrics](05-Logistic_Regression_and_Classification_Error_Metrics/Logistic_Regression_and_Classification_Error_Metrics_Exercises.ipynb)| + Explain cost functions, regularization, feature selection, and hyper-parameters<br> + | 120 min |
54
-
|[06-SVM_and_Kernels](06-SVM_and_Kernels/SVM_Kernels_Exercises.ipynb)| + Apply support vector machines (SVMs)—a popular algorithm used for classification problems<br> + Recognize SVM similarity to logistic regression<br> + Compute the cost function of SVMs<br> + Apply regularization in SVMs and some tips to obtain non-linear classifications with SVMs<nbr> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware | 120 min |
55
-
|[07-Decision_Trees](07-Decision_Trees/Decision_Trees_Exercises.ipynb)| + Recognize Decision trees and how to use them for classification problems<br> + Recognize how to identify the best split and the factors for splitting<br> + Explain strengths and weaknesses of decision trees<br> + Explain how regression trees help with classifying continuous values<br> + Describe motivation for choosing Random Forest Classifier over Decision Trees<br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware for Random Forest Classifier| 120 min |
56
-
|[08-Bagging](08-Bagging/Bagging_Exercises.ipynbb)| + Associate concepts of bootstrapping and aggregating (commonly known as “bagging”) to reduce variance<br> + Apply Random Forest algorithm that further reduces the correlation seen in bagging models<br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 120 min |
57
-
|[09-Boosting_and_Stacking](09-Boosting_and_Stacking/Boosting_and_Stacking_Exercises.ipynb)| + Explain how the boosting algorithm helps reduce variance and bias.<br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 120 min |
58
-
|[10-Introduction_to_Unsupervised_Learning_and_Clustering_Methods](10-Introduction_to_Unsupervised_Learning_and_Clustering_Methods/Clustering_Methods_Exercises.ipynb)| + Discuss unsupervised learning algorithms and how they can be applied<br> + Apply clustering<br> + Apply dimensionality reduction<br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 120 min |
59
-
|[11-Dimensionality_Reduction_and_Advanced_Topics](11-Dimensionality_Reduction_and_Advanced_Topics/Dimensionality_Reduction_Exercises.ipynb)| + Explain and Apply Principal Component Analysis (PCA)<br> + Explain Multidimensional Scaling (MDS)<br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 120 min |
60
-
61
-
#### Content Structure
62
-
63
-
Each module folder has a Jupyter Notebook file (`*.ipynb`), this can be opened in Jupyter Lab to view the training contant, edit code and compile/run.
64
-
65
-
## Install Directions
66
-
67
-
The training content can be accessed locally on the computer after installing necessary tools, or you can directly access using Intel DevCloud without any installation.
68
-
69
-
#### Local Installation of JupyterLab and oneAPI Tools
70
-
71
-
The Jupyter Notebooks can be downloaded locally to computer and accessed:
72
-
- Install Jupyter Lab on local computer: [Installation Guide](https://jupyterlab.readthedocs.io/en/stable/getting_started/installation.html)
73
-
- Install Intel oneAPI Base Toolkit on local computer: [Installation Guide](https://www.intel.com/content/www/us/en/developer/tools/oneapi/base-toolkit-download.html)
74
-
- git clone the repo and access the Notebooks using Jupyter Lab
75
-
76
-
77
-
#### Access using Intel DevCloud
78
-
79
-
The Jupyter notebooks are tested and can be run on Intel DevCloud without any installation necessary, below are the steps to access these Jupyter notebooks on Intel DevCloud:
80
-
1. Register on [Intel DevCloud](https://devcloud.intel.com/oneapi)
81
-
2. Login, Get Started and Launch Jupyter Lab
82
-
3. Open Terminal in Jupyter Lab and git clone the repo and access the Notebooks
Copy file name to clipboardExpand all lines: AI-and-Analytics/Jupyter/Introduction_to_Machine_Learning/02-Supervised_Learning_and_K_Nearest_Neighbors/README.md
+5-43Lines changed: 5 additions & 43 deletions
Original file line number
Diff line number
Diff line change
@@ -18,7 +18,11 @@ At the end of this course you will be able to:
18
18
19
19
- Explain supervised learning and how it can be applied to regression and classification problems
20
20
- Apply K-Nearest Neighbor (KNN) algorithm for classification
21
-
- Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware
21
+
- Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware using two lines of code
22
+
```python
23
+
from sklearnex import patch_sklearn
24
+
patch_sklearn()
25
+
```
22
26
23
27
## License
24
28
Code samples
@@ -35,46 +39,4 @@ Third party program Licenses can be found here: [third-party-programs.txt](https
35
39
- Statistics
36
40
37
41
38
-
## Syllabus
39
-
40
-
- 11 Modules (18 hours)
41
-
- 11 Lab Exercises
42
-
43
-
-----------------------
44
-
| Modules | Description | Duration |
45
-
| :--- | :------ | :------ |
46
-
|[01_Introduction_to_Machine_Learning_and_Tools](01_Introduction_to_Machine_Learning_and_Tools/Introduction_to_Machine_Learning_and_Toolkit.ipynb)| + Demonstrate supervised learning algorithms.<br> + Explain key concepts like under- and over-fitting, regularization, and cross-validation<br>: + Classify the type of problem to be solved <br> + Choose an algorithm, tune parameters, and validate a model <br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 60 min |
47
-
|[02-Introduction_to_Supervised_Learning_and_K_Nearest_Neighbors](02-Introduction_to_Supervised_Learning_and_K_Nearest_Neighbors/Supervised_Learning_and_K_Nearest_Neighbors_Exercises.ipynb)| + Explain supervised learning and how it can be applied to regression and classification problems.<br> + Apply K-Nearest Neighbor (KNN) algorithm for classification <br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 120 min |
48
-
|[03-Train_Test_Splits_Validation_Linear_Regression](03-Train_Test_Splits_Validation_Linear_Regression/Train_Test_Splits_Validation_Linear_Regression.ipynb)| + Explain the difference between over-fitting and under-fitting a model<br> + Describe Bias-variance tradeoffs<br> + Find the optimal training and test data set splits, cross-validation, and model complexity versus error<br> + Apply a linear regression model for supervised learning.<br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 120 min |
49
-
|[04-Regularization_and_Gradient_Descent](04-Regularization_and_Gradient_Descent/Regularization_and_Gradient_Descent_Exercises.ipynb)| + Describe Logistic regression and how it differs from linear regression<br> + Identify metrics for classification errors and scenarios in which they can be used<br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 120 min |
50
-
|[05-Logistic_Regression_and_Classification_Error_Metrics](05-Logistic_Regression_and_Classification_Error_Metrics/Logistic_Regression_and_Classification_Error_Metrics_Exercises.ipynb)| + Explain cost functions, regularization, feature selection, and hyper-parameters<br> + | 120 min |
51
-
|[06-SVM_and_Kernels](06-SVM_and_Kernels/SVM_Kernels_Exercises.ipynb)| + Apply support vector machines (SVMs)—a popular algorithm used for classification problems<br> + Recognize SVM similarity to logistic regression<br> + Compute the cost function of SVMs<br> + Apply regularization in SVMs and some tips to obtain non-linear classifications with SVMs<nbr> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware | 120 min |
52
-
|[07-Decision_Trees](07-Decision_Trees/Decision_Trees_Exercises.ipynb)| + Recognize Decision trees and how to use them for classification problems<br> + Recognize how to identify the best split and the factors for splitting<br> + Explain strengths and weaknesses of decision trees<br> + Explain how regression trees help with classifying continuous values<br> + Describe motivation for choosing Random Forest Classifier over Decision Trees<br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware for Random Forest Classifier| 120 min |
53
-
|[08-Bagging](08-Bagging/Bagging_Exercises.ipynbb)| + Associate concepts of bootstrapping and aggregating (commonly known as “bagging”) to reduce variance<br> + Apply Random Forest algorithm that further reduces the correlation seen in bagging models<br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 120 min |
54
-
|[09-Boosting_and_Stacking](09-Boosting_and_Stacking/Boosting_and_Stacking_Exercises.ipynb)| + Explain how the boosting algorithm helps reduce variance and bias.<br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 120 min |
55
-
|[10-Introduction_to_Unsupervised_Learning_and_Clustering_Methods](10-Introduction_to_Unsupervised_Learning_and_Clustering_Methods/Clustering_Methods_Exercises.ipynb)| + Discuss unsupervised learning algorithms and how they can be applied<br> + Apply clustering<br> + Apply dimensionality reduction<br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 120 min |
56
-
|[11-Dimensionality_Reduction_and_Advanced_Topics](11-Dimensionality_Reduction_and_Advanced_Topics/Dimensionality_Reduction_Exercises.ipynb)| + Explain and Apply Principal Component Analysis (PCA)<br> + Explain Multidimensional Scaling (MDS)<br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 120 min |
57
-
58
-
#### Content Structure
59
-
60
-
Each module folder has a Jupyter Notebook file (`*.ipynb`), this can be opened in Jupyter Lab to view the training contant, edit code and compile/run.
61
-
62
-
## Install Directions
63
-
64
-
The training content can be accessed locally on the computer after installing necessary tools, or you can directly access using Intel DevCloud without any installation.
65
-
66
-
#### Local Installation of JupyterLab and oneAPI Tools
67
-
68
-
The Jupyter Notebooks can be downloaded locally to computer and accessed:
69
-
- Install Jupyter Lab on local computer: [Installation Guide](https://jupyterlab.readthedocs.io/en/stable/getting_started/installation.html)
70
-
- Install Intel oneAPI Base Toolkit on local computer: [Installation Guide](https://www.intel.com/content/www/us/en/developer/tools/oneapi/base-toolkit-download.html)
71
-
- git clone the repo and access the Notebooks using Jupyter Lab
72
-
73
-
74
-
#### Access using Intel DevCloud
75
-
76
-
The Jupyter notebooks are tested and can be run on Intel DevCloud without any installation necessary, below are the steps to access these Jupyter notebooks on Intel DevCloud:
77
-
1. Register on [Intel DevCloud](https://devcloud.intel.com/oneapi)
78
-
2. Login, Get Started and Launch Jupyter Lab
79
-
3. Open Terminal in Jupyter Lab and git clone the repo and access the Notebooks
0 commit comments