Skip to content

Commit dbc81bc

Browse files
Bc update (oneapi-src#1155)
* updated sample.json and Teacherkit.ipynb * remove ANSWERS * update sample.json w pip install -r requirements.txt * fixes to bracnh bc-update: fix broken links, consolodate data to common data folder, answers in pw protected zip * fix broken links lower level READMEs * Update sample.json replace sample.json with Piotr's updates last month * Update sample.json Co-authored-by: Jimmy Wei <[email protected]>
1 parent 4eb1d11 commit dbc81bc

File tree

133 files changed

+375
-401402
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

133 files changed

+375
-401402
lines changed

AI-and-Analytics/Jupyter/Introduction_to_Machine_Learning/01_Introduction_to_Machine_Learning_and_Tools/Introduction_to_Machine_Learning_and_Toolkit-ANSWERS.ipynb

Lines changed: 0 additions & 680 deletions
This file was deleted.

AI-and-Analytics/Jupyter/Introduction_to_Machine_Learning/01_Introduction_to_Machine_Learning_and_Tools/Introduction_to_Machine_Learning_and_Toolkit.ipynb

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -95,7 +95,7 @@
9595
"source": [
9696
"from __future__ import print_function\n",
9797
"import os\n",
98-
"data_path = [ 'data']"
98+
"data_path = [ '../data']"
9999
]
100100
},
101101
{
@@ -366,7 +366,7 @@
366366
"metadata": {
367367
"anaconda-cloud": {},
368368
"kernelspec": {
369-
"display_name": "Python 3.8 (Intel® oneAPI)",
369+
"display_name": "Python 3 (Intel® oneAPI 2022.3)",
370370
"language": "python",
371371
"name": "c009-intel_distribution_of_python_3_oneapi-beta05-python"
372372
},
@@ -380,7 +380,7 @@
380380
"name": "python",
381381
"nbconvert_exporter": "python",
382382
"pygments_lexer": "ipython3",
383-
"version": "3.9.10"
383+
"version": "3.9.12"
384384
}
385385
},
386386
"nbformat": 4,

AI-and-Analytics/Jupyter/Introduction_to_Machine_Learning/01_Introduction_to_Machine_Learning_and_Tools/README.md

Lines changed: 7 additions & 45 deletions
Original file line numberDiff line numberDiff line change
@@ -19,8 +19,13 @@ At the end of this course you will be able to:
1919
- Demonstrate supervised learning algorithms
2020
- Explain key concepts like under- and over-fitting, regularization, and cross-validation
2121
- Classify the type of problem to be solved, choose the right algorithm, tune parameters, and validate a model
22-
- Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware
23-
- Download and Install Intel® Extension for Scikit-learn* powered by oneAPI
22+
- Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware using two lines of code
23+
``` python
24+
from sklearnex import patch_sklearn
25+
patch_sklearn()
26+
```
27+
- Reference describing: Download and Install Intel® Extension for Scikit-learn* powered by oneAPI
28+
2429

2530

2631
## License
@@ -38,46 +43,3 @@ Third party program Licenses can be found here: [third-party-programs.txt](https
3843
- Statistics
3944

4045

41-
## Syllabus
42-
43-
- 11 Modules (18 hours)
44-
- 11 Lab Exercises
45-
46-
-----------------------
47-
| Modules | Description | Duration |
48-
| :--- | :------ | :------ |
49-
|[01_Introduction_to_Machine_Learning_and_Tools](01_Introduction_to_Machine_Learning_and_Tools/Introduction_to_Machine_Learning_and_Toolkit.ipynb)| + Demonstrate supervised learning algorithms.<br> + Explain key concepts like under- and over-fitting, regularization, and cross-validation<br>: + Classify the type of problem to be solved <br> + Choose an algorithm, tune parameters, and validate a model <br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 60 min |
50-
|[02-Introduction_to_Supervised_Learning_and_K_Nearest_Neighbors](02-Introduction_to_Supervised_Learning_and_K_Nearest_Neighbors/Supervised_Learning_and_K_Nearest_Neighbors_Exercises.ipynb)| + Explain supervised learning and how it can be applied to regression and classification problems.<br> + Apply K-Nearest Neighbor (KNN) algorithm for classification <br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 120 min |
51-
|[03-Train_Test_Splits_Validation_Linear_Regression](03-Train_Test_Splits_Validation_Linear_Regression/Train_Test_Splits_Validation_Linear_Regression.ipynb)| + Explain the difference between over-fitting and under-fitting a model<br> + Describe Bias-variance tradeoffs<br> + Find the optimal training and test data set splits, cross-validation, and model complexity versus error<br> + Apply a linear regression model for supervised learning.<br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 120 min |
52-
|[04-Regularization_and_Gradient_Descent](04-Regularization_and_Gradient_Descent/Regularization_and_Gradient_Descent_Exercises.ipynb)| + Describe Logistic regression and how it differs from linear regression<br> + Identify metrics for classification errors and scenarios in which they can be used<br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 120 min |
53-
|[05-Logistic_Regression_and_Classification_Error_Metrics](05-Logistic_Regression_and_Classification_Error_Metrics/Logistic_Regression_and_Classification_Error_Metrics_Exercises.ipynb)| + Explain cost functions, regularization, feature selection, and hyper-parameters<br> + | 120 min |
54-
|[06-SVM_and_Kernels](06-SVM_and_Kernels/SVM_Kernels_Exercises.ipynb)| + Apply support vector machines (SVMs)—a popular algorithm used for classification problems<br> + Recognize SVM similarity to logistic regression<br> + Compute the cost function of SVMs<br> + Apply regularization in SVMs and some tips to obtain non-linear classifications with SVMs<nbr> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware | 120 min |
55-
|[07-Decision_Trees](07-Decision_Trees/Decision_Trees_Exercises.ipynb)| + Recognize Decision trees and how to use them for classification problems<br> + Recognize how to identify the best split and the factors for splitting<br> + Explain strengths and weaknesses of decision trees<br> + Explain how regression trees help with classifying continuous values<br> + Describe motivation for choosing Random Forest Classifier over Decision Trees<br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware for Random Forest Classifier| 120 min |
56-
|[08-Bagging](08-Bagging/Bagging_Exercises.ipynbb)| + Associate concepts of bootstrapping and aggregating (commonly known as “bagging”) to reduce variance<br> + Apply Random Forest algorithm that further reduces the correlation seen in bagging models<br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 120 min |
57-
|[09-Boosting_and_Stacking](09-Boosting_and_Stacking/Boosting_and_Stacking_Exercises.ipynb)| + Explain how the boosting algorithm helps reduce variance and bias.<br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 120 min |
58-
|[10-Introduction_to_Unsupervised_Learning_and_Clustering_Methods](10-Introduction_to_Unsupervised_Learning_and_Clustering_Methods/Clustering_Methods_Exercises.ipynb)| + Discuss unsupervised learning algorithms and how they can be applied<br> + Apply clustering<br> + Apply dimensionality reduction<br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 120 min |
59-
|[11-Dimensionality_Reduction_and_Advanced_Topics](11-Dimensionality_Reduction_and_Advanced_Topics/Dimensionality_Reduction_Exercises.ipynb)| + Explain and Apply Principal Component Analysis (PCA)<br> + Explain Multidimensional Scaling (MDS)<br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 120 min |
60-
61-
#### Content Structure
62-
63-
Each module folder has a Jupyter Notebook file (`*.ipynb`), this can be opened in Jupyter Lab to view the training contant, edit code and compile/run.
64-
65-
## Install Directions
66-
67-
The training content can be accessed locally on the computer after installing necessary tools, or you can directly access using Intel DevCloud without any installation.
68-
69-
#### Local Installation of JupyterLab and oneAPI Tools
70-
71-
The Jupyter Notebooks can be downloaded locally to computer and accessed:
72-
- Install Jupyter Lab on local computer: [Installation Guide](https://jupyterlab.readthedocs.io/en/stable/getting_started/installation.html)
73-
- Install Intel oneAPI Base Toolkit on local computer: [Installation Guide](https://www.intel.com/content/www/us/en/developer/tools/oneapi/base-toolkit-download.html)
74-
- git clone the repo and access the Notebooks using Jupyter Lab
75-
76-
77-
#### Access using Intel DevCloud
78-
79-
The Jupyter notebooks are tested and can be run on Intel DevCloud without any installation necessary, below are the steps to access these Jupyter notebooks on Intel DevCloud:
80-
1. Register on [Intel DevCloud](https://devcloud.intel.com/oneapi)
81-
2. Login, Get Started and Launch Jupyter Lab
82-
3. Open Terminal in Jupyter Lab and git clone the repo and access the Notebooks
83-

AI-and-Analytics/Jupyter/Introduction_to_Machine_Learning/02-Supervised_Learning_and_K_Nearest_Neighbors/README.md

Lines changed: 5 additions & 43 deletions
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,11 @@ At the end of this course you will be able to:
1818

1919
- Explain supervised learning and how it can be applied to regression and classification problems
2020
- Apply K-Nearest Neighbor (KNN) algorithm for classification
21-
- Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware
21+
- Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware using two lines of code
22+
``` python
23+
from sklearnex import patch_sklearn
24+
patch_sklearn()
25+
```
2226

2327
## License
2428
Code samples
@@ -35,46 +39,4 @@ Third party program Licenses can be found here: [third-party-programs.txt](https
3539
- Statistics
3640

3741

38-
## Syllabus
39-
40-
- 11 Modules (18 hours)
41-
- 11 Lab Exercises
42-
43-
-----------------------
44-
| Modules | Description | Duration |
45-
| :--- | :------ | :------ |
46-
|[01_Introduction_to_Machine_Learning_and_Tools](01_Introduction_to_Machine_Learning_and_Tools/Introduction_to_Machine_Learning_and_Toolkit.ipynb)| + Demonstrate supervised learning algorithms.<br> + Explain key concepts like under- and over-fitting, regularization, and cross-validation<br>: + Classify the type of problem to be solved <br> + Choose an algorithm, tune parameters, and validate a model <br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 60 min |
47-
|[02-Introduction_to_Supervised_Learning_and_K_Nearest_Neighbors](02-Introduction_to_Supervised_Learning_and_K_Nearest_Neighbors/Supervised_Learning_and_K_Nearest_Neighbors_Exercises.ipynb)| + Explain supervised learning and how it can be applied to regression and classification problems.<br> + Apply K-Nearest Neighbor (KNN) algorithm for classification <br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 120 min |
48-
|[03-Train_Test_Splits_Validation_Linear_Regression](03-Train_Test_Splits_Validation_Linear_Regression/Train_Test_Splits_Validation_Linear_Regression.ipynb)| + Explain the difference between over-fitting and under-fitting a model<br> + Describe Bias-variance tradeoffs<br> + Find the optimal training and test data set splits, cross-validation, and model complexity versus error<br> + Apply a linear regression model for supervised learning.<br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 120 min |
49-
|[04-Regularization_and_Gradient_Descent](04-Regularization_and_Gradient_Descent/Regularization_and_Gradient_Descent_Exercises.ipynb)| + Describe Logistic regression and how it differs from linear regression<br> + Identify metrics for classification errors and scenarios in which they can be used<br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 120 min |
50-
|[05-Logistic_Regression_and_Classification_Error_Metrics](05-Logistic_Regression_and_Classification_Error_Metrics/Logistic_Regression_and_Classification_Error_Metrics_Exercises.ipynb)| + Explain cost functions, regularization, feature selection, and hyper-parameters<br> + | 120 min |
51-
|[06-SVM_and_Kernels](06-SVM_and_Kernels/SVM_Kernels_Exercises.ipynb)| + Apply support vector machines (SVMs)—a popular algorithm used for classification problems<br> + Recognize SVM similarity to logistic regression<br> + Compute the cost function of SVMs<br> + Apply regularization in SVMs and some tips to obtain non-linear classifications with SVMs<nbr> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware | 120 min |
52-
|[07-Decision_Trees](07-Decision_Trees/Decision_Trees_Exercises.ipynb)| + Recognize Decision trees and how to use them for classification problems<br> + Recognize how to identify the best split and the factors for splitting<br> + Explain strengths and weaknesses of decision trees<br> + Explain how regression trees help with classifying continuous values<br> + Describe motivation for choosing Random Forest Classifier over Decision Trees<br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware for Random Forest Classifier| 120 min |
53-
|[08-Bagging](08-Bagging/Bagging_Exercises.ipynbb)| + Associate concepts of bootstrapping and aggregating (commonly known as “bagging”) to reduce variance<br> + Apply Random Forest algorithm that further reduces the correlation seen in bagging models<br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 120 min |
54-
|[09-Boosting_and_Stacking](09-Boosting_and_Stacking/Boosting_and_Stacking_Exercises.ipynb)| + Explain how the boosting algorithm helps reduce variance and bias.<br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 120 min |
55-
|[10-Introduction_to_Unsupervised_Learning_and_Clustering_Methods](10-Introduction_to_Unsupervised_Learning_and_Clustering_Methods/Clustering_Methods_Exercises.ipynb)| + Discuss unsupervised learning algorithms and how they can be applied<br> + Apply clustering<br> + Apply dimensionality reduction<br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 120 min |
56-
|[11-Dimensionality_Reduction_and_Advanced_Topics](11-Dimensionality_Reduction_and_Advanced_Topics/Dimensionality_Reduction_Exercises.ipynb)| + Explain and Apply Principal Component Analysis (PCA)<br> + Explain Multidimensional Scaling (MDS)<br> + Apply Intel® Extension for Scikit-learn* to leverage underlying compute capabilities of hardware| 120 min |
57-
58-
#### Content Structure
59-
60-
Each module folder has a Jupyter Notebook file (`*.ipynb`), this can be opened in Jupyter Lab to view the training contant, edit code and compile/run.
61-
62-
## Install Directions
63-
64-
The training content can be accessed locally on the computer after installing necessary tools, or you can directly access using Intel DevCloud without any installation.
65-
66-
#### Local Installation of JupyterLab and oneAPI Tools
67-
68-
The Jupyter Notebooks can be downloaded locally to computer and accessed:
69-
- Install Jupyter Lab on local computer: [Installation Guide](https://jupyterlab.readthedocs.io/en/stable/getting_started/installation.html)
70-
- Install Intel oneAPI Base Toolkit on local computer: [Installation Guide](https://www.intel.com/content/www/us/en/developer/tools/oneapi/base-toolkit-download.html)
71-
- git clone the repo and access the Notebooks using Jupyter Lab
72-
73-
74-
#### Access using Intel DevCloud
75-
76-
The Jupyter notebooks are tested and can be run on Intel DevCloud without any installation necessary, below are the steps to access these Jupyter notebooks on Intel DevCloud:
77-
1. Register on [Intel DevCloud](https://devcloud.intel.com/oneapi)
78-
2. Login, Get Started and Launch Jupyter Lab
79-
3. Open Terminal in Jupyter Lab and git clone the repo and access the Notebooks
8042

0 commit comments

Comments
 (0)