You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This sample code illustrates how to use Intel® Distribution of Modin for ETL operations and ridge regression algorithm from the Intel® oneAPI Data Analytics Library (oneDAL) accelerated scikit-learn library to build and run an end to end machine learning workload. Both Intel Distribution of Modin and oneDAL accelerated scikit-learn libraries are available together in [Intel AI Analytics Toolkit](https://software.intel.com/content/www/us/en/develop/tools/oneapi/ai-analytics-toolkit.html). This sample code demonstrates how to seamlessly run the end-to-end census workload using the toolkit, without any external dependencies.
3
+
This sample code illustrates how to use Intel® Distribution of Modin for ETL operations and ridge regression algorithm from the Intel® oneAPI Data Analytics Library (oneDAL) accelerated scikit-learn library to build and run an end to end machine learning workload. Both Intel Distribution of Modin and oneDAL accelerated scikit-learn libraries are available together in [Intel® oneAPI AI Analytics Toolkit](https://software.intel.com/content/www/us/en/develop/tools/oneapi/ai-analytics-toolkit.html). This sample code demonstrates how to seamlessly run the end-to-end census workload using the toolkit, without any external dependencies.
4
4
5
5
| Optimized for | Description
6
6
| :--- | :---
@@ -27,17 +27,21 @@ This end-to-end workload sample code is implemented for CPU using the Python lan
27
27
Code samples are licensed under the MIT license. See
28
28
[License.txt](https://github.com/oneapi-src/oneAPI-samples/blob/master/License.txt) for details.
29
29
30
-
Third party program Licenses can be found here: [third-party-programs.txt](https://github.com/oneapi-src/oneAPI-samples/blob/master/third-party-programs.txt)
30
+
Third party program Licenses can be found here: [third-party-programs.txt](https://github.com/oneapi-src/oneAPI-samples/blob/master/third-party-programs.txt).
31
+
32
+
## Running Samples on the Intel® DevCloud
33
+
If you are running this sample on the Intel® DevCloud, skip the Pre-requirements and go to the [Activate Conda Environment](#activate-conda) section.
31
34
32
35
## Building Intel® Distribution of Modin and Intel® oneAPI Data Analytics Library (oneDAL) for CPU to build and run end-to-end workload
33
36
37
+
### Pre-requirements (Local or Remote Host Installation)
34
38
Intel® Distribution of Modin and Intel® oneAPI Data Analytics Library (oneDAL) is ready for use once you finish the Intel AI Analytics Toolkit installation with the Conda Package Manager.
35
39
36
40
You can refer to the oneAPI [main page](https://software.intel.com/en-us/oneapi), and the Toolkit [Getting Started Guide for Linux](https://software.intel.com/content/www/us/en/develop/documentation/get-started-with-ai-linux/top.html) for installation steps and scripts.
37
41
38
-
### Activate conda environment With Root Access
42
+
### Activate conda environment With Root Access<aname="activate-conda"></a>
39
43
40
-
Please follow the Getting Started Guide steps (above) to set up your oneAPI environment with the `setvars.sh` script and [Intel® Distribution of Modin environment installation] (https://software.intel.com/content/www/us/en/develop/articles/installing-ai-kit-with-conda.html). Then navigate in Linux shellto your oneapi installation path, typically `/opt/intel/oneapi/` when installed as root or sudo, and `~/intel/oneapi/` when not installed as a super user. If you customized the installation folder, the `setvars.sh` file is in your custom folder.
44
+
In the Linux shell, navigate to your oneapi installation path, typically `/opt/intel/oneapi/` when installed as root or sudo, and `~/intel/oneapi/` when not installed as a super user.
41
45
42
46
Activate the conda environment with the following command:
### Activate conda environment Without Root Access (Optional)
50
54
51
-
By default, the Intel AI Analytics toolkit is installed in the `oneapi` folder, which requires root privileges to manage it. If you would like to bypass using root access to manage your conda environment, then you can clone your desired conda environment using the following command:
55
+
By default, the Intel oneAPI AI Analytics toolkit is installed in the `oneapi` folder, which requires root privileges to manage it. If you would like to bypass using root access to manage your conda environment, then you can clone your desired conda environment using the following command:
Launch Jupyter Notebook in the directory housing the code example
71
+
Launch Jupyter Notebook in the directory housing the code example.
68
72
69
73
```
70
74
conda install jupyter nb_conda_kernels
@@ -76,7 +80,7 @@ pip install jupyter
76
80
77
81
### Install wget package
78
82
79
-
Install wget package to retrieve the Census dataset using HTTPS
83
+
Install wget package to retrieve the Census dataset using HTTPS.
80
84
81
85
```
82
86
pip install wget
@@ -85,7 +89,7 @@ pip install wget
85
89
#### View in Jupyter Notebook
86
90
87
91
88
-
Launch Jupyter Notebook in the directory housing the code example
92
+
Launch Jupyter Notebook in the directory housing the code example.
89
93
90
94
```
91
95
jupyter notebook
@@ -112,3 +116,16 @@ Run the Program
112
116
##### Expected Printed Output:
113
117
Expected Cell Output shown for census_modin.ipynb:
114
118

119
+
120
+
121
+
### Request a Compute Node
122
+
In order to run on the DevCloud, you need to request a compute node using node properties such as: `gpu`, `xeon`, `fpga_compile`, `fpga_runtime` and others. For more information about the node properties, execute the `pbsnodes` command.
123
+
This node information must be provided when submitting a job to run your sample in batch mode using the qsub command. When you see the qsub command in the Run section of the [Hello World instructions](https://devcloud.intel.com/oneapi/get_started/aiAnalyticsToolkitSamples/), change the command to fit the node you are using. Nodes which are in bold indicate they are compatible with this sample:
0 commit comments