Skip to content

Commit 8203828

Browse files
authored
updated Census readme to add DevCloud instructions (oneapi-src#533)
1 parent 46b9936 commit 8203828

File tree

1 file changed

+27
-10
lines changed
  • AI-and-Analytics/End-to-end-Workloads/Census

1 file changed

+27
-10
lines changed

AI-and-Analytics/End-to-end-Workloads/Census/README.md

Lines changed: 27 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
1-
# End-to-end machine learning workload: `Census` Sample
1+
# End-to-end Machine Learning Workload: `Census` Sample
22

3-
This sample code illustrates how to use Intel® Distribution of Modin for ETL operations and ridge regression algorithm from the Intel® oneAPI Data Analytics Library (oneDAL) accelerated scikit-learn library to build and run an end to end machine learning workload. Both Intel Distribution of Modin and oneDAL accelerated scikit-learn libraries are available together in [Intel AI Analytics Toolkit](https://software.intel.com/content/www/us/en/develop/tools/oneapi/ai-analytics-toolkit.html). This sample code demonstrates how to seamlessly run the end-to-end census workload using the toolkit, without any external dependencies.
3+
This sample code illustrates how to use Intel® Distribution of Modin for ETL operations and ridge regression algorithm from the Intel® oneAPI Data Analytics Library (oneDAL) accelerated scikit-learn library to build and run an end to end machine learning workload. Both Intel Distribution of Modin and oneDAL accelerated scikit-learn libraries are available together in [Intel® oneAPI AI Analytics Toolkit](https://software.intel.com/content/www/us/en/develop/tools/oneapi/ai-analytics-toolkit.html). This sample code demonstrates how to seamlessly run the end-to-end census workload using the toolkit, without any external dependencies.
44

55
| Optimized for | Description
66
| :--- | :---
@@ -27,17 +27,21 @@ This end-to-end workload sample code is implemented for CPU using the Python lan
2727
Code samples are licensed under the MIT license. See
2828
[License.txt](https://github.com/oneapi-src/oneAPI-samples/blob/master/License.txt) for details.
2929

30-
Third party program Licenses can be found here: [third-party-programs.txt](https://github.com/oneapi-src/oneAPI-samples/blob/master/third-party-programs.txt)
30+
Third party program Licenses can be found here: [third-party-programs.txt](https://github.com/oneapi-src/oneAPI-samples/blob/master/third-party-programs.txt).
31+
32+
## Running Samples on the Intel® DevCloud
33+
If you are running this sample on the Intel® DevCloud, skip the Pre-requirements and go to the [Activate Conda Environment](#activate-conda) section.
3134

3235
## Building Intel® Distribution of Modin and Intel® oneAPI Data Analytics Library (oneDAL) for CPU to build and run end-to-end workload
3336

37+
### Pre-requirements (Local or Remote Host Installation)
3438
Intel® Distribution of Modin and Intel® oneAPI Data Analytics Library (oneDAL) is ready for use once you finish the Intel AI Analytics Toolkit installation with the Conda Package Manager.
3539

3640
You can refer to the oneAPI [main page](https://software.intel.com/en-us/oneapi), and the Toolkit [Getting Started Guide for Linux](https://software.intel.com/content/www/us/en/develop/documentation/get-started-with-ai-linux/top.html) for installation steps and scripts.
3741

38-
### Activate conda environment With Root Access
42+
### Activate conda environment With Root Access<a name="activate-conda"></a>
3943

40-
Please follow the Getting Started Guide steps (above) to set up your oneAPI environment with the `setvars.sh` script and [Intel® Distribution of Modin environment installation] (https://software.intel.com/content/www/us/en/develop/articles/installing-ai-kit-with-conda.html). Then navigate in Linux shell to your oneapi installation path, typically `/opt/intel/oneapi/` when installed as root or sudo, and `~/intel/oneapi/` when not installed as a super user. If you customized the installation folder, the `setvars.sh` file is in your custom folder.
44+
In the Linux shell, navigate to your oneapi installation path, typically `/opt/intel/oneapi/` when installed as root or sudo, and `~/intel/oneapi/` when not installed as a super user.
4145

4246
Activate the conda environment with the following command:
4347

@@ -48,7 +52,7 @@ source activate intel-aikit-modin
4852

4953
### Activate conda environment Without Root Access (Optional)
5054

51-
By default, the Intel AI Analytics toolkit is installed in the `oneapi` folder, which requires root privileges to manage it. If you would like to bypass using root access to manage your conda environment, then you can clone your desired conda environment using the following command:
55+
By default, the Intel oneAPI AI Analytics toolkit is installed in the `oneapi` folder, which requires root privileges to manage it. If you would like to bypass using root access to manage your conda environment, then you can clone your desired conda environment using the following command:
5256

5357
#### Linux
5458
```
@@ -62,9 +66,9 @@ conda activate intel-aikit-modin
6266
```
6367

6468

65-
### Install Jupyter Notebook
69+
### Install Jupyter Notebook*
6670

67-
Launch Jupyter Notebook in the directory housing the code example
71+
Launch Jupyter Notebook in the directory housing the code example.
6872

6973
```
7074
conda install jupyter nb_conda_kernels
@@ -76,7 +80,7 @@ pip install jupyter
7680

7781
### Install wget package
7882

79-
Install wget package to retrieve the Census dataset using HTTPS
83+
Install wget package to retrieve the Census dataset using HTTPS.
8084

8185
```
8286
pip install wget
@@ -85,7 +89,7 @@ pip install wget
8589
#### View in Jupyter Notebook
8690

8791

88-
Launch Jupyter Notebook in the directory housing the code example
92+
Launch Jupyter Notebook in the directory housing the code example.
8993

9094
```
9195
jupyter notebook
@@ -112,3 +116,16 @@ Run the Program
112116
##### Expected Printed Output:
113117
Expected Cell Output shown for census_modin.ipynb:
114118
![Output](Expected_output.jpg "Expected output for Jupyter Notebook")
119+
120+
121+
### Request a Compute Node
122+
In order to run on the DevCloud, you need to request a compute node using node properties such as: `gpu`, `xeon`, `fpga_compile`, `fpga_runtime` and others. For more information about the node properties, execute the `pbsnodes` command.
123+
This node information must be provided when submitting a job to run your sample in batch mode using the qsub command. When you see the qsub command in the Run section of the [Hello World instructions](https://devcloud.intel.com/oneapi/get_started/aiAnalyticsToolkitSamples/), change the command to fit the node you are using. Nodes which are in bold indicate they are compatible with this sample:
124+
125+
<!---Mark each compatible Node in BOLD-->
126+
| Node | Command |
127+
| ----------------- | ------------------------------------------------------- |
128+
| GPU | qsub -l nodes=1:gpu:ppn=2 -d . hello-world.sh |
129+
| CPU | qsub -l nodes=1:xeon:ppn=2 -d . hello-world.sh |
130+
| FPGA Compile Time | qsub -l nodes=1:fpga\_compile:ppn=2 -d . hello-world.sh |
131+
| FPGA Runtime | qsub -l nodes=1:fpga\_runtime:ppn=2 -d . hello-world.sh |

0 commit comments

Comments
 (0)