Proactive Data Containers (PDC) software provides an object-centric API and a runtime system with a set of data object management services. These services allow placing data in the memory and storage hierarchy, performing data movement asynchronously, and providing scalable metadata operations to find data objects. PDC revolutionizes how data is stored and accessed by using object-centric abstractions to represent data that moves in the high-performance computing (HPC) memory and storage subsystems. PDC manages extensive metadata to describe data objects to find desired data efficiently as well as to store information in the data objects.
PDC API, data types, and developer notes are available in docs/readme.md.
More information and publications of PDC is available at https://sdm.lbl.gov/pdc
The following instructions are for installing PDC on Linux and Cray machines. GCC version 7 or newer and a version of MPI are needed to install PDC.
Current PDC tests have been verified with MPICH. To install MPICH, follow the documentation in https://www.mpich.org/static/downloads/3.4.1/mpich-3.4.1-installguide.pdf
PDC also depends on libfabric and Mercury. We provide detailed instructions for installing libfabric, Mercury, and PDC below. Make sure to record the environmental variables (lines that contains the "export" commands). They are needed for running PDC and to use the libraries again.
PDC relies on libfabric
as well as mercury
. Therefore, let's prepare the dependencies.
Before installing the dependencies and downloading the code repository, we assume there is a directory created for your installation already, e.g. $WORK_SPACE
and now you are in $WORK_SPACE
.
export WORK_SPACE=/path/to/your/work/space
mkdir -p $WORK_SPACE/source
mkdir -p $WORK_SPACE/install
Now, let's download libfabric
, mercury
and pdc
into our source
directory.
cd $WORK_SPACE/source
git clone [email protected]:ofiwg/libfabric.git
git clone [email protected]:mercury-hpc/mercury.git
git clone [email protected]:hpc-io/pdc.git
export LIBFABRIC_SRC_DIR=$WORK_SPACE/source/libfabric
export MERCURY_SRC_DIR=$WORK_SPACE/source/mercury
export PDC_SRC_DIR=$WORK_SPACE/source/pdc
export LIBFABRIC_DIR=$WORK_SPACE/install/libfabric
export MERCURY_DIR=$WORK_SPACE/install/mercury
export PDC_DIR=$WORK_SPACE/install/pdc
mkdir -p $LIBFABRIC_SRC_DIR
mkdir -p $MERCURY_SRC_DIR
mkdir -p $PDC_SRC_DIR
mkdir -p $LIBFABRIC_DIR
mkdir -p $MERCURY_DIR
mkdir -p $PDC_DIR
echo "export LIBFABRIC_SRC_DIR=$LIBFABRIC_SRC_DIR" > $WORK_SPACE/pdc_env.sh
echo "export MERCURY_SRC_DIR=$MERCURY_SRC_DIR" >> $WORK_SPACE/pdc_env.sh
echo "export PDC_SRC_DIR=$PDC_SRC_DIR" >> $WORK_SPACE/pdc_env.sh
echo "export LIBFABRIC_DIR=$LIBFABRIC_DIR" >> $WORK_SPACE/pdc_env.sh
echo "export MERCURY_DIR=$MERCURY_DIR" >> $WORK_SPACE/pdc_env.sh
echo "export PDC_DIR=$PDC_DIR" >> $WORK_SPACE/pdc_env.sh
Remember, from now on, at any time, you can simply run the following to set the above environment variables so that you can run any of the following command for your installation.
export WORK_SPACE=/path/to/your/work/space
source $WORK_SPACE/pdc_env.sh
Check out tag v1.11.2
for libfabric
:
cd $LIBFABRIC_SRC_DIR
git checkout tags/v1.11.2
Configure, compile and install:
./autogen.sh
./configure --prefix=$LIBFABRIC_DIR CC=cc CFLAG="-O2"
make -j 32
make install
export LD_LIBRARY_PATH="$LIBFABRIC_DIR/lib:$LD_LIBRARY_PATH"
export PATH="$LIBFABRIC_DIR/include:$LIBFABRIC_DIR/lib:$PATH"
echo 'export LD_LIBRARY_PATH=$LIBFABRIC_DIR/lib:$LD_LIBRARY_PATH' >> $WORK_SPACE/pdc_env.sh
echo 'export PATH=$LIBFABRIC_DIR/include:$LIBFABRIC_DIR/lib:$PATH' >> $WORK_SPACE/pdc_env.sh
Note: On NERSC supercomputers, e.g. Cori and Perlmutter, we should add --disable-efa --disable-sockets
to the ./configure
command during the compilation on login nodes.
Now, you may check out a specific tag version of mercury
, for example, v2.2.0
:
cd $MERCURY_SRC_DIR
mkdir build
git checkout tags/v2.2.0
git submodule update --init
Configure, compile, test and install:
cd build
cmake ../ -DCMAKE_INSTALL_PREFIX=$MERCURY_DIR -DCMAKE_C_COMPILER=cc -DBUILD_SHARED_LIBS=ON -DBUILD_TESTING=ON -DNA_USE_OFI=ON -DNA_USE_SM=OFF -DNA_OFI_TESTING_PROTOCOL=tcp
make -j 32 && make install
ctest
export LD_LIBRARY_PATH="$MERCURY_DIR/lib:$LD_LIBRARY_PATH"
export PATH="$MERCURY_DIR/include:$MERCURY_DIR/lib:$PATH"
echo 'export LD_LIBRARY_PATH=$MERCURY_DIR/lib:$LD_LIBRARY_PATH' >> $WORK_SPACE/pdc_env.sh
echo 'export PATH=$MERCURY_DIR/include:$MERCURY_DIR/lib:$PATH' >> $WORK_SPACE/pdc_env.sh
Now, it's time to compile and install PDC.
- One can replace
mpicc
to other available MPI compilers. For example, on Cori,cc
can be used to replacempicc
. ctest
contains both sequential and MPI tests for the PDC settings. These can be used to perform regression tests.
cd $PDC_SRC_DIR
git checkout develop
mkdir build
cd build
cmake ../ -DBUILD_MPI_TESTING=ON -DBUILD_SHARED_LIBS=ON -DBUILD_TESTING=ON -DCMAKE_INSTALL_PREFIX=$PDC_DIR -DPDC_ENABLE_MPI=ON -DMERCURY_DIR=$MERCURY_DIR -DCMAKE_C_COMPILER=cc -DMPI_RUN_CMD=srun
make -j 32 && make install
Let's run ctest
now on a compute node:
salloc --nodes 1 --qos interactive --time 01:00:00 --constraint haswell
salloc --nodes 1 --qos interactive --time 01:00:00 --constraint cpu --account=mxxxx
Once you are on the compute node, you can run ctest
.
ctest
Note: On Cori, if you happen to see failures regarding libibverb
validation, login to one of the compute nodes by running an interactive job and re-compile all PDC's dependencies and PDC itself. Then problem will be solved.
If all the tests pass, you can now specify the environment variables.
export LD_LIBRARY_PATH="$PDC_DIR/lib:$LD_LIBRARY_PATH"
export PATH="$PDC_DIR/include:$PDC_DIR/lib:$PATH"
echo 'export LD_LIBRARY_PATH=$PDC_DIR/lib:$LD_LIBRARY_PATH' >> $WORK_SPACE/pdc_env.sh
echo 'export PATH=$PDC_DIR/include:$PDC_DIR/lib:$PATH' >> $WORK_SPACE/pdc_env.sh
One can also install PDC
with Spack
, with which the dependencies of PDC
can be easily managed and installed.
git clone -c feature.manyFiles=true https://github.com/spack/spack.git
cd spack/bin
./spack install pdc
Essentially, PDC is a typical client-server application.
To run PDC
, one needs to start the server processes first, and then the clients can be started to issue RPC requests handled by the Mercury
RPC framework.
We provide mpi_test.sh
utility for running MPI tests. For example, on a regular Linux machine, you may run the following:
export JOB_RUNNER=mpiexec
cd $PDC_DIR/bin
./mpi_test.sh ./pdc_init $JOB_RUNNER 2 4
This is test will start 2 processes for PDC servers. The client program ./pdc_init will start 4 processes. Similarly, one can run any of the client examples in ctest
.
Depending on the specific HPC environment where you run PDC
, the value of $JOB_RUNNER
variable can be changed to srun
(for NERSC), aprun
(for Theta), or jsrun
for Summit
, accordingly.
These source code will provide some knowledge of how to use PDC. For more reference, one may check the documentation folder in this repository.
If you are running PDC
on Cori supercomputer, here are some tips you would need to follow:
-
On Cori, it is recommended to use
cc
as the default compiler when compiling PDC and its dependencies. -
When preparing compilation for
PDC
usingCMake
, it is suggested to append console argument-DMPI_RUN_CMD=srun
so thatctest
can be executed on Cori. -
Sometimes, it might be helpful to unload
darshan
module before the installation. -
For opening an interactive job session on Cori, it is recommended to add
--gres=craynetwork:2
option to thesalloc
command:salloc -C haswell -N 4 -t 01:00:00 -q interactive --gres=craynetwork:2
-
To launch the PDC server and the client, add
--gres=craynetwork:1
before the executables, for example:-
Run 4 server processes, each on one node in background:
srun -N 4 -n 4 -c 2 --mem=25600 --cpu_bind=cores --gres=craynetwork:1 --overlap ./bin/pdc_server.exe &
-
Run 64 client processes that concurrently create 1000 objects in total:
srun -N 4 -n 64 -c 2 --mem=25600 --cpu_bind=cores --gres=craynetwork:1 --overlap ./bin/create_obj_scale -r 1000
-