在 Slurm 中运行多节点MPI作业 AWS PCS - AWS PCS

本文属于机器翻译版本。若本译文内容与英语原文存在差异,则一律以英文原文为准。

在 Slurm 中运行多节点MPI作业 AWS PCS

这些说明演示了如何使用 Slurm 在中运行消息传递接口 (MPI) 作业。 AWS PCS

在登录节点的 shell 提示符下运行以下命令。

  • 成为默认用户。切换到其主目录。

    sudo su - ec2-user cd ~/
  • 使用 C 编程语言创建源代码。

    cat > hello.c << EOF // * mpi-hello-world - https://www.mpitutorial.com // Released under MIT License // // Copyright (c) 2014 MPI Tutorial. // // Permission is hereby granted, free of charge, to any person obtaining a copy // of this software and associated documentation files (the "Software"), to // deal in the Software without restriction, including without limitation the // rights to use, copy, modify, merge, publish, distribute, sublicense, and/or // sell copies of the Software, and to permit persons to whom the Software is // furnished to do so, subject to the following conditions: // The above copyright notice and this permission notice shall be included in // all copies or substantial portions of the Software. // // THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR // IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, // FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE // AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER // LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING // FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER // DEALINGS IN THE SOFTWARE. #include <mpi.h> #include <stdio.h> #include <stddef.h> int main(int argc, char** argv) { // Initialize the MPI environment. The two arguments to MPI Init are not // currently used by MPI implementations, but are there in case future // implementations might need the arguments. MPI_Init(NULL, NULL); // Get the number of processes int world_size; MPI_Comm_size(MPI_COMM_WORLD, &world_size); // Get the rank of the process int world_rank; MPI_Comm_rank(MPI_COMM_WORLD, &world_rank); // Get the name of the processor char processor_name[MPI_MAX_PROCESSOR_NAME]; int name_len; MPI_Get_processor_name(processor_name, &name_len); // Print off a hello world message printf("Hello world from processor %s, rank %d out of %d processors\n", processor_name, world_rank, world_size); // Finalize the MPI environment. No more MPI calls can be made after this MPI_Finalize(); } EOF
  • 加载打开MPI模块。

    module load openmpi
  • 编译 C 程序。

    mpicc -o hello hello.c
  • 编写 Slurm 作业提交脚本。

    cat > hello.sh << EOF #!/bin/bash #SBATCH -J multi #SBATCH -o multi.out #SBATCH -e multi.err #SBATCH --exclusive #SBATCH --nodes=4 #SBATCH --ntasks-per-node=1 srun $HOME/hello EOF
  • 切换到共享目录。

    cd /shared
  • 提交作业脚本。

    sbatch -p demo ~/hello.sh
  • squeue用于监视作业直至其完成。

  • 检查以下内容multi.out

    cat multi.out

    输出类似于以下内容。请注意,每个等级都有自己的 IP 地址,因为它运行在不同的节点上。

    Hello world from processor ip-10-3-133-204, rank 0 out of 4 processors
    Hello world from processor ip-10-3-128-219, rank 2 out of 4 processors
    Hello world from processor ip-10-3-141-26, rank 3 out of 4 processors
    Hello world from processor ip-10-3-143-52, rank 1 out of 4 processor