问如何在slurm中设置ntask或ntasks per-node,以便在pytorch中运行多节点分布式训练?EN在数据越来越多的...
问使用SLURM进行单GPU Pytorch训练-如何设置“ntasks per-node”?EN“ 大家好哇!前面我们对slurm作业...
Not sure if there is a valid use case forSLURM_NTASKS < SLURM_NTASKS_PER_NODE. But if there is not it would be awesome if Lightning could raise an error in this scenario. The same error also happens if--ntasks-per-nodeis not set. In this case Lightning assumes 2 devices (I guess...
UserWarning: You requested 2 GPUs but launched 1 slurm tasks. We will launch 2 processes for you. We recommend you let slurm manage the processes by setting: --ntasks-per-node=2 If you're not using SLURM, ignore this message! I made the suggested change, but I still get the warning...