mirror of
https://github.com/smyalygames/FiniteVolumeGPU.git
synced 2025-05-17 22:14:14 +02:00
Removed nvtx and MPI options for job script
This commit is contained in:
parent
a588948e77
commit
a260230002
@ -27,7 +27,7 @@ from mpi4py import MPI
|
||||
import time
|
||||
|
||||
import pycuda.driver as cuda
|
||||
import nvtx
|
||||
#import nvtx
|
||||
|
||||
|
||||
|
||||
@ -314,14 +314,14 @@ class MPISimulator(Simulator.BaseSimulator):
|
||||
|
||||
def substep(self, dt, step_number):
|
||||
|
||||
nvtx.mark("substep start", color="yellow")
|
||||
#nvtx.mark("substep start", color="yellow")
|
||||
|
||||
self.profiling_data_mpi["start"]["t_mpi_step"] += time.time()
|
||||
|
||||
nvtx.mark("substep external", color="blue")
|
||||
#nvtx.mark("substep external", color="blue")
|
||||
self.sim.substep(dt, step_number, external=True, internal=False) # only "internal ghost cells"
|
||||
|
||||
nvtx.mark("substep internal", color="red")
|
||||
#nvtx.mark("substep internal", color="red")
|
||||
self.sim.substep(dt, step_number, internal=True, external=False) # "internal ghost cells" excluded
|
||||
|
||||
#nvtx.mark("substep full", color="blue")
|
||||
@ -331,7 +331,7 @@ class MPISimulator(Simulator.BaseSimulator):
|
||||
|
||||
self.profiling_data_mpi["end"]["t_mpi_step"] += time.time()
|
||||
|
||||
nvtx.mark("exchange", color="blue")
|
||||
#nvtx.mark("exchange", color="blue")
|
||||
self.full_exchange()
|
||||
|
||||
#nvtx.mark("download", color="blue")
|
||||
@ -343,10 +343,10 @@ class MPISimulator(Simulator.BaseSimulator):
|
||||
#nvtx.mark("upload", color="blue")
|
||||
#self.upload_for_exchange(self.sim.u0)
|
||||
|
||||
nvtx.mark("sync start", color="blue")
|
||||
#nvtx.mark("sync start", color="blue")
|
||||
self.sim.stream.synchronize()
|
||||
self.sim.internal_stream.synchronize()
|
||||
nvtx.mark("sync end", color="blue")
|
||||
#nvtx.mark("sync end", color="blue")
|
||||
|
||||
self.profiling_data_mpi["n_time_steps"] += 1
|
||||
|
||||
|
@ -6,6 +6,7 @@
|
||||
#SBATCH -t 0-00:10 # time (D-HH:MM)
|
||||
#SBATCH -o slurm.%N.%j.out # STDOUT
|
||||
#SBATCH -e slurm.%N.%j.err # STDERR
|
||||
#SBATCH --reservation=martinls_6
|
||||
|
||||
ulimit -s 10240
|
||||
module load slurm/20.02.7
|
||||
@ -26,7 +27,8 @@ cp -r . /work/$USER/$SLURM_JOB_ID/ShallowWaterGPU
|
||||
cd /work/$USER/$SLURM_JOB_ID/ShallowWaterGPU
|
||||
#mpirun --mca btl_openib_if_include mlx5_0 --mca btl_openib_warn_no_device_params_found 0 $HOME/miniconda3/envs/ShallowWaterGPU_HPC/bin/python3 mpiTesting.py -nx $NX -ny $NY --profile
|
||||
#nsys profile -t nvtx,cuda mpirun -np $SLURM_NTASKS numactl --cpunodebind=0 --localalloc $HOME/miniconda3/envs/ShallowWaterGPU_HPC/bin/python3 mpiTesting.py -nx $NX -ny $NY --profile
|
||||
mpirun -np $SLURM_NTASKS numactl --cpunodebind=0 --localalloc $HOME/miniconda3/envs/ShallowWaterGPU_HPC/bin/python3 mpiTesting.py -nx $NX -ny $NY --profile
|
||||
#mpirun -np $SLURM_NTASKS numactl --cpunodebind=0 --localalloc $HOME/miniconda3/envs/ShallowWaterGPU_HPC/bin/python3 mpiTesting.py -nx $NX -ny $NY --profile
|
||||
mpirun -np $SLURM_NTASKS $HOME/miniconda3/envs/ShallowWaterGPU_HPC/bin/python3 mpiTesting.py -nx $NX -ny $NY --profile
|
||||
cd $HOME/src/ShallowWaterGPU
|
||||
|
||||
## Copy files from work directory:
|
||||
|
Loading…
x
Reference in New Issue
Block a user