Fluids

Fluids

Topics related to Fluent, CFX, Turbogrid and more.

Query understanding fluent batch mode .out file for parallelization.

    • cviveknair
      Subscriber

      Running fluent on a large cluster results in the following outfile section with The ID hostname Core PID etc..

      ------------------------------------------------------------------------------------------

      ID   Hostname       Core   O.S.    PID       Vendor

      ------------------------------------------------------------------------------------------

      n0-67 c455-051.xx.t  68/272  Linux-64  151468-151535  Intel(R) Xeon Phi(TM) 7250

      host  c455-051xx.t      Linux-64  148637      Intel(R) Xeon Phi(TM) 7250


      MPI Option Selected: intel

      Selected system interconnect: mpi-auto-selected

      ------------------------------------------------------------------------------------------

      I am used to seeing this section in the following way:

      #example parallelization 16 cores: 

      ID Hostname Core OS PID Vendor 

      n##. XX. 1/16 X. ##. XX 

      n##. XX. 2/16 X. ##. XX 

      n##. XX. 3/16 X. ##. XX 

      n##. XX. 4/16 X. ##. XX 

      and so on...

      While the ID n0-67 and the PID 151468-151535 point to 68 parallel subprocesses being created (as requested) It seems like they are all assigned to a single node 68/272. Is this an incorrect assumption on my part and is the process actually being divided across 68 cores are well ?

      Any help would be appreciated. Thank you.

    • DrAmine
      Ansys Employee
      How are you starting (command) and how many cores do exist on each node? Also print the proc and sys stats in Fluent.
    • cviveknair
      Subscriber
      Hello Dr. Amine An error on my part in the original question- "It seems like they are all assigned to a single node 68/272." I meant to say It seems like they are all assigned to a single CORE 68/272.
      each node has 68 cores with 4 threads each and In the run that produced the out file I am assigning the subprocess to run across 68 cores all within a single node only.
      The job scheduler is SLURM and this is the command used to start fluent within the script submitted for the job.
      #SBATCH -N 1# number of nodes
      #SBATCH -n 68#number of cores
      .
      .
      module load Ansys
      MYFILEHOSTS="hostlist.$SLURM_JOB_ID"
      srun hostname | sort > $MYFILEHOSTS
      # Run ANSYS Job
      "/home1/apps/ANSYS/2021R2/v212/fluent/bin/fluent" 3ddp -g -ssh -t$SLURM_NTASKS -mpi=intel -cnf=$MYFILEHOSTS -i myinpfile.inp
    • cviveknair
      Subscriber
      SYS-PROC AND SYS-STAT OUTPUTS.
    • cviveknair
      Subscriber
      Hello Dr. Amine An error on my part in the original question- "It seems like they are all assigned to a single node 68/272." I meant to say It seems like they are all assigned to a single CORE 68/272.
      each node has 68 cores with 4 threads each and In the run that produced the out file I am assigning the subprocess to run across 68 cores all within a single node only.
      The job scheduler is SLURM and this is the command used to start fluent within the script submitted for the job.
      #SBATCH -N 1# number of nodes
      #SBATCH -n 68#number of cores
      .
      .
      module load Ansys
      MYFILEHOSTS="hostlist.$SLURM_JOB_ID"
      srun hostname | sort > $MYFILEHOSTS
      # Run ANSYS Job
      "/home1/apps/ANSYS/2021R2/v212/fluent/bin/fluent" 3ddp -g -ssh -t$SLURM_NTASKS -mpi=intel -cnf=$MYFILEHOSTS -i myinpfile.inp

      PROC-STATS:
      ------------------------------------------------------------------------------
      | Virtual Mem Usage (GB) | Resident Mem Usage(GB) |
      ID | CurrentPeak| CurrentPeak| Page Faults
      ------------------------------------------------------------------------------
      host | 0.685974 0.685974| 0.125336 0.125336| 5
      n0 | 1.458281.45828 | 0.109219 0.109219| 2
      n1 | 1.458281.45828 | 0.106804 0.106804| 2
      n2 | 1.458281.45828 | 0.108505 0.108505| 7
      n3 | 1.458291.45829 | 0.107037 0.107037| 1
      n4 | 1.458281.45828 | 0.105835 0.105835| 2
      n5 | 1.458281.45828 | 0.107105 0.107105| 1
      n6 | 1.458281.45828 | 0.103867 0.103867| 5
      n7 | 1.458281.45828 | 0.109097 0.109097| 3
      n8 | 1.458281.45828 | 0.106197 0.106197| 3
      n9 | 1.458291.45829 | 0.108322 0.108322| 1
      n10| 1.458281.45828 | 0.108925 0.108925| 4
      n11| 1.458291.45829 | 0.106041 0.106041| 5
      n12| 1.458281.45828 | 0.109291 0.109291| 4
      n13| 1.458281.45828 | 0.108879 0.108879| 5
      n14| 1.458281.45828 | 0.108509 0.108509| 6
      n15| 1.458281.45828 | 0.108433 0.108433| 7
      n16| 1.458281.45828 | 0.104279 0.104279| 4
      n17| 1.458281.45828 | 0.103687 0.103687| 5
      n18| 1.458281.45828 | 0.108402 0.108402| 6
      ------------------------------------------------------------------------------
      | Virtual Mem Usage (GB) | Resident Mem Usage(GB) |
      ID | CurrentPeak| CurrentPeak| Page Faults
      ------------------------------------------------------------------------------
      n19| 1.458281.45828 | 0.103561 0.103561| 2
      n20| 1.458281.45828 | 0.106049 0.106049| 2
      n21| 1.458281.45828 | 0.108795 0.108795| 2
      n22| 1.458281.45828 | 0.108913 0.108913| 2
      n23| 1.458281.45828 | 0.108597 0.108597| 5
      n24| 1.458281.45828 | 0.106255 0.106255| 2
      n25| 1.458281.45828 | 0.108864 0.108864| 1
      n26| 1.458291.45829 | 0.105907 0.105907| 2
      n27| 1.458281.45828 | 0.107899 0.107899| 3
      n28| 1.458291.45829 | 0.103806 0.103806| 5
      n29| 1.458281.45828 | 0.105881 0.105881| 4
      n30| 1.458281.45828 | 0.106823 0.106823| 2
      n31| 1.458281.45828 | 0.103672 0.103672| 2
      n32| 1.458281.45828 | 0.109081 0.109081| 3
      n33| 1.458291.45829 | 0.108501 0.108501| 6
      n34| 1.458281.45828 | 0.105602 0.105602| 3
      n35| 1.458291.45829 | 0.107803 0.107803| 2
      n36| 1.458281.45828 | 0.108860.10886 | 2
      n37| 1.458291.45829 | 0.106445 0.106445| 3
      n38| 1.458281.45828 | 0.108925 0.108925| 3
      ------------------------------------------------------------------------------
      | Virtual Mem Usage (GB) | Resident Mem Usage(GB) |
      ID | CurrentPeak| CurrentPeak| Page Faults
      ------------------------------------------------------------------------------
      n39| 1.458281.45828 | 0.108524 0.108524| 4
      n40| 1.458281.45828 | 0.109142 0.109142| 2
      n41| 1.458281.45828 | 0.108612 0.108612| 4
      n42| 1.458291.45829 | 0.103645 0.103645| 3
      n43| 1.458281.45828 | 0.103748 0.103748| 3
      n44| 1.458281.45828 | 0.108551 0.108551| 3
      n45| 1.458281.45828 | 0.108677 0.108677| 2
      n46| 1.458291.45829 | 0.110538 0.110538| 6
      n47| 1.458281.45828 | 0.107948 0.107948| 8
      n48| 1.458291.45829 | 0.108994 0.108994| 4
      n49| 1.458291.45829 | 0.106754 0.106754| 3
      n50| 1.458281.45828 | 0.109081 0.109081| 1
      n51| 1.458281.45828 | 0.108223 0.108223| 7
      n52| 1.458281.45828 | 0.104256 0.104256| 1
      n53| 1.458281.45828 | 0.109032 0.109032| 2
      n54| 1.458281.45828 | 0.108379 0.108379| 1
      n55| 1.458281.45828 | 0.103695 0.103695| 2
      n56| 1.458291.45829 | 0.104046 0.104046| 5
      n57| 1.458281.45828 | 0.108395 0.108395| 3
      n58| 1.458281.45828 | 0.106735 0.106735| 2
      ------------------------------------------------------------------------------
      | Virtual Mem Usage (GB) | Resident Mem Usage(GB) |
      ID | CurrentPeak| CurrentPeak| Page Faults
      ------------------------------------------------------------------------------
      n59| 1.458281.45828 | 0.108227 0.108227| 2
      n60| 1.458281.45828 | 0.106087 0.106087| 1
      n61| 1.458281.45828 | 0.108635 0.108635| 2
      n62| 1.458281.45828 | 0.108311 0.108311| 2
      n63| 1.458281.45828 | 0.106277 0.106277| 6
      n64| 1.458281.45828 | 0.107216 0.107216| 2
      n65| 1.458281.45828 | 0.108475 0.108475| 1
      n66| 1.458281.45828 | 0.107117 0.107117| 1
      n67| 1.458291.45829 | 0.106918 0.106918| 2
      ------------------------------------------------------------------------------
      Total| 99.849299.8492 | 7.418257.41825 | 219
      ------------------------------------------------------------------------------

      ------------------------------------------------------------------------------------------------
      | Virtual Mem Usage (GB)| Resident Mem Usage(GB)| System Mem (GB)
      Hostname| CurrentPeak | CurrentPeak |
      ------------------------------------------------------------------------------------------------
      c455-012.xx.t| 99.849299.8492| 7.418517.41851| 94.1257
      ------------------------------------------------------------------------------------------------
      Total | 99.849299.8492| 7.418517.41851|
      ------------------------------------------------------------------------------------------------
      SYS_STATS:
      ---------------------------------------------------------------------------------------
      | CPU | System Mem (GB)
      Hostname| Sock x Core x HTClock (MHz) Load | TotalAvailable
      ---------------------------------------------------------------------------------------
      c455-012.xx.t| 1 x 68 x 40 17.91| 94.125781.3468
      ---------------------------------------------------------------------------------------
      Total | 272 - -| 94.125781.3468
      ---------------------------------------------------------------------------------------



    • cviveknair
      Subscriber
      Hello Dr. Amine An error on my part in the original question- "It seems like they are all assigned to a single node 68/272." I meant to say It seems like they are all assigned to a single CORE 68/272.
      each node has 68 cores with 4 threads each and In the run that produced the out file I am assigning the subprocess to run across 68 cores all within a single node only.
      The job scheduler is SLURM and this is the command used to start fluent within the script submitted for the job.
      #SBATCH -N 1# number of nodes
      #SBATCH -n 68#number of cores
      .
      .
      module load Ansys
      MYFILEHOSTS="hostlist.$SLURM_JOB_ID"
      srun hostname | sort > $MYFILEHOSTS
      # Run ANSYS Job
      "/home1/apps/ANSYS/2021R2/v212/fluent/bin/fluent" 3ddp -g -ssh -t$SLURM_NTASKS -mpi=intel -cnf=$MYFILEHOSTS -i myinpfile.inp

      PROC-STATS:
      ------------------------------------------------------------------------------
      | Virtual Mem Usage (GB) | Resident Mem Usage(GB) |
      ID | CurrentPeak| CurrentPeak| Page Faults
      ------------------------------------------------------------------------------
      host | 0.685974 0.685974| 0.125336 0.125336| 5
      n0 | 1.458281.45828 | 0.109219 0.109219| 2
      n1 | 1.458281.45828 | 0.106804 0.106804| 2
      n2 | 1.458281.45828 | 0.108505 0.108505| 7
      n3 | 1.458291.45829 | 0.107037 0.107037| 1
      n4 | 1.458281.45828 | 0.105835 0.105835| 2
      n5 | 1.458281.45828 | 0.107105 0.107105| 1
      n6 | 1.458281.45828 | 0.103867 0.103867| 5
      n7 | 1.458281.45828 | 0.109097 0.109097| 3
      n8 | 1.458281.45828 | 0.106197 0.106197| 3
      n9 | 1.458291.45829 | 0.108322 0.108322| 1
      n10| 1.458281.45828 | 0.108925 0.108925| 4
      n11| 1.458291.45829 | 0.106041 0.106041| 5
      n12| 1.458281.45828 | 0.109291 0.109291| 4
      n13| 1.458281.45828 | 0.108879 0.108879| 5
      n14| 1.458281.45828 | 0.108509 0.108509| 6
      n15| 1.458281.45828 | 0.108433 0.108433| 7
      n16| 1.458281.45828 | 0.104279 0.104279| 4
      n17| 1.458281.45828 | 0.103687 0.103687| 5
      n18| 1.458281.45828 | 0.108402 0.108402| 6
      ------------------------------------------------------------------------------
      | Virtual Mem Usage (GB) | Resident Mem Usage(GB) |
      ID | CurrentPeak| CurrentPeak| Page Faults
      ------------------------------------------------------------------------------
      n19| 1.458281.45828 | 0.103561 0.103561| 2
      n20| 1.458281.45828 | 0.106049 0.106049| 2
      n21| 1.458281.45828 | 0.108795 0.108795| 2
      n22| 1.458281.45828 | 0.108913 0.108913| 2
      n23| 1.458281.45828 | 0.108597 0.108597| 5
      n24| 1.458281.45828 | 0.106255 0.106255| 2
      n25| 1.458281.45828 | 0.108864 0.108864| 1
      n26| 1.458291.45829 | 0.105907 0.105907| 2
      n27| 1.458281.45828 | 0.107899 0.107899| 3
      n28| 1.458291.45829 | 0.103806 0.103806| 5
      n29| 1.458281.45828 | 0.105881 0.105881| 4
      n30| 1.458281.45828 | 0.106823 0.106823| 2
      n31| 1.458281.45828 | 0.103672 0.103672| 2
      n32| 1.458281.45828 | 0.109081 0.109081| 3
      n33| 1.458291.45829 | 0.108501 0.108501| 6
      n34| 1.458281.45828 | 0.105602 0.105602| 3
      n35| 1.458291.45829 | 0.107803 0.107803| 2
      n36| 1.458281.45828 | 0.108860.10886 | 2
      n37| 1.458291.45829 | 0.106445 0.106445| 3
      n38| 1.458281.45828 | 0.108925 0.108925| 3
      ------------------------------------------------------------------------------
      | Virtual Mem Usage (GB) | Resident Mem Usage(GB) |
      ID | CurrentPeak| CurrentPeak| Page Faults
      ------------------------------------------------------------------------------
      n39| 1.458281.45828 | 0.108524 0.108524| 4
      n40| 1.458281.45828 | 0.109142 0.109142| 2
      n41| 1.458281.45828 | 0.108612 0.108612| 4
      n42| 1.458291.45829 | 0.103645 0.103645| 3
      n43| 1.458281.45828 | 0.103748 0.103748| 3
      n44| 1.458281.45828 | 0.108551 0.108551| 3
      n45| 1.458281.45828 | 0.108677 0.108677| 2
      n46| 1.458291.45829 | 0.110538 0.110538| 6
      n47| 1.458281.45828 | 0.107948 0.107948| 8
      n48| 1.458291.45829 | 0.108994 0.108994| 4
      n49| 1.458291.45829 | 0.106754 0.106754| 3
      n50| 1.458281.45828 | 0.109081 0.109081| 1
      n51| 1.458281.45828 | 0.108223 0.108223| 7
      n52| 1.458281.45828 | 0.104256 0.104256| 1
      n53| 1.458281.45828 | 0.109032 0.109032| 2
      n54| 1.458281.45828 | 0.108379 0.108379| 1
      n55| 1.458281.45828 | 0.103695 0.103695| 2
      n56| 1.458291.45829 | 0.104046 0.104046| 5
      n57| 1.458281.45828 | 0.108395 0.108395| 3
      n58| 1.458281.45828 | 0.106735 0.106735| 2
      ------------------------------------------------------------------------------
      | Virtual Mem Usage (GB) | Resident Mem Usage(GB) |
      ID | CurrentPeak| CurrentPeak| Page Faults
      ------------------------------------------------------------------------------
      n59| 1.458281.45828 | 0.108227 0.108227| 2
      n60| 1.458281.45828 | 0.106087 0.106087| 1
      n61| 1.458281.45828 | 0.108635 0.108635| 2
      n62| 1.458281.45828 | 0.108311 0.108311| 2
      n63| 1.458281.45828 | 0.106277 0.106277| 6
      n64| 1.458281.45828 | 0.107216 0.107216| 2
      n65| 1.458281.45828 | 0.108475 0.108475| 1
      n66| 1.458281.45828 | 0.107117 0.107117| 1
      n67| 1.458291.45829 | 0.106918 0.106918| 2
      ------------------------------------------------------------------------------
      Total| 99.849299.8492 | 7.418257.41825 | 219
      ------------------------------------------------------------------------------

      ------------------------------------------------------------------------------------------------
      | Virtual Mem Usage (GB)| Resident Mem Usage(GB)| System Mem (GB)
      Hostname| CurrentPeak | CurrentPeak |
      ------------------------------------------------------------------------------------------------
      c455-012.xx.t| 99.849299.8492| 7.418517.41851| 94.1257
      ------------------------------------------------------------------------------------------------
      Total | 99.849299.8492| 7.418517.41851|
      ------------------------------------------------------------------------------------------------
      SYS_STATS:
      ---------------------------------------------------------------------------------------
      | CPU | System Mem (GB)
      Hostname| Sock x Core x HTClock (MHz) Load | TotalAvailable
      ---------------------------------------------------------------------------------------
      c455-012.xx.t| 1 x 68 x 40 17.91| 94.125781.3468
      ---------------------------------------------------------------------------------------
      Total | 272 - -| 94.125781.3468
      ---------------------------------------------------------------------------------------



    • cviveknair
      Subscriber
      PROC-STATS:
      ------------------------------------------------------------------------------
      | Virtual Mem Usage (GB) | Resident Mem Usage(GB) |
      ID | CurrentPeak| CurrentPeak| Page Faults
      ------------------------------------------------------------------------------
      host | 0.685974 0.685974| 0.125336 0.125336| 5
      n0 | 1.458281.45828 | 0.109219 0.109219| 2
      n1 | 1.458281.45828 | 0.106804 0.106804| 2
      n2 | 1.458281.45828 | 0.108505 0.108505| 7
      .
      .
      n64| 1.458281.45828 | 0.107216 0.107216| 2
      n65| 1.458281.45828 | 0.108475 0.108475| 1
      n66| 1.458281.45828 | 0.107117 0.107117| 1
      n67| 1.458291.45829 | 0.106918 0.106918| 2
      ------------------------------------------------------------------------------
      Total| 99.849299.8492 | 7.418257.41825 | 219
      ------------------------------------------------------------------------------

      ------------------------------------------------------------------------------------------------
      | Virtual Mem Usage (GB)| Resident Mem Usage(GB)| System Mem (GB)
      Hostname| CurrentPeak | CurrentPeak |
      ------------------------------------------------------------------------------------------------
      c455-012.xx.t| 99.849299.8492| 7.418517.41851| 94.1257
      ------------------------------------------------------------------------------------------------
      Total | 99.849299.8492| 7.418517.41851|
      ------------------------------------------------------------------------------------------------
      SYS_STATS:
      ---------------------------------------------------------------------------------------
      | CPU | System Mem (GB)
      Hostname| Sock x Core x HTClock (MHz) Load | TotalAvailable
      ---------------------------------------------------------------------------------------
      c455-012.xx.t| 1 x 68 x 40 17.91| 94.125781.3468
      ---------------------------------------------------------------------------------------
      Total | 272 - -| 94.125781.3468
      ---------------------------------------------------------------------------------------

    • DrAmine
      Ansys Employee
      I do not issue and recommend switching hyper-threading off.
    • cviveknair
      Subscriber
      Hi Dr. Amine,
      Im sorry I didn't quite understand your first statement. Just for my clarity, are the subprocesses being run in 68 threads (or cores) and not just a single thread (or core)?
      Im fairly certain that this is the case but I just want a confirmation from someone.

      I will also turn off hyper threading in subsequent runs.
      Thank you.

Viewing 8 reply threads
  • The topic ‘Query understanding fluent batch mode .out file for parallelization.’ is closed to new replies.