Next Gen Ascend Early User Program

The Ascend cluster, which includes 24 Quad GPU nodes deployed in 2022 and 250 Dual GPU nodes deployed this spring, is expected to be made available to all OSC clients on March 31, 2025, tentatively.

Who is eligible to participate in the Early User Program?

OSC is deploying the Next Gen Ascend cluster in collaboration with The Ohio State University Wexner Medical Center and the Ohio State University College of Medicine.

Current OSC users from OSU College of Medicine, as well as projects which already have Ascend access, are participating in the early user program. Notifications are sent to PIs and College of Medicine users in late February.

Early user period

March 3 - 31, 2025 (tentative)

Hardware

Detailed system specifications of Dual GPU nodes on the Next Gen Ascend:

  •  250 Dell R7525 server nodes, each with:
    • 2 AMD EPYC 7H12 processors (2.60 GHz, each with 60 usable cores) 
    • 2 NVIDIA A100 GPUs with 40GB memory each, PCIe, 250W
    • 512GB Memory with 472GB usable Memory
    • 1.92TB NVMe internal storage
    • HDR100 Infiniband (100 Gbps)

In addition, some Quad GPU nodes are also included for testing:

  •  2 Power Edge XE 8545 nodes, each with:
    • 2 AMD EPYC 7643 (Milan) processors (2.3 GHz, each with 44 usable cores) 
    • 4 NVIDIA A100 GPUs with 80GB memory each, supercharged by NVIDIA NVLink
    • 921GB usable RAM
    • 12.8TB NVMe internal storage
    • Mellanox/NVIDA 200 Gbps HDR InfiniBand​

Software environment

During the early access period, the programming environment and software packages will keep being updated; and the system may go down or jobs may be killed with little or no warning. If your work won't tolerate this level of instability, we recommend that you use other clusters instead.

The Next Gen Ascend cluster is now running on Red Hat Enterprise Linux (RHEL) 9, introducing several software-related changes compared to the RHEL 7/8 environment used on the Pitzer and original Ascend cluster. These updates provide access to modern tools and libraries but may also require adjustments to your workflows. Please refer to the Ascend Software Environment page for key software changes and available software.

Key change

A key change is that you are now required to specify the module version when loading any modules. For example, instead of using module load intel, you must use module load intel/2021.10.0. Failure to specify the version will result in an error message. 

Below is an example message when loading gcc without specifying the version:

$ module load gcc
Lmod has detected the following error:  These module(s) or extension(s) exist but cannot be loaded as requested: "gcc".

You encountered this error for one of the following reasons:
1. Missing version specification: On Ascend, you must specify an available version.
2. Missing required modules: Ensure you have loaded the appropriate compiler and MPI modules.

Try: "module spider gcc" to view available versions or required modules.

If you need further assistance, please contact oschelp@osc.edu with the subject line "lmod error: gcc"

Programming environment

The Next Gen Ascend supports programming in C, C++, and Fortran. The available compiler suites include Intel, oneAPI, and GCC.  Please refer to the Ascend Programming Environment page for details on compiler commands, parallel and GPU computing.

How to log into Next Gen Ascend

  • SSH Method

To login to Next Gen Ascend at OSC, ssh to the following hostname:

ascend-nextgen.osc.edu 

You can either use an ssh client application or execute ssh on the command line in a terminal window as follows:

ssh <username>@ascend-nextgen.osc.edu

From there, you are connected to the Next Gen Ascend login node and have access to the compilers and other software development tools on the RHEL 9 environment. You can run programs interactively or through batch requests. Please use batch jobs for any compute-intensive or memory-intensive work. 

  • OnDemand Method

You can also login to Next Gen Ascend at OSC with our OnDemand tool. The first step is to log into OnDemand. Then once logged in you can access it by clicking on "Clusters," and then selecting ">_Ascend Nextgen Shell Access."

How to test

  • For users NOT from the Ohio State University College of Medicine:

You will have access to:

  • Quad GPU nodes on the RHEL 8 environment: No special partition is required. The nodes are identical to those in the Ascend cluster you’ve been using, with the same software and programming environment. You can continue to run your previous Ascend jobs here, and refer to the original Ascend page for more details. Please note that, on March 31, the environment will be migrated to RHEL 9, along with a new suite of softwares. We recommend testing your workflows under the RHEL 9 environment (as outlined in the next two bullets) as soon as possible to ensure compatibility.
  • Quad GPU nodes on the RHEL 9 environment: You need to include --partition=quad to test it.  The nodes are identical to those in the Ascend cluster you’ve been using, but on RHEL 9, along with a new suite of softwares, as discussed in 'Software environment' on this page.
  • Dual GPU nodes on the RHEL 9 environment: You need to include --partition=nextgen to test it. It includes Dual GPU nodes discussed in 'Hardware' on this page, on RHEL 9, along with a new suite of softwares, as discussed in 'Software environment' on this page.
  • For users from the Ohio State University College of Medicine:

You will only have access to:

  • Dual GPU nodes on the RHEL 9 environment: You need to include --partition=nextgen to test it. It includes Dual GPU nodes discussed in 'Hardware' on this page, on RHEL 9, along with a new suite of softwares, as discussed in 'Software environment' on this page.
  • Please contact OSC Help if you need access to Qual GPU nodes. 

Scheduling policy 

  • Memory limit

It is strongly suggested to consider the memory use to the available per-core memory when users request OSC resources for their jobs.

Partition # of gpus per node Usable cores per node  default memory per core max usable memory per node
nextgen 2 120 4,027 MB 471.91 GB
quad 4 88 10,724 MB 921.59 GB
batch 4 88 10,724 MB 921.59 GB
  • CPU-only jobs

We reserve 1 core per 1 GPU. The CPU-only job can be scheduled but can only request up to 118 cores per dual GPU node and up to 84 cores per quad GPU node. You can also request multiple nodes for one CPU-only job. 

  • Batch limit

Partition Max walltime limit Min job size Max job size Note
nextgen 7-00:00:00 (168 hours) 1 core 16 nodes

Can request multiple partial nodes

quad 7-00:00:00 (168 hours) 1 core  2 nodes  Can request multiple partial nodes
debug-nextgen 1 hour 1 core 2 nodes  
debug-quad 1 hour 1 core 2 nodes  
  • Job limits on Ascend (including all partitions)

  Max # of cores in use Max # of GPUs in use Max # of running jobs  Max # of jobs to submit
Per user 5,632 96 256 1000
Per project 5,632 96 512 n/a

How do the jobs get charged?

  • For users NOT from the Ohio State University College of Medicine:

Jobs on Ascend (including Quad GPU nodes and Dual GPU nodes) are eligible for the early user program and will not be charged. All queued jobs submitted during the early user program will be deleted from the system at the end of the early user program to avoid any unwanted charges.

All jobs submitted after the early user program will be charged. The charge for core-hour and GPU-hour on Ascend is the same as the Standard compute core-hour and GPU-hour on Pitzer and Cardinal. Academic users can check the service costs page for more information. Please contact OSC Help if you have any questions about the charges. 

  • For users from the Ohio State University College of Medicine:

All College of Medicine users only have access to Dual GPU nodes. Jobs eligible for the early user program will not be charged. All queued jobs submitted during the early user program will be deleted from the system at the end of the early user program to avoid any unwanted charges.

Clients may continue to access the Pitzer cluster condo and other OSC resources during this time and get charged based on the existing service agreement, but may access only the Ascend cluster after its formal launch. HPC jobs on Ascend will be zero cost and have priority scheduling. We will update this CoM service page for more updated information.   

How do I find my jobs submitted during the early user program?

For any queued or running jobs, you can check the job information with either Slurm commands (which are discussed here) or the OSC OnDemand Jobs app by clicking "Active Jobs" and choosing "Ascend NextGen" as the cluster name.

For any completed jobs, you can check the job information using the OSC XDMoD Tool. Choose "Ascend" as "Resource." Check here for more information on how to use XDMoD.  

How do I get help?

Please feel free to contact OSC Help if you have any questions. 

Supercomputer: