Policies: Difference between revisions
No edit summary |
No edit summary |
||
Line 27: | Line 27: | ||
|} | |} | ||
In some exceptional cases users may be granted a higher storage quota in their home directories. An increased limit must be requested via Helpdesk's ticketing system. Such requests are reviewed on a individual basis and approved only in exceptional cases. | In some exceptional cases users may be granted a higher storage quota in their home directories. An increased limit must be requested via Helpdesk's ticketing system. Such requests are reviewed on a individual basis and approved only in exceptional cases. | ||
=== Checking your storage quota === | |||
One can use the following terminal commands to check your or your group member storage quota in home directory as well as to see how much is actually being used. | |||
<code>beegfs-ctl --getquota --uid $(id -u)</code> | |||
<code>beegfs-ctl --getquota --uid $(id -u <username>)</code> | |||
=== Additional storage - zdisk === | === Additional storage - zdisk === |
Revision as of 12:03, 17 September 2024
Important Note: Software configurations on NU HPC facilities are updated on a continuous basis. Minor policy changes also occur regularly. Some of these changes might not be immediately reflected on this website. The limits on job execution and maximum storage allocations are subject to change based on decisions made by the NU HPC Committee and actual system utilization.
Acceptable Use
The HPC system is a unique resource for NU researchers and the community. It has special characteristics, such as a large amount of RAM and the capability for massive parallelism. Due to its uniqueness and expense, its use is supervised by the HPC team to ensure efficient and fair utilization.
Storage quotas
Home directory
Users’ home directories are physically stored on fast SSD arrays that have very high bandwidth and endurance.
In the case of Shabyt cluster, the main storage servers are connected to the system via Infiniband interfaces (100 Mbit/s). All compute nodes are also connected via Infiniband. This provides very high bandwidth for users both when they access their data from the login node and when running their jobs on compute nodes using SLURM.
In Muon cluster the main SSD storage is in the login node with all SSD connected via fast u.2 interfaces. However, keep in mind that Muon's compute nodes have limited bandwidth with the login node (1 Mbit/s Ethernet). Therefore, batch jobs cannot read and write data faster than this.
System | Path | Default storage limit |
---|---|---|
Shabyt cluster | /shared/home/<username>
|
100 GB |
Muon cluster | /shared/home/<username>
|
100 GB |
In some exceptional cases users may be granted a higher storage quota in their home directories. An increased limit must be requested via Helpdesk's ticketing system. Such requests are reviewed on a individual basis and approved only in exceptional cases.
Checking your storage quota
One can use the following terminal commands to check your or your group member storage quota in home directory as well as to see how much is actually being used.
beegfs-ctl --getquota --uid $(id -u)
beegfs-ctl --getquota --uid $(id -u <username>)
Additional storage - zdisk
In Shabyt cluster, users can store larger amounts of data in their group directory on a slower HDD array. Keep in mind that this array does not have an Infiniband connection. Therefore, data access and transfer speeds from both the login node and compute nodes are limited to the standard 1 Mbit/s Ethernet speeds. In zdisk, each research group has a shared allocation. This can be particularly handy when the data needs to be transferred or shared within a single research group.
System | Path | Default storage limit |
---|---|---|
Shabyt cluster | /zdisk/<researchgroupname>
|
1 TB |
Muon cluster | /zdisk/<researchgroupname>
|
1 TB |
Again, in exceptional cases users/group may be granted an increased quota. Such requests are reviewed on an individual basis upon receiving a ticket by the PI via NU Helpdesk.
Data integrity and backup
Please be advised that users take full responsibility for the integrity and safety of their data stored on NU HPC facilities. While our clusters feature enterprise level hardware, failures are still a possibility. We do backup data in user home directories automatically several times a week (note that this applies only to your home directory in /shared/home, not to the group storage allocations in /zdisk). However, if a major hardware failure takes place, even if your data is eventually restored, you may not have access to it for a prolonged period of time while the system is being repaired. In some unfortunate situations it might take many days or even weeks to get everything back. Moreover, no system or storage solution is 100% reliable. Therefore we highly recommend that you backup most important and precious data on your personal computer from time to time.
Queues and the number of jobs
Currently, Shabyt has two partitions for user jobs. While at this time, when the system is still being configured and fine-tuned, there is no hardcoded limit on the number of jobs by any individual user, it will likely change in the near future.
Acknowledgments
If the computational resources provided by NU HPC facilities were an essential tool in your research that resulted in a publication, we ask that you include an acknowledgment in it. A natural place for it is the same section where you would typically acknowledge funding sources. Two of many possible formats of this acknowledgement are as follows:
The authors acknowledge the use of computational facilities provided by the Nazarbayev University Research Computing.
A.B. and C. D. (author initials) acknowledge the use of Shabyt HPC cluster at Nazarbayev University Research Computing.