diff --git a/doc.zih.tu-dresden.de/docs/application/misc/active_application.png b/doc.zih.tu-dresden.de/docs/application/misc/active_application.png new file mode 100644 index 0000000000000000000000000000000000000000..fa2cc0d172499744384e7a760796a13aadfd3f9d Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/active_application.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/app_data.png b/doc.zih.tu-dresden.de/docs/application/misc/app_data.png new file mode 100644 index 0000000000000000000000000000000000000000..e2d9412cb30109dd7061113d16dde358b3c5f9f0 Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/app_data.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/app_finalize.png b/doc.zih.tu-dresden.de/docs/application/misc/app_finalize.png new file mode 100644 index 0000000000000000000000000000000000000000..c409c27e462f5bd0f6da7c0f1cab89e84c63123e Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/app_finalize.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/app_pi.png b/doc.zih.tu-dresden.de/docs/application/misc/app_pi.png new file mode 100644 index 0000000000000000000000000000000000000000..9819bbbad81d92712c1552beae87d1f6fb3c897d Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/app_pi.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/app_project_data.png b/doc.zih.tu-dresden.de/docs/application/misc/app_project_data.png new file mode 100644 index 0000000000000000000000000000000000000000..4018dadc38db87b07ce36c322cd0739f96d858d1 Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/app_project_data.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/app_project_type.png b/doc.zih.tu-dresden.de/docs/application/misc/app_project_type.png new file mode 100644 index 0000000000000000000000000000000000000000..a953b3d1094f8c37f38a8e31945ad4fe3836173b Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/app_project_type.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/app_resource_cpu.png b/doc.zih.tu-dresden.de/docs/application/misc/app_resource_cpu.png new file mode 100644 index 0000000000000000000000000000000000000000..c2ea8ca750c0b5ec2c7ff40b8c198eb2400eb2a5 Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/app_resource_cpu.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/app_resource_gpu.png b/doc.zih.tu-dresden.de/docs/application/misc/app_resource_gpu.png new file mode 100644 index 0000000000000000000000000000000000000000..e2d6b93279c8ca779e29b07d9ad13eecd4d32e1a Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/app_resource_gpu.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/app_resources.png b/doc.zih.tu-dresden.de/docs/application/misc/app_resources.png new file mode 100644 index 0000000000000000000000000000000000000000..45a70852f2de68c0d1f7ae01dc3e0820cca7cc7a Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/app_resources.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/app_software.png b/doc.zih.tu-dresden.de/docs/application/misc/app_software.png new file mode 100644 index 0000000000000000000000000000000000000000..d0318561ee019108118143b4e241a92709e02dde Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/app_software.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/application_1.png b/doc.zih.tu-dresden.de/docs/application/misc/application_1.png new file mode 100644 index 0000000000000000000000000000000000000000..d44bee573c20ce312aadcb4ffc6e8a1ec1d6c203 Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/application_1.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/email_link.png b/doc.zih.tu-dresden.de/docs/application/misc/email_link.png new file mode 100644 index 0000000000000000000000000000000000000000..f513abcb04730ff5d6d710f49ea179c751fe9035 Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/email_link.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/email_login.png b/doc.zih.tu-dresden.de/docs/application/misc/email_login.png new file mode 100644 index 0000000000000000000000000000000000000000..e45b24a0ce3805d9d66c367fc5e9fc8e2b03d924 Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/email_login.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/finalized_application.png b/doc.zih.tu-dresden.de/docs/application/misc/finalized_application.png new file mode 100644 index 0000000000000000000000000000000000000000..9e77f8b1a3a9db4afd9eab58a77d3aae585ec6d4 Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/finalized_application.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/idm_login.png b/doc.zih.tu-dresden.de/docs/application/misc/idm_login.png new file mode 100644 index 0000000000000000000000000000000000000000..a6142848df2a9a5cb7b4f4ff1f6aef191b0957be Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/idm_login.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/new_application.png b/doc.zih.tu-dresden.de/docs/application/misc/new_application.png new file mode 100644 index 0000000000000000000000000000000000000000..dba92b385bc54dfdd2b0b187f340092d1b96b3e4 Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/new_application.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/nhr_10a_data_mgmt.png b/doc.zih.tu-dresden.de/docs/application/misc/nhr_10a_data_mgmt.png new file mode 100644 index 0000000000000000000000000000000000000000..336aa89a4663dc87479ae8e2e676d10081c0b95f Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/nhr_10a_data_mgmt.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/nhr_10b_data_mgmt.png b/doc.zih.tu-dresden.de/docs/application/misc/nhr_10b_data_mgmt.png new file mode 100644 index 0000000000000000000000000000000000000000..adfa98cc281762a7424dc63b6757b2672dcd5869 Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/nhr_10b_data_mgmt.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/nhr_11_upload_regular.png b/doc.zih.tu-dresden.de/docs/application/misc/nhr_11_upload_regular.png new file mode 100644 index 0000000000000000000000000000000000000000..6fe33c3d3ccc3163b0e41b4287fc2284ad4ed7c8 Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/nhr_11_upload_regular.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/nhr_11_upload_trial_project.png b/doc.zih.tu-dresden.de/docs/application/misc/nhr_11_upload_trial_project.png new file mode 100644 index 0000000000000000000000000000000000000000..d96ababa48a4d92c30035c2be49ec5876a21baa4 Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/nhr_11_upload_trial_project.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/nhr_1a_applicationlist.png b/doc.zih.tu-dresden.de/docs/application/misc/nhr_1a_applicationlist.png new file mode 100644 index 0000000000000000000000000000000000000000..4160ee87606f1e154d2545658d8f78a8afdd1ea5 Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/nhr_1a_applicationlist.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/nhr_1b_applicationlist.png b/doc.zih.tu-dresden.de/docs/application/misc/nhr_1b_applicationlist.png new file mode 100644 index 0000000000000000000000000000000000000000..d48ac06d0208d80183374dc1f16745e3caec2604 Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/nhr_1b_applicationlist.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/nhr_2_projecttype.png b/doc.zih.tu-dresden.de/docs/application/misc/nhr_2_projecttype.png new file mode 100644 index 0000000000000000000000000000000000000000..9e426c81a74da22bc20f34f146aecbff42385b10 Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/nhr_2_projecttype.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/nhr_3_choose_pi_pc.png b/doc.zih.tu-dresden.de/docs/application/misc/nhr_3_choose_pi_pc.png new file mode 100644 index 0000000000000000000000000000000000000000..09203845b315a5a29f79aeb6ebda92fba2fc8628 Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/nhr_3_choose_pi_pc.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/nhr_4_data_pi.png b/doc.zih.tu-dresden.de/docs/application/misc/nhr_4_data_pi.png new file mode 100644 index 0000000000000000000000000000000000000000..f7270622851bc0d99bb98687c9034ec590617194 Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/nhr_4_data_pi.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/nhr_5_data_pc.png b/doc.zih.tu-dresden.de/docs/application/misc/nhr_5_data_pc.png new file mode 100644 index 0000000000000000000000000000000000000000..eb6597b4de500179b29086987f312c4861e6f8a7 Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/nhr_5_data_pc.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/nhr_6_resources.png b/doc.zih.tu-dresden.de/docs/application/misc/nhr_6_resources.png new file mode 100644 index 0000000000000000000000000000000000000000..67fca210ebf7f7d3c7946ca57b7617b15b7c9b02 Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/nhr_6_resources.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/nhr_7a_CPUresources.png b/doc.zih.tu-dresden.de/docs/application/misc/nhr_7a_CPUresources.png new file mode 100644 index 0000000000000000000000000000000000000000..dc87c9d98e0c4f4c44d826944561badaf75e4c8d Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/nhr_7a_CPUresources.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/nhr_7b_GPUresources.png b/doc.zih.tu-dresden.de/docs/application/misc/nhr_7b_GPUresources.png new file mode 100644 index 0000000000000000000000000000000000000000..c9b6632e426ebf9d2b7219b645c18109631043cb Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/nhr_7b_GPUresources.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/nhr_8a_project_data.png b/doc.zih.tu-dresden.de/docs/application/misc/nhr_8a_project_data.png new file mode 100644 index 0000000000000000000000000000000000000000..7053cd648e5dc48f3f441e5ed89ec7897d3fe1db Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/nhr_8a_project_data.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/nhr_8b_project_data.png b/doc.zih.tu-dresden.de/docs/application/misc/nhr_8b_project_data.png new file mode 100644 index 0000000000000000000000000000000000000000..cf63628985004c1c67a444533eef952bdcb98289 Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/nhr_8b_project_data.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/nhr_8c_project_data.png b/doc.zih.tu-dresden.de/docs/application/misc/nhr_8c_project_data.png new file mode 100644 index 0000000000000000000000000000000000000000..b04dd327b58ef0f4e05d643e754f7d16d4b5cad1 Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/nhr_8c_project_data.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/nhr_9a_software.png b/doc.zih.tu-dresden.de/docs/application/misc/nhr_9a_software.png new file mode 100644 index 0000000000000000000000000000000000000000..36ff231983c4d5eee256c166d3b94ff15a34bd80 Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/nhr_9a_software.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/nhr_9b_software.png b/doc.zih.tu-dresden.de/docs/application/misc/nhr_9b_software.png new file mode 100644 index 0000000000000000000000000000000000000000..58be917837815d921a1f55ac97ded8d624d58fa3 Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/nhr_9b_software.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/overview_application.png b/doc.zih.tu-dresden.de/docs/application/misc/overview_application.png new file mode 100644 index 0000000000000000000000000000000000000000..0053853345ff837b0406696dc9d0614da10b28d5 Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/overview_application.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/project_application.png b/doc.zih.tu-dresden.de/docs/application/misc/project_application.png new file mode 100644 index 0000000000000000000000000000000000000000..1acd5c7d0235b3e0f3143429aeefbdcb9b030dc3 Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/project_application.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/misc/project_application_small.png b/doc.zih.tu-dresden.de/docs/application/misc/project_application_small.png new file mode 100644 index 0000000000000000000000000000000000000000..e172875e1da773300e16e21fdd9ebdd63a7585d4 Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/application/misc/project_application_small.png differ diff --git a/doc.zih.tu-dresden.de/docs/application/overview.md b/doc.zih.tu-dresden.de/docs/application/overview.md index ac8a94a3c78eadbd1e18fa5b104169ceb05f5816..df98127158af3af36724cb3f9d20acf483eeb8af 100644 --- a/doc.zih.tu-dresden.de/docs/application/overview.md +++ b/doc.zih.tu-dresden.de/docs/application/overview.md @@ -14,14 +14,14 @@ To learn more about applying for a project or a course, check the following page ### Get Access without an existing ZIH Login -To use the resources, you need a [valid ZIH login][2]. To get a ZIH login and the access to HPC, -please use the [HPC login application form][3]. +To use the resources, you need a valid ZIH login. To get a ZIH login and the access to HPC, +please use the [HPC login application form][2]. ### Get Access with a valid ZIH Login When you have a valid ZIH login, there are two possibilities for you: Either the manager or administrator of a HPC project needs to add you as a member of this project via the -[manager view](project_management.md#manage-project-members-dis-enable), or you have to +[manager view](project_management.md), or you have to [apply for a project](project_request_form.md). ## Acknowledgment in Publications @@ -33,5 +33,5 @@ publications that are based on granted HPC resources of the TU Dresden. We provide some [acknowledgment examples](acknowledgement.md) that show you how to do that. [1]: https://tu-dresden.de/zih/hochleistungsrechnen/zugang -[2]: https://tu-dresden.de/zih/dienste/service-katalog/zugangsvoraussetzung +[2]: https://selfservice.zih.tu-dresden.de/index.php/hpclogin/noLogin [3]: https://selfservice.zih.tu-dresden.de/l/index.php/hpclogin diff --git a/doc.zih.tu-dresden.de/docs/application/project_request_form_jards.md b/doc.zih.tu-dresden.de/docs/application/project_request_form_jards.md new file mode 100644 index 0000000000000000000000000000000000000000..f81773abce2ef7e954250a88509c8163c0fe0612 --- /dev/null +++ b/doc.zih.tu-dresden.de/docs/application/project_request_form_jards.md @@ -0,0 +1,431 @@ +# Project Request Form Jards + +This page describes the steps to fill the form on +[https://hpcprojekte.zih.tu-dresden.de/](https://hpcprojekte.zih.tu-dresden.de/application/). + +Please be aware of, with the changeover to JARDS, the approval +period lasts between 4 and 8 weeks. Except for trial projects and courses, +regular projects are only approved after a positive review. + +If you have not already reached the login form through a specialized page, +here you have an overview of the possible project applications. +There are technical and scientific reviews of the project applications, +which vary depending on the type of application. + +Since 2021, HPC at universities has been restructured by the NHR network. +The network consists of nine centers, which operate the systems and offer +a coordinated consulting service on the methodological competence of scientific HPC. +The aim is to provide scientists at German universities with computing capacity +for their research and to strengthen their skills in the efficient use of this resource. + +Due to the structuring there are different ways to access [HPC resources](https://doc.zih.tu-dresden.de/jobs_and_resources/hardware_overview/). + +## Workflow + +There are several steps for a project application ... + +- Selection which kind of project proposal is applicable → [Kind of Application](#kind-of-application) +- Login to the application portal → [Login](#login) +- Managing one or several applications in the application portal → [Jards Frontend](#jards-frontend) +- Details for each project application → [Application](#application) +- Finalize and submit your application → [Finalized Application](#finalized-application) + +Afterwards, the submitted application undergoes a technical and scientific +review and you will be informed about the result in 1 to N work days. +If a project is granted, you can use the project management portal +[https://hpcprojekte.zih.tu-dresden.de/](https://hpcprojekte.zih.tu-dresden.de/application/). + +All steps for an application are documented in detail below. + +### Kind of Application + +Since January 2021 ZIH, TU Dresden is a NHR-center (Nationales Hochleistungsrechnen). +More details can be found in [https://tu-dresden.de/zih/hochleistungsrechnen/nhr-center](https://tu-dresden.de/zih/hochleistungsrechnen/nhr-center). + +At ZIH, TU Dresden we have 3 different kinds of application +for applying for HPC resources: NHR, Saxony and TUD/TIER3. + +- [NHR](https://hpcprojekte.zih.tu-dresden.de/application/login.php?appkind=nhr) +- [Saxony](https://hpcprojekte.zih.tu-dresden.de/application/login.php?appkind=saxony) +- [TUD/Tier3](https://hpcprojekte.zih.tu-dresden.de/application/login.php?appkind=tier3) + +![HPC Project Application][37] + +??? abstract "National NHR Application:" + + #### NHR + + This application kind is recommended for: + + * for researchers from all over Germany there is the possibility + to apply for HPC resources at the NHR Center of the TU Dresden + * for all research projects related to our focus topics + * Life Sciences + * Earth System Science + * Methods for big data and data analysis and management + * Machine Learning + * Tiered storage architectures and I/O optimization + * Performance and energy efficiency analysis and optimization + * Further information can be found in [https://tu-dresden.de/zih/hochleistungsrechnen/nhr-center](https://tu-dresden.de/zih/hochleistungsrechnen/nhr-center). + If your research topic fits the above focus topics, the application kind 'NHR' is recommended. + * for other focuses other NHR centers are available [https://www.nhr-gs.de/ueber-uns/nhr-verein](https://www.nhr-gs.de/ueber-uns/nhr-verein) + + Application Login: + + * [https://hpcprojekte.zih.tu-dresden.de/application/login.php?appkind=nhr](https://hpcprojekte.zih.tu-dresden.de/application/login.php?appkind=nhr) + +??? abstract "Saxony Application:" + + #### Saxony + + This application is: + + * only applicable for researchers from universities in Saxony. + * If you are from Saxony and your project matches the NHR + focus topics mentioned above, then either application kind + is feasible and there are no differences for the resulting + project between the two. We recommend to choose the application kind "NHR". + * to apply at the NHR Center of the TU-Dresden for further focus topics + which are not covered by an NHR application + + Application Login: + + * [https://hpcprojekte.zih.tu-dresden.de/application/login.php?appkind=saxony](https://hpcprojekte.zih.tu-dresden.de/application/login.php?appkind=saxony) + +??? abstract "TU Dresden (Tier 3) Application:" + + #### TU Dresden + + This application is: + + * only for researchers at TU Dresden entitled + to special HPC contingents, either from TU Dresdens + own contribution to the funding of the NHR center or from own HPC funding. + + Application Login: + + * [https://hpcprojekte.zih.tu-dresden.de/application/login.php?appkind=tier3](https://hpcprojekte.zih.tu-dresden.de/application/login.php?appkind=tier3) + +!!! hint "ScaDS.AI Application:" + + ### ScaDS.AI + + This is a separate kind of application for ScaDS.AI Dresden/Leipzig + [https://scads.ai/research/resources-for-research/](https://scads.ai/research/resources-for-research/), + only applicable for members, assiciated members and partners of ScaDS.AI. + The following application steps are the same as for the others, though. + +### Login + +!!! info ZIH login + For the application of a project no ZIH Login is necessary. + + You only need a ZIH login if you want to access + the resources you need for your project. + + If you have a TU Dresden account already, then use the first option. Otherwise use the E-Mail callback option. + +#### With TU Dresden Account + +If you have a ZIH Login, you can use it in the form. + +![Login via TU Dresden Account][2] + +#### With E-Mail Callback + +If you do not have a ZIH login, you can have a login session emailed to you, +just enter your e-mail address in the form. + +![Login via E-Mail Callback][3] + +Click on the link in the mail. + +![Callback Mail][4] + +### Jards Frontend + +After logging in, you will land on the overview page, which is divided into 3 parts. + +#### Overview + +In the first part you will get information about the site. +Please pay attention to the title of the page and what project +type of request you are making. + +![Overview][7] + +#### Active Applications + +In this section you can see all applications that have not yet been sent. +You can edit these applications at any time. +![Active Application][6] + +##### New Application + +To make a new request, press the button at the end of the listing. + +!!! hint "Create New Application" + Be aware of, when you press the 'New Project Application' button, + your project will be assigned a unique project ID. + If you are not sure about some details, you can save them and edit them later. + +If you decide to create a project proposal for the same project again, +you will receive a new unique project ID. + +![New Application][8] + +!!! info New Application Kind + Please pay attention to the title of the page and what kind of request you are making. + +#### Finalized Applications + +Applications that have already been sent can be viewed in this section. +These applications can no longer be updated and are under +review and will be activated in case of a positive result. + +![Finalized Application][5] + +### Application + +- [Project Type](#project-type) +- [Principal Investigator (PI) and Person to Contact (PC)](#principal-investigator) +- [Resources](#resources) +- [Project Data](#project-data) +- [Software](#software) +- [Data Management](#data_management) +- [Upload Full Project description](#upload-full-project-description) + +No matter which application you submit, the procedure is always identical. + +??? abstract "Project Type" + + ### Project Type + + After a new project application is created you can choose a + + * **Trial project**: With a trial projects you can test + whether the HPC resources meet your requirements. + Trial projects are all applications with up to 3500 CPUh/month + or 250 GPUh/month. No detailed project description is required. + * **Regular project**: A regular project are for data intensive and + compute intensive HPC applications, e.g. FEM, CFD and MD simulations. + Regular projects have higher resource requirements than the trial projects. + A detailed project description is required in addition to the online application. + * **Courses**: Courses is a project type for teaching courses, + where you can enable up to 5000 CPUh/month, without additional + communication. If there is a higher resource requirement needed, + it can be enabled after consultation the HPC support. + No detailed project description is required. + Please be aware, that in exceptional cases, ZIH reserves the right + to request a more detailed abstract. + + The project application differs depending on the type of + application - and thus which components of our HPC system - resources are applied for. + + ![Project Type Selection][21] + +??? abstract "Principal Investigator (PI) and Person to Contact (PC)" + + ### Principal Investigator (PI) and Person to Contact (PC) + + Choose the principal investigator and the person to contact. + On the Button 'change my role ...' below, you can change your organisational part in the project. + + Keep in mind that the principal investigator (PI) has to be: + + * institute director + * Professors + * research group leader + + ![Principal Investigator (PI)][22] + + If you have not yet submitted an application to us, you will + be asked to provide your personal information. + + ![Principal Investigator (PI)][23] + ![Principal Investigator (PI)][24] + +??? abstract "Resources" + + ### Resources + + Choose the needed resources. Typically, you can choose between GPU and CPU. + If you are unsure about some of the detail questions here and below, + please provide a rough estimate or your best guess instead of specifying 0 or no data. + + !!! info + You can only choose after you have specified a project type. + + For each choosen resource there are some questions after the project data. + + ![Resource Selection][12] + + Questions for CPU resources could be: + + - How much computing time is required for your project in the upcoming + compute time period of one year (in core-hours)? + - What is your job size and how many cores do you need? + - Define your memory: How many GB per core do you need? + + !!! hint "Hint" + If you have chosen only CPU resources, please be aware to specify + here only resources for pure CPU nodes. Resources needed on accelerators + (GPUs) must be specified in the corresponding section. If you like + to change your computing time requirements go one slide back. + + ![CPU][13] + + Questions for CPU resources could be: + + - How much computing time is required on GPU nodes for your HPC project (in GPU-hours)? + - How many GPU do you use for a single job? + - How much GB of memory do you use for your typical job? + - Which GPU programming scheme (CUDA, OpenACC, etc.) do you use? + - Which software do you deploy on GPUs (own code, third party)? + + !!! hint "Hint" + If you have chosen both CPU and GPU resources, but you do not need + compute time on GPUs type in 0 here to be able to continue with your + project application in Jards. + + ![GPU][14] + +??? abstract "Project Data" + + ### Project Data + + In this step, the basic conditions of the project must be specified. + + Depending on the selected project type, these can be different. + + For example: + + - How much time do you plan for your HPC project (add ending date)? + - What is your title of your HPC project (max. 250 char)? + - Define your keywords to explain your HPC project best. + - Please add a short project description (max. 6000 char). + - Explain your plans for this granting period (max. 500 char). + - Are you doing commissioned research (yes/no)? + - Who is sponsoring your HPC project? + - Classify your HPC project according to the following main and sub categories. + - Which methods (physical, mathematical, numerical, ...) are used? + - Which reasons make the use of this supercomputer necessary? + + !!! hint "Commissioned Research" + Are you doing commissioned research? + + Commissioned research means direct order by private third parties to the + TU Dresden as well as indirect assignments by private third parties on a + promoting institution (university , research institution, or similar institutions). + All research projects sponsored by public institutions such as state, country DFG, + EU etc. are excluded. + + This specification is very important. In the end you must confirm this information + in writing. For this purpose, a PDF is generated, which you must send back to us. + + ![Project Data][28] + ![Project Data][29] + ![Project Data][30] + +??? abstract "Software" + + ### Software + + Next you need to specify which software you want to use. + + Here are the following categories, you should answer: + + - Commercial Software Packages + - Compilers + - Programming Languages + - Other packages, not named here. + - Use own-developed packages, please name here and provide the links to + the open source software. + - Parallelization strategy + + !!!hint "Information About Required Software" + This is important for us, because we need to see if the software is installed, there is a valid license and it is available under the appropriate architecture. The biggest problem is the imagination of the license writers. We have to abide by these conditions. Even if we try to provide the common HPC software including licenses, not every software vendor offers appropriate licenses. + + ![Software][31] + ![Software][32] + +??? abstract "Data Management" + + ### Data Management + + The last thing to specify is the I/O behavior. All this information helps us + to check and, if necessary, to create the necessary conditions. + + For example: + + * How much memory do you need. + * How do your calculations behave (many small files or one big file)? + * Fetch data from external? How? + + ![Data][33] + ![Data][34] + +??? abstract "Upload Full Project description" + + ### Upload Full Project description + + If you are choosing the project type 'Regular project' a + detailed project description is needed to complete the application. + You can save the application form and add the detailed description later, + before submitting and finalize the HPC project application. + + You can find a template on this page: + [https://tu-dresden.de/zih/hochleistungsrechnen/zugang/projektantrag?set_language=en](https://tu-dresden.de/zih/hochleistungsrechnen/zugang/projektantrag?set_language=en). + +### Finalized Application + +The final step is to complete the application process. +The application is submitted to us and approved for review. Editing is no longer possible. + +In the next step you will receive an email with the data of the application +as a PDF. In it you will find a template for the confirmation of the commissioned research. + +![Finalize][18] + +!!! danger "Important" + Please send the confirmation of commissioned research signed back to us, + even if it is not commissioned research. + +[1]: misc/application_1.png "Select Application" +[2]: misc/idm_login.png "Login via TU Dresden Account" +[3]: misc/email_login.png "Login via E-Mail Callback" +[4]: misc/email_link.png "Login via E-Mail Callback" +[5]: misc/finalized_application.png "Finalize" +[6]: misc/active_application.png +[7]: misc/overview_application.png +[8]: misc/new_application.png +[10]: misc/app_project_type.png "Project Type Selection" +[11]: misc/app_pi.png "PI Selection" +[12]: misc/app_resources.png "Resource Selection" +[13]: misc/app_resource_cpu.png "Resource CPU" +[14]: misc/app_resource_gpu.png "Resource GPU" +[15]: misc/app_project_data.png "Project Data" +[16]: misc/app_software.png "Software" +[17]: misc/app_data.png "Data" +[18]: misc/app_finalize.png "Finalize" + +[19]: misc/nhr_1a_applicationlist.png +[20]: misc/nhr_1b_applicationlist.png +[21]: misc/nhr_2_projecttype.png +[22]: misc/nhr_3_choose_pi_pc.png +[23]: misc/nhr_4_data_pi.png +[24]: misc/nhr_5_data_pc.png +[25]: misc/nhr_6_resources.png +[26]: misc/nhr_7a_CPUresources.png +[27]: misc/nhr_7b_GPUresources.png +[28]: misc/nhr_8a_project_data.png +[29]: misc/nhr_8b_project_data.png +[30]: misc/nhr_8c_project_data.png +[31]: misc/nhr_9a_software.png +[32]: misc/nhr_9b_software.png +[33]: misc/nhr_10a_data_mgmt.png +[34]: misc/nhr_10b_data_mgmt.png +[35]: misc/nhr_11_upload_regular.png +[36]: misc/nhr_11_upload_trial_project.png +[37]: misc/project_application_small.png +[38]: misc/project_application.png diff --git a/doc.zih.tu-dresden.de/docs/application/project_request_form_jards2.md b/doc.zih.tu-dresden.de/docs/application/project_request_form_jards2.md new file mode 100644 index 0000000000000000000000000000000000000000..a51d6a96d69e5ce9d9dbd1fbd5dd516778ae4b05 --- /dev/null +++ b/doc.zih.tu-dresden.de/docs/application/project_request_form_jards2.md @@ -0,0 +1,337 @@ +# Project Request Form Jards + +This page describes the steps to fill the form on +[https://hpcprojekte.zih.tu-dresden.de/](https://hpcprojekte.zih.tu-dresden.de/application/). + +If you have not already reached the login form through a specialized page, +here you have an overview of the possible project applications. + +Since 2021, HPC at universities has been restructured by the NHR network. +The network consists of nine centers, which operate the systems and offer +a coordinated consulting service on the methodological competence of scientific HPC. +The aim is to provide scientists at German universities with computing capacity +for their research and to strengthen their skills in the efficient use of this resource. + +Due to the structuring there are different ways to access [HPC resources](https://doc.zih.tu-dresden.de/jobs_and_resources/hardware_overview/). + +## Workflow + + ``` mermaid + graph LR + A(NHR Type) --> |National| B(Login); + A --> |Saxony| B; + A --> |TU Dresden| B; + B --> |ZIH Login| H(Jards Frontend); + B --> |E-Mail Callback| H; + J --> |show| J; + J--> |copy| J; + H --> |create| J(Application); + J --> |edit| J; + J --> | submit | L(Finalized Application); + + click A href "http://localhost:8000/application/project_request_form_jards/#application-type" _self; + click E href "http://localhost:8000/application/project_request_form_jards/#nhr" _self; + click F href "http://localhost:8000/application/project_request_form_jards/#saxony" _self; + click G href "http://localhost:8000/application/project_request_form_jards/#tu-dresden" _self; + click B href "http://localhost:8000/application/project_request_form_jards/#login" _self; + click C href "http://localhost:8000/application/project_request_form_jards/#with-tu-dresden-idm-account" _self; + click D href "http://localhost:8000/application/project_request_form_jards/#with-tu-dresden-idm-account" _self; + ``` + +{: align="center"} + +=== "NHR Type" + + ### NHR Type + + Since January 2021 ZIH, TU Dresden is a NHR-center (Nationales Hochleistungsrechnen). + More details can be found in [https://tu-dresden.de/zih/hochleistungsrechnen/nhr-center](https://tu-dresden.de/zih/hochleistungsrechnen/nhr-center). + + At ZIH, TU Dresden we have 3 different application categories for applying for HPC resources: NHR, Saxony and TUD/TIER3. + + ![Picture 1: Select Application][1] + {: align="center"} + + ??? abstract "National NHR Application:" + + #### NHR + + This application is: + + * for researchers from all over Germany there is the possibility + to apply for HPC resources at the NHR Center of the TU Dresden + * for all research projects related to our focus topics + [https://tu-dresden.de/zih/hochleistungsrechnen/nhr-center](https://tu-dresden.de/zih/hochleistungsrechnen/nhr-center), application to our NHR center is recommended + * for other focuses other NHR centers are available [https://www.nhr-gs.de/ueber-uns/nhr-verein](https://www.nhr-gs.de/ueber-uns/nhr-verein) + + Application Login: + + * [https://hpcprojekte.zih.tu-dresden.de/application/login.php?appkind=nhr](https://hpcprojekte.zih.tu-dresden.de/application/login.php?appkind=nhr) + + ??? abstract "Saxony Application:" + + #### Saxony + + This application is: + + * for researchers from Saxon universities + * to apply at the NHR Center of the TU-Dresden for further focus topics + which are not covered by an NHR application + + Application Login: + + * [https://hpcprojekte.zih.tu-dresden.de/application/login.php?appkind=saxony](https://hpcprojekte.zih.tu-dresden.de/application/login.php?appkind=saxony) + + ??? abstract "TU Dresden (Tier 3) Application:" + + #### TU Dresden + + This application is: + + * for researchers of the TU Dresden with entitlement to special HPC contingents of the TU Dresden. + (This includes, for example, appointment commitments) + + Application Login: + + * [https://hpcprojekte.zih.tu-dresden.de/application/login.php?appkind=tier3](https://hpcprojekte.zih.tu-dresden.de/application/login.php?appkind=tier3) + + !!! hint "ScaDS.AI Application:" + + ### ScaDS.AI + + ScaDS.AI is not a NHR Type. If you need an application for ScaDS.AI, you know how. + +=== "Login" + + ### Login + + !!! info ZIH login + For the application of a project no ZIH Login is necessary. Only if you want to use the resources yourself, a login is required: [https://tu-dresden.de/zih/hochleistungsrechnen/hpc](https://tu-dresden.de/zih/hochleistungsrechnen/hpc). + + + #### With TU Dresden Account + + If you have a ZIH Login, you can use it in the form. + + ![Login via TU Dresden Account][2] + + #### With E-Mail Callback + + If you do not have a ZIH login, you can have a login session emailed to you, just enter your e-mail address in the form. + + ![Login via E-Mail Callback][3] + + Click on the link in the mail. + + ![Calback Mail][4] + +=== "Jards Frontend" + + ### Jards Frontend + + After logging in, you will land on the overview page, which is divided into 3 parts. + + #### Overview + + In the first part you will get information about the site. Please pay attention to the title of the page and what type of request you are making. + + ![Overview][7] + + #### Active Applications + + In this section you can see all applications that have not yet been sent. You can edit these applications at any time. + ![Active Application][6] + + ##### New Application + + To make a new request, press the button at the end of the listing. + + !!! hint "Create New Application" + Be aware of, when you press the 'New Project Application' button, your project will be assigned a unique project ID. + If you are not sure about some details, you can save them and edit them later. + + If you decide to create a project proposal for the same project again, you will receive a new unique project ID. + + ![New Application][8] + + !!! info New Application Type + Please pay attention to the title of the page and what type of request you are making + + #### Finalized Applications + + Applications that have already been sent can be viewed in this section. These applications can no longer be updated and are under review and will be activated in case of a positive result. + + ![Finalized Application][5] + +=== "Application" + + ### Application + + No matter which application you submit, the procedure is always identical. + + ??? abstract "Project Type" + + Choose the project type. Please note the information about the project types on the page. + + The project application differs depending on the type of application - and thus which components of our HPC system - resources are applied for. + + ![Project Type Selection][10] + + ??? abstract "Principal Investigator (PI) and Person to Contact (PC)" + + Choose the principal investigator and the person to contact. On the Button 'change my role ...' below, you can change your organisational part in the project. + + Keep in mind that the principal investigator (PI) has to be: + + * institute director + * chair holder + * research group leader + + ![Principal Investigator (PI)][11] + + If you have not yet submitted an application to us, you will be asked to provide your information. + + ??? abstract "Resources" + Choose the needed resources. Typically, you can choose between GPU and CPU. + + !!! info + You can only choose after you have specified a project type. + + For each choosen resource there are some questions after the project data. + + ![Resource Selection][12] + + Questions for CPU resources could be: + + - How much computing time is required for your project in the upcoming compute time period of one year (in core-hours)? + - What is your job size and how many cores do you need? + - Define your memory: How many GB per core do you need? + + !!! hint "Hint" + If you have chosen only CPU resources, please be aware to specify here only resources for pure CPU nodes. Resources needed on accelerators (GPUs) must be specified in the corresponding section. If you like to change your computing time requirements go one slide back. + + ![CPU][13] + + Questions for CPU resources could be: + + - How much computing time is required on GPU nodes for your HPC project (in GPU-hours)? + - How many GPU do you use for a single job? + - How much GB of memory do you use for your typical job? + - Which GPU programming scheme (CUDA, OpenACC, etc.) do you use? + - Which software do you deploy on GPUs (own code, third party)? + + !!! hint "Hint" + If you have chosen both CPU and GPU resources, but you do not need compute time on GPUs type in 0 here to be able to continue with your project application in Jards. + + ![GPU][14] + + ??? abstract "Project Data" + In this step, the basic conditions of the project must be specified. + + Depending on the selected project type, these can be different. + + For example: + + - How much time do you plan for your HPC project (add ending date)? + - What is your title of your HPC project (max. 250 char)? + - Define your keywords to explain your HPC project best. + - Please add a short project description (max. 6000 char). + - Explain your plans for this granting period (max. 500 char). + - Are you doing commissioned research (yes/no)? + - Who is sponsoring your HPC project? + - Classify your HPC project according to the following main and sub categories. + - Which methods (physical, mathematical, numerical, ...) are used? + - Which reasons make the use of this supercomputer necessary? + + !!! hint "Commissioned Research" + Are you doing commissioned research? + + Commissioned research means direct order by private third parties to the TU Dresden as well as indirect assignments by private third parties on a promoting institution (university , research institution, or similar institutions). All research projects sponsored by public institutions such as state, country DFG , EU etc. are excluded. + + This specification is very important. In the end you must confirm this information in writing. For this purpose, a PDF is generated, which you must send back to us. + + ![Project Data][15] + + ??? abstract "Software" + Next you need to specify which software you want to use. + + Here are the following categories, you should answer: + + - Commercial Software Packages + - Compilers + - Programming Languages + - Other packages, not named here. + - Use own-developed packages, please name here and provide the links to the open source software. + - Parallelization strategy + + !!!hint "Information About Required Software" + This is important for us, because we need to see if the software is installed, there is a valid license and it is available under the appropriate architecture. The biggest problem is the imagination of the license writers. We have to abide by these conditions. Even if we try to provide the common HPC software including licenses, not every software vendor offers appropriate licenses. + + ![Software][16] + + ??? abstract "Data Management" + The last thing to specify is the I/O behavior. All this information helps us to check and, if necessary, to create the necessary conditions. + + For example: + + * How much memory do you need. + * How do your calculations behave (many small files or one big file)? + * Fetch data from external? How? + + ![Data][17] + + ??? abstract "Upload Full Projectdescription" + On some project types a detailed project description is needed to complete the application. + + You can find a Template on this page: + [https://tu-dresden.de/zih/hochleistungsrechnen/zugang/projektantrag?set_language=en](https://tu-dresden.de/zih/hochleistungsrechnen/zugang/projektantrag?set_language=en). + +=== "Finalized Application" + + ### Finalized Application + + The final step is to complete the application process. The application is submitted to us and approved for review. Editing is no longer possible. + + In the next step you will receive an email with the data of the application as a PDF. In it you will find a template for the confirmation of the commissioned research. + + ![Finalize][18] + + !!! danger "Important" + Please send the confirmation of commissioned research signed back to us, even if it is not commissioned research. + +[1]: misc/application_1.png "Select Application" +[2]: misc/idm_login.png "Login via TU Dresden Account" +[3]: misc/email_login.png "Login via E-Mail Callback" +[4]: misc/email_link.png "Login via E-Mail Callback" +[5]: misc/finalized_application.png "Finalize" +[6]: misc/active_application.png +[7]: misc/overview_application.png +[8]: misc/new_application.png +[10]: misc/app_project_type.png "Project Type Selection" +[11]: misc/app_pi.png "PI Selection" +[12]: misc/app_resources.png "Resource Selection" +[13]: misc/app_resource_cpu.png "Resource CPU" +[14]: misc/app_resource_gpu.png "Resource GPU" +[15]: misc/app_project_data.png "Project Data" +[16]: misc/app_software.png "Software" +[17]: misc/app_data.png "Data" +[18]: misc/app_finalize.png "Finalize" + +[19]: misc/nhr_1a_applicationlist.png +[20]: misc/nhr_1b_applicationlist.png +[21]: misc/nhr_2_projecttype.png +[22]: misc/nhr_3_choose_pi_pc.png +[23]: misc/nhr_4_data_pi.png +[24]: misc/nhr_5_data_pc.png +[25]: misc/nhr_6_resources.png +[26]: misc/nhr_7a_CPUresources.png +[27]: misc/nhr_7b_GPUresources.png +[28]: misc/nhr_8a_project_data.png +[29]: misc/nhr_8b_project_data.png +[30]: misc/nhr_8c_project_data.png +[31]: misc/nhr_9a_software.png +[32]: misc/nhr_9b_software.png +[33]: misc/nhr_10a_data_mgmt.png +[34]: misc/nhr_10b_data_mgmt.png +[35]: misc/nhr_11_upload_regular.png +[36]: misc/nhr_11_upload_trial_project.png diff --git a/doc.zih.tu-dresden.de/docs/archive/overview.md b/doc.zih.tu-dresden.de/docs/archive/overview.md index 7600ef01e81d7f623f616d28d70abbf73cb07ed2..dfcb393a253c916a86ab21649aa75eb509ee2862 100644 --- a/doc.zih.tu-dresden.de/docs/archive/overview.md +++ b/doc.zih.tu-dresden.de/docs/archive/overview.md @@ -3,4 +3,4 @@ A warm welcome to the **archive**. You probably got here by following a link from within the compendium or by purpose. The archive holds outdated documentation for future reference. -Hence, documentation in the archive, is not further updated. +Hence, documentation in the archive is not further updated. diff --git a/doc.zih.tu-dresden.de/docs/archive/systems_switched_off.md b/doc.zih.tu-dresden.de/docs/archive/systems_switched_off.md index c4f9890ac3ad36580c617b6fb5292cb0b1ceffcb..0a9a50bb72c25b6920d41a66d2828ec57cd4c7b1 100644 --- a/doc.zih.tu-dresden.de/docs/archive/systems_switched_off.md +++ b/doc.zih.tu-dresden.de/docs/archive/systems_switched_off.md @@ -10,3 +10,29 @@ Documentation on former systems for future reference can be found on the followi - [Windows-HPC-Server Titan](system_titan.md) - [PC-Cluster Triton](system_triton.md) - [Shared-Memory-System Venus](system_venus.md) + +## Historical Overview + +| Year | System | +|------|--------| +| 1968 | Zeiss-Rechenautomat Nr. 1 (ZRA1) Performance: 150 to 200 instructions/s, 4096 storage cells with 48 bit each (magnetic drum) | +| 1970 | Commissioning of large stand-alone computers, coupled stand-alone computers and terminals (BESM, ESER) | +| 1976 | Computer network DELTA and graphic workstations | +| 1981 | Deployment of the first microcomputers; experimental testing of local area networks (LAN) | +| 1986 | Workstation computers replace mechanical devices; the first PC pools for teaching and studying are set up | +| 1991 | Short-term operation of used mainframe computers | +| 1993 | VP200-EX (857 MFlop/s Peak) | +| 1996 | Development of infrastructure on the TU campus | +| 1997 | SGI Origin2000 (21,8 GFlop/s, 56 CPUs, 17 GB RAM, 350 GB disk capacity)| +| 1998 | Cray T3E (38,4 GFlop/s, 64 CPUs, 8 GB RAM, 100 GB disk capacity)| +| 2001/02 | SGI Origin3800 (51,2 + 102,4 GFlop/s, 64 + 128 CPUs, 64 + 64 GB RAM, 790 GB disk capacity)| +| 2004 | Itanium-Cluster Castillo| +| 2005/06 | Hochleistungsrechner/Speicherkomplex: <br/> SGI Altix 4700: 13 TFlop/s, 6,5 TB RAM <br/> PC-Farm: 13 TFlop/s, 5,5 TB RAM <br/> SAN capacity: 136 TB <br/> Tape archive: 1 PB, 2500 tapes | +| 2007 | Setup PC-SAN <br/> NEC SX6: 72 GFlop/s | +| 2008 | Microsoft HPC-System | +| 2010 | IBM-Cluster iDataPlex | +| 2012 | GPU-Cluster <br/> HPC-Cluster Atlas: 50 TFlop/s Peak, 13 TB RAM | +| 2012/13| SGI UV 2000: 10,6 TFlop/s Peak, 8 TB RAM | +| 2013 | HPC-Cluster Taurus (HRSK-II): 135 TFlop/s Peak, 18 TB RAM | +| 2015 | HRSK-II Extension: 1,64 PFlop/s und 139 TB RAM | +| 2017/18| HPC-DA (HRSK-II Extension) | diff --git a/doc.zih.tu-dresden.de/docs/contrib/contribute_container.md b/doc.zih.tu-dresden.de/docs/contrib/contribute_container.md index 0855205da14267213b01552618cc0fbdb8734a5d..f0d5fe56c8835867bda545948ad7ae0f6bbd09c5 100644 --- a/doc.zih.tu-dresden.de/docs/contrib/contribute_container.md +++ b/doc.zih.tu-dresden.de/docs/contrib/contribute_container.md @@ -42,11 +42,37 @@ description of your changes. If you work on an issue, please also add "Closes 17 `git push origin 174-check-contribution-documentation`. 1. As an output you get a link to create a merge request against the preview branch. 1. When the merge request is created, a continuous integration (CI) pipeline automatically checks -your contributions. - -When you contribute, please follow our [content rules](content_rules.md) to make incorporating your -changes easy. We also check these rules via continuous integration checks and/or reviews. -You can find the details and commands to preview your changes and apply checks in the next sections. +your contributions. If you forked the repository, these automatic checks are not available, but you +can [run checks locally](#run-the-proposed-checks-inside-container). + +!!! tip + + When you contribute, please follow our [content rules](content_rules.md) to make incorporating + your changes easy. We also check these rules via continuous integration checks and/or reviews. + You can find the details and commands to [preview your changes](#start-the-local-web-server) and + [apply checks](#run-the-proposed-checks-inside-container). + +## Merging of Forked Repositories + +When you have forked the repository as mentioned above, the process for merging is a bit different +from internal merge requests. Because branches of forks are not automatically checked by CI, +someone with at least developer access needs to do some more steps to incorporate the changes of +your MR: + +1. The developer informs you about the start of merging process. +1. The developer needs to review your changes to make sure that your changes are specific and don't introduce +problems, such as changes in the Dockerfile or any script could. +1. The developer needs to create a branch in our repository. Let's call this "internal MR branch". +1. The developer needs to change the target branch of your MR from "preview" to "internal MR branch". +1. The developer needs to merge it. +1. The developer needs to open another MR from "internal MR branch" to "preview" to check whether + the changes pass the CI checks. +1. The developer needs to fix things that were found by CI. +1. The developer informs you about the MR or asks for your support while fixing the CI. + +When you follow our [content rules](content_rules.md) and +[run checks locally](#run-the-proposed-checks-inside-container), you are making this process +faster. ## Tools to Ensure Quality diff --git a/doc.zih.tu-dresden.de/docs/data_lifecycle/intermediate_archive.md b/doc.zih.tu-dresden.de/docs/data_lifecycle/intermediate_archive.md index 2ce7e0a16ee6edeaa4d966cb624932f97635d2db..894626208947186e48ba7d08b439cf6aace48655 100644 --- a/doc.zih.tu-dresden.de/docs/data_lifecycle/intermediate_archive.md +++ b/doc.zih.tu-dresden.de/docs/data_lifecycle/intermediate_archive.md @@ -27,6 +27,9 @@ For storing and restoring your data in/from the "Intermediate Archive" you can u ```console marie@login$ dtcp -r /<directory> /archiv/<project or user>/<directory> # or marie@login$ dtrsync -av /<directory> /archiv/<project or user>/<directory> +# example: +marie@login$ dtcp -r /scratch/marie/results /archiv/marie/ # or +marie@login$ dtrsync -av /scratch/marie/results /archiv/marie/results ``` ### Restore Data @@ -34,11 +37,16 @@ marie@login$ dtrsync -av /<directory> /archiv/<project or user>/<directory> ```console marie@login$ dtcp -r /archiv/<project or user>/<directory> /<directory> # or marie@login$ dtrsync -av /archiv/<project or user>/<directory> /<directory> +# example: +marie@login$ dtcp -r /archiv/marie/results /scratch/marie/ # or +marie@login$ dtrsync -av /archiv/marie/results /scratch/marie/results ``` -### Examples +!!! note "Listing files in archive" -```console -marie@login$ dtcp -r /scratch/rotscher/results /archiv/rotscher/ # or -marie@login$ dtrsync -av /scratch/rotscher/results /archiv/rotscher/results -``` + The intermediate archive is not mounted on the login nodes, but only on the [export nodes](../data_transfer/export_nodes.md). + + In order to list the user's files in the archive use the `dtls` command + ```console + marie@login$ dtls /archiv/$USER/ + ``` diff --git a/doc.zih.tu-dresden.de/docs/data_lifecycle/longterm_preservation.md b/doc.zih.tu-dresden.de/docs/data_lifecycle/longterm_preservation.md index de04504fdd68766f01c5b37887d7f4f03e45e4a3..3b1ad0c9c595fa4d09c0e113b65c82a71b274a35 100644 --- a/doc.zih.tu-dresden.de/docs/data_lifecycle/longterm_preservation.md +++ b/doc.zih.tu-dresden.de/docs/data_lifecycle/longterm_preservation.md @@ -91,7 +91,7 @@ require some data preparation (e.g. packaging), but serve different use cases: ### Storing very infrequently used data during the course of the project The intermediate archive is a tape storage easily accessible as a directory -(`/archive/<HRSK-project>/` or `/archive/<login>/`) using the +(`/archiv/<HRSK-project>/` or `/archiv/<login>/`) using the [export nodes](../data_transfer/export_nodes.md) and [Datamover tools](https://doc.zih.tu-dresden.de/data_transfer/datamover/) to move your data to. diff --git a/doc.zih.tu-dresden.de/docs/data_lifecycle/workspaces.md b/doc.zih.tu-dresden.de/docs/data_lifecycle/workspaces.md index 6180e5db831c8faf69a8752247ad5b8ee5ef6313..141ccef84a8ffe098e42ab9bea02afd560590901 100644 --- a/doc.zih.tu-dresden.de/docs/data_lifecycle/workspaces.md +++ b/doc.zih.tu-dresden.de/docs/data_lifecycle/workspaces.md @@ -99,13 +99,13 @@ days with an email reminder for 7 days before the expiration. The lifetime of a workspace is finite. Different filesystems (storage systems) have different maximum durations. A workspace can be extended multiple times, depending on the filesystem. -| Filesystem (use with parameter `-F`) | Duration, days | Extensions | Remarks | -|:------------------------------------:|:----------:|:-------:|:-----------------------------------:| -| `ssd` | 30 | 2 | High-IOPS filesystem (`/lustre/ssd`, symbolic link: `/ssd`) on SSDs. | -| `beegfs_global0` (deprecated) | 30 | 2 | High-IOPS filesystem (`/beegfs/global0`) on NVMes. | -| `beegfs` | 30 | 2 | High-IOPS filesystem (`/beegfs`) on NVMes. | -| `scratch` | 100 | 10 | Scratch filesystem (`/lustre/ssd`, symbolic link: `/scratch`) with high streaming bandwidth, based on spinning disks | -| `warm_archive` | 365 | 2 | Capacity filesystem based on spinning disks | +| Filesystem (use with parameter `-F`) | Duration, days | Extensions | [Filesystem Feature](../jobs_and_resources/slurm.md#filesystem-features) | Remarks | +|:-------------------------------------|---------------:|-----------:|:-------------------------------------------------------------------------|:--------| +| `ssd` | 30 | 2 | `fs_lustre_ssd` | High-IOPS filesystem (`/lustre/ssd`, symbolic link: `/ssd`) on SSDs. | +| `beegfs_global0` (deprecated) | 30 | 2 | `fs_beegfs_global0` | High-IOPS filesystem (`/beegfs/global0`) on NVMes. | +| `beegfs` | 30 | 2 | `fs_beegfs` | High-IOPS filesystem (`/beegfs`) on NVMes. | +| `scratch` | 100 | 10 | `fs_lustre_scratch2` | Scratch filesystem (`/lustre/ssd`, symbolic link: `/scratch`) with high streaming bandwidth, based on spinning disks | +| `warm_archive` | 365 | 2 | `fs_warm_archive_ws` | Capacity filesystem based on spinning disks | To extent your workspace use the following command: diff --git a/doc.zih.tu-dresden.de/docs/data_transfer/datamover.md b/doc.zih.tu-dresden.de/docs/data_transfer/datamover.md index 0891ca2a66f49b5e2f5c243fe4e86cdf07e1e2e9..28aba7bbfdcec8411f6510061d509c949d128f34 100644 --- a/doc.zih.tu-dresden.de/docs/data_transfer/datamover.md +++ b/doc.zih.tu-dresden.de/docs/data_transfer/datamover.md @@ -37,7 +37,7 @@ To identify the mount points of the different filesystems on the data transfer m | | `/warm_archive/ws` | `/warm_archive/ws` | | | `/home` | `/home` | | | `/projects` | `/projects` | -| **Archive** | | `/archive` | +| **Archive** | | `/archiv` | | **Group storage** | | `/grp/<group storage>` | ## Usage of Datamover diff --git a/doc.zih.tu-dresden.de/docs/jobs_and_resources/slurm.md b/doc.zih.tu-dresden.de/docs/jobs_and_resources/slurm.md index f7456662f5ca54887b20b075e58dd25517aa3c96..2ba282a77bff4ff03a0c5ba92e4ef1e19494c669 100644 --- a/doc.zih.tu-dresden.de/docs/jobs_and_resources/slurm.md +++ b/doc.zih.tu-dresden.de/docs/jobs_and_resources/slurm.md @@ -428,13 +428,13 @@ constraints, please refer to the [Slurm documentation](https://slurm.schedmd.com A feature `fs_*` is active if a certain filesystem is mounted and available on a node. Access to these filesystems are tested every few minutes on each node and the Slurm features set accordingly. -| Feature | Description | -|:-------------------|:---------------------------------------------------------------------| -| `fs_lustre_scratch2` | `/scratch` mounted read-write (mount point is `/lustre/scratch2`) | -| `fs_lustre_ssd` | `/ssd` mounted read-write (mount point is `/lustre/ssd`) | -| `fs_warm_archive_ws` | `/warm_archive/ws` mounted read-only | -| `fs_beegfs_global0` | `/beegfs/global0` mounted read-write | -| `fs_beegfs` | `/beegfs` mounted read-write | +| Feature | Description | [Workspace Name](../data_lifecycle/workspaces.md#extension-of-a-workspace) | +|:---------------------|:-------------------------------------------------------------------|:---------------------------------------------------------------------------| +| `fs_lustre_scratch2` | `/scratch` mounted read-write (mount point is `/lustre/scratch2`) | `scratch` | +| `fs_lustre_ssd` | `/ssd` mounted read-write (mount point is `/lustre/ssd`) | `ssd` | +| `fs_warm_archive_ws` | `/warm_archive/ws` mounted read-only | `warm_archive` | +| `fs_beegfs_global0` | `/beegfs/global0` mounted read-write | `beegfs_global0` | +| `fs_beegfs` | `/beegfs` mounted read-write | `beegfs` | For certain projects, specific filesystems are provided. For those, additional features are available, like `fs_beegfs_<projectname>`. diff --git a/doc.zih.tu-dresden.de/docs/software/containers.md b/doc.zih.tu-dresden.de/docs/software/containers.md index ca9e3a14d360521660604a372dba15241c1e4d78..cca62479279afef568fbd69c3e0c6ecc56120e1a 100644 --- a/doc.zih.tu-dresden.de/docs/software/containers.md +++ b/doc.zih.tu-dresden.de/docs/software/containers.md @@ -46,7 +46,7 @@ instructions from the official documentation to install Singularity. 1. Check if `go` is installed by executing `go version`. If it is **not**: ```console - marie@local$ wget <https://storage.googleapis.com/golang/getgo/installer_linux> && chmod +x + marie@local$ wget 'https://storage.googleapis.com/golang/getgo/installer_linux' && chmod +x installer_linux && ./installer_linux && source $HOME/.bash_profile ``` @@ -88,7 +88,9 @@ instructions from the official documentation to install Singularity. There are two possibilities: 1. Create a new container on your local workstation (where you have the necessary privileges), and - then copy the container file to ZIH systems for execution. + then copy the container file to ZIH systems for execution. Therefore you also have to install + [Singularity](https://sylabs.io/guides/3.0/user-guide/quick_start.html#quick-installation-steps) + on your local workstation. 1. You can, however, import an existing container from, e.g., Docker. Both methods are outlined in the following. @@ -103,10 +105,11 @@ You can create a new custom container on your workstation, if you have root righ which is different to the x86 architecture in common computers/laptops. For that you can use the [VM Tools](singularity_power9.md). -Creating a container is done by writing a **definition file** and passing it to +Creating a container is done by writing a definition file, such as `myDefinition.def`, and passing +it to `singularity` via ```console -marie@local$ singularity build myContainer.sif <myDefinition.def> +marie@local$ singularity build myContainer.sif myDefinition.def ``` A definition file contains a bootstrap @@ -167,7 +170,7 @@ https://github.com/singularityware/singularity/tree/master/examples. You can import an image directly from the Docker repository (Docker Hub): ```console -marie@local$ singularity build my-container.sif docker://ubuntu:latest +marie@login$ singularity build my-container.sif docker://ubuntu:latest ``` Creating a singularity container directly from a local docker image is possible but not @@ -284,7 +287,7 @@ While the `shell` command can be useful for tests and setup, you can also launch inside the container directly using "exec": ```console -marie@login$ singularity exec my-container.img /opt/myapplication/bin/run_myapp +marie@login$ singularity exec my-container.sif /opt/myapplication/bin/run_myapp ``` This can be useful if you wish to create a wrapper script that transparently calls a containerized @@ -299,7 +302,7 @@ if [ "z$X" = "z" ] ; then exit 1 fi -singularity exec /scratch/p_myproject/my-container.sif /opt/myapplication/run_myapp "$@" +singularity exec /projects/p_number_crunch/my-container.sif /opt/myapplication/run_myapp "$@" ``` The better approach is to use `singularity run`, which executes whatever was set in the `%runscript` diff --git a/doc.zih.tu-dresden.de/docs/software/fem_software.md b/doc.zih.tu-dresden.de/docs/software/fem_software.md index 8b8eb4cfe10c4476e48c4b30ac7f16b83589a38d..af6b9fb80986e2bc727ae88e97b2cca614ffd629 100644 --- a/doc.zih.tu-dresden.de/docs/software/fem_software.md +++ b/doc.zih.tu-dresden.de/docs/software/fem_software.md @@ -8,7 +8,9 @@ ```console marie@login$ module avail ANSYS [...] - marie@login$ module load ANSYS/<version> + marie@login$ # module load ANSYS/<version> + marie@login$ # e.g. + marie@login$ module load ANSYS/2020R2 ``` The section [runtime environment](modules.md) provides a comprehensive overview @@ -105,7 +107,9 @@ all data via `-C`. ```console # SSH connection established using -CX -marie@login$ module load ANSYS/<version> +marie@login$ # module load ANSYS/<version> +marie@login$ # e.g. +marie@login$ module load ANSYS/2020R2 marie@login$ runwb2 ``` @@ -113,7 +117,9 @@ If more time is needed, a CPU has to be allocated like this (see [batch systems Slurm](../jobs_and_resources/slurm.md) for further information): ```console -marie@login$ module load ANSYS/<version> +marie@login$ # module load ANSYS/<version> +marie@login$ # e.g. +marie@login$ module load ANSYS/2020R2 marie@login$ srun --time=00:30:00 --x11=first [SLURM_OPTIONS] --pty bash [...] marie@login$ runwb2 @@ -153,7 +159,9 @@ parameter (for batch mode), `-F` for your project file, and can then either add unset SLURM_GTIDS # Odd, but necessary! - module load ANSYS/<version> + # module load ANSYS/<version> + # e.g. + module load ANSYS ANSYS/2020R2 runwb2 -B -F Workbench_Taurus.wbpj -E 'Project.Update' -E 'Save(Overwrite=True)' #or, if you wish to use a workbench replay file, replace the -E parameters with: -R mysteps.wbjn @@ -208,7 +216,7 @@ firewall of ZIH. For further information, please refer to the COMSOL manual. ```console marie@login$ module load COMSOL - marie@login$ srun --ntasks=1 --cpus-per-task=4 --mem-per-cpu=2500 --time=8:00 comsol -np 4 server + marie@login$ srun --ntasks=1 --cpus-per-task=4 --mem-per-cpu=2500 --time=08:00:00 comsol -np 4 server ``` ??? example "Interactive Job" @@ -218,7 +226,7 @@ firewall of ZIH. For further information, please refer to the COMSOL manual. ```console marie@login$ module load COMSOL - marie@login$ srun --ntasks=1 --cpus-per-task=4 --mem-per-cpu=2500 --time=8:00 --pty --x11=first comsol -np 4 + marie@login$ srun --ntasks=1 --cpus-per-task=4 --mem-per-cpu=2500 --time=08:00:00 --pty --x11=first comsol -np 4 ``` Please make sure, that the option *Preferences* --> Graphics --> *Renedering* is set to *software @@ -264,10 +272,10 @@ You need a job file (aka. batch script) to run the MPI version. srun mpp-dyna i=neon_refined01_30ms.k memory=120000000 ``` - Submit the job file to the batch system via + Submit the job file named `job.sh` to the batch system via ```console - marie@login$ sbatch <filename> + marie@login$ sbatch job.sh ``` Please refer to the section [Slurm](../jobs_and_resources/slurm.md) for further details and diff --git a/doc.zih.tu-dresden.de/docs/software/mathematics.md b/doc.zih.tu-dresden.de/docs/software/mathematics.md index 909970debf879b3b3c227fd80dbd26e06463309e..66fdc7050b050b77b899c83133d73758ac2dced6 100644 --- a/doc.zih.tu-dresden.de/docs/software/mathematics.md +++ b/doc.zih.tu-dresden.de/docs/software/mathematics.md @@ -105,7 +105,7 @@ marie@compute$ matlab ``` With following command you can see a list of installed software - also -the different versions of matlab. +the different versions of MATLAB. ```console marie@login$ module avail @@ -123,7 +123,7 @@ Or use: marie@login$ module load MATLAB ``` -(then you will get the most recent Matlab version. +(then you will get the most recent MATLAB version. [Refer to the modules section for details.](../software/modules.md#modules)) ### Interactive @@ -135,7 +135,7 @@ with command marie@login$ srun --pty --x11=first bash ``` -- now you can call "matlab" (you have 8h time to work with the matlab-GUI) +- now you can call "matlab" (you have 8h time to work with the MATLAB-GUI) ### Non-interactive @@ -218,9 +218,347 @@ marie@login$ srun ./run_compiled_executable.sh $EBROOTMATLAB Please refer to the documentation `help parfor` for further information. -## Octave +### MATLAB Parallel Computing Toolbox -GNU [Octave](https://www.gnu.org/software/octave/index) is a high-level language, primarily intended -for numerical computations. It provides a convenient command line interface for solving linear and -nonlinear problems numerically, and for performing other numerical experiments using a language that -is mostly compatible with Matlab. It may also be used as a batch-oriented language. +In the following, the steps to configure MATLAB to submit jobs to a cluster, retrieve results, and +debug errors are outlined. + +#### Configuration – MATLAB client on the cluster + +After logging into the HPC system, you configure MATLAB to run parallel jobs on the HPC system by +calling the shell script `configCluster.sh`. This only needs to be called once per version of +MATLAB. + +```console +marie@login$ module load MATLAB +marie@login$ configCluster.sh +``` + +Jobs will now default to the HPC system rather than submit to the local machine. + +#### Installation and Configuration – MATLAB client off the cluster + +The MATLAB support package for ZIH Systems can be found as follows: + +* Windows: + * [tud.nonshared.R2021b.zip](misc/tud.nonshared.R2021b.zip) + * [tud.nonshared.R2022a.zip](misc/tud.nonshared.R2022a.zip) +* Linux/macOS: + * [tud.nonshared.R2021b.tar.gz](misc/tud.nonshared.R2021b.tar.gz) + * [tud.nonshared.R2022a.tar.gz](misc/tud.nonshared.R2022a.tar.gz) + +Download the appropriate archive file and start MATLAB. The archive file should be extracted +in the location returned by calling + +```matlabsession +>> userpath +``` + +Configure MATLAB to run parallel jobs on ZIH Systems by calling `configCluster`. `configCluster` +only needs to be called once per version of MATLAB. + +```matlabsession +>> configCluster +``` + +Submission to the remote cluster requires SSH credentials. You will be prompted for your SSH +username and password or identity file (private key). The username and location of the private key +will be stored in MATLAB for future sessions. Jobs will now default to the cluster rather than +submit to the local machine. + +!!! note + + If you would like to submit to the local machine then run the following command: + + ```matlab + >> % Get a handle to the local resources + >> c = parcluster('local'); + ``` + +#### Configuring Jobs + +Prior to submitting the job, you can specify various parameters to pass to your jobs, such as queue, +e-mail, walltime, etc. *Only `MemPerCpu` and `QueueName` are required*. + +```matlabsession +>> % Get a handle to the cluster +>> c = parcluster; + +[REQUIRED] + +>> % Specify memory to use, per core (default: 2gb) +>> c.AdditionalProperties.MemPerCpu = '4gb'; + +>> % Specify the walltime (e.g., 5 hours) +>> c.AdditionalProperties.WallTime = '05:00:00'; + +[OPTIONAL] + +>> % Specify the account to use +>> c.AdditionalProperties.Account = 'account-name'; + +>> % Request constraint +>> c.AdditionalProperties.Constraint = 'a-constraint'; + +>> % Request job to run on exclusive node(s) (default: false) +>> c.AdditionalProperties.EnableExclusive = true; + +>> % Request email notification of job status +>> c.AdditionalProperties.EmailAddress = 'user-id@tu-dresden.de'; + +>> % Specify number of GPUs to use (GpuType is optional) +>> c.AdditionalProperties.GpusPerNode = 1; +>> c.AdditionalProperties.GpuType = 'gpu-card'; + +>> % Specify the queue to use +>> c.AdditionalProperties.Partition = 'queue-name'; + +>> % Specify a reservation to use +>> c.AdditionalProperties.Reservation = 'a-reservation'; +``` + +Save changes after modifying `AdditionalProperties` for the above changes to persist between MATLAB +sessions. + +```matlabsession +>> c.saveProfile +``` + +To see the values of the current configuration options, display `AdditionalProperties`. + +```matlabsession +>> % To view current properties +>> c.AdditionalProperties +``` + +You can unset a value when no longer needed. + +```matlabsession +>> % Turn off email notifications +>> c.AdditionalProperties.EmailAddress = ''; +>> c.saveProfile +``` + +#### Interactive Jobs - MATLAB Client on the Cluster + +To run an interactive pool job on the ZIH systems, continue to use `parpool` as you’ve done before. + +```matlabsession +>> % Get a handle to the cluster +>> c = parcluster; + +>> % Open a pool of 64 workers on the cluster +>> pool = c.parpool(64); +``` + +Rather than running local on your machine, the pool can now run across multiple nodes on the +cluster. + +```matlabsession +>> % Run a parfor over 1000 iterations +>> parfor idx = 1:1000 + a(idx) = … + end +``` + +Once you are done with the pool, delete it. + +```matlabsession +>> % Delete the pool +>> pool.delete +``` + +#### Independent Batch Job + +Use the batch command to submit asynchronous jobs to the HPC system. The `batch` command will return +a job object which is used to access the output of the submitted job. See the MATLAB documentation +for more help on `batch`. + +```matlabsession +>> % Get a handle to the cluster +>> c = parcluster; + +>> % Submit job to query where MATLAB is running on the cluster +>> job = c.batch(@pwd, 1, {}, ... + 'CurrentFolder','.', 'AutoAddClientPath',false); + +>> % Query job for state +>> job.State + +>> % If state is finished, fetch the results +>> job.fetchOutputs{:} + +>> % Delete the job after results are no longer needed +>> job.delete +``` + +To retrieve a list of currently running or completed jobs, call `parcluster` to retrieve the cluster +object. The cluster object stores an array of jobs that were run, are running, or are queued to +run. This allows us to fetch the results of completed jobs. Retrieve and view the list of jobs as +shown below. + +```matlabsession +>> c = parcluster; +>> jobs = c.Jobs; +``` + +Once you have identified the job you want, you can retrieve the results as done previously. + +`fetchOutputs` is used to retrieve function output arguments; if calling `batch` with a script, use +`load` instead. Data that has been written to files on the cluster needs be retrieved directly +from the filesystem (e.g. via ftp). To view results of a previously completed job: + +```matlabsession +>> % Get a handle to the job with ID 2 +>> job2 = c.Jobs(2); +``` + +!!! note + + You can view a list of your jobs, as well as their IDs, using the above `c.Jobs` command. + + ```matlabsession + >> % Fetch results for job with ID 2 + >> job2.fetchOutputs{:} + ``` + +#### Parallel Batch Job + +You can also submit parallel workflows with the `batch` command. Let’s use the following example +for a parallel job, which is saved as `parallel_example.m`. + +```matlab +function [t, A] = parallel_example(iter) + +if nargin==0 + iter = 8; +end + +disp('Start sim') + +t0 = tic; +parfor idx = 1:iter + A(idx) = idx; + pause(2) + idx +end +t = toc(t0); + +disp('Sim completed') + +save RESULTS A + +end +``` + +This time when you use the `batch` command, to run a parallel job, you will also specify a MATLAB +Pool. + +```matlabsession +>> % Get a handle to the cluster +>> c = parcluster; + +>> % Submit a batch pool job using 4 workers for 16 simulations +>> job = c.batch(@parallel_example, 1, {16}, 'Pool',4, ... + 'CurrentFolder','.', 'AutoAddClientPath',false); + +>> % View current job status +>> job.State + +>> % Fetch the results after a finished state is retrieved +>> job.fetchOutputs{:} +ans = + 8.8872 +``` + +The job ran in 8.89 seconds using four workers. Note that these jobs will always request N+1 CPU +cores, since one worker is required to manage the batch job and pool of workers. For example, a +job that needs eight workers will consume nine CPU cores. + +You might run the same simulation but increase the Pool size. This time, to retrieve the results later, +you will keep track of the job ID. + +!!! note + + For some applications, there will be a diminishing return when allocating too many workers, as + the overhead may exceed computation time. + + ```matlabsession + >> % Get a handle to the cluster + >> c = parcluster; + + >> % Submit a batch pool job using 8 workers for 16 simulations + >> job = c.batch(@parallel_example, 1, {16}, 'Pool', 8, ... + 'CurrentFolder','.', 'AutoAddClientPath',false); + + >> % Get the job ID + >> id = job.ID + id = + 4 + >> % Clear job from workspace (as though you quit MATLAB) + >> clear job + ``` + +Once you have a handle to the cluster, you can call the `findJob` method to search for the job with +the specified job ID. + +```matlabsession +>> % Get a handle to the cluster +>> c = parcluster; + +>> % Find the old job +>> job = c.findJob('ID', 4); + +>> % Retrieve the state of the job +>> job.State +ans = + finished +>> % Fetch the results +>> job.fetchOutputs{:}; +ans = + 4.7270 +``` + +The job now runs in 4.73 seconds using eight workers. Run code with different number of workers to +determine the ideal number to use. Alternatively, to retrieve job results via a graphical user +interface, use the Job Monitor (Parallel > Monitor Jobs). + + +{: summary="Retrieve job results via GUI using the Job Monitor." align="center"} + +#### Debugging + +If a serial job produces an error, call the `getDebugLog` method to view the error log file. When +submitting independent jobs, with multiple tasks, specify the task number. + +```matlabsession +>> c.getDebugLog(job.Tasks(3)) +``` + +For Pool jobs, only specify the job object. + +```matlabsession +>> c.getDebugLog(job) +``` + +When troubleshooting a job, the cluster admin may request the scheduler ID of the job. This can be +derived by calling `schedID`. + +```matlabsession +>> schedID(job) +ans = + 25539 +``` + +#### Further Reading + +To learn more about the MATLAB Parallel Computing Toolbox, check out these resources: + +* [Parallel Computing Coding + Examples](https://www.mathworks.com/help/parallel-computing/examples.html) +* [Parallel Computing Documentation](http://www.mathworks.com/help/distcomp/index.html) +* [Parallel Computing Overview](http://www.mathworks.com/products/parallel-computing/index.html) +* [Parallel Computing + Tutorials](http://www.mathworks.com/products/parallel-computing/tutorials.html) +* [Parallel Computing Videos](http://www.mathworks.com/products/parallel-computing/videos.html) +* [Parallel Computing Webinars](http://www.mathworks.com/products/parallel-computing/webinars.html) diff --git a/doc.zih.tu-dresden.de/docs/software/misc/matlab_monitor_jobs.png b/doc.zih.tu-dresden.de/docs/software/misc/matlab_monitor_jobs.png new file mode 100644 index 0000000000000000000000000000000000000000..c91906e819495e345da69f80192ea3b8fee0a248 Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/software/misc/matlab_monitor_jobs.png differ diff --git a/doc.zih.tu-dresden.de/docs/software/misc/tud.nonshared.R2021b.tar.gz b/doc.zih.tu-dresden.de/docs/software/misc/tud.nonshared.R2021b.tar.gz new file mode 100644 index 0000000000000000000000000000000000000000..a4a30943f36ee4ebb5ad94c635be49a016f1eadd Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/software/misc/tud.nonshared.R2021b.tar.gz differ diff --git a/doc.zih.tu-dresden.de/docs/software/misc/tud.nonshared.R2021b.zip b/doc.zih.tu-dresden.de/docs/software/misc/tud.nonshared.R2021b.zip new file mode 100644 index 0000000000000000000000000000000000000000..02118ef5354a1f972321bde558a3e2bb08a5b6af Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/software/misc/tud.nonshared.R2021b.zip differ diff --git a/doc.zih.tu-dresden.de/docs/software/misc/tud.nonshared.R2022a.tar.gz b/doc.zih.tu-dresden.de/docs/software/misc/tud.nonshared.R2022a.tar.gz new file mode 100644 index 0000000000000000000000000000000000000000..a4160a5f33f094f340eb20c3e140687170864609 Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/software/misc/tud.nonshared.R2022a.tar.gz differ diff --git a/doc.zih.tu-dresden.de/docs/software/misc/tud.nonshared.R2022a.zip b/doc.zih.tu-dresden.de/docs/software/misc/tud.nonshared.R2022a.zip new file mode 100644 index 0000000000000000000000000000000000000000..481ab9a1d1a18515abcda82fbdb7d1ab3b580a5e Binary files /dev/null and b/doc.zih.tu-dresden.de/docs/software/misc/tud.nonshared.R2022a.zip differ diff --git a/doc.zih.tu-dresden.de/docs/software/python_virtual_environments.md b/doc.zih.tu-dresden.de/docs/software/python_virtual_environments.md index 13b623174f21016084917fb2cd424b500727e5f3..a999c5596693a25c90bfd74c108551001c2294e8 100644 --- a/doc.zih.tu-dresden.de/docs/software/python_virtual_environments.md +++ b/doc.zih.tu-dresden.de/docs/software/python_virtual_environments.md @@ -66,6 +66,37 @@ the environment as follows: (env) marie@compute$ deactivate #Leave the virtual environment ``` +??? example + + This is an example on partition Alpha. The example creates a conda virtual environment, and + installs the package `torchvision` with conda. + ```console + marie@login$ srun --partition=alpha-interactive --nodes=1 --gres=gpu:1 --time=01:00:00 --pty bash + marie@alpha$ ws_allocate -F scratch my_python_virtualenv 100 # use a workspace for the environment + marie@alpha$ cd /scratch/ws/1/marie-my_python_virtualenv + marie@alpha$ module load modenv/hiera GCC/10.2.0 CUDA/11.1.1 OpenMPI/4.0.5 PyTorch/1.9.0 + Module GCC/10.2.0, CUDA/11.1.1, OpenMPI/4.0.5, PyTorch/1.9.0 and 54 dependencies loaded. + marie@alpha$ which python + /sw/installed/Python/3.8.6-GCCcore-10.2.0/bin/python + marie@alpha$ pip list + [...] + marie@alpha$ virtualenv --system-site-packages my-torch-env + created virtual environment CPython3.8.6.final.0-64 in 42960ms + creator CPython3Posix(dest=[...]/my-torch-env, clear=False, global=True) + seeder FromAppData(download=False, pip=bundle, setuptools=bundle, wheel=bundle, via=copy, app_data_dir=~/.local/share/virtualenv) + added seed packages: pip==21.1.3, setuptools==57.2.0, wheel==0.36.2 + activators BashActivator,CShellActivator,FishActivator,PowerShellActivator,PythonActivator,XonshActivator + marie@alpha$ source my-torch-env/bin/activate + (my-torch-env) marie@alpha$ pip install torchvision==0.10.0 + [...] + Installing collected packages: torchvision==0.10.0 + Successfully installed torchvision-0.10.0 + [...] + (my-torch-env) marie@alpha$ python -c "import torchvision; print(torchvision.__version__)" + 0.10.0+cu102 + (my-torch-env) marie@alpha$ deactivate + ``` + ### Persistence of Python Virtual Environment To persist a virtualenv, you can store the names and versions of installed @@ -134,34 +165,41 @@ can deactivate the conda environment as follows: (conda-env) marie@compute$ conda deactivate #Leave the virtual environment ``` +!!! warning + When installing conda packages via `conda install`, ensure to have enough main memory requested + in your job allocation. + +!!! hint + We do not recommend to use conda environments together with EasyBuild modules due to + dependency conflicts. Nevertheless, if you need EasyBuild modules, consider installing conda + packages via `conda install --no-deps [...]` to prevent conda from installing dependencies. + ??? example - This is an example on partition Alpha. The example creates a virtual environment, and installs - the package `torchvision` with pip. + This is an example on partition Alpha. The example creates a conda virtual environment, and + installs the package `torchvision` with conda. ```console marie@login$ srun --partition=alpha-interactive --nodes=1 --gres=gpu:1 --time=01:00:00 --pty bash - marie@alpha$ mkdir python-environments # please use workspaces - marie@alpha$ module load modenv/hiera GCC/10.2.0 CUDA/11.1.1 OpenMPI/4.0.5 PyTorch - Module GCC/10.2.0, CUDA/11.1.1, OpenMPI/4.0.5, PyTorch/1.9.0 and 54 dependencies loaded. - marie@alpha$ which python - /sw/installed/Python/3.8.6-GCCcore-10.2.0/bin/python - marie@alpha$ pip list + marie@alpha$ ws_allocate -F scratch my_conda_virtualenv 100 # use a workspace for the environment + marie@alpha$ cd /scratch/ws/1/marie-my_conda_virtualenv + marie@alpha$ module load Anaconda3 + Module Anaconda3/2021.11 loaded. + marie@alpha$ conda create --prefix my-torch-env python=3.8 + Collecting package metadata (current_repodata.json): done + Solving environment: done [...] - marie@alpha$ virtualenv --system-site-packages python-environments/my-torch-env - created virtual environment CPython3.8.6.final.0-64 in 42960ms - creator CPython3Posix(dest=~/python-environments/my-torch-env, clear=False, global=True) - seeder FromAppData(download=False, pip=bundle, setuptools=bundle, wheel=bundle, via=copy, app_data_dir=~/.local/share/virtualenv) - added seed packages: pip==21.1.3, setuptools==57.2.0, wheel==0.36.2 - activators BashActivator,CShellActivator,FishActivator,PowerShellActivator,PythonActivator,XonshActivator - marie@alpha$ source python-environments/my-torch-env/bin/activate - (my-torch-env) marie@alpha$ pip install torchvision + Proceed ([y]/n)? y [...] - Installing collected packages: torchvision - Successfully installed torchvision-0.10.0 + marie@alpha$ conda activate my-torch-env + (my-torch-env) marie@alpha$ conda install -c pytorch torchvision + Collecting package metadata (current_repodata.json): done [...] + Preparing transaction: done + Verifying transaction: done + (my-torch-env) marie@alpha$ which python # ensure to use the correct Python (my-torch-env) marie@alpha$ python -c "import torchvision; print(torchvision.__version__)" - 0.10.0+cu102 - (my-torch-env) marie@alpha$ deactivate + 0.12.0 + (my-torch-env) marie@alpha$ conda deactivate ``` ### Persistence of Conda Virtual Environment diff --git a/doc.zih.tu-dresden.de/mkdocs.yml b/doc.zih.tu-dresden.de/mkdocs.yml index 7ff0d21b471c8fb5f9397fd980a56b1fa4a22b89..64f60c813cd992499f6519128383a584d025ba6e 100644 --- a/doc.zih.tu-dresden.de/mkdocs.yml +++ b/doc.zih.tu-dresden.de/mkdocs.yml @@ -102,7 +102,7 @@ nav: - Binding And Distribution Of Tasks: jobs_and_resources/binding_and_distribution_of_tasks.md - Support: - How to Ask for Support: support/support.md - - Archive of the Old Wiki: + - Archive: - Overview: archive/overview.md - Bio Informatics: archive/bioinformatics.md - CXFS End of Support: archive/cxfs_end_of_support.md @@ -239,4 +239,4 @@ extra: - link: /data_protection_declaration name: "Data Protection Declaration / Datenschutzerklärung" - link: https://tu-dresden.de/zertifikate - name: "Certificates" + name: "Certificates" \ No newline at end of file diff --git a/doc.zih.tu-dresden.de/tud_theme/stylesheets/extra.css b/doc.zih.tu-dresden.de/tud_theme/stylesheets/extra.css index 98c54afa6b64cd8b38b576f52ff6f8117a75cf16..6b855870a148cc8a08d731561c5ee82d1d527695 100644 --- a/doc.zih.tu-dresden.de/tud_theme/stylesheets/extra.css +++ b/doc.zih.tu-dresden.de/tud_theme/stylesheets/extra.css @@ -36,6 +36,8 @@ /* interaction color */ --tud-red-90: rgb(221, 29, 29); --fg-color--light: rgba(0, 0, 0, 0.6); + + --icon-external-link: url('data:image/svg+xml, <svg xmlns="http://www.w3.org/2000/svg" viewBox="0 0 20 20"> <g style="stroke:rgb(35,82,124);stroke-width:1"> <line x1="5" y1="5" x2="5" y2="14" /> <line x1="14" y1="9" x2="14" y2="14" /> <line x1="5" y1="14" x2="14" y2="14" /> <line x1="5" y1="5" x2="9" y2="5" /> <line x1="10" y1="2" x2="17" y2="2" /> <line x1="17" y1="2" x2="17" y2="9" /> <line x1="10" y1="9" x2="17" y2="2" style="stroke-width:1.5" /> </g> </svg>'); } .md-typeset h1, @@ -192,7 +194,26 @@ p { margin: 0.2em; } /* main */ - +/* style external links as found on https://stackoverflow.com/questions/5379752/css-style-external-links */ +.md-content a[href]:where([href*="\."]):not(:where( + /* exclude hash only links */ + [href^="#"], + /* exclude relative but not double slash only links */ + [href^="/"]:not([href^="//"]), + /* exclude page itself */ + [href*="//doc.zih.tu-dresden.de"], + /* exclude relative links beginning with ../ */ + [href^="\.\./"], + [href^="misc/"], + /* exclude buttons */ + .md-content__button, + /* exclude icons */ + .md-icon +)):after { + content: ''; + background: no-repeat var(--icon-external-link); + padding-right: 1em; +} /* footer */ .md-footer * { justify-content: flex-start; diff --git a/doc.zih.tu-dresden.de/wordlist.aspell b/doc.zih.tu-dresden.de/wordlist.aspell index a808318d64a38981956ed1ac5fa5a7d1c05e703d..1dc7ca99f0490683a7eb630bf74089527cb7a818 100644 --- a/doc.zih.tu-dresden.de/wordlist.aspell +++ b/doc.zih.tu-dresden.de/wordlist.aspell @@ -6,8 +6,8 @@ ALLREDUCE Altix Amber Amdahl's -analytics Analytics +analytics anonymized Ansys APIs @@ -15,6 +15,7 @@ AVX awk BeeGFS benchmarking +BESM BLAS BMC broadwell @@ -33,23 +34,23 @@ CLI CMake COMSOL conda -config CONFIG -cpu +config CPU +cpu CPUID -cpus CPUs +cpus crossentropy css CSV CUDA cuDNN CXFS -dask Dask -dataframes +dask DataFrames +dataframes Dataheap Datamover DataParallel @@ -83,6 +84,7 @@ engl english env EPYC +ESER Espresso ESSL facto @@ -91,18 +93,20 @@ FFT FFTW filesystem filesystems -flink Flink +flink FlinkExample FMA foreach Fortran +Frontend Galilei Gauss Gaussian GBit GDB GDDR +GFlop GFLOPS gfortran GiB @@ -113,8 +117,8 @@ GitLab's glibc Gloo gnuplot -gpu GPU +gpu GPUs gres GROMACS @@ -128,16 +132,19 @@ HDF HDFS HDFView hiera -horovod +Hochleistungsrechnen +Hochleistungsrechner Horovod +horovod horovodrun hostname Hostnames -hpc HPC +hpc hpcsupport HPE HPL +HRSK html hvd hyperparameter @@ -145,6 +152,7 @@ hyperparameters hyperthreading icc icpc +iDataPlex ifort ImageNet img @@ -159,11 +167,12 @@ ipython IPython ISA Itanium +Jards jobqueue jpg jss -jupyter Jupyter +jupyter JupyterHub JupyterLab Jupytext @@ -176,22 +185,24 @@ LAPACK lapply Leichtbau LINPACK -linter Linter +linter lmod LoadLeveler localhost lsf lustre +macOS markdownlint Mathematica MathKernel MathWorks -matlab +MATLAB Matplotlib MEGWARE mem Memcheck +MFlop MiB Microarchitecture MIMD @@ -207,6 +218,7 @@ Montecito mortem Mortem mountpoint +Mpi mpi mpicc mpiCC @@ -215,20 +227,23 @@ mpif mpifort mpirun multicore -multiphysics Multiphysics +multiphysics multithreaded Multithreading NAMD +Nationales natively nbgitpuller nbsp NCCL +NEC Neptun NFS NGC -nodelist +NHR NODELIST +nodelist NRINGS Nsight ntasks @@ -254,25 +269,26 @@ OpenBLAS OpenCL OpenGL OpenMP -openmpi OpenMPI +openmpi OpenSSH Opteron OST OTF overfitting -pandarallel Pandarallel +pandarallel PAPI parallelization parallelize parallelized parfor pdf -perf Perf +perf performant PESSL +PFlop PGI PiB Pika @@ -281,8 +297,8 @@ PMI png PowerAI ppc -pre Pre +pre preload Preload preloaded @@ -299,19 +315,20 @@ pty PuTTY pymdownx PythonAnaconda -pytorch PyTorch +pytorch Quantum queue -quickstart Quickstart +quickstart randint reachability README +Rechenautomat reproducibility requeueing -resnet ResNet +resnet RHEL Rmpi rome @@ -322,8 +339,8 @@ RStudio rsync Rsync runnable -runtime Runtime +runtime sacct salloc Sandybridge @@ -356,6 +373,7 @@ SMP SMT SparkExample spawner +Speicherkomplex spython squeue srun @@ -375,22 +393,25 @@ SXM TBB TCP TensorBoard -tensorflow TensorFlow +tensorflow +TFlop TFLOPS Theano tmp -todo ToDo +todo toolchain toolchains -torchvision Torchvision +torchvision tracefile tracefiles tracepoints transferability Trition +TUD +und undistinguishable unencrypted uplink @@ -411,6 +432,7 @@ VMSize VNC VPN VRs +walltime WebVNC WinSCP WML @@ -424,7 +446,8 @@ XLC XLF Xming yaml -zih +Zeiss ZIH ZIH's +ZRA ZSH