#acl EditorGroup:read,write,revert Known:read All:read #pragma section-numbers 2 = IFCA Datacenter usage guidelines = <> {{{#!wiki note If you find any information that is out-dated, incorrect or incomplete, do not hesitate to [[#Support|Open a ticket]]. }}} == Introduction == The '''GridUI''' (Grid User Interface) cluster is the interactive gateway to the [[http://grid.ifca.es|Advanced Computing and e-Science]] resources at IFCA. This cluster is comprised of a pool of machines reachable through a single entry point. The connections to the internal machines are managed by a director node that tries to ensure that proper balancing is made across the available nodes at a given moment. Please note that this cluster is ''not intended for the execution of CPU intensive tasks'', for this purpose use any of the available computing resources. Every process spawned is limited to a maximum CPU time of 2 hours. Login on these machines is provided via [[http://en.wikipedia.org/wiki/Secure_Shell|Secure Shell]]. Outgoing SSH connections are not allowed by default from this cluster. Inactive SSH sessions may be closed after 12h. It is highly recommended that you set up [[Cluster/Usage/SSHKeyManagement| SSH Keys]] for authentication, instead of using your username and password. || '''Hostname''' || '''Operating System''' || '''SSH server key fingerprint''' || || `gridui.ifca.es`, `griduisl6.ifca.es` || Scientific Linux 6.X || `29:80:9b:28:e7:8a:00:fe:6c:60:ef:e6:a6:71:33:bd` || == Authentication and user accounts == See [[Cluster/SSO]]. == Access to Scientific Linux 5 machines == After the [[https://grid.ifca.es/sl5-user-interfaces-deprecation-plan2.html|Scientific Linux 5 deprecation]] interactive access to Scientific Linux 5 is still possible trough the batch system. In order to request a SLC5 machine you must append the complex `scientificlinux5` to your request: {{{#!highlight console numbers=disable user@cloudprv-10-0:~ $ qsub -l scientificlinux5=true (...) }}} If you want an interactive session, append the complex to your `qlogin` request: {{{#!highlight console numbers=disable user@cloudprv-10-0:~ $ qlogin -l scientificlinux5=true (...) JSV "/nfs4/opt/gridengine/util/resources/jsv/jsv-IFCA.tcl" has been started JSV "/nfs4/opt/gridengine/util/resources/jsv/jsv-IFCA.tcl" has been stopped Your job 1822278 ("QLOGIN") has been submitted waiting for interactive job to be scheduled ... Your interactive job 1822278 has been successfully scheduled. Establishing builtin session to host cloudprv-02-9.ifca.es ... user@cloudprv-02-9:~$ cat /etc/redhat-release Scientific Linux SL release 5.5 (Boron) user@cloudprv-02-9:~$ }}} == SGE Cluster == The SGE Cluster is based on `Scientific Linux CERN SLC release 6.2` machines, running on x86_64. Local submission is allowed for certain projects. As stated below, there are some shared areas that can be accessed from the computing nodes. The underlying batch system is [[https://arc.liv.ac.uk/trac/SGE|Son of Grid Engine]] 8.0.0d. Refer to the following sources for information: * Using [[/GridEngine]]. * Some useful [[http://arc.liv.ac.uk/SGE/howto/|HOWTOS]]. * SGE [[http://grid.ifca.es/wiki/Cluster/SGE/howto/basic_usage.html|basic usage]]. {{{#!wiki important '''IFCA Gridengine documentation has moved''' The specific documentation for IFCA has been moved [[/GridEngine|to a separate section]]. }}} == Shared areas == The `$HOME` directories are shared between the UIs and the computing nodes. There is a ''projects'' shared area (located at `/gpfs/csic_projects/`), also accessible from the UI and the computing nodes. If your group does not have this area, please open an [[http://support.ifca.es|Incidence ticket]]. === Usage === The shared directories '''are not intended for scratch''', use the temporal areas of the local filesystems instead. In other words, instruct every job you send to copy the input from the shared directory to the local scratch (`$TMPDIR`), execute all operations there, then copy the output back to some shared area where you will be able to retrieve it comfortably from the UI. As mentioned above, the contents of `$TMPDIR` are removed after job execution. === Disk quotas === Disk quotas are enabled on both user and projects filesystems. A message with this information should be shown upon login. If you need more quota on your user space (not in the project shared area), please contact the system administrators explaining your reasons. If you wish to check your quota at a later time, you can use the commands `mmlsquota gpfs_csic` (for user quotas) and `mmlsquota -g `id -g` gpfs_projects` (for group quotas). A script reporting both quotas is located on `/nfs4/usr/bin/rep-user-quotas.py`. A sample output of the latter could be: {{{ ********************************************************************** INFORMATION ABOUT YOUR CURRENT DISK USAGE USER Used Soft Hard Doubt Grace Space (GB): 3.41 20.00 0.00 0.06 none Files (x1000): 64 0 0 0 none GROUP Used Soft Hard Doubt Grace Space (GB): 0.00 1000.00 1500.00 0.00 none Files (x1000): 0 0 0 0 none ********************************************************************** }}} For a basic interpretation of this output, note that the "Used" column will tell you about how much disk space you are using, whereas "Soft" denotes the limit this "Used" amount should not exceed. The "Hard" column is the value of the limit "Used" plus "Doubt" should not cross. A healthy disk space management would require that you periodically delete unused files in your `$HOME` directory, keeping its usage below the limits at all times. In the event that the user exceeds a limit, a grace period will be shown in the "Grace" column. If the user does not correct the situation within the grace period, she will be banned from writing to the disk. For further information you can read the [[http://www.nersc.gov/vendor_docs/ibm/gpfs/am3admst119.html|mmlsquota command manual page]]. == Extra utils/Software == Some extra packages can be found at `/nfs4/opt/`. This is the location for some pieces of software commonly used like: * Matlab's like `octave` language for numerical analysis. * Data plotting `gnuplot` program. * Profiling and debugging `valgrind` tools. Please note that these packages are provided as-is, without further support from IFCA staff. == Support == Before opening a new incidence, please check the [[Cluster/FAQ|Frequently Asked Questions page]] Questions, support and/or feedback should be directed through the use the [[https://support.ifca.es|Helpdesk]]. ---- CategoryUserSupport