Difference between revisions of "Vital-IT"
Line 55: | Line 55: | ||
903580 avalses RUN normal devfrt01 cpt166 695889 Feb 26 14:22 | 903580 avalses RUN normal devfrt01 cpt166 695889 Feb 26 14:22 | ||
'' | '' | ||
+ | |||
* RUN : The job is currently running. | * RUN : The job is currently running. | ||
* PEND : The job is pending, that is, it has not yet been started. | * PEND : The job is pending, that is, it has not yet been started. | ||
+ | * DONE : The job has terminated with status of 0. | ||
+ | * EXIT : The job has terminated with a non-zero status - it may have been aborted due to an error in its execution, or killed by its owner or the LSF administrator. | ||
+ | |||
+ | Special status : | ||
* PSUSP : The job has been suspended, either by its owner or the LSF administrator, while pending. | * PSUSP : The job has been suspended, either by its owner or the LSF administrator, while pending. | ||
* USUSP : The job has been suspended, either by its owner or the LSF administrator, while running. | * USUSP : The job has been suspended, either by its owner or the LSF administrator, while running. | ||
Line 62: | Line 67: | ||
** The load conditions on the execution host or hosts have exceeded a threshold according to the loadStop vector defined for the host or queue. | ** The load conditions on the execution host or hosts have exceeded a threshold according to the loadStop vector defined for the host or queue. | ||
** The run window of the job’s queue is closed. | ** The run window of the job’s queue is closed. | ||
− | |||
− | |||
* UNKWN : mbatchd has lost contact with the sbatchd on the host on which the job runs. | * UNKWN : mbatchd has lost contact with the sbatchd on the host on which the job runs. | ||
* WAIT : For jobs submitted to a chunk job queue, members of a chunk job that are waiting to run. | * WAIT : For jobs submitted to a chunk job queue, members of a chunk job that are waiting to run. | ||
Line 74: | Line 77: | ||
You might also use | You might also use | ||
− | bjobs -r # list all running jobs | + | bjobs -r # list all running jobs |
− | bjobs -d # list all finished jobs (either successfully completed or failed ones) | + | bjobs -d # list all finished jobs (either successfully completed or failed ones) |
− | bjobs -u marcel # list all jobs for this user | + | bjobs -u marcel # list all jobs for this user |
− | bjobs -q normal# list my jobs on this queue | + | bjobs -q normal # list my jobs on this queue |
− | + | ||
Bkill is your best friend, we something goes wrong, you can kill your job(s) with : | Bkill is your best friend, we something goes wrong, you can kill your job(s) with : | ||
bkill 007 # kill job id's 007 | bkill 007 # kill job id's 007 | ||
bkill 0 # kill all my jobs | bkill 0 # kill all my jobs | ||
bkill -q normal 0 # kill all my jobs from the normal queue | bkill -q normal 0 # kill all my jobs from the normal queue | ||
+ | bkill -J "toto" # kill job called toto | ||
=== Building nicer bsub === | === Building nicer bsub === | ||
Line 90: | Line 94: | ||
Then you can use the -w bsub option | Then you can use the -w bsub option | ||
bsub -J a "sh a.sh" | bsub -J a "sh a.sh" | ||
− | bsub -J b -w '(done "a")' "sh b.sh" # start b when a is successfully done | + | bsub -J b -w '(done "a")' "sh b.sh" # start b when a is successfully done |
− | bsub -J c -w '(exit "b")' "sh b.sh" # start c if b has failed | + | bsub -J c -w '(exit "b")' "sh b.sh" # start c if b has failed |
And here we go, we have a mini-pipeline. | And here we go, we have a mini-pipeline. | ||
Revision as of 16:05, 26 February 2009
Contents
How to run jobs on Vital-IT, hints and good pratice
When you have many many jobs to run, running them on the Vital-IT cluster might be better than running them on shoshana or maya.
Simply because running 300+ jobs on a 16 processor machine, will make your jobs competing with each other. (i.e. each job will not be using 100% of a processor, but will be sharing the resources with others).
If your jobs take few minutes to complete, that might not be an issue though.
Any jobs that do not require huge amount of memory (i.e. more than 2Gb) can be easily run on the Vital-it machines. For huge-memory there are few machines available, although only one (rserv) competing with shoshana or maya.
Prerequisites
Before working or crashing vital-it, you will need an account.
You can ask for one there [1]
Ways to submit jobs
You can submit jobs through :
- a web interface [2]
- or you can use a python script (wsub.py), documentation available at wsub-python[3]
- or you can log on to a front-end node (dev.vital-it.ch or prd.vital-it.ch) and submit jobs using the bsub command.[4]
Being nice
PLEASE DO NOT RUN ANY COMPUTATION ON THE FRONT_END NODES (dev,prd) !!!
These front-end nodes are only to submit jobs and do not have the resources to allow you running your jobs interatively.
For interactive and/or heavy computation, you can log on rserv.vital-it.ch or noko01.vital-it.ch .
The jobs on these machines will share the resources (RAM, CPU, I/O) with all other user's jobs.
Bsub in a nutshell
Submitting a simple job.
bsub "sh myscript.sh > mylog" Job <903956> is submitted to default queue <normal>.
That will submit it to the cluster and return you its job id. Here outputs will be redirected to mylog. But you can separate STDOUT and STDERR messages in distinct files with :
bsub -e myerrorfile -o myoutputfile "sh myscript.sh"
Monitoring jobs
You can check it status by doing
bjobs JOBID USER STAT QUEUE FROM_HOST EXEC_HOST JOB_NAME SUBMIT_TIME 903583 avalses RUN normal devfrt01 cpt176 698099 Feb 26 14:22 903581 avalses RUN normal devfrt01 cpt167 695923 Feb 26 14:22 903580 avalses RUN normal devfrt01 cpt166 695889 Feb 26 14:22
- RUN : The job is currently running.
- PEND : The job is pending, that is, it has not yet been started.
- DONE : The job has terminated with status of 0.
- EXIT : The job has terminated with a non-zero status - it may have been aborted due to an error in its execution, or killed by its owner or the LSF administrator.
Special status :
- PSUSP : The job has been suspended, either by its owner or the LSF administrator, while pending.
- USUSP : The job has been suspended, either by its owner or the LSF administrator, while running.
- SSUSP : The job has been suspended by LSF. Either because
- The load conditions on the execution host or hosts have exceeded a threshold according to the loadStop vector defined for the host or queue.
- The run window of the job’s queue is closed.
- UNKWN : mbatchd has lost contact with the sbatchd on the host on which the job runs.
- WAIT : For jobs submitted to a chunk job queue, members of a chunk job that are waiting to run.
- ZOMBI : A job becomes ZOMBI if:
- A non-rerunnable job is killed by bkill while the sbatchd on the execution host is unreachable and the job is shown as UNKWN.
- The host on which a rerunnable job is running is unavailable and the job has been requeued by LSF with a new job ID, as if the job were submitted as a new job.
- After the execution host becomes available, LSF tries to kill the ZOMBI job. Upon successful termination of the ZOMBI job, the job’s status is changed to EXIT.
- With MultiCluster, when a job running on a remote execution cluster becomes a ZOMBI job, the execution cluster treats the job the same way as local ZOMBI jobs. In addition, it notifies the submission cluster that the job is in ZOMBI state and the submission cluster requeues the job.
You might also use
bjobs -r # list all running jobs bjobs -d # list all finished jobs (either successfully completed or failed ones) bjobs -u marcel # list all jobs for this user bjobs -q normal # list my jobs on this queue
Bkill is your best friend, we something goes wrong, you can kill your job(s) with :
bkill 007 # kill job id's 007 bkill 0 # kill all my jobs bkill -q normal 0 # kill all my jobs from the normal queue bkill -J "toto" # kill job called toto
Building nicer bsub
You can submit many jobs and ensure some start after the completion of some other. i.e. if you want to run a,b,c and b needs the output from a, and c is to do only when b failed, Then you can use the -w bsub option
bsub -J a "sh a.sh" bsub -J b -w '(done "a")' "sh b.sh" # start b when a is successfully done bsub -J c -w '(exit "b")' "sh b.sh" # start c if b has failed
And here we go, we have a mini-pipeline.
FAQ
- Can I submit LSF jobs from rserv or noko01?
No, use dev or prd instead.
- My ls is painfully slow, that's because of SFS. Why?
Use /bin/ls or ls --color=none instead.
- How to I check the space left?
Please note, that Vital-it will crash if the space left is less than 1Tb !!! Because, there are some webservices relying on this minimal free space.
df -h . Filesystem Size Used Avail Use% Mounted on client_o2ib 16T 13T 2.7T 83% /sfs1
Known limitations
Vital-it uses the SFS file system [5], files are stripped to many discs for backup reasons.
But this means that any file stat operation (i.e. a simple ls), needs to query the various discs where the data stripes are. This can be painfully slow...
This means that a job doing lots of I/O operations will be slower compared to a NFS file system. Still running in parallel 200+ jobs will be much faster than one by one or by small batches on maya/shoshana.