asc:tutorials:2014october_connect
Differences
This shows you the differences between two versions of the page.
Both sides previous revisionPrevious revisionNext revision | Previous revision | ||
asc:tutorials:2014october_connect [2014/10/09 19:23] – [Lesson 6: Working on the ANL farm] asc | asc:tutorials:2014october_connect [2014/10/10 15:28] (current) – [Lesson 6: Using HTCondor and Tier2] asc | ||
---|---|---|---|
Line 205: | Line 205: | ||
====== Lesson 5: Running on multiple cores ====== | ====== Lesson 5: Running on multiple cores ====== | ||
- | This example is not needed for ATLAS connect. | + | This example is not needed for ATLAS connect. |
+ | look at [[asc: | ||
- | ====== Lesson 6: Working on the ANL farm ====== | + | ====== Lesson 6: Using HTCondor and Tier2 ====== |
- | This example is ANL-specific. If you ran analysis code for RUN I, you should be a familiar with it. We used [[http:// | + | Lesson 5: Working |
- | <note important> | + | In this example |
+ | Start from the new shell and set up environment, | ||
- | First, create a directory for this example: | + | <file bash startJob.sh> |
+ | # | ||
+ | export RUCIO_ACCOUNT=YOUR_CERN_USERNAME | ||
+ | export ATLAS_LOCAL_ROOT_BASE=/ | ||
+ | source ${ATLAS_LOCAL_ROOT_BASE}/ | ||
+ | localSetupFAX | ||
+ | source $AtlasSetup/ | ||
+ | export X509_USER_PROXY=x509up_u21183 | ||
+ | unzip payload.zip | ||
+ | ls | ||
+ | rcSetup -u; rcSetup Base, | ||
+ | rc find_packages | ||
+ | rc compile | ||
+ | cd MyAnalysis/ | ||
+ | rm submitDir | ||
+ | |||
+ | echo $1 | ||
+ | sed -n $1, | ||
+ | cp Inp_$1.txt inputdata.txt | ||
+ | cat inputdata.txt | ||
+ | echo " | ||
+ | testRun submitDir | ||
+ | echo " | ||
+ | </ | ||
- | Prepare a fresh directory: | + | Make sure the RUCIO_ACCOUNT variable is properly set. Make this file executable and create the file that describes our job needs and that we will give to condor: |
- | < | + | |
- | mkdir lesson_6 | + | |
- | </ | + | |
- | Copy the needed directories with RootCore example: | + | <file bash job.sub> |
+ | Jobs=10 | ||
+ | getenv | ||
+ | executable | ||
+ | output | ||
+ | error = MyAnal_$(Jobs).$(Process).error | ||
+ | log = MyAnal_$(Jobs).$(Process).log | ||
+ | arguments = $(Process) $(Jobs) | ||
+ | environment = " | ||
+ | transfer_input_files = payload.zip,/ | ||
+ | universe | ||
+ | # | ||
+ | queue $(Jobs) | ||
+ | </ | ||
- | <code bash> | + | To access files using FAX the jobs need a valid grid proxy. That's why we send it with each job. Proxy is the file starting with " |
- | cp -r / | + | |
- | </ | + | |
- | Then check that you can compile | + | You need to pack all of the working directory into a payload.zip file: |
- | <code bash> | + | <file bash startJob.sh> |
- | cd lesson_6/ | + | startJob.sh |
- | source setup.sh | + | rc clean |
- | </code> | + | rm -rf RootCoreBin |
+ | zip -r payload.zip * | ||
+ | </file> | ||
- | As usual, our analysis is in " | + | Now you may submit your task for the execution and follow its status |
- | Now we want to submit jobs to the data distributed on several computers of the farm. Go to the upper directory and setup the farm: | ||
- | |||
- | <code bash> | ||
- | cd ../; source s_asc; | ||
- | </ | ||
- | |||
- | We will send jobs using the directory " | ||
- | |||
- | <code bash> | ||
- | cd submit | ||
- | arc_ls -s / | ||
- | arc_ls | ||
- | </ | ||
- | |||
- | The first command shows the summary of distributed data (12 files per server), while the second lists all available data on each node. | ||
- | Now we will send the job to the farm. | ||
- | Change the line: | ||
< | < | ||
- | package_dir=/users/chakanau/ | + | chmod 755 ./startJob.sh; |
</ | </ | ||
- | inside " | ||
- | This sends jobs to the farm (2 jobs per server). Check the status as: | ||
<code bash> | <code bash> | ||
- | condor_q | + | ~> condor_submit job.sub |
+ | Submitting job(s).......... | ||
+ | 10 job(s) submitted to cluster 49677. | ||
+ | |||
+ | ~> condor_q | ||
+ | -- Submitter: login.atlas.ci-connect.net : < | ||
+ | | ||
+ | 49677.0 | ||
+ | 49677.1 | ||
+ | 49677.2 | ||
+ | 49677.3 | ||
+ | 49677.4 | ||
+ | 49677.5 | ||
+ | 49677.6 | ||
+ | 49677.7 | ||
+ | 49677.8 | ||
+ | 49677.9 | ||
+ | |||
+ | 10 jobs; 0 completed, 0 removed, 0 idle, 10 running, 0 held, 0 suspended | ||
</ | </ | ||
- | |||
- | When the jobs are done, the output files will be inside " | ||
- | |||
- | <code bash> | ||
- | arc_add | ||
- | </ | ||
- | This will create the final output file " | ||
- | |||
- | |||
- |
asc/tutorials/2014october_connect.1412882610.txt.gz · Last modified: 2014/10/09 19:23 by asc