5.6.1 Running CMSSW code on the Grid using CRAB

Complete: 5
Detailed Review status

WARNING

  • You should always use latest production CRAB version
  • This tutorial may be outdated since it was prepared for a live lesson at a specific time and thus refers to a particular dataset and CMSSW version that may not be available when you read this (and where you try it).
    • as of July 2012 you should be able to kickstart your Crab work using CMSSW 5_2_5 and the dataset /RelValBeamHalo/CMSSW_5_2_5_cand1-START52_V9-v1/GEN-SIM-DIGI-RAW as MC data and /SingleMu/Run2012B-TOPMuPlusJets-PromptSkim-v1/AOD as real data.

Contents:

Prerequisites to run the tutorial

  • to have a valid Grid certificate
  • to be registered to the CMS virtual organization
  • to be registered to the siteDB
  • to have access to lxplus machines or to an SLC5 User Interface

Recipe for the tutorial

For this tutorial we will refer to CMS software:

  • CMSSW_5_2_5

and we will use an already prepared CMSSW analysis code to analyze the sample:

  • The tutorial will focus on the basic workflow using the dataset: /RelValBeamHalo/CMSSW_5_2_5_cand1-START52_V9-v1/GEN-SIM-DIGI-RAW (MC dataset) and /SingleMu/Run2012B-TOPMuPlusJets-PromptSkim-v1/AOD (real data): CRAB configuration file for real data with lumi mask

We will use the central installation of CRAB available at CERN:

  • CRAB_2_8_1

The example is written to use the csh shell family. If you want to use the Bourne Shell replace csh with sh.

Legend of colors for this tutorial

BEIGE background for the commands to execute  (cut&paste)
GREEN background for the output sample of the executed commands (nearly what you should see in your terminal)
BLUE background for the configuration files  (cut&paste)

Setup local Environment and prepare user analysis code

In order to submit jobs to the Grid, you must have an access to a LCG User Interface (LCG UI). It will allow you to access WLCG-affiliated resources in a fully transparent way. LXPLUS users can get an LCG UI via AFS by:

source /afs/cern.ch/cms/LCG/LCG-2/UI/cms_ui_env.csh

Install CMSSW project in a directory of your choice. In this case we create a "Tutorial" directory:

mkdir TUTORIAL
cd TUTORIAL
cmsrel CMSSW_5_2_5
#cmsrel is an alias of scramv1 project CMSSW CMSSW_5_2_5
cd CMSSW_5_2_5/src/
cmsenv
#cmsenv is an alias for scramv1 runtime -csh

For this tutorial we are going to use as CMSSW configuration file, the tutorial.py:

import FWCore.ParameterSet.Config as cms
process = cms.Process('Slurp')

process.source = cms.Source("PoolSource", fileNames = cms.untracked.vstring())
process.maxEvents = cms.untracked.PSet( input       = cms.untracked.int32(10) )
process.options   = cms.untracked.PSet( wantSummary = cms.untracked.bool(True) )

process.output = cms.OutputModule("PoolOutputModule",
    outputCommands = cms.untracked.vstring("drop *", "keep recoTracks_*_*_*"),
    fileName = cms.untracked.string('outfile.root'),
)
process.out_step = cms.EndPath(process.output)

CRAB setup

Setup on lxplus:

In order to setup and use CRAB from any directory, source the script crab.(c)sh located in /afs/cern.ch/cms/ccs/wm/scripts/Crab/, which always points to the latest version of CRAB. After the source of the script it's possible to use CRAB from any directory (typically use it on your CMSSW working directory).

source /afs/cern.ch/cms/ccs/wm/scripts/Crab/crab.csh

Warning: in order to have the correct environment, the order to source env files has always to be

  • source of UI env
  • setup of CMSSW software
  • source of CRAB env

Locate the dataset and prepare CRAB submission

In order to run our analysis over a whole dataset, we have to find first the data name and then put it on the crab configuration file.

Data selection

To select data you want to access, use the DBS web page where available datasets are listed Data Aggregation Service (DAS) . For this tutorial we'll use :

/RelValBeamHalo/CMSSW_5_2_5_cand1-START52_V9-v1/GEN-SIM-DIGI-RAW
 (MC data)
  • Beware: datasets availability as sites changes with time, if you are trying to follow this tutorial after the date it was given, you may need to use another one

CRAB configuration

Modify the CRAB configuration file crab.cfg according to your needs: a fully documented template is available at $CRABPATH/full_crab.cfg, a template with essential parameters is available at $CRABPATH/crab.cfg . The default name of configuration file is crab.cfg, but you can rename it as you want.

Copy one of these files in your local area.

For guidance, see the list and description of configuration parameters in the on-line crab documentation (first point of this page) . For this tutorial, the only relevant sections of the file are [CRAB], [CMSSW] and [USER] .

Configuration parameters

The list of the main parameters you need to specify on your crab.cfg:
  • pset: the CMSSW configuration file name;
  • output_file: the output file name produced by your pset; if in the CMSSW pset the output is defined in TFileService, the file is automatically handled by CRAB, and there is no need to specify it on this parameter;
  • datasetpath: the full dataset name you want to analyze;
  • Jobs splitting:
    • By event: only for MC data. You need to specify 2 of these parameters: total_number_of_events, number_of_jobs, events_per_job
      • specify the total_number_of_events and the number_of_jobs: this will assing to each jobs a number of events equal to total_number_of_events/number_of_jobs;
      • specify the total_number_of_events and the events_per_job: will assing to each jobs the number events_per_job and will calculate the number of jobs by total_number_of_events/events_per_job;
      • or you can specify the number_of_jobs and the events_per_job...;
    • By lumi: real data require it. You need to specify 2 of these parameters: total_number_of_lumis, lumis_per_job, number_of_jobs
      • because jobs in split by lumi mode process entire rather than partial files, you will often end up with fewer jobs processing more lumis than expected. Additionally, a single job cannot analyze files from multiple blocks in DBS. So these parameters are "advice" to CRAB rather than determinative.
      • specify the lumis_per_job and the number_of_jobs the total number of lumis processed will be number_of_jobs x lumis_per_job.
      • or you can specify the total_number_of_lumis and the number_of_jobs .....
      • lumi_mask: the filename of a JSON file that describes which runs and lumis to process. CRAB will skip luminosity blocks not listed in the file.
  • return_data: this can be 0 or 1; if it is one you will retrieve your output files to your local working area;
  • copy_data: this can be 0 or 1; if it is one you will copy your output files to a remote Storage Element;
  • local_stage_out: this can be 0 or 1; if this is one your produced output is copied the closeSE in the case of failure of the copy to the SE specified in your crab.cfg
  • publish_data: this can be 0 or 1; if it is one you can publish your produced data to a local DBS;
  • use_server: one of the available servers will be used depending on the client release;
  • scheduler: the name of the scheduler you want to use;
  • jobtype: the type of the jobs.

Run CRAB on MonteCarlo data copying the output to an SE

The chance to copy the output to an existing Storage Element allows to bypass the output size limit constraint, to publish the data on a local DBS and then to easily re-run over the published data. In order to make CRAB copies to a Storage Element you have to add the following information on the Crab configuration file:
  • that we want to copy our results adding copy_data=1 and return_data=0 (it is not allowed to have both at 1);
  • add the official CMS site name where we are going to copy our results (examples just for this session with Legnaro T2_IT_Legnaro StorageElement); the name of other official sites can be found in the siteDB

CRAB configuration file for MonteCarlo data

You can find more details on this at the corresponding link on the Crab FAQ page.

The CRAB configuration file (default name crab.cfg) should be located at the same location as the CMSSW parameter-set to be used by CRAB with the following content:

[CMSSW]
total_number_of_events  = 10
number_of_jobs          = 5
pset                    = tutorial.py
datasetpath             = /RelValProdTTbar/JobRobot-MC_3XY_V24_JobRobot-v1/GEN-SIM-DIGI-RECO 
output_file              = outfile.root

[USER]
return_data             = 0
email                   = yourEmailAddressHere@cern.ch   

copy_data               = 1
storage_element        = T2_IT_Legnaro
user_remote_dir         = TutGridSchool

[CRAB]
scheduler               = glite
jobtype                 = cmssw
use_server              = 1

Run Crab

Once your crab.cfg is ready and the whole underlying environment is set up, you can start running CRAB. CRAB supports command line help which can be useful for the first time. You can get it via:
crab -h

Job Creation

The job creation checks the availability of the selected dataset and prepares all the jobs for submission according to the selected job splitting specified in the crab.cfg

  • By default the creation process creates a CRAB project directory (default: crab_0_date_time) in the current working directory, where the related crab configuration file is cached for further usage, avoiding interference with other (already created) projects

  • Using the [USER] ui_working_dir parameter in the configuration file CRAB allows the user to chose the project name, so that it can be used later to distinguish multiple CRAB projects in the same directory.

crab -create  
that takes by default the configuration file called crab.cfg associated for this tutorial with MC data.

The creation command could ask for proxy/myproxy passwords the first time you use it and it should produce a similar screen output like:

[lxplus444] $ crab -create
crab:  Version 2.8.1 running on Tue Jul 24 17:59:34 2012 CET (15:59:34 UTC)

crab. Working options:
        scheduler           glite
        job type            CMSSW
        server              ON (use_server)
        working directory   /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120724_175934/

Enter GRID pass phrase:
Your identity: /DC=ch/DC=cern/OU=Organic Units/OU=Users/CN=fanzago/CN=610896/CN=Federica Fanzago
Creating temporary proxy .............................. Done
Contacting  lcg-voms.cern.ch:15002 [/DC=ch/DC=cern/OU=computers/CN=lcg-voms.cern.ch] "cms" Failed

Error: cms: Problems in DB communication.

Trying next server for cms.
Creating temporary proxy ................................................................. Done
Contacting  voms.cern.ch:15002 [/DC=ch/DC=cern/OU=computers/CN=voms.cern.ch] "cms" Done
Creating proxy ........................ Done
Your proxy is valid until Wed Aug  1 17:59:40 2012
crab:  Contacting Data Discovery Services ...
crab:  Accessing DBS at: http://cmsdbsprod.cern.ch/cms_dbs_prod_global/servlet/DBSServlet
crab:  Requested dataset: /RelValBeamHalo/CMSSW_5_2_5_cand1-START52_V9-v1/GEN-SIM-DIGI-RAW has 9000 events in 1 blocks.

crab:  May not create the exact number_of_jobs requested.
crab:  5 job(s) can run on 10 events.

crab:  List of jobs and available destination sites:

Block     1: jobs                  1-5: sites: T2_CH_CERN

crab:  Checking remote location
crab:  Creating 5 jobs, please wait...
crab:  Total of 5 jobs created.

Log file is /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120724_175934/log/crab.log

  • the project directory called crab_0_120724_175934 is created

#CMS.JobSubmission

Job Submission

With the submission command it's possible to specify a combination of jobs and job-ranges separated by comma (e.g.: =1,2,3-4), the default is all. To submit all jobs of the last created project with the default name, it's enough to execute the following command:

crab -submit 
to submit a specific project:
crab -submit -c  <dir name>

which should produce a similar screen output like:


[lxplus444]  $ crab -submit
crab:  Version 2.8.1 running on Tue Jul 24 18:02:39 2012 CET (16:02:39 UTC)

crab. Working options:
        scheduler           glite
        job type            CMSSW
        server              ON (default)
        working directory   /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120724_175934/

crab:  Registering credential to the server : vocms21.cern.ch
crab:  Credential successfully delegated to the server.

crab:  Starting sending the project to the storage vocms21.cern.ch...
crab:  Task crab_0_120724_175934 successfully submitted to server vocms21.cern.ch

crab:  Total of 5 jobs submitted
Log file is /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120724_175934/log/crab.log

Job Status Check

Check the status of the jobs in the latest CRAB project with the following command:
crab -status 
to check a specific project:
crab -status -c  <dir name>

which should produce a similar screen output like:

[lxplus444] $ crab -status
crab:  Version 2.8.1 running on Tue Jul 24 18:09:23 2012 CET (16:09:23 UTC)

crab. Working options:
        scheduler           glite
        job type            CMSSW
        server              ON (default)
        working directory   /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120724_175934/

crab:  
ID    END STATUS            ACTION       ExeExitCode JobExitCode E_HOST
----- --- ----------------- ------------  ---------- ----------- ---------
1     N   Running           SubSuccess                           ce203.cern.ch
2     N   Ready             SubSuccess                           ce208.cern.ch
3     N   Ready             SubSuccess                           ce203.cern.ch
4     N   Running           SubSuccess                           ce204.cern.ch
5     N   Ready             SubSuccess                           ce208.cern.ch
                  

crab:   5 Total Jobs 
 >>>>>>>>> 5 Jobs Submitted 
        List of jobs Submitted: 1-5 

crab:  You can also follow the status of this task on :
        CMS Dashboard: http://dashb-cms-job-task.cern.ch/taskmon.html#task=fanzago_crab_0_120724_175934_0ml7t2
        Server page: http://vocms21.cern.ch:8888/logginfo
        Your task name is: fanzago_crab_0_120724_175934_0ml7t2 

Log file is /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120724_175934/log/crab.log

Also, you can have a look at the web page of the server where you can see the status progress of you job. Simply, execute the command:

crab -printId
crab -printId -c <dir name>
And you will get the unique id of your jobs:

[lxplus444] $ crab -printId -c crab_0_120724_175934/
crab:  Version 2.8.1 running on Tue Jul 24 18:56:36 2012 CET (16:56:36 UTC)

crab. Working options:
        scheduler           glite
        job type            CMSSW
        server              ON (default)
        working directory   /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120724_175934/

crab:  
Task Id = fanzago_crab_0_120724_175934_0ml7t2     
--------------------------------------------------------------------------------------------

crab:  You can also follow the status of this task on :
        CMS Dashboard: http://dashb-cms-job-task.cern.ch/taskmon.html#task=fanzago_crab_0_120724_175934_0ml7t2
        Server page: http://vocms21.cern.ch:8888/logginfo
        Your task name is: fanzago_crab_0_120724_175934_0ml7t2 

Log file is /afs/cern.ch/user/f/fanzago/scratch0/TEST_RELEASE/TUTORIAL/crab_0_120724_175934//log/crab.log

Job Output Retrieval

For the jobs which are in the "Done" status it is possible to retrieve the log files of the jobs (just the log files, because the output files are copied to the Storage Element associated to the T2 specified on the crab.cfg and infact return_data is 0). The following command retrieves the log files of all "Done" jobs of the last created CRAB project:
crab -getoutput 
to get the output of a specific project:
crab -getoutput -c  <dir name>

the job results (CMSSW_n.stdout, CMSSW_n.stderr and crab_fjr_n.xml) will be copied in the res subdirectory of your crab project:

[lxplus444] $ crab -get -c crab_0_120724_175934/
crab:  Version 2.8.1 running on Tue Jul 24 19:17:40 2012 CET (17:17:40 UTC)

crab. Working options:
        scheduler           glite
        job type            CMSSW
        server              ON (default)
        working directory   /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120724_175934/

crab:  Starting retrieving output from server vocms21.cern.ch...
crab:  Results of Jobs # 1 are in /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120724_175934//res/
crab:  Results of Jobs # 2 are in /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120724_175934//res/
crab:  Results of Jobs # 3 are in /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120724_175934//res/
crab:  Results of Jobs # 4 are in /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120724_175934//res/
crab:  Results of Jobs # 5 are in /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120724_175934//res/
Log file is /afs/cern.ch/user/f/fanzago/scratch0//TUTORIAL/crab_0_120724_175934//log/crab.log

The stderr is an empty file, the stdout is the output of the wrapper of your analysis code (the output of CMSSW.sh script created by CRAB) and the crab_fjr.xml is the FrameworkJobReport created by your analysis code.

Use the -report option

Print a short report about the task, namely the total number of events and files processed/requested/available, the name of the dataset path, a summary of the status of the jobs, and so on. A summary file of the runs and luminosity sections processed is written to res/. In principle -report should generate all the info needed for an analysis. Command to execute:

crab -report
Example of execution:

[lxplus444] $ crab -report -c crab_0_120724_175934/
crab:  Version 2.8.1 running on Tue Jul 24 19:20:14 2012 CET (17:20:14 UTC)

crab. Working options:
        scheduler           glite
        job type            CMSSW
        server              ON (default)
        working directory   /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120724_175934/

crab:  --------------------
Dataset: /RelValBeamHalo/CMSSW_5_2_5_cand1-START52_V9-v1/GEN-SIM-DIGI-RAW
Remote output :
SE: T2_IT_Legnaro t2-srm-02.lnl.infn.it  srmPath: srm://t2-srm-02.lnl.infn.it:8443/srm/managerv2?SFN=/pnfs/lnl.infn.it/data/cms/store/user/fanzago/TutGridSchool/
Total Events read: 10
Total Files read: 5
Total Jobs : 5
Luminosity section summary file: /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120724_175934/res/lumiSummary.json
   # Jobs: Cleared:5

----------------------------

Log file is /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120724_175934//log/crab.log

The full srm path will allow you to know where your data has been stored and to perform operations by hand on it. As example you can delete the data using srmrm command and check the content of the remote directory through srmls. In this case the remote directory is:

srm://t2-srm-02.lnl.infn.it:8443/srm/managerv2?SFN=/pnfs/lnl.infn.it/data/cms/store/user/fanzago/TutGridSchool/

it could be necessary to substitute the ? with the "?" in the srm path, depending on the shell you are using.

Here the content of the file containing the luminosity summary /crab_0_101028_123033/res/lumiSummary.json:

{"1": [[666666, 666666]]}

Copy the output from the SE to the local User Interface

Option that can be used only if your output have been previously copied by CRAB on a remote SE. By default the copyData copies your output from the remote SE locally on the current CRAB working directory (under res). Otherwise you can copy the output from the remote SE to another one, specifying either -dest_se= or -dest_endpoint=. If dest_se is used, CRAB finds the correct path where the output can be stored. The command to execute in order to retrieve locally the remote output files to your local user interface is: Command to execute:
crab -copyData ## or crab -copyData -c <dir name>
An example of execution:
[lxplus444] $ crab -copyData -c crab_0_120724_175934/crab:  Version 2.8.1 running on Tue Jul 24 19:23:41 2012 CET (17:23:41 UTC)

crab. Working options:
        scheduler           glite
        job type            CMSSW
        server              ON (default)
        working directory   /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120724_175934/

crab:  Copy file locally.
        Output dir: /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120724_175934//res/
crab:  Starting copy...
directory/afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120724_175934//res/already exists
crab:  Copy success for file: outfile_3_1_OnZ.root 
crab:  Copy success for file: outfile_2_1_tyW.root 
crab:  Copy success for file: outfile_1_1_JIQ.root 
crab:  Copy success for file: outfile_4_1_rVx.root 
crab:  Copy success for file: outfile_5_1_TmT.root 
Log file is /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120724_175934//log/crab.log

Publish your result in DBS

The publication of the produced data to a DBS allows to re-run over the produced data that has been published. Below the instruction to follow and here the link to the how to. You have to add to the Crab configuration file more information specifying the data name to publish and the DBS url instance where to register the output results.
[USER]
....
publish_data            = 1
publish_data_name       = what_you_want
dbs_url_for_publication = url_local_dbs
....
Warning:
  • all the parameters related publication have to be added in the configuration file before creation of jobs, even if the publication step is executed after retrieving of job output.
  • for this tutorial we will publish the data to the test DBS instance https://cmsdbsprod.cern.ch:8443/cms_dbs_ph_analysis_02_writer/servlet/DBSServlet. This instance is only for publication test, so the maintaining of published data is not guarantee for long time and the publication here doesn't require writing authorization. If you belong to a PAG groups, you have to publish your data to the DBS associated to your group, checking at the DBS access twiki page the correct DBS url and which role in voms you need to be an allowed user..
  • remember to change the ui_working_dir value in the configuration file to create a new project (if you don't use the default name of crab project), otherwise the creation step will fail with the error message "project already exists, please remove it before create new task "

Run Crab publishing your results

You can also run your analysis code publishing the results copied to a remote Storage Element. Here below an example of the CRAB configuration file, coherent with this tutorial:

For MC data (crab.cfg)

[CMSSW]
total_number_of_events  = 100
number_of_jobs          = 10
pset                    = tutorial.py
datasetpath             = /RelValProdTTbar/JobRobot-MC_3XY_V24_JobRobot-v1/GEN-SIM-DIGI-RECO 
output_file              = outfile.root

[USER]
return_data             = 0
email                   = yourEmailAddressHere@cern.ch

copy_data               = 1
storage_element         = T2_IT_Legnaro
publish_data            = 1
publish_data_name       = publish_TutGridSchool
dbs_url_for_publication = https://cmsdbsprod.cern.ch:8443/cms_dbs_ph_analysis_02_writer/servlet/DBSServlet

[CRAB]
scheduler               = glite
jobtype                 = cmssw
use_server              = 1

And with this crab.cfg you can re-do complete the workflow as described before plus the publication step:

  • creation
  • submission
  • status progress monitoring
  • output retrieval
  • publish the results

Use the -publish option

After having done the previous workflow untill the retrieval of you jobs, you can publish the output data that have been stored in the Storage Element indicated in the crab.cfg using

   crab -publish
or
   crab -publish -c <dir name>
to publish outputs of a specific project. It is not necessary all the jobs are done and retrieved. You can publish your output at different time.

It will look for all the FrameworkJobReport ( crab-project-dir/res/crab_fjr_*.xml ) produced by each jobs and will extract from there the information (i.e. number of events, LFN,....) to publish.

Publication output example


[lxplus428] $ crab -publish
crab:  Version 2.8.1 running on Wed Jul 25 01:24:00 2012 CET (23:24:00 UTC)
crab. Working options:
        scheduler           glite
        job type            CMSSW
        server              ON (default)
        working directory   /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120725_002327/

crab:  <dbs_url_for_publication> = https://cmsdbsprod.cern.ch:8443/cms_dbs_ph_analysis_02_writer/servlet/DBSServlet
file_list =  ['/afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120725_002327/res//crab_fjr_1.xml', '/afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120725_002327/res//crab_fjr_2.xml', '/afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120725_002327/res//crab_fjr_3.xml', '/afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120725_002327/res//crab_fjr_4.xml', '/afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120725_002327/res//crab_fjr_5.xml']
crab:  --->>> Start dataset publication
crab:  --->>> Importing parent dataset in the dbs: /RelValBeamHalo/CMSSW_5_2_5_cand1-START52_V9-v1/GEN-SIM-DIGI-RAW
crab:  --->>> Importing all parents level
-----------------------------------------------------------------------------------
Transferring path /RelValBeamHalo/CMSSW_5_2_1-START52_V4-v1/GEN-SIM 
            block /RelValBeamHalo/CMSSW_5_2_1-START52_V4-v1/GEN-SIM#0e9236ec-24fb-4671-b46f-60d78005f138 
-----------------------------------------------------------------------------------

-----------------------------------------------------------------------------------
Transferring path /RelValBeamHalo/CMSSW_5_2_5_cand1-START52_V9-v1/GEN-SIM-DIGI-RAW 
            block /RelValBeamHalo/CMSSW_5_2_5_cand1-START52_V9-v1/GEN-SIM-DIGI-RAW#eb664126-d232-43ef-bf51-ab38ccfd807c 
-----------------------------------------------------------------------------------

crab:  --->>> duration of all parents import (sec): 12.8555440903
crab:  Import ok of dataset /RelValBeamHalo/CMSSW_5_2_5_cand1-START52_V9-v1/GEN-SIM-DIGI-RAW
crab:  PrimaryDataset = RelValBeamHalo
crab:  ProcessedDataset = fanzago-publish_TutGridSchool-f30a6bb13f516198b2814e83414acca1
crab:  <User Dataset Name> = /RelValBeamHalo/fanzago-publish_TutGridSchool-f30a6bb13f516198b2814e83414acca1/USER
crab:  --->>> End dataset publication
crab:  --->>> Start files publication
crab:  --->>> End files publication
crab:  --->>> End dataset publication
crab:  --->>> Start files publication
DEBUG:crab::FJR = /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120725_002327/res//crab_fjr_1.xml
DEBUG:crab::--->>> LFN of file to publish =  /store/user/fanzago/RelValBeamHalo/publish_TutGridSchool/f30a6bb13f516198b2814e83414acca1/outfile_1_1_lJ7.root
DEBUG:crab::--->>> Inserting file in blocks = ['/RelValBeamHalo/fanzago-publish_TutGridSchool-f30a6bb13f516198b2814e83414acca1/USER#acc7c18e-9820-4e90-b883-52292d24cdb1']
debug_verbose:crab::BlocksList = ['/RelValBeamHalo/fanzago-publish_TutGridSchool-f30a6bb13f516198b2814e83414acca1/USER#acc7c18e-9820-4e90-b883-52292d24cdb1']
....
debug_verbose:crab::closeBlock True
crab:  --->>> End files publication
crab:  --->>> Check data publication: dataset /RelValBeamHalo/fanzago-publish_TutGridSchool-f30a6bb13f516198b2814e83414acca1/USER in DBS url https://cmsdbsprod.cern.ch:8443/cms_dbs_ph_analysis_02_writer/servlet/DBSServlet

=== dataset /RelValBeamHalo/fanzago-publish_TutGridSchool-f30a6bb13f516198b2814e83414acca1/USER
=== dataset description =  
===== File block name: /RelValBeamHalo/fanzago-publish_TutGridSchool-f30a6bb13f516198b2814e83414acca1/USER#acc7c18e-9820-4e90-b883-52292d24cdb1
      File block located at:  ['t2-srm-02.lnl.infn.it']
      File block status: 0
      Number of files: 5
      Number of Bytes: 7178518
      Number of Events: 10

 total events: 10 in dataset: /RelValBeamHalo/fanzago-publish_TutGridSchool-f30a6bb13f516198b2814e83414acca1/USER

crab:  You can obtain more info about files of the dataset using: crab -checkPublication -USER.dataset_to_check=/RelValBeamHalo/fanzago-publish_TutGridSchool-f30a6bb13f516198b2814e83414acca1/USER -USER.dbs_url_for_publication=https://cmsdbsprod.cern.ch:8443/cms_dbs_ph_analysis_02_writer/servlet/DBSServlet -debug
Log file is /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120725_002327/log/crab.log

crab:  --->>> Check data publication: dataset /RelValBeamHalo/fanzago-publish_TutGridSchool-f30a6bb13f516198b2814e83414acca1/USER in DBS url https://cmsdbsprod.cern.ch:8443/cms_dbs_ph_analysis_02_writer/servlet/DBSServlet

=== dataset /RelValBeamHalo/fanzago-publish_TutGridSchool-f30a6bb13f516198b2814e83414acca1/USER
=== dataset description =  
===== File block name: /RelValBeamHalo/fanzago-publish_TutGridSchool-f30a6bb13f516198b2814e83414acca1/USER#acc7c18e-9820-4e90-b883-52292d24cdb1
      File block located at:  ['t2-srm-02.lnl.infn.it']
      File block status: 0
      Number of files: 5
      Number of Bytes: 7178518
      Number of Events: 10

 total events: 10 in dataset: /RelValBeamHalo/fanzago-publish_TutGridSchool-f30a6bb13f516198b2814e83414acca1/USER

crab:  You can obtain more info about files of the dataset using: crab -checkPublication -USER.dataset_to_check=/RelValBeamHalo/fanzago-publish_TutGridSchool-f30a6bb13f516198b2814e83414acca1/USER -USER.dbs_url_for_publication=https://cmsdbsprod.cern.ch:8443/cms_dbs_ph_analysis_02_writer/servlet/DBSServlet -debug
Log file is /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120725_002327/log/crab.log

Warning: some versions of CMSSW switch off the debug mode of crab, so a lot of duplicated info can be reported at screen level.

Check the result of data publication and analyze your published data

Note that:
  • CRAB by default publishes all files finished correctly, including files with 0 events
  • CRAB by default imports all dataset parents of your dataset

To check if your data have been published you can use the option:

crab -checkPublication -USER.dataset_to_check=your_dataset_path -USER.dbs_url_for_publication=url_local_dbs -debug
where dbs_url_for_publication is the dbs_url you have written in the crab.cfg file and name_of_your_dataset is the name of dataset published by CRAB primarydataset/publish_data_name/USER (it is also printed by CRAB in corrispondence of the line "User Dataset Name" when you run the crab -publish command).

The output is:

[lxplus428] $ crab -checkPublication -USER.dataset_to_check=/RelValBeamHalo/fanzago-publish_TutGridSchool-f30a6bb13f516198b2814e83414acca1/USER -USER.dbs_url_for_publication=https://cmsdbsprod.cern.ch:8443/cms_dbs_ph_analysis_02_writer/servlet/DBSServlet -debug

crab:  Version 2.8.1 running on Wed Jul 25 01:42:30 2012 CET (23:42:30 UTC)

crab. Working options:
        scheduler           glite
        job type            CMSSW
        server              ON (default)
        working directory   /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120725_002327/

crab:  Downloading file [http://cmsdoc.cern.ch/cms/LCG/crab/config/] to [/afs/cern.ch/user/f/fanzago/.cms_crab/allowed_releases.conf].
....
crab:  Enforced black list: <Downloader.Downloader instance at 0x1bc9cea8> 
crab:  --->>> Check data publication: dataset /RelValBeamHalo/fanzago-publish_TutGridSchool-f30a6bb13f516198b2814e83414acca1/USER in DBS url https://cmsdbsprod.cern.ch:8443/cms_dbs_ph_analysis_02_writer/servlet/DBSServlet

PrimaryDataset =  RelValBeamHalo
ProcessedDataset =  fanzago-publish_TutGridSchool-f30a6bb13f516198b2814e83414acca1
DataTier =  USER
datasets matching your requirements=  [{'RunsList': [], 'Name': 'fanzago-publish_TutGridSchool-f30a6bb13f516198b2814e83414acca1', 'PathList': ['/RelValBeamHalo/fanzago-publish_TutGridSchool-f30a6bb13f516198b2814e83414acca1/USER'], 'LastModifiedBy': '/DC=ch/DC=cern/OU=Organic Units/OU=Users/CN=fanzago/CN=610896/CN=Federica Fanzago', 'AlgoList': [{'ExecutableName': 'cmsRun', 'ApplicationVersion': 'CMSSW_5_2_5', 'ParameterSetID': {'Hash': 'f30a6bb13f516198b2814e83414acca1'}, 'ApplicationFamily': 'cmsRun'}], 'XtCrossSection': 0.0, 'Status': 'VALID', 'ParentList': [], 'AcquisitionEra': '', 'PhysicsGroup': 'NoGroup', 'Description': '', 'GlobalTag': '', 'PrimaryDataset': {'Name': 'RelValBeamHalo'}, 'TierList': ['USER'], 'CreatedBy': '/DC=ch/DC=cern/OU=Organic Units/OU=Users/CN=fanzago/CN=610896/CN=Federica Fanzago', 'PhysicsGroupConverner': 'NO_CONVENOR', 'CreationDate': '1343172257', 'LastModificationDate': '1343172257'}]

=== dataset /RelValBeamHalo/fanzago-publish_TutGridSchool-f30a6bb13f516198b2814e83414acca1/USER
=== dataset description =  
===== File block name: /RelValBeamHalo/fanzago-publish_TutGridSchool-f30a6bb13f516198b2814e83414acca1/USER#acc7c18e-9820-4e90-b883-52292d24cdb1
      File block located at:  ['t2-srm-02.lnl.infn.it']
      File block status: 0
      Number of files: 5
      Number of Bytes: 7178518
      Number of Events: 10
--------- info about files --------
 Size    Events          LFN     FileStatus 
1432882 2 /store/user/fanzago/RelValBeamHalo/publish_TutGridSchool/f30a6bb13f516198b2814e83414acca1/outfile_1_1_lJ7.root 
1432672 2 /store/user/fanzago/RelValBeamHalo/publish_TutGridSchool/f30a6bb13f516198b2814e83414acca1/outfile_2_1_Qux.root 
1430002 2 /store/user/fanzago/RelValBeamHalo/publish_TutGridSchool/f30a6bb13f516198b2814e83414acca1/outfile_3_1_UOE.root 
1437024 2 /store/user/fanzago/RelValBeamHalo/publish_TutGridSchool/f30a6bb13f516198b2814e83414acca1/outfile_4_1_to9.root 
1445938 2 /store/user/fanzago/RelValBeamHalo/publish_TutGridSchool/f30a6bb13f516198b2814e83414acca1/outfile_5_1_eU5.root 

 total events: 10 in dataset: /RelValBeamHalo/fanzago-publish_TutGridSchool-f30a6bb13f516198b2814e83414acca1/USER

Log file is /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120725_002327/log/crab.log



If you want to analyze your published data you have to modify your crab.cfg specifying the datasetpath name of your dataset and the dbs_url where data are published

[CMSSW]
....
datasetpath=your_dataset_path
dbs_url=url_local_dbs
If you found that data of your interest is in the DBS instance "https://cmsdbsprod.cern.ch:8443/cms_dbs_ph_analysis_02_writer/servlet/DBSServlet" you can specify
https://cmsdbsprod.cern.ch:8443/cms_dbs_ph_analysis_02_writer/servlet/DBSServlet

The creation output will be something similar to:

[lxplus428] $ crab -create -cfg crab_5_2_5_analisi_pubbl.cfg
crab:  Version 2.8.1 running on Wed Jul 25 01:49:24 2012 CET (23:49:24 UTC)

crab. Working options:
        scheduler           glite
        job type            CMSSW
        server              ON (default)
        working directory   /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120725_014924/

crab:  Contacting Data Discovery Services ...
crab:  Accessing DBS at: https://cmsdbsprod.cern.ch:8443/cms_dbs_ph_analysis_02_writer/servlet/DBSServlet
crab:  Requested dataset: /RelValBeamHalo/fanzago-publish_TutGridSchool-f30a6bb13f516198b2814e83414acca1/USER has 10 events in 1 blocks.

crab:  May not create the exact number_of_jobs requested.
crab:  5 job(s) can run on 10 events.

crab:  List of jobs and available destination sites:

Block     1: jobs                  1-5: sites: T2_IT_Legnaro

crab:  Checking remote location
crab:  WARNING: The stageout directory already exists. Be careful not to accidentally mix outputs from different tasks
crab:  Creating 5 jobs, please wait...
crab:  Total of 5 jobs created.

Log file is /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_120725_014924/log/crab.log

Run CRAB on real data copying the output to an SE

Running CRAB on real data has no major difference with running CRAB on MonteCarlo data. The main difference is related on the configuration preparation for the CRAB workflow, as showed in the next section.

CRAB configuration file for real data with lumi mask

You can find more details on this at the corresponding link on the Crab FAQ page.

The CRAB configuration file (default name crab.cfg) should be located at the same location as the CMSSW parameter-set to be used by CRAB. The dataset used is: /Mu/Run2010A-Nov4ReReco_v1/RECO

In this example it is specified the user working directory name crab_lumi. Here it is an example for this tutorial:

For real data (crab_lumi.cfg)

[CMSSW]
lumis_per_job  = 50
number_of_jobs          = 10
pset                    = tutorial.py
datasetpath             = /SingleMu/Run2012B-TOPMuPlusJets-PromptSkim-v1/AOD
lumi_mask               =  Cert_190456-195947_8TeV_PromptReco_Collisions12_JSON_v2.txt
output_file              = outfile.root

[USER]
ui_working_dir         =  crab_lumi
return_data             = 0
email                   = yourEmailAddressHere@cern.ch
copy_data               = 1
storage_element         = T2_IT_Legnaro
user_remote_dir         = TutGridSchool_real

[CRAB]
scheduler               = glite
jobtype                 = cmssw
use_server              = 1

where the lumi_mask file can be downloaded with

wget --no-check-certificate https://cms-service-dqm.web.cern.ch/cms-service-dqm/CAF/certification/Collisions12/8TeV/Prompt/Cert_190456-195947_8TeV_PromptReco_Collisions12_JSON_v2.txt

For the tutorial we are using a subset of run and lumi. The lumi_mask file contains:

("194643": [[1, 287]], "194644": [[1, 168], [171, 181], [184, 185], [187, 319], [321, 424]], "194691": [[150, 155], [271, 272], [275, 289]], "194699": [[1, 30], [32, 52], [55, 64], [67, 71], [73, 154], [157, 215], [218, 238], [241, 259]], "194702": [[1, 138], [141, 191]], "194704": [[1, 41], [44, 545], [548, 592]], "194711": [[1, 7], [9, 619]], "194712": [[1, 56], [61, 418], [420, 625], [627, 759]])

Job Creation

Creating jobs for real data is analogue then to montecarlo data. To not overwrite previous run for this tutorial, it is suggested to use a dedicated cfg:

crab -create -cfg crab_lumi.cfg  
that takes as configuration file the file name specified with the option -cfg, in this case the crab_lumi.cfg associated for this tutorial with real data.

[lxplus444] $ crab -create -cfg crab_lumi.cfg
crab:  Version 2.8.1 running on Tue Jul 24 19:09:34 2012 CET (17:09:34 UTC)

crab. Working options:
        scheduler           glite
        job type            CMSSW
        server              ON (use_server)
        working directory   /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_lumi

crab:  Contacting Data Discovery Services ...
crab:  Accessing DBS at: http://cmsdbsprod.cern.ch/cms_dbs_prod_global/servlet/DBSServlet
crab:  Requested (A)DS /SingleMu/Run2012B-TOPMuPlusJets-PromptSkim-v1/AOD has 13 block(s).
crab:  Each job will process about 10 lumis.
crab:  Requested number of lumis reached.
crab:  6 jobs created to run on 100 lumis
crab:  Checking remote location
crab:  WARNING: The stageout directory already exists. Be careful not to accidentally mix outputs from different tasks
crab:  Creating 6 jobs, please wait...
crab:  Total of 6 jobs created.

Log file is /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_lumi/log/crab.log


  • The project directory called crab_0_120724_190934 is created.
  • As explained the number of created jobs can not match the number of jobs required in the configuration file (6 created but 10 required jobs).

#CMS.JobSubmission2

Job Submission

Job submission is always analogue:

[lxplus444] $ crab -submit -c crab_lumi
crab:  Version 2.8.1 running on Tue Jul 24 19:09:52 2012 CET (17:09:52 UTC)

crab. Working options:
        scheduler           glite
        job type            CMSSW
        server              ON (default)
        working directory   /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_lumi/

crab:  Registering credential to the server : vocms22.cern.ch
crab:  Your proxy lacks of retrieval and renewal policies for the requested server.
crab:  Renew your myproxy credentials.
crab:  Please renew MyProxy delegated proxy:

Your identity: /DC=ch/DC=cern/OU=Organic Units/OU=Users/CN=fanzago/CN=610896/CN=Federica Fanzago
Enter GRID pass phrase for this identity:
Creating proxy ................................... Done
Proxy Verify OK
Your proxy is valid until: Tue Jul 31 19:09:58 2012
A proxy valid for 168 hours (7.0 days) for user /DC=ch/DC=cern/OU=Organic Units/OU=Users/CN=fanzago/CN=610896/CN=Federica Fanzago now exists on myproxy.cern.ch.
crab:  Credential successfully delegated to the server.

crab:  Starting sending the project to the storage vocms22.cern.ch...
crab:  Task crab_lumi successfully submitted to server vocms22.cern.ch

crab:  Total of 6 jobs submitted
Log file is /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_lumi/log/crab.log

Job Status Check

Check the status of the jobs in the latest CRAB project with the following command:
crab -status 
to check a specific project:
crab -status -c  <dir name>

which should produce a similar screen output like:

[lxplus428] $ crab -status -c crab_lumi
crab:  Version 2.8.1 running on Wed Jul 25 00:40:52 2012 CET (22:40:52 UTC)

crab. Working options:
        scheduler           glite
        job type            CMSSW
        server              ON (default)
        working directory   /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_lumi

crab:  
ID    END STATUS            ACTION       ExeExitCode JobExitCode E_HOST
----- --- ----------------- ------------  ---------- ----------- ---------
1     Y   Done              Terminated    0          0           red-gw2.unl.edu
2     Y   Done              Terminated    0          0           red.unl.edu
3     Y   Done              Terminated    0          0           red-gw2.unl.edu
4     Y   Done              Terminated    0          0           red-gw1.unl.edu
5     Y   Done              Terminated    0          0           red.unl.edu
6     Y   Done              Terminated    0          0           red-gw2.unl.edu

crab:  ExitCodes Summary
 >>>>>>>>> 6 Jobs with Wrapper Exit Code : 0 
         List of jobs: 1-6 
        See https://twiki.cern.ch/twiki/bin/view/CMS/JobExitCodes for Exit Code meaning

crab:   6 Total Jobs 

crab:  You can also follow the status of this task on :
        CMS Dashboard: http://dashb-cms-job-task.cern.ch/taskmon.html#task=fanzago_crab_lumi_t456uk
        Server page: http://vocms22.cern.ch:8888/logginfo
        Your task name is: fanzago_crab_lumi_t456uk 

Log file is /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_lumi/log/crab.log

Also, you can have a look at the web page of the server where you can see the status progress of you job. Simply, execute the command:
crab -printId
crab -printId -c <dir name>

And you will get the unique id of your jobs:

[lxplus428] $ crab -printId -c crab_lumi/
crab:  Version 2.8.1 running on Wed Jul 25 00:43:25 2012 CET (22:43:25 UTC)

crab. Working options:
        scheduler           glite
        job type            CMSSW
        server              ON (default)
        working directory   /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_lumi/

crab:  
Task Id = fanzago_crab_lumi_t456uk     
--------------------------------------------------------------------------------------------

crab:  You can also follow the status of this task on :
        CMS Dashboard: http://dashb-cms-job-task.cern.ch/taskmon.html#task=fanzago_crab_lumi_t456uk
        Server page: http://vocms22.cern.ch:8888/logginfo
        Your task name is: fanzago_crab_lumi_t456uk 

Log file is /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_lumi/log/crab.log

Copy the unique id of your task (in the above example: fanzago_crab_lumi_t456uk), go to the link of the server being used (in this case the CERN Server, paste the unique id in the text field and press the "Show" button.

Job Output Retrieval

For the jobs which are in the "Done" status it is possible to retrieve the log files of the jobs (just the log files, because the output files are copied to the Storage Element associated to the T2 specified on the crab.cfg and infact return_data is 0). The following command retrieves the log files of all "Done" jobs of the last created CRAB project:
crab -getoutput 
to get the output of a specific project:
crab -getoutput -c  <dir name>

the job results will be copied in the res subdirectory of your crab project:

[lxplus428] $ crab -get -c crab_lumi
crab:  Version 2.8.1 running on Wed Jul 25 00:40:38 2012 CET (22:40:38 UTC)

crab. Working options:
        scheduler           glite
        job type            CMSSW
        server              ON (default)
        working directory   /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_lumi

crab:  Starting retrieving output from server vocms22.cern.ch...
crab:  Results of Jobs # 1 are in /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_lumi//res/
crab:  Results of Jobs # 2 are in /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_lumi//res/
crab:  Results of Jobs # 3 are in /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_lumi//res/
crab:  Results of Jobs # 4 are in /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_lumi//res/
crab:  Results of Jobs # 5 are in /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_lumi//res/
crab:  Results of Jobs # 6 are in /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_0_lumi//res/
Log file is /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_lumi//log/crab.log

Use the -report option

As for the MonteCarlo data example, it is possible to run the report command:

crab -report -c <dir name>

[lxplus428] $ crab -report -c crab_lumi
crab:  Version 2.8.1 running on Wed Jul 25 00:53:02 2012 CET (22:53:02 UTC)

crab. Working options:
        scheduler           glite
        job type            CMSSW
        server              ON (default)
        working directory   /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_lumi/

crab:  --------------------
Dataset: /SingleMu/Run2012B-TOPMuPlusJets-PromptSkim-v1/AOD
Remote output :
SE: T2_IT_Legnaro t2-srm-02.lnl.infn.it  srmPath: srm://t2-srm-02.lnl.infn.it:8443/srm/managerv2?SFN=/pnfs/lnl.infn.it/data/cms/store/user/fanzago/TutGridSchool_real/
Total Events read: 9178
Total Files read: 13
Total Jobs : 6
Luminosity section summary file: /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_lumi/res/lumiSummary.json
   # Jobs: Cleared:6

----------------------------

Log file is /afs/cern.ch/user/f/fanzago/scratch0/TUTORIAL/crab_lumi/log/crab.log

And the content of the file containing the luminosity summary /crab_lumi/res/lumiSummary.json:

{"194691": [[150, 155], [271, 272], [275, 289]], "194699": [[1, 30], [45, 45], [57, 64], [67, 69], [71, 71], [79, 80], [85, 85], [93, 93], [95, 103], [105, 106], [108, 108], [110, 110], [112, 116], [126, 127], [247, 253], [256, 258]]}

Run Crab retrieving your output (without copying to a Storage Element)

You can also run your analysis code without interacting with a remote Storage Element, but retrieving the outputs to your workspace area (under the res dir of the project). Here below an example of the CRAB configuration file, coerent with this tutorial:

[CMSSW]
total_number_of_events  = 100
number_of_jobs          = 10
pset                    = tutorial.py
datasetpath             =  /RelValBeamHalo/CMSSW_5_2_5_cand1-START52_V9-v1/GEN-SIM-DIGI-RAW
output_file              = outfile.root

[USER]
return_data             = 1
email                   = yourEmailAddressHere@cern.ch

[CRAB]
scheduler               = glite
jobtype                 = cmssw
use_server              = 1

And with this crab.cfg on place you can re-do de workflow as described before (a part of the publication step):

  • creation
  • submission
  • status progress monitoring
  • output retrieval (in this step you'll be able to retrieve directly the real output produced by your pset file)

Where to find more on CRAB

Note also that all CMS members using the Grid must subscribe to the Grid Annoucements CMS.HyperNews forum.

Review status

Reviewer/Editor and Date (copy from screen) Comments
NitishDhingra - 2012-04-07 See detailed comments below.
MattiaCinquilli - 2010-04-15 Update for tutorial
FedericaFanzago - 18 Feb 2009 Update for tutorial
AndriusJuodagalvis - 2009-08-21 Added an instance of url_local_dbs

Complete Review, Minor Changes. Page gives a good idea of doing a physics analysis using CRAB

Responsible: FedericaFanzago

Topic attachments
I Attachment History Action Size Date Who Comment
Cascading Style Sheet filecss tutorial.css r1 manage 0.3 K 2010-04-14 - 10:19 MattiaCinquilli  
Edit | Attach | Watch | Print version | History: r120 | r104 < r103 < r102 < r101 | Backlinks | Raw View | Raw edit | More topic actions...
Topic revision: r102 - 2012-10-16 - FedericaFanzago
 
    • Cern Search Icon Cern Search
    • TWiki Search Icon TWiki Search
    • Google Search Icon Google Search

    CMSPublic All webs login

This site is powered by the TWiki collaboration platform Powered by PerlCopyright & 2008-2022 by the contributing authors. All material on this collaboration platform is the property of the contributing authors.
or Ideas, requests, problems regarding TWiki? use Discourse or Send feedback