TWiki
>
LHCb Web
>
LHCbComputing
>
CCRC08
(2011-10-21,
PhilippeCharpentier
)
(raw view)
E
dit
A
ttach
P
DF
---+!! LHCb CCRC08 information %TOC% ---+++ <a name="Planned_tasks"></a> Planned tasks * Raw data distribution from pit to T0 centre * Use of rfcp into CASTOR from pit - T1D0 * Raw data distribution from T0 to T1 centres * Use of FTS - T1D0 * Recons of raw data at CERN & T1 centres * Production of rDST data - T1D0 * Use of SRM 2.2 * Stripping of data at CERN & T1 centres * Input data: RAW & rDST - T1D0 * Output data: DST - T1D1 * Use SRM 2.2 * Distribution of DST data to all other centres * Use of FTS - T0D1 (except CERN T1D1) More details given in the preGDB October meeting (see <a target="_top" href="https://twiki.cern.ch/twiki/pub/LHCb/CCRC08/LHCbVision.pdf">slides</a>) Updated (more detailed) resource requirements given <a target="_top" href="https://twiki.cern.ch/twiki/pub/LHCb/CCRC08/LHCb-CMStable.pdf">here</a> - latest info 28th October 2007. With a breakdown by site, for February, given <a target="_top" href="https://twiki.cern.ch/twiki/pub/LHCb/CCRC08/DressRehearsal071113.pdf">here</a> (latest update 13th November 2007). The above documents were updated following the January'08 face-to-face LCG CCRC08 meeting and feedback from some sites can be found <a target="_top" href="https://twiki.cern.ch/twiki/pub/LHCb/CCRC08/LHCb_CCRCJan08_1.pdf">here</a> (22nd January 2008) High level planning for the CCRC08 is given below: * status of 16th Feb is given <a target="_top" href="https://twiki.cern.ch/twiki/pub/LHCb/CCRC08/LHCbPlans_080215.htm">here</a>. * status of 25th Feb is given <a target="_top" href="https://twiki.cern.ch/twiki/pub/LHCb/CCRC08/LHCbPlans_080225.htm">here</a>. The presentation to the CCRC08 F2F meeting (in April 08) can be found <a target="_top" href="https://twiki.cern.ch/twiki/pub/LHCb/CCRC08/LHCb_CCRCApr08.pdf">here</a>. It gives the current resource estimates for the May'08 phase of CCRC08 and the services expected at the T1 sites. Since the presentation to the F2F meeting LHCb have identified space requirements under the LHCb_USER space token. The requirements for CCRC08 (and for data taking) are given <a target="_top" href="https://twiki.cern.ch/twiki/pub/LHCb/CCRC08/LHCb_User_Space.pdf">here</a>. ---+++ <a name="Daily_Meetings"></a> Daily Meetings Monday 11:00 operation meeting - 1st item should be CCRC08 activities Tuesday 9:30 PASTE meeting - 1st item should be CCRC08 activities Wednesday 11:00 CCRC08 specific meeting Thursday 11:00 operation meeting - 1st item should be CCRC08 activities Friday 11:00 CCRC08 specific meeting If you need the phone details contact Nick Brook ---+++ <a name="Logical_Namespace"></a> ---+++ <a name="Critical_service_list"></a> Critical service list The list below is the situation at the time of CCRC08.The current set of [[https://twiki.cern.ch/twiki/bin/view/LHCb/CriticalServices][critical services]] should be looked at for an up-to-date list. | *<a rel="nofollow" href="https://twiki.cern.ch/twiki/bin/view/LHCb/CCRC08?hlppage3=-1&hlppage7=0.5&hlppage4=-3&sortcol=0;table=1;up=0#sorted_table" style="color: rgb(255, 255, 255);" title="Sort by this column">Rank</a>* | *<a rel="nofollow" href="https://twiki.cern.ch/twiki/bin/view/LHCb/CCRC08?hlppage3=-1&hlppage7=0.5&hlppage4=-3&sortcol=1;table=1;up=0#sorted_table" style="color: rgb(255, 255, 255);" title="Sort by this column">Definition</a>* | *<a rel="nofollow" href="https://twiki.cern.ch/twiki/bin/view/LHCb/CCRC08?hlppage3=-1&hlppage7=0.5&hlppage4=-3&sortcol=2;table=1;up=0#sorted_table" style="color: rgb(255, 255, 255);" title="Sort by this column">Max downtime (hrs)</a>* | *<a rel="nofollow" href="https://twiki.cern.ch/twiki/bin/view/LHCb/CCRC08?hlppage3=-1&hlppage7=0.5&hlppage4=-3&sortcol=3;table=1;up=0#sorted_table" style="color: rgb(255, 255, 255);" title="Sort by this column">Comment</a>* | | 10 | Critical | 0.5 | | | 7 | Serious disruption | 8 | | | 5 | Major reduction in effectiveness | 8 | | | 3 | Reduced effectiveness | 24 | | | 1 | Not critical | 72 | | | *<a rel="nofollow" href="https://twiki.cern.ch/twiki/bin/view/LHCb/CCRC08?hlppage3=-1&hlppage7=0.5&hlppage4=-3&sortcol=0;table=2;up=0#sorted_table" style="color: rgb(255, 255, 255);" title="Sort by this column">Service</a>* | *<a rel="nofollow" href="https://twiki.cern.ch/twiki/bin/view/LHCb/CCRC08?hlppage3=-1&hlppage7=0.5&hlppage4=-3&sortcol=1;table=2;up=0#sorted_table" style="color: rgb(255, 255, 255);" title="Sort by this column">Rank</a>* | *<a rel="nofollow" href="https://twiki.cern.ch/twiki/bin/view/LHCb/CCRC08?hlppage3=-1&hlppage7=0.5&hlppage4=-3&sortcol=2;table=2;up=0#sorted_table" style="color: rgb(255, 255, 255);" title="Sort by this column">Comment</a>* | | CERN VO boxes | 10 | | | CERN LFC service (Master) | 10 | | | T0 SE | 10 | | | CERN network Campus | 10 | | | CERN AFS | 10 | | | VOMS proxy service | 7 | | | T1 SE | 7 | | | T0/T1 FTS | 7 | both CERN to/from T1 & inter T1 | | WN misconfig | 7 | | | T0/T1 CE | 7 | | | Conditions DB | 7 | | | LHCb Bookkeeping service | 7 | | | Oracle streaming from CERN | 7 | | | SAMDB service | 7 | We should rely on this to OK a site ? | | LHCb gLite WMS | 5 | at CERN (at T1 3) | | T1 LFC service (mirrors) | 3 | | | Dashboard | 3 | | | T1 VOboxes | 3 | | the table above represents a first sketch presented beginning 2008. A detailed document describing all LHCb critical services, metrics,monitoring tests and criticality is available <a target="_top" href="https://twiki.cern.ch/twiki/pub/LHCb/CCRC08/LHCb_critical_Services2.pdf">here</a><br /> ---+++ <a name="Planning_Document"></a> Planning Document The latest version (3rd April 2008) of LHCb milestone document can be found <a target="_top" href="https://twiki.cern.ch/twiki/pub/LHCb/CCRC08/LHCb_CCRC08_Plans_080403.pdf">here</a> ---+++ <a name="Storage_requirements_and_space_m"></a> Storage requirements and space management Each site should mind to provide a configuration like that <a target="_top" href="https://twiki.cern.ch/twiki/bin/view/LCG/GSSDLHCBCCRC08">LHCb Space Management</a> and the middleware should be able to cope with that requirements. An updated document describing the space token connectivity can be found here [[%ATTACHURL%/LHCbSpaces.xls][LHCbSpaces.xls]]. A detailed description of Storage Class requirements is available <a target="_self" href="https://twiki.cern.ch/twiki/bin/view/LCG/GSSDLHCB">here.</a> <b><i><a name="Logical_Namespace"></a></i></b> __Logical Namespace__ <br />Information on the LHCb namespace can be found [[NameSpace][here]]. Only the real data configuration will be used in February. For CCRC, the proposed name for <year> is CCRC08.<br />In case the SAPATH terminates with "/lhcb", the "/lhcb" should not be repeated, otherwise it will. If site need different SAPATH for different spaces, we don't care so much (this concerns what is _before_ the LFN in the PFN). Note however there will be problems with space migration, but unimportant for CCRC'08. ---+++ <a name="Critical_service_list"></a> <b><i><a name="Storage Requirements"></a></i></b> __Storage Requirements__ <br /> Information on the LHCb space requirements can be found at the following document [[%ATTACHURL%/PostCCRC_storage.pdf][PostCCRC_storage.pdf]] <b><i><a name="Space tokens deployment and monitoring "></a></i></b> __Space tokens deployment and monitoring__ <br /> Information on the LHCb [[https://twiki.cern.ch/twiki/bin/view/LHCb/SpaceTokensDeployment][Space tokens deployment status]] at Tier1 sites <a name="Information from sites"></a> Collection from the sites of their hardware description [[https://twiki.cern.ch/twiki/bin/view/LHCb/T1StorageSetup][T1 Storage System setup]] :updated at September the 4th ---+++ E-logbook Issues experienced during CCRC are available at the following <a target="_top" href="http://lblogbook.cern.ch/CCRC08/">e-log</a> ---+++ <a name="What_To_Do_When_WTDW"></a><a name="What_To_Do_When_WTDW_"></a> What To Do When... (WTDW) This [[LHCbWtdw][Twiki page]] should give you some guidance on trying to solve some of the problems observed during CCRC08. ---+++ <a name="February_Phase_summary"></a> February Phase summary *15-17th February* Exercise of pit-T0-T1 machinery at 1 file/minute. An issue with several data management agents polling volhcb03 with the consequence a re-boot was necessary. In general everything seem to proceed successfully *19th February* Small scale production with low number of events run successfully. Some of the jobs observed to have gsidcap issues with lost connection at IN2P3. There were issues around a corrupted software area at GridKa. *20th February* Transfer recommenced at a rate 1 file per minute. It was observed that that file removal from CERN using srm only removed from namespace and not the files from the cache leading to the trnasfer issue to CERN. Ths was a consequence of the CASTOR software deployed for LHCb. Autojob submission was started. Transfer (pit-T0-T1)running constantly *21st February* Issues with SRM reporting space full at IN2P3. This was associated of cleaning SE with SRM not reclaiming space. There were issues surrounding the use of role=pilot that meant jobs ran as sgm in test production. Role=production will be used for CCRC08. ANother test production submitted. *22nd February* Issues were observed with upload/creating directories at IN2P3 and RAL (associated with dirac_directory creation.) Transfer now automatically running 6 hours on 6 hours off to mimic LHC operations. *23rd February* Re-occurrence of the IN2P3 SRM space problem. Transfer to CNAF failed - associated with not enough LSF slots on diskservers( increased to 200 per server) Analysis of weekend jobs: _General_ Brunel application problem associated with TransportSvc. Proxy retrieval issues associated with relying on the voms server to extend proxy lifetime. Incorrect error flag on reaching EOF for the MDF file. _CERN_ There was a temporary software area hitch. A failure to copy and register output files; this was associated with files greater than ~2 GB and the subsequet checking of file size after the data was uploaded. _CNAF_ A failure to copy and register output files; the same issue as associated with CERN. Jobs hug waiting for data from CASTOR this is associated with the LSF job slot problems observed in the transfers. _RAL_ A failure to copy and register output files; the same issue as associated with CERN and CNAF. There were timeouts associated with the bookkeeping. _IN2P3_ There was a gPlazma authorisation issue that occurred rarely. Major issue with gsidcap doors failure; this is due for a 2 hour timeout for any connection being set by the site. There were timeouts associated with the bookkeeping. _NIKHEF_ As the number of jobs increase at NIKHEF all jobs failed trying to access through dCache due to failing to create a control line. This seemed to be a load issue. NIKHEF will move the gsidcap server from the srm server - not scheduled yet. Issue still open. There were timeouts associated with the bookkeeping. _PIC_ All seemed fine. _GridKa_ Some stalled jobs associated with dCache file access problems. *27th February* Transfer issues to NIKHEF; problems with one of the pools. *29th February* The bookkeeping problems observed at IN2P3, NIKHEF and RAL were solved after the appropriate firewall ports were opened. <br /> * [[%ATTACHURL%/PostCCRC_storage.pdf][PostCCRC_storage.pdf]]: Storage requirements per ST * [[%ATTACHURL%/LHCb_critical_Services2.pdf][LHCb_critical_Services2.pdf]]: LHCb_critical_Services2.pdf
Attachments
Attachments
Topic attachments
I
Attachment
History
Action
Size
Date
Who
Comment
pdf
LHCb_critical_Services2.pdf
r3
r2
r1
manage
98.4 K
2009-05-19 - 15:36
RobertoSantinel
E
dit
|
A
ttach
|
Watch
|
P
rint version
|
H
istory
: r29
<
r28
<
r27
<
r26
<
r25
|
B
acklinks
|
V
iew topic
|
WYSIWYG
|
M
ore topic actions
Topic revision: r29 - 2011-10-21
-
PhilippeCharpentier
Log In
LHCb
LHCb Web
LHCb Web Home
Changes
Index
Search
LHCb webs
LHCbComputing
LHCb FAQs
LHCbOnline
LHCbPhysics
LHCbVELO
LHCbST
LHCbOT
LHCbPlume
LHCbRICH
LHCbMuon
LHCbTrigger
LHCbDetectorAlignment
LHCbTechnicalCoordination
LHCbUpgrade
Public webs
Public webs
ABATBEA
ACPP
ADCgroup
AEGIS
AfricaMap
AgileInfrastructure
ALICE
AliceEbyE
AliceSPD
AliceSSD
AliceTOF
AliFemto
ALPHA
Altair
ArdaGrid
ASACUSA
AthenaFCalTBAna
Atlas
AtlasLBNL
AXIALPET
CAE
CALICE
CDS
CENF
CERNSearch
CLIC
Cloud
CloudServices
CMS
Controls
CTA
CvmFS
DB
DefaultWeb
DESgroup
DPHEP
DM-LHC
DSSGroup
EGEE
EgeePtf
ELFms
EMI
ETICS
FIOgroup
FlukaTeam
Frontier
Gaudi
GeneratorServices
GuidesInfo
HardwareLabs
HCC
HEPIX
ILCBDSColl
ILCTPC
IMWG
Inspire
IPv6
IT
ItCommTeam
ITCoord
ITdeptTechForum
ITDRP
ITGT
ITSDC
LAr
LCG
LCGAAWorkbook
Leade
LHCAccess
LHCAtHome
LHCb
LHCgas
LHCONE
LHCOPN
LinuxSupport
Main
Medipix
Messaging
MPGD
NA49
NA61
NA62
NTOF
Openlab
PDBService
Persistency
PESgroup
Plugins
PSAccess
PSBUpgrade
R2Eproject
RCTF
RD42
RFCond12
RFLowLevel
ROXIE
Sandbox
SocialActivities
SPI
SRMDev
SSM
Student
SuperComputing
Support
SwfCatalogue
TMVA
TOTEM
TWiki
UNOSAT
Virtualization
VOBox
WITCH
XTCA
Welcome Guest
Login
or
Register
Cern Search
TWiki Search
Google Search
LHCb
All webs
Copyright &© 2008-2023 by the contributing authors. All material on this collaboration platform is the property of the contributing authors.
or Ideas, requests, problems regarding TWiki? use
Discourse
or
Send feedback