TWiki> HEPTape Web>Survey>Triumf_Survey (revision 3)EditAttachPDF


Question ________________________________Response__________________________________
Site and Endpoints
What is the site name? TRIUMF
Which endpoint URLs do your archival systems expose? srm://
How is tape storage selected for a write (choice of endpoint, specification of a spacetoken, namespace prefix). endpoint same as above, ATLAS only
What limits should clients respect? We do accept any kind of recalls, but prefer bulk requests. We purposely delay requests to get processed in order to get bulk requests
---> Max number of outstanding requests No exactly number, there was one peak number more than hundreds of k during ATLAS test, no problem for us
---> Min/Max requests submitted at one time 5k-30k is good, even 1k is ok, but few at a time is not welcomed
---> Min/Max bulk request size? few TB - 200TB
Should clients back off under certain circumstances? Ideally no, our HSM is able to handle hundreds of k requests without load problem, however there is a hard limit from disk buffer size, we don't use any extra disk buffer for tape operations, the disk buffer that tape use is also used for ATLAS((dcache hsm pools), space is limited to that, so realistically speaking, few TB-200TB a day is good enough, though can be reached to 500TB
---> How is this signalled to client? through SRM
---> For which operations? Depends on ATLAS how launch and check requests
Is it advantageous to group requests by a particular criterion (e.g. tape family, date)? tape family grouped by datatype, dataset, and date
---> What criterion? data will wait for at least 45 hours within a dataset if the dataset size not exceed a tape capacity, or will be processed when the dataset size > a single tape capacity, different datasets will be packed together by project, data type, for example: data17_900GeV, mc15_5TeV, further grouped by datatype, datatape, mctape etc..
Can you handle priority requests? quite often tape is quiet, no need yet, also no priority flat in ATLAS operations, can be implemented if particular circumstance is identified
---> How is this requested?  
Protocol support
Are there any unsupported or partially supported operations (e.g. pinning) ? we can always pin or restore a file through dcache HSM interface, bulk or individual
What timeouts do you recommend? All tape requests will be served . 24 hours or even longer if large amounts data requested, few hours if requests are small number
Do you have hardcoded or default timeouts? No
Operations and metrics
Can you provide total sum of data stored by VO in the archive to 100TB accuracy? Yes
Can you provide space occupied on tapes by VO (includes deleted data, but not yet reclaimed space) to 100TB accuracy? Yes,all deleted data on tape is logcal delete
How do you allocate free tape space to VOs? Through info publish
What is the frequency with which you run repack operations to reclaim space on tapes after data deletion? we do repack when space is needed or media upgrade, just done LTO5-> lto7 migration this early year, 2600 LTO5 tapes, 2.5PB(on tape since 2012) migrated to LTO7 tapes at new site, no data lost
Recommendations for clients
Recommendation 1 Bulk requests by dataset, you will get your data quicker, our tape system pick the tapes has most requests first
---> Information required by users to follow advice  
Recommendation 2 Let us know how you use tape data, then we can tweak our tape data packing policy, you will get fast readback and data safe protection, tape has mounting limits
-- OliverKeeble - 2018-01-30
Edit | Attach | Watch | Print version | History: r4 < r3 < r2 < r1 | Backlinks | Raw View | Raw edit | More topic actions...
Topic revision: r3 - 2018-04-13 - OliverKeeble
    • Cern Search Icon Cern Search
    • TWiki Search Icon TWiki Search
    • Google Search Icon Google Search

    HEPTape All webs login

This site is powered by the TWiki collaboration platform Powered by PerlCopyright & 2008-2021 by the contributing authors. All material on this collaboration platform is the property of the contributing authors.
or Ideas, requests, problems regarding TWiki? use Discourse or Send feedback