Question |
________________________________Response__________________________________ |
Site and Endpoints |
What is the site name? |
TRIUMF |
Which endpoint URLs do your archival systems expose? |
srm://triumf.ca/atlas/tape/ |
How is tape storage selected for a write (choice of endpoint, specification of a spacetoken, namespace prefix). |
endpoint same as above, ATLAS only |
Queue |
What limits should clients respect? |
We do accept any kind of recalls, but prefer bulk requests. We purposely delay requests to get processed in order to get bulk requests |
---> Max number of outstanding requests |
No exactly number, there was one peak number more than hundreds of k during ATLAS test, no problem for us |
---> Min/Max requests submitted at one time |
5k-30k is good, even 1k is ok, but few at a time is not welcomed |
---> Min/Max bulk request size? |
few TB - 200TB |
Should clients back off under certain circumstances? |
Ideally no, our HSM is able to handle hundreds of k requests without load problem, however there is a hard limit from disk buffer size, we don't use any extra disk buffer for tape operations, the disk buffer that tape use is also used for ATLAS((dcache hsm pools), space is limited to that, so realistically speaking, few TB-200TB a day is good enough, though can be reached to 500TB |
---> How is this signalled to client? |
through SRM |
---> For which operations? |
Depends on ATLAS how launch and check requests |
Is it advantageous to group requests by a particular criterion (e.g. tape family, date)? |
tape family grouped by datatype, dataset, and date |
---> What criterion? |
data will wait for at least 45 hours within a dataset if the dataset size not exceed a tape capacity, or will be processed when the dataset size > a single tape capacity, different datasets will be packed together by project, data type, for example: data17_900GeV, mc15_5TeV, further grouped by datatype, datatape, mctape etc.. |
Prioritisation |
Can you handle priority requests? |
quite often tape is quiet, no need yet, also no priority flat in ATLAS operations, can be implemented if particular circumstance is identified |
---> How is this requested? |
|
Protocol support |
SRM |
Are there any unsupported or partially supported operations (e.g. pinning) ? |
we can always pin or restore a file through dcache HSM interface, bulk or individual |
Timeouts |
All tape requests will be served |
What timeouts do you recommend? |
24 hours or even longer if large amounts data requested, few hours if requests are small number |
Do you have hardcoded or default timeouts? |
No |
Operations and metrics |
Can you provide total sum of data stored by VO in the archive to 100TB accuracy? |
Yes |
Can you provide space occupied on tapes by VO (includes deleted data, but not yet reclaimed space) to 100TB accuracy? |
Yes,all deleted data on tape is logcal delete |
How do you allocate free tape space to VOs? |
Through info publish |
What is the frequency with which you run repack operations to reclaim space on tapes after data deletion? |
we do repack when space is needed or media upgrade, just done LTO5-> lto7 migration this early year, 2600 LTO5 tapes, 2.5PB(on tape since 2012) migrated to LTO7 tapes at new site, no data lost |
Recommendations for clients |
Recommendation 1 |
Bulk requests by dataset, you will get your data quicker, our tape system pick the tapes has most requests first |
---> Information required by users to follow advice |
|
Recommendation 2 |
Let us know how you use tape data, then we can tweak our tape data packing policy, you will get fast readback and data safe protection, tape has mounting limits |