Various initiatives were started in order to help computational tasks required for COVID-19 research to run on the WLCG infrastructure.
are those applications which have been already tried on the WLCG infrastructure. Moreover, Fight against COVID TF at CERN indicated that
is an application WLCG might consider. WLCG Operations Coordination put together community experts to provide support and coordination for this activity.
Country/Federation |
Application_run_on the infrastructure |
Job_submission_framework |
Resource_type |
Specific_requirements |
Scale_of_the_activity |
Possible_issues |
Comments |
Contacts |
Italy T1/T2. CNAF and Tier-2s in Pisa, Rome, Bari, Legnaro, Frascati, Napoli and Milano. |
Custom application provided by an INFN spinoff which operates in the field of drugs design.It comes in the form of a compiled executable, which was expected to have a run time of ~ 11 days on 32 threads. |
Not possible to use LHC experiment frameworks since no WLCG queue allows for 32 threads and 10-15 days of execution time |
mainly CPU, though GPU could be used |
AVX2 is needed. Compiled executable needs cc7 and gcc9. Solved by using gcc9.2 from the CMS environment on CVMFS. On older nodes (sl6), we operated via a singularity container |
Topped at 30k CPU cores (so close to 1000 32 thread jobs) . Were assuming to be busy like that for ~ 2 weeks, however the jobs are already close to be done after 8-10 days. |
The variability between sites (sl6 vs cc7, condor vs LSF, shared posix disk vs dcache/dpm, ...) was the biggest obstacle |
|
Tommaso.Boccali@cernNOSPAMPLEASE.ch, luca.dellagnello@cnafNOSPAMPLEASE.infn.it |
UK |
FAH https://stats.foldingathome.org/team/246309 . Also the "Ferguson code" from IC has been installed/tested/debugged. Other code from the national RAMP initiative containerised and tested. Working with FAH to develop RUCIO usage and test data transfers |
DIRAC, direct submission and ATLAS covid work |
CPU + GPUs both local and as part of ATLAS submission |
|
5k cores (3% in terms of HS06 in May) + some fraction of ATLAS workload running FAH |
Talked with quite a number of different projects but only FAH were really ready/able to use large scale resources |
|
Alessandra.Forti@cernNOSPAMPLEASE.ch |
CERN |
F@H with a little Rosetta@Home |
None, static |
Run on old, about to be retired hardware, due to be retired soon |
|
8192 cores |
Limited to external connectivity. |
|
tim.bell@cernNOSPAMPLEASE.ch |
France Tier1 |
Application use by Laboratory for Therapeutic Innovation (LIT) which operate in the field of drug |
Submission to batch system directly |
CPU |
This community already use since many year the CC IN2P3 computing resource. But they requested a huge increase of the cpu available |
10k slots for some weeks |
Ramping up from few jobs submitted to many thousand per day ? |
10k slots has been requested to CC IN2P3 but on same time 30 k slots has been requested on another French computing center |
eric.fede@ccNOSPAMPLEASE.in2p3.fr |
France Tier1 |
Application use by Laboratory for Therapeutic Innovation (LIT) which operate in the field of drug |
Submission to batch system directly |
CPU |
This community already use since many year the CC IN2P3 computing resource. But they requested a huge increase of the cpu available |
10k slots for some weeks |
Ramping up from few jobs submitted to many thousand per day ? |
10k slots has been requested to CC IN2P3 but on same time 30 k slots has been requested on another French computing center |
eric.fede@ccNOSPAMPLEASE.in2p3.fr |
Spain |
|
|
|
|
|
|
PIC is running F@H on GPUs, installed and running. UAM T2 is running Rosetta@Home. We have asked our funding agency and it's ok to run these initiatives as background in national WLCG resources, following the WLCG MB recent request. We could accept these types of Workloads sent centrally by LHC VOs. Last Update: 8/4/2020 |
jflix@picNOSPAMPLEASE.es |
NDGF |
Custom application for population spread |
Direct to batch system |
CPU |
MPI |
1-10M corehours |
Turning proof-of-concept python into scalable C, O(n^2) scaling on population hard when doing full country population studies |
Running on the underlying HPC resources in Sweden with higher priority than other communities (including WLCG) |
maswan@ndgfNOSPAMPLEASE.org |
Canada |
TRIUMF Tier-1 is running F@H inside docker container on the compute nodes which are out of warranty |
|
CPU |
|
~500 cores and can scale up to 4500 cores if the F@H client can get work units |
Most of time F@H client could not get work unit to run |
TRIUMF deployed the FAH suite on all of the SUN blades which is equivalent to about ~500 cores. We are part of the TRIUMF_CANADA team. The goal is to ramp up slowly with a few kcores using FAH and possibly also Rosetta@Home. |
dqing@triumfNOSPAMPLEASE.ca |
Compute Canada is running FaH in VMs on the Arbutus cloud (which also hosts CA-VICTORIA-WESTGRID-T2) |
independent VMs running FAHClient, managed by terrafold (Ansible and Terraform) |
CPU and GPU |
Openstack VMs |
~ 10K vCPUs, 300 vGPUs |
|
VMs are consistently fully occupied with available work units |
rptaylor@uvicNOSPAMPLEASE.ca |
Germany Tier 1 (GridKa) |
Folding@Home and Rosetta@Home and WeNMR |
COBalD/TARDIS as job factory. Configuration available at COBalD/TARDIS Folding@Home configurations . Stats available at GridKa Grafana . Following the last Ops Coordination Meeting, we have contacted WeNMR and are now receiving jobs via DIRAC. |
CPUs and GPUs |
- |
~10000 cores |
- |
Ensuring good CPU utilization required some tuning. See configuration for details. |
Manuel.Giffels@kitNOSPAMPLEASE.edu and Andreas.Petzold@kitNOSPAMPLEASE.edu |
Netherlands LHC/Tier-1 (SURFsara and Nikhef) |
Rosetta@Home and WeNMR |
See Rosetta@Home and WeNMR for info; Rosetta@Home stats here https://boinc.bakerlab.org/rosetta/top_users.php (Team Nifhack) |
CPUs |
- |
~5000 cores |
- |
Rosetta@Home is running as burn-in test of a new cluster |
templon@nikhefNOSPAMPLEASE.nl |
TW-ASGC |
1. Taiwan T2 supports F@H by Atlas infrastructure. 2. CryoEM by local DiCOS system. |
1. PanDa. 2. DiCOS web app. DiCOS |
1. CPU for F@H 2. GPU for CryoEM |
|
1. CPU: up to 664 cores, 2. GPU: 104 |
- |
- |
felix@twgridNOSPAMPLEASE.org |