Report Open Access

HEP Application Delivery on HPC Resources

Shaffer, Tim; Blomer, Jakob; Ganis, Gerardo


DCAT Export

<?xml version='1.0' encoding='utf-8'?>
<rdf:RDF xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:adms="http://www.w3.org/ns/adms#" xmlns:cnt="http://www.w3.org/2011/content#" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:dct="http://purl.org/dc/terms/" xmlns:dctype="http://purl.org/dc/dcmitype/" xmlns:dcat="http://www.w3.org/ns/dcat#" xmlns:duv="http://www.w3.org/ns/duv#" xmlns:foaf="http://xmlns.com/foaf/0.1/" xmlns:frapo="http://purl.org/cerif/frapo/" xmlns:geo="http://www.w3.org/2003/01/geo/wgs84_pos#" xmlns:gsp="http://www.opengis.net/ont/geosparql#" xmlns:locn="http://www.w3.org/ns/locn#" xmlns:org="http://www.w3.org/ns/org#" xmlns:owl="http://www.w3.org/2002/07/owl#" xmlns:prov="http://www.w3.org/ns/prov#" xmlns:rdfs="http://www.w3.org/2000/01/rdf-schema#" xmlns:schema="http://schema.org/" xmlns:skos="http://www.w3.org/2004/02/skos/core#" xmlns:vcard="http://www.w3.org/2006/vcard/ns#" xmlns:wdrs="http://www.w3.org/2007/05/powder-s#">
  <rdf:Description rdf:about="https://doi.org/10.5281/zenodo.61157">
    <rdf:type rdf:resource="http://www.w3.org/ns/dcat#Dataset"/>
    <dct:type rdf:resource="http://purl.org/dc/dcmitype/Text"/>
    <dct:identifier rdf:datatype="http://www.w3.org/2001/XMLSchema#anyURI">https://doi.org/10.5281/zenodo.61157</dct:identifier>
    <foaf:page rdf:resource="https://doi.org/10.5281/zenodo.61157"/>
    <dct:creator>
      <rdf:Description>
        <rdf:type rdf:resource="http://xmlns.com/foaf/0.1/Agent"/>
        <foaf:name>Shaffer, Tim</foaf:name>
        <foaf:givenName>Tim</foaf:givenName>
        <foaf:familyName>Shaffer</foaf:familyName>
        <org:memberOf>
          <foaf:Organization>
            <foaf:name>CERN openlab Summer Student</foaf:name>
          </foaf:Organization>
        </org:memberOf>
      </rdf:Description>
    </dct:creator>
    <dct:creator>
      <rdf:Description>
        <rdf:type rdf:resource="http://xmlns.com/foaf/0.1/Agent"/>
        <foaf:name>Blomer, Jakob</foaf:name>
        <foaf:givenName>Jakob</foaf:givenName>
        <foaf:familyName>Blomer</foaf:familyName>
        <org:memberOf>
          <foaf:Organization>
            <foaf:name>Summer Student Supervisor</foaf:name>
          </foaf:Organization>
        </org:memberOf>
      </rdf:Description>
    </dct:creator>
    <dct:creator>
      <rdf:Description>
        <rdf:type rdf:resource="http://xmlns.com/foaf/0.1/Agent"/>
        <foaf:name>Ganis, Gerardo</foaf:name>
        <foaf:givenName>Gerardo</foaf:givenName>
        <foaf:familyName>Ganis</foaf:familyName>
        <org:memberOf>
          <foaf:Organization>
            <foaf:name>Summer Student Supervisor</foaf:name>
          </foaf:Organization>
        </org:memberOf>
      </rdf:Description>
    </dct:creator>
    <dct:title>HEP Application Delivery on HPC Resources</dct:title>
    <dct:publisher>
      <foaf:Agent>
        <foaf:name>Zenodo</foaf:name>
      </foaf:Agent>
    </dct:publisher>
    <dct:issued rdf:datatype="http://www.w3.org/2001/XMLSchema#gYear">2016</dct:issued>
    <dcat:keyword>CERN openlab summer student</dcat:keyword>
    <dct:issued rdf:datatype="http://www.w3.org/2001/XMLSchema#date">2016-08-30</dct:issued>
    <owl:sameAs rdf:resource="https://zenodo.org/record/61157"/>
    <adms:identifier>
      <adms:Identifier>
        <skos:notation rdf:datatype="http://www.w3.org/2001/XMLSchema#anyURI">https://zenodo.org/record/61157</skos:notation>
      </adms:Identifier>
    </adms:identifier>
    <dct:isPartOf rdf:resource="https://zenodo.org/communities/cernopenlab"/>
    <dct:description>&lt;p&gt;Project Specification&lt;/p&gt; &lt;p&gt;High-performance computing (HPC) contributes a significant and growing share of&amp;nbsp;resource to high-energy physics (HEP). Individual supercomputers such as&amp;nbsp;Edison or&amp;nbsp;Titan in the U.S. or SuperMUC in Europe deliver a raw performance of the same order of&amp;nbsp;magnitude than the Worldwide LHC Computing Grid. As we have seen with codes from&amp;nbsp;ALICE and ATLAS, it is notoriously difficult to deploy high-energy physics applications&amp;nbsp;on supercomputers, even though they often run a standard Linux on Intel x86_64 CPUs.&lt;/p&gt; &lt;p&gt;The three main problems are:&lt;/p&gt; &lt;p&gt;1. Limited or no Internet access;&lt;/p&gt; &lt;p&gt;2. The lack of privileged local system rights;&lt;/p&gt; &lt;p&gt;3. The concept of cluster submission or whole-node submission of jobs in contrast to&amp;nbsp;single CPU slot submission in HEP.&lt;/p&gt; &lt;p&gt;Generally, the delivery of applications to hardware resources in high-energy physics is&amp;nbsp;done by CernVM-FS [1]. CernVM-FS is optimized for high-throughput resources.&amp;nbsp;Nevertheless, some successful results on HPC resources where&amp;nbsp;achieved using the Parrot&amp;nbsp;system[2] that allows to use CernVM-FS without special privileges. Building on these&amp;nbsp;results, the project aims to prototype a toolkit for application delivery that seamlessly&amp;nbsp;integrates with HEP experiments job submission systems, for instance with ALICE AliEn&amp;nbsp;or ATLAS PanDA. The&amp;nbsp;task includes a performance study of the parrot-induced&amp;nbsp;overhead which will be used to guide performance tuning for both CernVM-FS and&amp;nbsp;Parrot on typical&amp;nbsp;supercomputers. The project should further deliver a lightweight&amp;nbsp;scheduling shim that translates HEP&amp;rsquo;s job slot allocation to a whole&amp;nbsp;node or cluster-based&amp;nbsp;allocation. Finally, in order to increase the turn-around of the evaluation of new&amp;nbsp;supercomputers, a set of &amp;quot;canary jobs&amp;quot; should be&amp;nbsp;collected that validate HEP codes on&amp;nbsp;new resources.&lt;/p&gt; &lt;p&gt;[1] http://cernvm.cern.ch/portal/filesystem&lt;/p&gt; &lt;p&gt;[2] http://cernvm.cern.ch/portal/filesystem/parrot&lt;/p&gt; &lt;p&gt;Abstract&lt;/p&gt; &lt;p&gt;On high performance computing (HPC) resources, users have less control over&amp;nbsp;worker&amp;nbsp;nodes than in the grid. Using HPC resources for high energy physics&amp;nbsp;applications&amp;nbsp;becomes more complicated because individual nodes often&amp;nbsp;don&amp;#39;t have Internet&amp;nbsp;connectivity or a filesystem configured to use as a local&amp;nbsp;cache. The current solution in&amp;nbsp;CVMFS preloads the cache from a gateway node onto the shared cluster file system.&amp;nbsp;This approach works but does not scale&amp;nbsp;well into large production environments. In this&amp;nbsp;project, we develop an in&amp;nbsp;memory cache for CVMFS, and assess approaches to running&amp;nbsp;jobs without&amp;nbsp;special privilege on the worker nodes. We propose using Parrot and CVMFS&amp;nbsp;with RAM cache as a viable approach to HEP application delivery on&amp;nbsp;HPC resources.&lt;/p&gt;</dct:description>
    <dct:accessRights rdf:resource="http://publications.europa.eu/resource/authority/access-right/PUBLIC"/>
    <dct:accessRights>
      <dct:RightsStatement rdf:about="info:eu-repo/semantics/openAccess">
        <rdfs:label>Open Access</rdfs:label>
      </dct:RightsStatement>
    </dct:accessRights>
    <dcat:distribution>
      <dcat:Distribution>
        <dct:license rdf:resource="http://creativecommons.org/licenses/by/4.0/legalcode"/>
        <dcat:accessURL rdf:resource="https://doi.org/10.5281/zenodo.61157"/>
      </dcat:Distribution>
    </dcat:distribution>
  </rdf:Description>
</rdf:RDF>
31
20
views
downloads
All versions This version
Views 3131
Downloads 2020
Data volume 9.1 MB9.1 MB
Unique views 3030
Unique downloads 1919

Share

Cite as