lulesh-ad
Artifact Description (AD) / Artifact Evaluation (AE)
Science Score: 49.0%
This score indicates how likely this project is to be science-related based on various indicators:
-
○CITATION.cff file
-
✓codemeta.json file
Found codemeta.json file -
✓.zenodo.json file
Found .zenodo.json file -
✓DOI references
Found 8 DOI reference(s) in README -
✓Academic publication links
Links to: arxiv.org -
○Academic email domains
-
○Institutional organization owner
-
○JOSS paper metadata
-
○Scientific vocabulary similarity
Low similarity (10.6%) to scientific vocabulary
Repository
Artifact Description (AD) / Artifact Evaluation (AE)
Basic Info
Statistics
- Stars: 0
- Watchers: 8
- Forks: 0
- Open Issues: 0
- Releases: 1
Metadata Files
README.md
Artifact Description (AD) / Artifact Evaluation (AE)
Title: Analytic Roofline Modeling and Energy Analysis of the LULESH Proxy Application on Multi-Core Clusters
Table of Contents
- A. Abstract
- B. Description
- C. Installation
- D. Experiment workflow
- E. Evaluation and expected result
- D. Experiment workflow
- F. Experiment customization
- G. Results Analysis Discussion
- H. Summary
- I. Notes
A. Abstract
B. Description
To allow a third party to duplicate the findings, this article provides reproducibility initiative dependencies (Artifact Description or Artifact Evaluation or Computational Results Analysis) appendix at https://doi.org/10.5281/zenodo.14056332. In addition to our extensive performance data artifact, it further describes details regarding the software environments, experimental design, and methodology employed for the results shown in the paper. The computational artifacts will enable experienced performance engineers to reproduce and interpret the data shown in the paper in the appropriate way and to follow the conclusions we draw from it.
B.1 Check-list (artifact meta information)
- B1.1.1 Algorithms and Programs: We employed MPI and OpenMP parallel programming models for Livermore Unstructured Lagrangian Explicit Shock Hydrodynamics (LULESH) application.
- B1.1.3 Binary: x86
B1.1.4 Hardware
- ClusterA: 36 core Intel Xeon Ice Lake (Platinum 8360Y) CPUs and HDR-100 InfiniBand
- ClusterB: 52 core Intel Xeon Sapphire (Platinum 8470) Rapids CPUs and HDR-100 InfiniBand
Further information on hardware is available in Table 1 of the paper.
B1.1.5 Run-time environment and state: A thorough state description of the two systems that were utilized to conduct the experiments can be found in
machine-states. This lists comprehensive hardware information on- libraries and compilers along with their versions
- operating system kernel, version and other details
- CPUset
- topology (cores, cache, NUMA)
- NUMA balancing
- general memory details
- transparent huge pages
- performance energy bias
- hardware power limits
B1.1.6 Output
Navigate to the output-data-perf-power-energy for additional thoroughly investigated performance, power and energy results, which are inside and outside the results published in the paper:
performance [z/s](total, scalar, vectorized)runtime [s]andspeedupFP arithmetic instructions(scalar, packed 128 B, packed 256 B, packed 512 B)verall instructions retiredmemory bandwidth [GBytes/s](total, read, write)memory data volume [GByte](total, read, write)operational intensity [F/B]cycles per instruction, CPIpower [Watt](total, chip, DRAM)energy [Joule](total, chip, DRAM)
- B1.1.7 Publicly available?
yes
B.2. How software can be obtained (if available)
To download softwares, check out the following website.
- Original proxy application: https://asc.llnl.gov/codes/proxy-apps/lulesh
- Modified code: modified-code
- Intel C++ compiler: https://software.intel.com/content/www/us/en/develop/tools/oneapi/components/dpc-compiler.html
- Intel MPI library: https://intel.com/content/www/us/en/developer/tools/oneapi/mpi-library.html
- Intel VTune: [https://www.intel.com/content/www/us/en/developer/tools/oneapi/vtune-profiler.html)
- LIKWID: https://github.com/RRZE-HPC/likwid
B.3. Hardware dependencies
Unless specified otherwise, experiments were conducted on ClusterA (Intel Xeon Ice Lake CPUs) at a base clock-frequency of 2.4 GHz (fixed, turbo disabled), and on ClusterB (Intel Xeon Sapphire Rapid CPUs) at a base clock frequency of 2.0 GHz (fixed, turbo disabled). The reproducibility of experiments requires mapping consecutive MPI processes to consecutive cores and fixing the frequency and switching-off the turbo mode. For node-level analysis, as the RAPL measurements often differ between nodes, all code versions employed the same node.
B.4. Software dependencies
- Livermore Unstructured Lagrangian Explicit Shock Hydrodynamics (LULESH), version 2.0
- Intel C++ compiler, version 2023, update 2
- Intel MPI library, version 2021, update 7
- Analyzing results metrics
- Intel VTune, version 2023, update 2
- LIKWID, version 2024, update 5.3.0 and 5.3.0/saprap1
- Monitoring results metrics: ClusterCockpit, version 2023, update 1.0.0
B.5. Datasets
Table 1 of the paper contains further information on input setup.
C. Installation
Please install the above-mentioned software dependencies.
D. Experiment workflow
To reproduce the experimental results, git clone the following repository and download modified code from the modified-code:
git clone https://github.com/RRZE-HPC/LULESH-AD && cd LULESH-AD/
To run OpenMP- and MPI-parallel LULESH code and to generate performance, power and energy results, the description for compiling and running can be found in the files available at the builds and scripts.
Outputs can be compared with results available in output-data-perf-power-energy.
E. Evaluation and expected result
See paper.
F. Experiment customization
See section 2 of the paper.
G. Results analysis discussion
See paper.
H. Summary
Please see the upshots and Section 7 of the paper that presents the summary.
I. Notes
Please cite the work as:
- A. Afzal, G. Hager, and G. Wellein: Analytic Roofline Modeling and Energy Analysis of the LULESH Proxy Application on Multi-Core Clusters. DOI:..
Bibtex:
@INPROCEEDINGS{SPEC2023,
author={Afzal, Ayesha and Hager, Georg and Wellein, Gerhard},
booktitle={arxiv},
title={Analytic Roofline Modeling and Energy Analysis of the LULESH Proxy Application on Multi-Core Clusters},
year={2024},
doi={...}}
- A. Afzal, G. Hager, and G. Wellein: Analytic Roofline Modeling and Energy Analysis of the LULESH Proxy Application on Multi-Core Clusters -- Performance Data Artifact Appendix. DOI: 10.5281/zenodo.14056332
Bibtex:
@INPROCEEDINGS{SPECAD2023,
author={Afzal, Ayesha and Hager, Georg and Wellein, Gerhard},
booktitle={[online]},
title={Analytic Roofline Modeling and Energy Analysis of the LULESH Proxy Application on Multi-Core Clusters {--} Performance Data Artifact Appendix},
year={2024},
doi={10.5281/zenodo.14056332.}}
Owner
- Name: RRZE-HPC
- Login: RRZE-HPC
- Kind: organization
- Email: hpc@fau.de
- Location: Erlangen, Germany
- Website: http://www.hpc.fau.de/
- Repositories: 50
- Profile: https://github.com/RRZE-HPC
GitHub Events
Total
- Release event: 1
- Push event: 8
- Create event: 3
Last Year
- Release event: 1
- Push event: 8
- Create event: 3
Issues and Pull Requests
Last synced: 10 months ago
All Time
- Total issues: 0
- Total pull requests: 0
- Average time to close issues: N/A
- Average time to close pull requests: N/A
- Total issue authors: 0
- Total pull request authors: 0
- Average comments per issue: 0
- Average comments per pull request: 0
- Merged pull requests: 0
- Bot issues: 0
- Bot pull requests: 0
Past Year
- Issues: 0
- Pull requests: 0
- Average time to close issues: N/A
- Average time to close pull requests: N/A
- Issue authors: 0
- Pull request authors: 0
- Average comments per issue: 0
- Average comments per pull request: 0
- Merged pull requests: 0
- Bot issues: 0
- Bot pull requests: 0