Introducing MINCER’s Performance Measurement and Reproducibility Appliance
Building a Standardized Framework for Reproducible Performance Measurements Across Diverse HPC Architectures
- Oct. 31, 2025 by
 - Gonzalez, Jhonny
 
A common occurrence in computer systems research is the struggle with reproducibility, which is the ability to rerun experiments while obtaining exact or similar results. Experiments involve differences in hardware, software, and configurations which make reproducibility difficult especially for complex high performance computing environments. Our project, MINCER (Monitoring Infrastructure for Network and Computing Environment Research), addresses the issue with a developing performance and reproducibility appliance on Chameleon Cloud. MINCER provides utility for monitoring, collecting, and comparing performance data across CPUs and GPUs (NVIDIA and AMD) in a consistent manner. The goal is to make it easier for researchers to replicate results and understand how system level factors influence performance. A push towards improved reproducibility and performance visibility with MINCER helps ensure scientific results are trustworthy, portable, and comparable across computing platforms.
Research Overview
MINCER automates the setup and monitoring of experiments using Docker containers that include performance measurement tools like PAPI (Performance Application Programming Interface). With PAPI, we are able to log system metadata (hardware, OS, GPUs) and performance metrics such as power usage, memory operations, and instruction rates. In the appliance, we include various example proxy applications such as XSBench, LULESH, Gunrock, and 3D-UNET to demonstrate reproducibility on real workloads. Each proxy application runs on different architectures for CPUs, AMD GPUs, and NVIDIA GPUs available on Chameleon cloud with the same scripts to follow proper comparisons. We use PAPI to collect metrics for performance, energy, etc., to help users identify bottlenecks or variations across runs. Reducing complexity across experiment setups with MINCER enables Chameleon users to collect accurate and repeatable measurements without needing to manually configure each environment. The HPC community benefits by making reproducible performance studies more accessible and standardized.
Experiment Implementation
We use Chameleon-bare-metal instances to run experiments since they allow full control of operating systems and hardware environment as it is significant for accurate performance measurements. We use the command line to download the MINCER appliance, build docker images based on the hardware being CPU, AMD, or NVIDIA, and execute our experiments with a python script. The results are automatically logged with timestamps, system metadata, and PAPI metrics. The hardware catalog is essential since it lets us determine the right nodes for our AMD and NVIDIA GPU experiments. To start, it may be best to try the CPU based tests before CPU and move to GPU from there.
MINCER relies on Chameleon's bare-metal reconfiguration to ensure consistent hardware performance across runs. The ability to customize the software environment using Docker and having OS-level access allows us to use specific versions of PAPI, CUDA, and ROCm. Without Chameleon, this experiment would be harder since Chameleon already provides a great level of control and various resources available to test and use.
Artifacts
Our files and code are available on GitHub. We are currently cleaning and updating files to have a complete version of this experiment to share. This repository will include setup scripts, performance measurement examples, and the data collection utilities for PAPI on CPU, AMD, and NVIDIA platforms. We show how to collect performance and energy metrics in the XSBench, LULESH, Gunrock, and 3D-UNET proxy apps. We are currently working on setting up an artifact to make it easier to share with other Chameleon users.
User Interview

How do you stay motivated through a long research project?
There are a few things I do to stay motivated. One is to focus on breaking the project down into small milestones. It motivates me when I know I am progressing through those tasks. I also accept that hitting a wall is a part of the process. I know that I will be reaching out to my mentor or peers with any questions I have. Letting them know about these issues helps guide me to new attempts to get through the frustration.
What has been a tough moment for you either in your life or throughout your career? How did you deal with it? How did it influence your future work direction?
The toughest moment was not academic, but instead personal. I had chosen that I would be leaving my home and family to pursue a Master's degree. I had avoided making plans that would take me far from my hometown, which is where I completed my undergraduate studies. At first it was difficult, but over time, I got used to calling my family and talking to them about how it had been and what I was up to. It helped to talk to them and I enjoyed planning what to do when I was back. This was also accompanied by exploring the new city I was in and having fun with friends. This experience has helped me accept being far from home and even look forward to the new places I will get to see.
Why did you choose this direction of research?
My research direction in performance measurement came from a great internship I attended during summer 2024. I enjoyed learning more about how metrics are gathered and the types of metrics used. I later met my current mentor, Dr. Shirley Moore who helped me use the skills I learned into my current research. Now, I get to continue learning more about performance and the tools that help record it.
Thank you for sharing your knowledge with the Chameleon community!
REPETO Releases Report on Challenges of Practical Reproducibility for Systems and HPC Computer Science
Findings from the November 2024 Community Workshop on Practical Reproducibility in HPC
- May 1, 2025 by
 - Marc Richardson
 
View or contribute to the experiment packaging and style checklists (appendix A and B) on our GitHub repository here.
Download the report here.
We’re excited to announce the publication of the NSF-sponsored REPETO Report on Challenges of Practical Reproducibility for Systems and HPC Computer Science, a culmination of our Community Workshop on Practical Reproducibility in HPC, held in November 2024 in Atlanta, GA (reproduciblehpc.org).
AutoAppendix: Towards One-Click Reproduction of Computational Artifacts
Streamlining Scientific Validation Through Automated Reproducibility Infrastructure
- Jan. 27, 2025 by
 - Klaus Kraßnitzer
 
The AutoAppendix project evaluates computational artifact reproducibility across SC24 conference submissions, revealing that most researchers struggle with creating truly replicable experiments despite their importance to scientific validity. By developing one-click reproduction templates for the Chameleon Cloud platform, this research aims to transform how computational scientists share and validate their work, potentially saving countless hours of frustration for both authors and reviewers.
Building MPI Clusters on Chameleon: A Practical Guide
Simplifying Distributed Computing Setup with Jupyter, Ansible, and Python
- Nov. 18, 2024 by
 - Michael Sherman
 
Running distributed applications across multiple nodes is a common need in scientific computing, but setting up MPI clusters can be challenging, especially in cloud environments. In this post, we'll explore a template for creating MPI clusters on Chameleon that handles the key configuration steps automatically, letting you focus on your research rather than infrastructure setup.
No comments