PoPETs 2026 Artifact Evaluation

PoPETs reviews and publishes digital artifacts related to its accepted papers. This process aids in the reproducibility of results and allows others to build on the work described in the papers. Artifact submissions are requested from authors of all accepted papers, and although they are optional, we strongly encourage you to submit your artifacts for review.

Possible artifacts include (but are not limited to):

Artifacts are evaluated by the artifact evaluation committee. Issues considered include software bugs, readability of the documentation, appropriate licensing, and the reproducibility of the results presented in the paper. After your artifact has been approved by the committee, the paper link on petsymposium.org will be accompanied by a link to the artifact along with the obtained artifact badges so that interested readers can find and build upon your work.

Outline

Artifact Submission Steps for Authors

  1. Please include the content of the ARTIFACT-APPENDIX.md file either within your README.md file or as a separate file. The file is important for not only reviewers during the evaluation process, but also for future researchers attempting to re-use your artifact. This file has different, marked sections for different badges.
  2. Decide which badges you want to apply for. As we describe below, in general, all submitted artifact should apply at least for “Artifact Available” badge, unless doing so would endanger someone. Authors should apply to all the badges for which they believe that their artifact meets the respective badge requirements.
  3. Ensure that you have filled out all sections of the ARTIFACT-APPENDIX.md file that are relevant for the badges that you apply for.
  4. For your submission on HotCRP, you will need to provide a copy of your paper and a direct link to the ARTIFACT-APPENDIX.md file.

Artifact Badges

Each accepted artifact can be granted up to three badges. During submission, authors must select which badges they want their artifacts to be evaluated against. As we detail below, in general, all submitted artifact should apply at least for "Artifact Available" badge, unless doing so would endanger someone. To ease reviewing effort, we encourage authors to apply appropriately to all the badges for which they believe requirements are met by their artifact. Our interpretation of the individual badges is aligned with the one of other conferences. If you have questions about which badges to apply for, first go through the FAQ below and then please contact the artifact evaluation chairs directly.

Artifact Available

For the "Artifact Available" badge, the reviewers check that the artifact is publicly accessible, has a license and is relevant to the paper.

Your artifact should be publicly accessible at a permanent location; it should not be behind any kind of paywall or restricted access or private repository. If the artifact requires you as an author to manually approve requests for access, it is not public and will not qualify for the “Artifact Available” badge. Note that all components of your artifact should be publicly available (e.g. source code, datasets, etc.).

Valid hosting options are institutional and third-party digital repositories (e.g., GitHub, Gitlab, BitBucket, Zenodo, Figshare, etc.). Do not use personal web pages or cloud storage services like Google Drive, Dropbox, etc.

All submitted artifacts should apply at least for “Artifact Available”, unless for some specific unusual situations when doing so could endanger someone. For instance, if the artifact demonstrates exploiting a vulnerability and the possible harms from releasing a proof of concept would outweigh the benefits of making it available as part of a responsible disclosure process. In this case, authors could apply for just the "Functional and Reproduced Badges". If you wish to commercialize your project, you can and should still submit your artifact for this badge, under restricted licensing, as discussed in the FAQ below.

The FAQ also provides resources on licensing, dealing with large files, and linking to multiple repositories.

Checklist for “Available Badge”:

This badge does not mean that the reviewers have checked that the code executes or that they have reproduced the results of the paper.

Artifact Functional

To be awarded the "Artifact Functional" badge the artifact should satisfy these criteria:

Documentation: The ARTIFACT-APPENDIX.md file within all source code artifacts should describe how to build and/or run the code. Reviewers will provide feedback on the clarity of the instructions and attempt to follow them and build and run the code.

Completeness: Consider the experiments in your artifact as arranged in a pipeline of multiple stages, such as data collection, data processing, and producing plots or tables for the paper. The "Completeness" criteria requires each stage to be represented. For instance, an artifact may have a proprietary machine learning model as a key component of the system, and so, achieving completeness may be difficult. If you are unable to represent any stage, then represent it in either a simplified manner or run it on dummy data, in order for reviewers to check the functionality of the stage. Provide the expected outcome of the fully run stage such that preceding stages are performed on ‘real’ data. Under the FAQ, we have examples on how authors can still prepare their artifact for the “Artifact Functional” badge in cases that involve licensing issues, time, or resource constraints.

Exercisability: All source code should be accompanied by a build environment such as a Dockerfile or a virtual machine (VM) install script, with all the dependencies (software and datasets) necessary to build the code. Include and pin the versions of your software dependencies. If the code is in a compiled language, the code should compile in the provided build environment by running the provided instructions. Compilation and setup should be automated as much as possible. Ideally, there will be one script that builds your software, runs your tests, and produces the results in a comprehensible way.

To receive this badge, artifacts are not required to be able to run on all hardware and OSes.

Checklist for “Functional Badge”:

Artifact Reproduced

The "Artifact Reproduced" badge requires the main claims of the paper to be reproduced by the reviewers. Implicitly, an artifact awarded the "Artifact Reproduced" badge needs to also meet the requirements of the "Artifact Functional" badge.

Authors must specify the commands to run the artifacts clearly and describe how to reproduce each main claim of the paper. Best practice is to point out which part of the paper is reproduced by a given script, e.g., name the table or figure. Also, the authors must highlight which results of the paper are not reproducible with the given artifact and explain why. Authors are encouraged to contemplate how their artifact could be re-used by others in the future and describe potential ways for improvement, etc.

Authors are asked to automate as much of the execution of the experiments as possible; manual effort from reviewers to run and interpret the results should be minimized. For instance, if an experiment is performing a swap across different parameters, a script automating it should be provided. Ideally, results should also be automatically parsed and output in a comprehensible way, as close to the output in the paper as possible (table or figure, etc.).

To award the “Artifact Reproduced” badge, reviewers must be able to reproduce the main claims of the paper with the provided artifact. As a rule of thumb, a quantitative claim should generally be considered reproducible if the results obtained by reviewers are within 5% of the reported value in the paper. That being said, some artifact-specific factors may prevent this; in these cases, artifact reviewers should also consider if the results that they obtain align qualitatively with the claims made in the paper. It is the reviewer’s role to enforce that these quantitative and/or qualitative expectations are met before awarding the “Artifact Reproduced” badge.

Additionally, some experiments may by nature be harder to fully reproduced during the timeframe of the artifact evaluation: e.g., take a while to run, need several iterations, train a model on a large dataset, etc. In these cases, authors should still provide the instructions and expected results for the “long” version of the experiment, and also for a simplified one (e.g., fewer iterations, smaller dataset, etc.). Indeed, even on a simplified version or fewer runs, reviewers should still somewhat be able to look at the results and the standard deviation, and check that results from the paper can be reproduced.

Finally, some artifacts, such as longitudinal studies or hardware-based contributions, may be infeasible for the “Artifact Reproduced” badge (see FAQ below), as reviewers have limited time and only commodity hardware available. Nevertheless, these authors can and should still prepare their artifacts for the “Artifact Functional” badge.

Checklist for “Reproduced Badge”:

When the artifact evaluation is over, a persistent and stable link pointing to the final evaluated version of the artifact will be collected for each accepted artifact. Depending on the hosting option picked, this is likely a link to a specific Git commit/tag or a DoI for a Zenodo record, etc. This link and the awarded badge(s) will be added to the website next to the corresponding paper title. As updates to the artifact are likely to occur to address reviewers’ feedback, we will only collect this link after a final decision is made.

Frequently Asked Questions (FAQ)

I want to commercialize my artifact. Should I still apply for any badges?

IP protections and commercialization prospects should not inhibit authors from applying for the "Artifact Available" badge. For instance, authors can choose restrictive licenses that prohibit others from using their code. Alternately, authors can design a smaller working prototype to demonstrate reproducibility of the contributions of their paper.

Which license should I choose for my artifact?

I need to upload a file that is larger than 100MB, but GitHub does not allow that. How can I make my file available?

We will need a single link to put on the PETS website and all artifacts associated with your paper should be discoverable from that one link.

I have a user study. How can I get the “Artifact Available” badge?

An example of a good user study artifact is available here. Authors of papers with user studies can generally achieve the “Artifact Available” badge, by including the following in their artifact:

I am designing a course or a game to teach privacy. How can I get the “Artifact Available” badge?

Here are examples of artifacts regarding an undergraduate course and a game to teach privacy.

For courses, thoroughly document the course structure, including the number of lessons or modules, lesson titles, number and types of assessments. You should also consider including:

For games, document the game mechanics, game materials, setup instructions as well as per-round instructions. You may include videos to supplement, but not replace, your documentation.

While in general we expect both game and course-related artifacts to be awarded the “Artifact Available” badge, the aforementioned artifact included programming exercises that could be fully exercised by the reviewers and was thus awarded the “Artifact Functional” badge.

I don’t have time to write a Dockerfile to build my project. Do you have examples? Should I upload a Dockerfile or a Docker image or both?

Authors are encouraged to check out the repository examples that have been put together by the artifact evaluation chairs.

These examples are in the form of GitHub repositories that include:

Authors can fork these repositories, and use the fork as a starting point for their artifact. For example, for Python-based projects, authors should modify the Dockerfile and add the pinned versions of their dependencies to a requirements.txt (or similar) file.

Note that these resources are not comprehensive, so authors and reviewers are not to interpret them as the only way to package an artifact; we also welcome suggestions to these resources in the form of issues, pull requests, or direct contributions.

Should I go for VMs or Docker?

For most artifacts, we have found that a Dockerfile suffices. As in our response to the question above, please check our repository examples on GitHub. If you use Docker: - Always include the Dockerfile and other scripts used to build the Docker image in your artifact. - Pin versions of dependencies as much as possible to avoid future breaking changes (e.g., specify a specific hash rather than a loose latest tag for the base image, same for dependencies versions). - Note that if you do the above points, you do not need to include both a Dockerfile and a Docker image; we strongly prefer using a Dockerfile with pinned versions. - Finally, our example GitHub repository automatically generates a Docker image from our (example) Dockerfile and publishes it as a “GitHub release”, so you do not need to worry about building and hosting the Docker image.

VMs could be a better fit for artifacts that require multiple nodes communicating with each other (you could also explore Docker Compose if that fits your use case). If you include a VM:

My artifact runs on cloud computing platforms such as AWS EC2, etc., and requires access credentials. How can I prepare my artifact for review?

Since PETS 2026, we have a dedicated submission field on HotCRP to allow authors to specify account credentials, API access keys, etc. Authors should state the amount of money/credits required to run the experiments and provide account credentials with enough credits. Our reviewers are not expected to invest their money or provide their credit cards to set up accounts.

If any form of credentials cannot be provided, provide an alternative way of running your artifact; you may communicate with the artifact chairs. If this is not possible, reconsider your choices of badges, as it may be impossible to assess your artifact for the “Artifact Functional or Reproduced” badges.

My paper includes a dataset. How should I prepare the dataset?

Please refer to the instructions under one of the previous FAQs on how to upload large files to your repository.

My paper involves a large machine learning (ML) model, or other such large files that are difficult to share. How can I get the “Artifact Functional” badge?

If a large ML model, or other file is required to execute the presented tool, the authors should include it within their artifact, unless it is proprietary. If the dataset or model is not included in the artifact, authors must share a synthetic dataset or dummy model, which may, perhaps perform worse, but which can be used by other researchers to test the principle functionality of the artifact. Authors should provide the code to train the original model, though depending on the contributions of the paper, it need not be executed.

Please refer to the instructions under one of the previous FAQs on how to upload large files to your repository.

My experiment has a lengthy runtime or requires a large amount of compute resources. How can I get the "Artifact Functional" badge?

Although experiments may require days or weeks of compute time on commodity hardware, the "Artifact Functional" badge can usually still be achieved, by following each of the steps below.

My paper involves a longitudinal study or crawl. How can I get the “Artifact Functional and Reproduced” badges?

Authors should provide:

My paper involves a hardware-based contribution. How can I prepare my artifact?

What makes a Good Submission

To ensure a smooth submission process, please follow these important guidelines:

Your cooperation in adhering to these guidelines will greatly contribute to the efficiency and effectiveness of your submission and review process. We eagerly anticipate receiving your high-quality contributions and look forward to showcasing your research!

Artifact Award

Since PETS 2022, distinguished artifacts are recognized by an artifact award. The main objective through that award is to reward authors who put a lot of effort into the release of their artifact and to showcase exemplar submissions that contribute to the open-science and reproducibility efforts of our community.

Since PETS 2026, we provide explicit criteria for reviewers to judge whether an artifact should be nominated for this award. Reviewers should consider the artifact quality, completeness, documentation, ease of reuse, artifact maturity and scope of its target audience, interactivity and responsiveness of authors, etc. Ultimately, nominations are reviewed and ranked.

Our suggestion to authors is to simply prepare and release their artifact in the same way that they wish anyone in their field would do to facilitate adoption and reproducibility by others.

What Makes a Good Review

Artifact reviewers should familiarize themselves with the artifact call, the aforementioned guidelines, and the format of the ARTIFACT-APPENDIX.md file. Reviewers should reach out to artifact chairs with any questions.

Towards the goal of contributing to open-science and reproducibility, the artifact evaluation process is designed to be interactive; authors are expected to take into account reviewers’ comments and modify their artifact accordingly. As such, reviewers are kindly asked to start their evaluation as early as possible and to post reviews or comments regularly. Once authors communicate to reviewers that issues were resolved, reviewers should then take another look and either approve the artifact or provide additional comments until a final decision is made.

We provide practical tips for reviewers below:

Distinguished Artifact Reviewers

Since PETS 2025, we recognize members of the artifact evaluation committee as distinguished artifact reviewers, based on the following criteria:

Volunteer for the Artifact Evaluation Committee

We are looking for volunteers to serve on the artifact evaluation committee. As a committee member, you will perform review of artifacts according to the guidelines above. We are looking for volunteers who will be interested in providing feedback on documentation and instructions, trying to get source code to build, or have experience with re-using published datasets. Please fill out the reviewer nomination form linked in the menu of the PETS website.