ICSA 2022 will be the second ICSA featuring an Artifact Evaluation Track. An artifact evaluation track aims to review and promote the research artifacts of accepted papers.
Artifacts can be software systems, scripts, or datasets. High-quality artifacts of published research papers are a foundation for the research results to be reproduced by other researchers and are thus a desirable part of the publication itself.
The artifact evaluation system is based on the upcoming NISO Recommended Practice on Reproducibility Badging and Definitions, which is supported by our publisher, IEEE, and the evaluation criteria are also inspired by ACM’s artifact review and the badging system as well as the criteria used by ICSE 2021.
Who can submit?
Authors of the papers accepted to the following ICSA 2022 tracks are invited to submit an artifact for evaluation for the Research Object Reviewed (ROR) – reusable badge and the Open Research Object (ORO) badge: Technical Track, Journal-First Track, Software Architecture in Practice Track, and New and Emerging Ideas Track. All authors of papers related to the topics mentioned in the call for papers of the ICSA technical track are invited to submit studies for the Results Reproduced (ROR-R) and Results Replicated (RER) badges.
In addition, authors of any prior software architecture research are invited to submit an artifact for evaluation as Results Reproduced (ROR-R) and Results Replicated (RER).
Please note that we require one author of each artifact submission to peer-review 2-3 other artifacts!
Artifacts of interest include (but are not limited to) the following:
- Software, which are implementations of systems or algorithms potentially useful in other studies
- Data repositories, which are data (e.g., logging data, system traces, survey raw data) that can be used for multiple software engineering approaches
- Frameworks, which are tools and services illustrating new approaches to software engineering that could be used by other researchers in different contexts
This list is not exhaustive, so the authors are asked to email the chairs before submitting if their proposed artifact is not on this list. Further information on data sharing principles and approaches are further introduced along with an introduction of the general notion of open science in the book chapter Open Science in Software Engineering by Méndez, Graziotin, Wagner, and Seibold: https://arxiv.org/abs/1904.06499.
The best artifact selected by the reviewers will be awarded the best artifact award.
For accepted ICSA 2022 papers, we plan to integrate the badge on the paper in the official IEEE proceedings.
Evaluation Criteria and Badges
Evaluation criteria for badges and additional information have been taken from ICSE 2021, which are based on the ACM policy on Artifact Review and Badging Version 1.1, and from the upcoming NISO Recommended Practice on Reproducibility Badging and Definitions, which is supported by our publisher, IEEE.
The ICSA artifact evaluation track uses a single-blind review process. The reviewers for the artifacts will be a combination of artifact authors and ICSA program committee members.
Artifacts will be evaluated using the criteria in the last row of the table below. The goal of this track is to encourage reusable research products.
In order to obtain any badge every artifact will be required to be Functional, that is, it must be properly documented, consistent, complete, exercisable, and include appropriate evidence of verification and validation. The badges to be awarded as well as the evaluation criteria are the following ones:
|RESEARCH OBJECT REVIEWED (ROR) – REUSABLE||OPEN RESEARCH OBJECT (ORO)||RESULTS REPRODUCED (ROR-R)||RESULTS REPLICATED (RER)|
|Functional + very carefully documented and well-structured to the extent that reuse and repurposing is facilitated. In particular, norms and standards of the research community for artifacts of this type are strictly adhered to.||Functional + placed on a publicly accessible archival repository. A DOI or link to this repository along with a unique identifier for the object is provided (this matches the ACM “Available” badge).||ROR + ORO + main results of the paper have been obtained in a subsequent study by a person or team other than the authors, using, in part, artifacts provided by the author.||ROR + ORO + the main results of the paper have been independently obtained in a subsequent study by a person or team other than the authors, without the use of author-supplied artifacts.|
Papers with such badges contain functional and reusable products that other researchers can use to bootstrap their own research. Experience shows that such papers earn increased citations and greater prestige in the research community.
General information and further suggested reading about artifact evaluation can be viewed in a public draft standard by Baldassarre, Ernst, Hermann, and Menzies.
Artifact Evaluation Registration Deadline:
January 11th, 2022 (mandatory) January 31th, 2022 (mandatory)
Artifact Evaluation Submissions Deadline:
January 14th, 2022 January 31th, 2022
Artifact Evaluation Notification:
January 31st, 2022 February 14th, 2022
Notes: All deadlines are 23:59h AoE (anywhere on Earth)
Submission and Review
The submission and review criteria are mainly taken from ICSE 2021, with some extensions. Different badges have different submission procedures as described below. First, these special instructions are described, then, the submission process that all authors must follow are detailed.
Submitting Research Object Reviewed (ROR) and Open Research Object (ORO) Badges
Only authors of papers accepted to the following ICSA 2022 tracks are invited to submit an artifact for evaluation for the Research Object Reviewed (ROR) – reusable badge and the Open Research Object (ORO) badge: Technical Track, Journal-First Track, Software Architecture in Practice Track, and New and Emerging Ideas Track.
Submitting Results Reproduced (ROR-R) and Results Replicated (RER) Badges
For Results Reproduced (ROR-R) and Results Replicated (RER) badges, authors will need to offer appropriate documentation in their abstract that their artifacts have reached that stage.
The abstract should include the paper title, the purpose of the research artifact, the badge(s) you are claiming, and the technology skills assumed by the reviewer evaluating the artifact. Please also mention if running your artifact requires specific Operating Systems or other environments.
TITLE: A (Partial)? (Replication|Reproduction) of XYZ. Please add the term partial to your title if only some of the original work could be replicated/reproduced.
- WHO: name the original authors (and paper) and the authors that performed the replication/reproduction. Include contact information and mark one author as the corresponding author
- WHERE: include a web link to a publically available URL directory containing (a) the original paper (that is being reproduced) and (b) any subsequent paper(s)/documents/reports that do the reproduction
- WHAT: describe the “thing” being replicated/reproduced
- WHY: clearly state why that “thing” is interesting/important
- HOW: say how it was done first
- STATE: describe the replication/reproduction. If the replication/reproduction was only partial, then explain what parts could be achieved or had to be missed
- DISCUSSION: What aspects of this “thing” made it easier/harder to replicate/reproduce. What are the lessons learned from this work that would enable more replication/reproduction in the future for other kinds of tasks or other kinds of research
- Two reviewers will review each abstract, possibly reaching out to the authors of the abstract or original paper. Abstracts will be ranked as follows
- If the reviewers do not find sufficient substantive evidence for replication/reproduction, the abstract will be rejected
- Any abstract that is judged to be unnecessarily critical of prior work will be rejected (*)
- The remaining abstracts will be sorted according to (a) how interesting they are to the community and (b) their correctness
- The top ranked abstracts will be invited to give lightning talks
(*) Our goal is to foster a positive environment that supports and rewards researchers for conducting replications and reproductions. To that end, we require that all abstracts and presentations pay due respect to the work they are reproducing/replicating. Criticism of prior work is acceptable only as part of a balanced and substantive discussion of prior accomplishments.
Submission Process for any badge
Authors must perform the following steps to submit an artifact:
- Preparing the artifact
- Making the artifact publicly available (by using repositories granting public access)
- Documenting the artifact
- Submitting the artifact
- Clarification period
1. Preparing the Artifact
There are two options depending on the nature of the artifacts: Installation Package or Simple Package. In both cases, the configuration and installation for the artifact should take less than 30 minutes. Otherwise, the artifact is unlikely to be endorsed simply because the committee will not have sufficient time to evaluate it.
Installation Package. If the artifact consists of a tool or software system, then the authors need to prepare an installation package so that the tool can be installed and run in the evaluator’s environment. Provide enough associated instruction, code, and data such that some CS person with a reasonable knowledge of scripting, build tools, etc. could install, build, and run the code. If the artifact contains or requires the use of a special tool or any other non-trivial piece of software the authors must provide a virtual machine in the Open Virtual Appliance (OVA) format (e.g., using Oracle Virtual Box or VMware), or a Docker container image with a working environment containing the artifact and all the necessary tools. Alternatively, authors can provide a link to a webservice that allows reviewers to execute the artifact (e.g., using Jupyter notebooks). Otherwise, the proposal may be desk-rejected.
Simple Package. If the artifact only contains documents that can be used with a simple text editor, a PDF viewer, or some other common tool (e.g., a spreadsheet program in its basic configuration) the authors can just save all documents in a single package file (zip or tar.gz).
2. Making the Artifact Available
The authors need to make the packaged artifact (installation package or simple package) available so that the Evaluation Committee can access it. We suggest a link to a public repository (e.g., GitHub) or to a single archive file in a widely available archive format.
If the authors are aiming for the badges Open Research Object (ORO) and beyond, the artifact needs to be publicly accessible in an archival repository that guarantees long-time storage. We suggest using Zenodo, figshare, or similar free services that provide Digital Object Identifiers (DOIs).
In other cases, the artifacts do not necessarily have to be publicly accessible for the review process. In this case, the authors are asked to provide a private link or a password-protected link. In any case, we encourage the authors to use permanent repositories dedicated to data sharing where no registration is necessary for those accessing the artifacts (e.g., please avoid using services such as GoogleDrive).
3. Documenting the Artifact
The authors need to write and submit documentation explaining how to obtain the artifact package, how to unpack the artifact, how to get started, and how to use the artifacts in more detail. The artifact submission must only describe the technicalities of the artifacts and uses of the artifact that are not already described in the paper.
The submission should contain the following documents (in plain text or pdf format) in a zip archive:
- A README main file describing what the artifact does and where it can be obtained (with hidden links and access password if necessary). Also, there should be a clear step-by-step description about how to reproduce the results presented in the paper (i.e., tables, figures, and other reported results), including links to all necessary scripts and input data. If the results presented in the paper take more than 1 hour on a standard laptop to be reproduced, authors are expected to additionally provide simplified examples with runtimes of less than one hour. If the results in the paper have been generated with an artifact using data that underlies non-disclosure agreements, authors should provide artificial replacement data that allows reviewers to assess the artifact independently of that data.
- A STATUS file stating what kind of badge(s) the authors are applying for as well as the reasons why the authors believe that the artifact deserves that badge(s).
- A LICENSE file describing the distribution rights. Note that to score Open Research Object (ORO) or higher, then that license needs to reflect some form of open source license.
- An INSTALL file with step-by-step installation instructions (if any). These instructions should include notes illustrating a very basic usage example or a method to test the installation. This could be, for instance, on what output to expect that confirms that the code is installed and working; and the code is doing something interesting and useful.
- A copy of the accepted paper in pdf format.
Note that for the badge Research Object Reviewed (ROR) – reusable, extensive documentation of the artifact is needed that allows other researchers to repurpose the tool. For software artifacts, this requires to include an overview of the artifact source code and its architecture.
4. Submitting the Artifact
By January 31, 2022, register your research artifact at the ICSA EasyChair site. Please use the submission form fields as follows:
- The title field should contain the name of the original paper to which the artifact belongs, preceded or followed by the name of the artifact if any.
- The abstract should describe research artifact so that reviewers can bid for it, including, at the end the required OS (if any) and used virtualization technique.
The remaining submission fields are explained at the submission site.
By January 31, 2022, complete your artifact submission by uploading the zip archive containing the documentation and providing the link to the software.
Before the actual evaluation, reviewers will check the integrity of the artifact and look for any possible setup problems that may prevent it from being properly evaluated (e.g., corrupted or missing files, VM won’t start, immediate crashes on the simplest example, etc.).
5. Clarification period
During the review, the reviewers may contact the authors to request clarifications on the basic installation and start-up procedures, to resolve simple installation problems or other questions. More details on this procedure can be found below in Section Communication during Clarification Period.
Please bid on artifacts during the bidding phase using Easychair. Please consider the platform and OS in the information provided by the authors when making your bids.
Please assess the artifacts based on the criteria for the badges the authors have applied for (see table above). If in doubt about how to assess a property, please use the Easychair discussion (after submitting a draft review or even an almost empty review) to contact your fellow reviewers as well as the track chairs to discuss. Additionally, if you notice any potential ethical or legal issues such as unfit licensing (too restrictive? too permissive?), institutional review of human subjects research (the IRB certificate should form part of the submission), data provenance (is the dataset biased?), and privacy protection (can individual information be obtained?), please ask the authors to clarify and additionally contact the track chairs.
If you need clarifications on the basic installation and start-up procedures, to resolve simple installation problems or other questions, please use the communication channel described below to contact the authors.
Please try to install the artifacts soon after the assignment but before February 13, so that the authors have time to respond to potential problems.
Clarification period and channel
For each submission, a Slack Channel will be set up by the artifact evaluation chairs. Reviewers will receive an invitation to the Slack ICSA Workspace and the private channel related to its assigned paper so that they can interact with authors anonymously.
Authors will be invited during artifact bidding with their e-mail address, so their names will be visible in their edits. Authors are expected to allow notifications of Slack for their artifact to get informed about reviewer questions.
Given the short review time available, the authors are expected to respond within a 48-hour period. Authors may update their research artifact after submission only for changes requested by reviewers in the clarification period.
Authors submitting an open-source repository link, are expected to give a tag to time-stamp the submissions.
Reviewers will receive a link to Slack Workspace and its assigned channel after artifact assignment. They will use such channel to interact anonymously with the authors of their assigned paper.
Reviewers are requested to consider at least one update to the artifact or to the usage instructions by the authors but are invited to consider as many updates and replies by the authors as needed to clarify an issue.
Final artifact review
Please enter your final review into Easychair and comment on whether the artifact shall receive the badges the authors have applied for. If the authors have applied for multiple badges, please comment on these individually. A possible outcome of your reviews may be that an artifact only receives a subset of the badges the authors applied for.
Artifact Evaluation Committee
- Andres Diaz Pace, UNICEN University (Argentina)
- Claudine Allen, University of The West Indies (Jamaica)
- Pablo Oliveira Antonino, Fraunhofer Institute for Experimental Software Engineering (Germany)
- Rami Bahsoon, University of Birmingham (United Kingdom)
- Luciano Baresi, Politecnico di Milano (Italy)
- Thais Batista, Federal University of Rio Grande do Norte (Brazil)
- Javier Berrocal, University of Extremadura (Spain)
- Tomas Bures, Charles University (Czech Republic)
- Javier Cámara, University of York (United Kingdom)
- Rafael Capilla, Universidad Rey Juan Carlos (Spain)
- Carlos Cuesta, Universidad Rey Juan Carlos (Spain)
- Daniel Feitosa, University of Groningen (The Netherlands)
- Andrei Furda, Queensland University of Technology (USA)
- Ilias Gerostathopoulos, TU Munich (Germany)
- Anne Koziolek, Karlsruhe Institute of Technology (Germany)
- Marcelo Maia, Universidade Federal De Uberlândia (Brazil)
- Henry Muccini, University of L’Aquila (Italy)
- Elisa Yumi, Nakagawa University of São Paulo (Brazil)
- Cesare Pautasso, USI Lugano (Switzerland)
- Jennifer Perez, Universidad Politécnica de Madrid (Spain)
- Claudia Raibulet, University of Milano-Bicocca (Italy)
- Bradley Schmerl, Carnegie Mellon University (USA)
- Mohamed Soliman, University of Groningen (Netherlands)
- Romina Spalazzese, Malmö University (Sweden)
- Bedir Tekinerdogan, Wageningen University (The Netherlands)
- Massimo Tivoli, University of L’Aquila (Italy)
- Perla Velasco-Elizondo, Autonomous University of Zacatecas (Mexico)
- Uwe Zdun, University of Vienna (Austria)