Registration

To register with the challenge, open a pull request (PR) to the SuperElastix GitHub repository with your method (how to make a PR). You have officially signed up for the challenge when your first PR is merged into SuperElastix. Both individual and team participations are welcome.

You are free to withdraw from the challenge at any point, but note that your code will be part of the git history of SuperElastix even if you submit a PR that deletes your code.

Participation

To participate in the challenge you must:

  • Implement your method in SuperElastix OR submit a parameter file that uses an existing component (guidelines). Only code in the official SuperElastix repository on Github will be 1) considered for participation and 2) benchmarked and published on the leaderboards.
  • Describe your method in the README.md file in your submission directory in the GitHub repository.
  • Use open source software. You are free to include any library in your method as long as it is compatible with the Apache 2.0 License and can be built as part of the SuperElastix build system.

In addition:

  • You have a maximum of one hour of wall clock time for registration of a pair of images.
  • External data (e.g. a pretrained TensorFlow model) is allowed and should be downloaded at build time via the module‚Äôs CMake script.
  • Participants can submit multiple parameter files (within reasonable limits).
  • Leaderboards will be updated continuously.
  • All participants will be invited to collaborate on a journal paper summarizing the challenge outcomes at the time of paper submission. We aim to submit the results to one of the leading journals in the field.

Rights

The challenge organization does not claim any ownership or any rights to the developed works, but the code will have to be published under the Apache 2.0 License. This is because the code needs to be in SuperElastix to be benchmarked, and SuperElastix itself is Apache 2.0 Licensed.

Data

The submissions will be tested on 4 lung datasets and 4 brain data sets. All data sets are open access except for the SPREAD lung data set which will be kept private. We do not provide direct download links, but participants are encouraged to download the data from the original source (see the data page or click 'instructions' below for details). The data sets are:

Name

Type

Access

Type

Ground truth

POPI (download, instructions)

4DCT Lung

Public

6 EF intra-subject, same timepoint

Points

DIR-LAB (download, instructions)

4DCT Lung

Public

10 EF intra-subject, same timepoint

Points

SPREAD (hidden data)

CT Lung

Private

21 SF intra-subject, diff timepoint

Points

EMPIRE (download, instructions)

CT Lung

Public

30 intra-subject, same timepoint

Points

LPBA40 (download, instructions)

MR Brain

Public

40 inter-subject

Segmentation

ISBR18 (download, instructions)

MR Brain

Public

18 inter-subject

Segmentation

CUMC12 (download, instructions)

MR Brain

Public

12 inter-subject

Segmentation

MGH10 (download, instructions)

MR Brain

Public

10 inter-subject

Segmentation

BRAINS (hidden data for WBIR2018 evaluation)

MR Brain

Public

20 inter-subject

Segmentation

Evaluation

Evaluations are pooled over all datasets within a category. For example, a submission to the lung category will be evaluated on all lung data sets, and the ranking will subsequently be calculated based on mean performance on all images in all data sets. 

For lung data, we evaluate the following metrics:

  • Target Registration Error (TRE)
  • Mean Hausdorff distance
  • Ratio of singularities in the deformation field to the number of voxels
  • Inverse consistency error

All lung data sets use points for ground truth.

For brain data, we evaluate the following metrics:

  • Mean DICE score
  • Mean Hausdorff distance
  • Ratio of singularities in the deformation field to the number of voxels
  • Inverse consistency error

All brain data sets use manual segmentations for ground truth.

Within each category, we compute the rankings for individual metrics. We do not compute a pooled ranking. All results are displayed on the leaderboard. The leaderboard is updated weekly.