NIH | National Cancer Institute | NCI Wiki  

Error rendering macro 'rw-search'

null

Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

Infrastructure for Algorithm Comparisons, Benchmarks, and Challenges

...

in Medical Imaging

Author: Jayashree Kalpathy-Cramer

...


In this article, we review a few historical imaging challenges. We then list the requirements we believe to be necessary (and nice to have) to support large-scale multimodal imaging challenges. We then review existing systems and develop a matrix of features and tools. Finally, we make some recommendations for developing Medical Imaging Challenge Infrastructure (MedICI), a system to support medical imaging challenges.

Review of

...

Historical Challenges


Challenges have been popular in a number of scientific communities since the 1990s. In the text retrieval community, the Text REtrieval Conference (TREC), co-sponsored by NIST is an early example of evaluation campaigns where participants work on a common task using data provided by the organizers and evaluated with a common set of metrics. ChaLearn has organized challenges in machine learning since 2013.

...

  • A task is defined (the output). In our context, this could be segmentation of a lesion or organ, classification of an imaging study as being benign or malignant, prediction of survival, classification of a patients as being a responder or non-responder, pixel/voxel level classification of tissue or tumor grading.
  • A set of images are provided (the input). These images are chosen to be of a sufficient size and diversity to reflect the challenges of the clinical problem. Data is typically spilt up into training and test datasets. The "truth" is made available to the participants for the training data but not the test data. This reduces the risk of overfitting the data and ensures the integrity of the results.
  • An evaluation procedure is clearly defined; given the output of an algorithm on a the test images, one or more metrics are computed that measure the performance, usually a reference output is used in this process, but it could also be a visual evaluation of the results by human experts)
  • Participants apply their algorithm to all data in the public test dataset provided. They can estimate their performance on the training test.
  • Some challenges have an optional leaderboard phase where a subset of the test images is made available to the participants ahead of the final test. Participants can submit their results to the challenge system and have them evaluated or ranked but these are not considered the final standing.
  • The reference standard or "ground truth" is defined using methodology clearly described to the participants but is not made publicly available in order to ensure that algorithm results are submitted to the organizers for publication rather than retained privately.
  • Final valuation is carried out by the challenge organizers on the test set where the ground truth is sequestered from the participants.

Radiology and Pathology

...

Challenges for Brain Tumor Imaging at MICCAI 2014

MICCAI 2014 held a day-long cluster of event in brain tumor computation including challenges for brain tumor classification and segmentation. The challenge consisted of radiology as well as pathology images. A majority of the images in the training data were from TCIA. Infrastructure support for the radiology portion of the challenges was provided by Bern University and the Virtual Skeleton Database system. The PAIS system support was provided by Stony Brook University for the pathology imaging.

...


Challenge stakeholders and their tasks
Figure 4. Challenge stakeholders and their tasks

Existing

...

Challenge Infrastructure

A number of platforms exist for conducting challenges and crowdsourced efforts. Many of the popular platforms are commercial products, typically offering hosting and organizing services. Challenge organizers work with the company to set up the challenge. In some cases, the challenges are fairly trivial to set up and can be set up with the organizer without much support from the challenge platform company.

Commercial/

...

Hosted

We begin with a brief review a number of popular platforms used for challenges.

...

 KaggleSynapseHubZero (challenges/projects)COMICVISCERALCodaLab
Ease of setting up new challenge2/4 (if new metrics need to be used)22/5231

Cost (own server/hosting options)

$10-$25k/challenge
(free for class)

Free/hosted

Free/hosted

Free/hosted

Free/Azure costs

Free/hosted

License

Commercial

OS

OS

OS

OS

OS

Ease of extensibility

5

4

4

2

3

2

Cloud support for algorithms

4

3

3

4

1

3

Maturity

1

1

1/5

3

4

3

Flexibility

 

 

 

 

 

 

Number of users

1

1

1/5

3

3

3

Types of challenges

1

1

1

3

1

1

Native imaging support

No

No

No

Yes

Limited

No

API to access data, code

5

1

3

4

4

4

Components of

...

Challenge Infrastructure

We describe below the various components of challenge infrastructure that would be necessary to host joint radiology/pathology challenges.

...