Participate in AI challenges across the ACR including the inaugural Breast Density Crowdsoucing Contest. Check back in for future challenges.

ACR SIIM Pneumothorax

A Kaggle competition for the best algorithm to identify pneumothorax was announced by SIIM and ACR at the SIIM 2019 Annual Meeting on June 26 and runs until September 4, 2019. Challenge participants will compete for 10 cash prizes totaling $30,000 by developing high quality pneumothorax detection algorithms.
View more

Resident Mammography Challenge

A no coding required challenge for the best model to classify breast density images. Participants will annotate their own data, design and train their own AI models, and submit them to the leaderboard. Everyone is welcome to participate but only residents are eligible to win funding to RSNA 2019! Challenege ran Sept. 16th - Oct. 4th, 2019.View more

A crowd-sourcing contest to compare classifications of breast density among an AI algorithm, a DMIST reader, and participants at ACR 2019 starting May 19th - 21st, 2019.View more

help_outlineHelp

Welcome

The ACR Data Science Institute has developed the ACR AI-LAB™, a data science toolkit designed to democratize AI by empowering radiologists to develop algorithms at their own institutions, using their own patient data, to meet their own clinical needs.

Select a page

Define

Annotate

Create

Evaluate

Run

Publish

Assess

Challenges

Define

Welcome to the ACR AI-LAB™ Define. Browse defined AI use cases, comment on published use cases, and submit a new use case for review.

Instructions

1.Browse defined AI use cases

2.Search, filter, or sort through all of the use cases

3.Click on the use case name under the USE CASE column to comment on a published use case.

4.Click on REQUEST TO DRAFT A USE CASE under the REQUEST column to request to draft a use case that is not yet published.

5.Click on SUBMIT A NEW USE CASE and fill out the form to submit a new AI use case for the ACR to review.

Tutorial

Annotate

Welcome to the ACR AI-LAB™ Annotate. Within each use case, you can annotate multiple scans to develop a data set.

Instructions

1. Select a use case.

2. Select a data set to annotate.

3. Use the image viewing tools to examine the scan.

4. Classify each image by selecting one of the categories.

5. Click NEXT to save your annotation and view the next scan.

6. Click FINISH to finish annotating and return to the home page.

Annotate

Annotate Tutorial

My Annotations

Welcome to the ACR AI-LAB™ My Annotations. Annotate at least 100 of the 5,000 studies to use to build your own breast density classification model.

Instructions

1. Use the image viewing tools to examine the scan.

2. Classify each image by selecting one of the categories.

3. Click SAVE AND CONTINUE to save your annotation and view the next scan.

4. Click SKIP to skip a study. You can always go back to review your skipped studies by clicking on the first dropdown in the upper left corner.

5. View your annotated studies by clicking on the second dropdown in the upper left corner. Change your annotation by selecting a new classification and clicking SAVE AND CONTINUE.

6. Click FINISH to finish annotating and return to the leaderboard. You can return at any time to continue annotating.

My Annotations Tutorial

If you experience any issues during the challenge, please let us know by emailing ailabsupport@acr.org.

Run

Welcome to the ACR AI-LAB™ Run. You can run a model and see the prediction by using a prepopulated image or by uploading your own image.

Instructions

Select an AI use case.

Select a model to run.

Click MORE DETAIL to view more information about the model.

Select a prepopulated image.

Click RUN PREDICTION to view the model’s output.

Click RESET or select another prepopulated image to run another prediction.

Inference

Run Tutorial

Report Text Generation

Step 1: Define-AI generates CDEs for a specific use case.

arrow_downward

Step 2: An Assist Module is created for each Define-AI use case which encodes CDEs and associated Report Text as XML

arrow_downward

Step 3: The Assist module can be represented in any UI such as:

Simulator

Radiology Report

Viewer

arrow_downward

Step 4: When the algorithm generates output, the associated Report Text is also generated. The text can be customized to local preference (e.g., can be blank if result is negative).

Publish

Welcome to the ACR AI-LAB™ Publish. You can publish a model you designed on the Create page to a destination like an AI marketplace.

Instructions

Select an AI use case.

If there are no available models, go to the create page, create a model, and save it.

Select a destination to publish your model to.

Click PUBLISH. This will send your model to the destination.

Assess

Welcome to the ACR AI-LAB™ Assess. You can assess the performances of all deployed models.

Instructions

Select an AI use case.

To view just one model at a time, select a model from the drop down list.

Terminology

Kappa Score

Statistic that measures the agreement between two raters that classify items into mutually exclusive categories and takes into account the probability of random agreement.
If two raters are in complete agreement, then the kappa score could be close to 1. If two raters are in complete disagreement, then the kappa score could be close to 0 or even negative if they are agreeing less than the probability of random agreement.

Confusion Matrix

A table that is used to display the performance of a classification model on a test data set as compared to the ground truth. Each square has a unique row and column combination. The square in row 1 and column 2 says that the ground truth classified those test data points as category 1 but the classification model classified those same test data points as category 2. All of the squares along the diagonal of the table display the number of test cases when the ground truth and classification model are in agreement. All squares off the diagonal represent any disagreement between the ground truth and classification model; the further away from the diagonal, the larger the disagreement. If there is complete agreement between the ground truth and classification model, then all of the off diagonal squares will be zero. If there is complete disagreement between the ground truth and classification model, then all of the diagonal squares will be zero.

Binary Threshold

Change the minimum probability threshold value required to classify the study as the positive class. If the probability that a study is in the positive class is greater than or equal to the threshold value, then the study is classified as positive. A threshold of 0.75 requires that the model be 75% sure that the study is in the positive class in order for it to be classified as positive. Whereas a threshold of 0.1 requires that the model only be 10% sure that the study is in the positive class in order for it to be classified as positive. Moving the threshold slider changes the statistics calculated for the specific data element.

Confusion Matrix

A table that is used to display the performance of a classification model on a test data set as compared to the ground truth. Each square has a unique row and column combination. The square in row 1 and column 2 says that the ground truth classified those test data points as category 2 but the classification model classified those same test data points as category 1. All of the squares along the diagonal of the table display the number of test cases when the ground truth and classification model are in agreement. All squares off the diagonal represent any disagreement between the ground truth and classification model; the further away from the diagonal, the larger the disagreement. If there is complete agreement between the ground truth and classification model, then all of the off diagonal squares will be zero. If there is complete disagreement between the ground truth and classification model, then all of the diagonal squares will be zero.

Kappa Score

Statistic that measures the agreement between two raters that classify items into mutually exclusive categories and takes into account the probability of random agreement.
If two raters are in complete agreement, then the kappa score could be close to 1. If two raters are in complete disagreement, then the kappa score could be close to 0 or even negative if they are agreeing less than the probability of random agreement.

AUC Score

(Area Under the the Receiver Operating Characteristic (ROC) Curve or AUC, also known as the C-statistic) Performance measurement for classification problems that measures how well a model separates categories. An AUC score close to 1 means that the model is good at separating categories; it classifies 0s as 0s and 1s as 1s. An AUC score close to 0 means that the model is separating into opposite categories; it classifies 0s as 1s and 1s as 0s. An AUC score close to 0.5 means that the model is no better at separating categories than random chance.

Sensitivity

Statistic that measures the ability of a binary classification model to correctly identify the positive (present) cases. It is the ratio of true positives (both the radiologist and model classify the case as positive) to ground truth positives (all cases the radiologist classifies as positive). Also commonly referred to as true positive rate. A model with 100% sensitivity identifies all positive cases and has few false negatives. However, sensitivity does not factor in false positives.

Specificity

Statistic that measures the ability of a binary classification model to correctly identify the negative (absent) cases. It is the ratio of true negatives (both the radiologist and model classify the case as negative) to ground truth negatives (all cases the radiologist classifies as negative). Also commonly referred to as selectivity or true negative rate. A model with 100% specificity identifies all negative cases and has few false positives. However, specificity does not factor in false negatives.

Linnear Kappa

Statistic that measures the agreement between two raters that classify items into mutually exclusive categories and takes into account the probability of random agreement. If two raters are in complete agreement, then the kappa score could be close to 1. If two raters are in complete disagreement, then the kappa score could be close to 0 or even negative if they are agreeing less than the probability of random agreement.

Scatterplot of Volumes

A scatterplot that compares the size of the predicted volumes to the size of the ground truth volumes. If all of the points fall on the dotted line, then the size of the predicted volumes is equal to the size of the ground truth volumes. If the points fall below the dotted line, then the model is generally predicting smaller volumes than the ground truth. If the points fall above the dotted line, then the model is generally predicting larger volumes.

Bland-Altman Plot

The Bland-Altman plot shows the agreement between two quantitative measurements. The x-axis shows the average of the two measures and the y-axis shows the difference between the two measures. The mean difference is shown as a solid horizontal line and indicates the bias between the two measurements. If mean difference line is at 0, then on average, there is no bias between the measurements. Additionally, there are two more solid horizontal lines indicating the ±2 standard deviations of the mean difference (about 95% of the points should fall between the two lines). The Bland-Altman plot is helpful in discovering biases in quantitative measurements.

Mean Dice Coefficient

Statistic that measures the average similarity between two sets (segmentations or bounding boxes for example). A dice coefficient close to 1 signifies the sets are very similar; there is a large overlap between the predicted set and the ground truth set. A dice coefficient close to 0 signifies the sets are very dissimilar; there is very little overlap. The dice coefficient is positively correlated with Intersection over Union statistic. Generally, the mean dice coefficient will be higher for larger ground truth volumes (the radiologist's segmentation is large compared to the original image) than for smaller ground truth volumes (the radiologist's segmentation is small).

Mean Average Precision

Statistic that measures the Average Precision of the volumes at different Intersection over Union thresholds. Average Precision is the ratio of true positives to true positives, false positives, and false negatives. Average Precision is not concerned with true negatives. A Mean Average Precision close to 1 indicates that on average, there is a large overlap of predicted volumes and ground truth volumes. A Mean Average Precision close to 0 indicates that the predicted volumes are generally missing the ground truth volumes. Because this statistic is for volumes and not classes, another statistic (Intersection over Union) is used to calculate which volumes are labeled true positives. An IoU Threshold (T) is used as the minimum IoU required to label a volume as a true positive. For example, if the IoU Threshold is set to T=0.5, then a predicted volume that overlaps 75% with the ground truth (IoU = 0.75 which is greater than T) is considered a true positive. However, a predicted volume that overlaps 49% with the ground truth (IoU = 0.49 which is less than T) is considered a false negative. The Mean Average Precision statistic is the mean Average Precision of all images over the range of IoU Thresholds.

Mean Intersection over Union

Statistic that measures the average ratio of the intersection (volume of overlap) and the union (combined volumes) of the predicted and ground truth volumes. If the prediction matches the ground truth exactly, then the overlap and the union would be the same, resulting in an Intersection over Union (IoU)=1. If the prediction completely misses the ground truth, then the overlap is zero, resulting in an Intersection over Union (IoU)=0. The Mean IoU is the average intersection over union of all the bounding boxes.

Random Sampling

Sample the studies randomly, in the distribution they are found in the original dataset.

Positives Only

Sample only the positive cases in the training dataset.

Equal Class Ratios

Sample the studies so that each class is represented equally – 25% almost entirely fat,
25% scattered fibroglandular densities etc.

Sigmoid Focal Loss

Sigmoid Focal Loss reduces the importance of easy cases and increases the importance of the more difficult cases. This is very helpful when using imbalanced datasets - when there are a lot more of one class than the other. Sigmoid Focal Loss allows the model to focus on the more difficult cases.

Predicted Probability over Time

Predicted Probability over Time uses the Study Date field in the DICOM header to monitor the model's predictions over time. The y-axis is the probability that the case is in the positive class (present) according to the model. The horizontal line indicates the Binary Threshold; the model classifies all cases above the threshold line as positive (present) and all cases below as negative (absent). In instances where the radiologist disagrees with the model's classification, the color will indicate if it is a false positive or a false negative.

RoC Curve

Receiver operating Characteristic (RoC) Curve shows the trade-off between sensitivity and specificity for all thresholds. The blue circle will move depending on the Binary Threshold. A curve that is close to diagonal indicates that the model is no better than random chance. A curve that hugs the top left corner indicates good model performance. The area under the RoC curve is the AUC score.

Evaluate

Welcome to the ACR AI-LAB™ Evaluate. Within each use case, you can test multiple models on the same test data set and compare the performances.

Instructions

Select an AI use case.

Select a validation data set.

Click on the models you want to evaluate, you may click on more than one.

Click EVALUATE MODELS to view the test results.

Click SHOW MODELS to change your selections.

Kappa Score

Statistic that measures the agreement between two raters that classify items into mutually exclusive categories and takes into account the probability of random agreement.
If two raters are in complete agreement, then the kappa score could be close to 1. If two raters are in complete disagreement, then the kappa score could be close to 0 or even negative if they are agreeing less than the probability of random agreement.

AUC Score

(Area under the curve of the Receiver operating characteristic) Performance measurement for classification problems that measures how well a model separates categories. An AUC score close to 1 means that the model is good at separating categories; it classifies 0s as 0s and 1s as 1s. An AUC score close to 0 means that the model is separating into opposite categories; it classifies 0s as 1s and 1s as 0s. An AUC score close to 0.5 means that the model is no better at separating categories than random chance.

Confusion Matrix

A table that is used to display the performance of a classification model on a test data set as compared to the ground truth. Each square has a unique row and column combination. The square in row 1 and column 2 says that the ground truth classified those test data points as category 1 but the classification model classified those same test data points as category 2. All of the squares along the diagonal of the table display the number of test cases when the ground truth and classification model are in agreement. All squares off the diagonal represent any disagreement between the ground truth and classification model; the further away from the diagonal, the larger the disagreement. If there is complete agreement between the ground truth and classification model, then all of the off diagonal squares will be zero. If there is complete disagreement between the ground truth and classification model, then all of the diagonal squares will be zero.

Create

Welcome to the ACR AI-LAB™ Create. Within each use case, you can design your own AI models and train, test, and run them on data sets.

Instructions

Define the problem.

Select a body area.

Select a modality.

Select an AI use case.

Prepare the data.

Select the training data set.

Select the Augmentation.

Click START PREPROCESSING to prepare the data.

Configure your model.

Select the model architecture.

Select the data sampling method.

Select the model loss function.

Select if the model is pre-trained or not.

Select if the model has early stopping or not.

Click TRAIN AND TEST to train and test your model.

Click RUN PREDICTION to view the model’s prediction.

Repeat the process to create another model. You can continue to add models to your screen.

Click REMOVE to remove a model from your screen.

Terminology

Training Dataset Size

The size of the training dataset can affect the accuracy of the model.
In general, the model will be more accurate as the training dataset size increases.

Data Set

The dataset that contains the prepared images that you train the model with.

If no data sets are listed, you will need to prepare them first.

Augmentation

Apply additional transformations, such as rotations, flips, or stretching, to images input to the network.

None: Do not transform images.

Random Flips/Rotations: Add random flips and rotations to the scans before the model sees them.

Architectures

Structure of the neural network.

ResNet: Type of architecture developed in 2015 for the ImageNet Large Scale Visual Recognition
Challenge that introduced skipped connections between layers and won first place in image classification,
detection, and localization.

VGGNet: Type of architecture developed in 2014 for the ImageNet Large Scale Visual Recognition Challenge
that won first place on image localization.

InceptionNet: Type of architecture whose innovation was to attempt to process information at different spatial scales.
It also used new methods to make neural networks "deeper" (more complex) while still learning broad, generalizable
concepts about its chosen task.

DenseNet: Type of architecture whose innovation in neural networks allowed them to maintain high levels of complexity and
classification ability, while greatly reducing the computational time and memory required to classify a single case.

Architecture Videos

Sampling Methods

Technique for selecting the studies that the model sees at each training step.

Equal Class Ratios: Sample the studies so that each class is represented equally – 25% almost entirely fat,
25% scattered fibroglandular densities etc.

Random Sampling: Sample the studies randomly, in the distribution they are found in the original dataset.

{{option.text}}

Loss Functions

How the model knows what to learn

Mean Absolute Error

Treat all successes differently depending on confidence, treat all failures differently depending on confidence.

If the model got the right category and was very confident in the choice, then it was a large success. If the model got the right category and was not confident in the choice, then it was a small success. If the model got the wrong category and was not confident in the choice, then it was a small failure. If the model got the wrong category and was very confident in the choice, then it was a large failure.

Assume that the distance between the categories is the same.

Categorical Cross-Entropy

Treat all successes differently depending on confidence, treat all failures equally.

If the model got the right category and was very confident
in the choice, then it was a large success. If the model got the right category and was not confident in the choice, then it was a small success.
If the model got the wrong category, then it was a failure and all failures are treated the same.

Mean-Squared Error

Treat all successes differently depending on confidence, treat all failures differently depending on confidence.

If the model got the right category and was very confident in the choice, then it was a large success. If the model got the right category and was not confident in the choice, then it was a small success. If the model got the wrong category and was not confident in the choice, then it was a small failure. If the model got the wrong category and was very confident in the choice, then it was a large failure.

Heavily penalize large failures and assume that the distance between the categories is the same.

Loss function Videos

Pretraining

Whether or not the model has been trained on a different task before training on a new task

Random Initialization: Randomly start the model, it has not been pretrained on another task.

Pretrained Weights (ImageNet): Start with a pretrained model that was previously trained to classify natural images,
such as photos of cars, dogs, and buildings.

Batch Size

The number of training examples the model learns from during a single step. If the batch size is too small, it can lead to chaotic training because the model doesn't see enough examples at once. If the batch size is too large, the training may fail due to memory constraints of the server. If the training keeps failing, try the smallest batch size possible and increase until the memory limit is exceeded.

Initial Learning Rate

How much the model should change while learning from the training data. A learning rate that is too small can lead to long training times and could cause the model to get 'stuck'. A learning rate that is too high can lead to chaotic training because the model is changing too much.

Improvement Interval

Number of epochs the model will wait for improvement before early stopping. If early stopping is True, then the model will continue to train for a minimum improvement interval. If no improvement is made, then the model will stop training early. If there is improvement, the interval restarts. If early stopping is False, the model will not stop training early.

Early Stopping

If there is early stopping, the model will stop learning once its performance is no longer improving.

Performance Metric

The metric being used to measure the performance of the algorithm.

Mean Average Precision

AUC

Sigmoid Focal Loss

Confusion Matrix

Kappa

AUC

Sensitivity

Specificity

Kappa Score

Statistic that measures the agreement between two raters that classify items into mutually exclusive categories and takes into account the probability of random agreement.
If two raters are in complete agreement, then the kappa score could be close to 1. If two raters are in complete disagreement, then the kappa score could be close to 0 or even negative if they are agreeing less than the probability of random agreement.

AUC Score

(Area under the curve of the Receiver operating characteristic) Performance measurement for classification problems that measures how well a model separates categories. An AUC score close to 1 means that the model is good at separating categories; it classifies 0s as 0s and 1s as 1s. An AUC score close to 0 means that the model is separating into opposite categories; it classifies 0s as 1s and 1s as 0s. An AUC score close to 0.5 means that the model is no better at separating categories than random chance.

Confusion Matrix

A table that is used to display the performance of a classification model on a test data set as compared to the ground truth. Each square has a unique row and column combination. The square in row 1 and column 2 says that the ground truth classified those test data points as category 1 but the classification model classified those same test data points as category 2. All of the squares along the diagonal of the table display the number of test cases when the ground truth and classification model are in agreement. All squares off the diagonal represent any disagreement between the ground truth and classification model; the further away from the diagonal, the larger the disagreement. If there is complete agreement between the ground truth and classification model, then all of the off diagonal squares will be zero. If there is complete disagreement between the ground truth and classification model, then all of the diagonal squares will be zero.

Overfitting

Epochs

Pre-processing

Confusion Matrix Details vs. Accuracy

Kappa

Create Tutorial: Part1

Create Tutorial: Part2

Train My Models

Welcome to the ACR AI-LAB™ Train My Models. Design and train models using the data annotated on My Annotations. Submit models to the leaderboard.

Train My Models Tutorial

Instructions

Prepare the data.

Select your annotated data as the training data set. You need at least 100 annotations in order to train.

Select the Augmentation.

Configure your model.

Select the model architecture.

Select the data sampling method. You may only choose equal sampling if you have at least 10 annotations of each classification.

Select the model loss function.

Select to begin with a pre-trained model or a brand new model.

Select if the model has early stopping or not.

Choose the maximum number of epochs. If early stopping is true, then your model might not reach this maximum.

Click TRAIN AND TEST to send your model to ACR's GPUs for training and testing.

Click CANCEL to cancel a model's training. A model training can only be cancelled if it is queued or in progress.

Repeat the process to create another model.

Click SUBMIT to submit a model to the challenge leaderboard. You may submit as many models as you want before the end of the challenge but only your most recent submission will appear on the leaderboard.

Terminology

Training Dataset Size

The size of the training dataset can affect the accuracy of the model.
In general, the model will be more accurate as the training dataset size increases.

Augmentation

Apply additional transformations, such as rotations, flips, or stretching, to images input to the network.

None: Do not transform images.

Random Flips/Rotations: Add random flips and rotations to the scans before the model sees them.

Prepared Dataset

.......

Architectures

Structure of the neural network.

ResNet: Type of architecture developed in 2015 for the ImageNet Large Scale Visual Recognition
Challenge that introduced skipped connections between layers and won first place in image classification,
detection, and localization.

VGGNet: Type of architecture developed in 2014 for the ImageNet Large Scale Visual Recognition Challenge
that won first place on image localization.

InceptionNet: Type of architecture whose innovation was to attempt to process information at different spatial scales.
It also used new methods to make neural networks "deeper" (more complex) while still learning broad, generalizable
concepts about its chosen task.

DenseNet: Type of architecture whose innovation in neural networks allowed them to maintain high levels of complexity and
classification ability, while greatly reducing the computational time and memory required to classify a single case.

Architecture Videos

Sampling Methods

The ratio of each class that the model sees at each training step

Equal Class Ratios: Sample the studies so that each class is represented equally – 25% almost entirely fat,
25% scattered fibroglandular densities etc.

Random Sampling: Sample the studies randomly, in the distribution they are found in the original dataset.

Loss Functions

How the model knows what to learn

Mean Absolute Error

Treat all successes differently depending on confidence, treat all failures differently depending on confidence.

If the model got the right category and was very confident in the choice, then it was a large success. If the model got the right category and was not confident in the choice, then it was a small success. If the model got the wrong category and was not confident in the choice, then it was a small failure. If the model got the wrong category and was very confident in the choice, then it was a large failure.

Assume that the distance between the categories is the same.

Categorical Cross-Entropy

Treat all successes differently depending on confidence, treat all failures equally.

If the model got the right category and was very confident
in the choice, then it was a large success. If the model got the right category and was not confident in the choice, then it was a small success.
If the model got the wrong category, then it was a failure and all failures are treated the same.

Mean-Squared Error

Treat all successes differently depending on confidence, treat all failures differently depending on confidence.

If the model got the right category and was very confident in the choice, then it was a large success. If the model got the right category and was not confident in the choice, then it was a small success. If the model got the wrong category and was not confident in the choice, then it was a small failure. If the model got the wrong category and was very confident in the choice, then it was a large failure.

Heavily penalize large failures and assume that the distance between the categories is the same.

Loss function Videos

Pretraining

Whether or not the model has been trained on a different task before training on a new task

Random Initialization: Randomly start the model, it has not been pretrained on another task.

Pretrained Weights (ImageNet): Start with a pretrained model that was previously trained to classify natural images,
such as photos of cars, dogs, and buildings.

Early Stopping

If there is early stopping, the model will stop learning once its performance is no longer improving.

Epochs

The number of times the model learns from the training dataset.

Kappa Score

Statistic that measures the agreement between two raters that classify items into mutually exclusive categories and takes into account the probability of random agreement.
If two raters are in complete agreement, then the kappa score could be close to 1. If two raters are in complete disagreement, then the kappa score could be close to 0 or even negative if they are agreeing less than the probability of random agreement.

AUC Score

(Area under the curve of the Receiver operating characteristic) Performance measurement for classification problems that measures how well a model separates categories. An AUC score close to 1 means that the model is good at separating categories; it classifies 0s as 0s and 1s as 1s. An AUC score close to 0 means that the model is separating into opposite categories; it classifies 0s as 1s and 1s as 0s. An AUC score close to 0.5 means that the model is no better at separating categories than random chance.

Confusion Matrix

A table that is used to display the performance of a classification model on a test data set as compared to the ground truth. Each square has a unique row and column combination. The square in row 1 and column 2 says that the ground truth classified those test data points as category 1 but the classification model classified those same test data points as category 2. All of the squares along the diagonal of the table display the number of test cases when the ground truth and classification model are in agreement. All squares off the diagonal represent any disagreement between the ground truth and classification model; the further away from the diagonal, the larger the disagreement. If there is complete agreement between the ground truth and classification model, then all of the off diagonal squares will be zero. If there is complete disagreement between the ground truth and classification model, then all of the diagonal squares will be zero.

Overfitting

Epochs

Pre-processing

Confusion Matrix Details vs. Accuracy

Kappa

Confusion Matrix Tutorial

If you experience any issues during the challenge, please let us know by emailing ailabsupport@acr.org.

Crowdsourcing Contest

Welcome to the ACR AI-LAB™ BREAST DENSITY CROWDSOURCING CONTEST. Annotate as many breast density scans as you can over the course of the ACR's annual meeting.

Instructions

Use the image viewing tools to examine the scan.

Classify each image by selecting one of the categories.

Click SKIP if you do not want to annotate that specific scan.

Click NEXT to save your annotation and view the next scan.

Click FINISH to finish annotating. You may return at any time.

The number at the top right of the page denotes the total number of images you have annotated.

Overview

Participation Tutorial

Registration Tutorial

Instructions

Prepare the data.

Select the Usecase

Select the training data set.

Choose Prep for Annotation

Choose Prep for Model Creation

Click Prepare Data to process the dataset for Annotation and Model Creation

Use Cases

Select the use case for which you would like to prepare the data for.

Data Set

Select the data set that contains the images that you would like to prepate.

If no data sets are listed, you will need to create them first in Data Manager.

Prep for Annotation

Select to convert the Dicom images on the dataset into png images for annotation purposes.

Prep for Model Creation

Select to convert the Dicom images in the dataset into Numpy objects for model creation.

Crowdsourcing Contest

Welcome to the ACR AI-LAB™ BREAST DENSITY CROWDSOURCING CONTEST. Annotate as many breast density scans as you can over the course of the ACR's annual meeting.

Instructions

Use the image viewing tools to examine the scan.

Classify each image by selecting one of the categories.

Click SKIP if you do not want to annotate that specific scan.

Click NEXT to save your annotation and view the next scan.

Click FINISH to finish annotating. You may return at any time.

The number at the top right of the page denotes the total number of images you have annotated.

Overview

Participation Tutorial

Registration Tutorial

Residency Mammography Challenge

Welcome to the ACR AI-LAB™ Residency Mammography Challenge.

Instructions

Click MY ANNOTATIONS to annotate at least 100 of the 5,000 breast density studies. Your annotations will be used for the training data set. You can compare your overall annotation distribution with that of a DMIST reader in the confusion matrix on this page but will not be able to see the DMIST reader annotations on an image level.

Click TRAIN MY MODELS to design your own AI model by adjusting hyperparameters and training on your own annotated data.

Send in your design for training, also known as a job, to ACR's GPUs.

Submit a model to the leader board to compare your test results with other models. The test data set will be using DMIST reader annotations as ground truth.

You may submit as many models as you want before the end of the challenge but only your most recent submission will appear on the leaderboard. The leaderboard is ranked by kappa scores on the test results. At the end of the challenge, all submissions will be judged against a final test data set so the leaderboard is not necessarily indicative of who could win.