https://github.com/keiserlab/plaquebox-paper
Repo for Tang et al, bioRxiv 454793 (2018)
https://github.com/keiserlab/plaquebox-paper
alzheimers-disease amlyoid-plaques convolutional-neural-networks deep-learning dementia explainable-ai immunohistochemical-images interpretable-machine-learning machine-learning medical-image-analysis pathology plaques saliency-map
Last synced: 6 months ago
JSON representation
Repo for Tang et al, bioRxiv 454793 (2018)
- Host: GitHub
- URL: https://github.com/keiserlab/plaquebox-paper
- Owner: keiserlab
- License: mit
- Created: 2018-11-01T21:42:24.000Z (over 6 years ago)
- Default Branch: master
- Last Pushed: 2019-03-24T03:48:08.000Z (about 6 years ago)
- Last Synced: 2024-11-03T03:27:02.109Z (6 months ago)
- Topics: alzheimers-disease, amlyoid-plaques, convolutional-neural-networks, deep-learning, dementia, explainable-ai, immunohistochemical-images, interpretable-machine-learning, machine-learning, medical-image-analysis, pathology, plaques, saliency-map
- Language: Jupyter Notebook
- Homepage:
- Size: 31.6 MB
- Stars: 38
- Watchers: 6
- Forks: 25
- Open Issues: 2
-
Metadata Files:
- Readme: README.md
- License: LICENSE.txt
Awesome Lists containing this project
README
# Interpretable Classification of Alzheimer's Disease Pathologies with a Convolutional Neural Network Pipeline
## bioRxiv 454793
## DOI: https://doi.org/10.1101/454793
## Zenodo Data Available at: https://doi.org/10.5281/zenodo.1470797
This repository accompanies the publication above. Specifically, we include Jupyter notebooks to reproduce all image preprocessing and processing, training of convolutional neural networks, confidence visualizations, and saliency maps. The available code is provided as-is, and does not constitute a full-fledged software package for analysis.## Systems requirements
This code repository was developed on Linux CentOS 7 and Ubuntu 18 and has not been tested on on other systems (Windows, MacOS).
Code requires the following Python packages:
```
python 3.6.5 hc3d631a_2
ipython 6.4.0 py36_0
jupyter 1.0.0 py36_4
matplotlib 2.2.2 py36h0e671d2_1
numpy 1.14.3 py36hcd700cb_1
pandas 0.23.0 py36h637b7d7_0
scikit-learn 0.19.1 py36h7aa7ec6_0
scikit-image 0.13.1 py36h14c3975_1
scipy 1.1.0 py36hfc37229_0
pytorch 0.3.0 py35_cuda8.0.61_cudnn7.0.3hb362f6e_4 pytorch
torchvision 0.2.1 py36_1 pytorch
libopencv 3.4.1 h1a3b859_1
opencv 3.4.1 py36h6fd60c2_2
py-opencv 3.4.1 py36h0676e08_1
pyvips 2.1.2
tqdm 4.23.4 py36_0
```In addition, libvips (version 8.2.2-1) was used in this study for handling of whole slide images. The open-source python package pyvips is a wrapper for libvips ((https://jcupitt.github.io/libvips/), which can be installed in Linux with the following:
```
sudo apt-get install libvips
```### Hardware Requirements
Graphics Cards for Deep Learning - All deep learning models were trained using 4 X NVIDIA 1080 GPUs. As indicated above, PyTorch requires CUDA 8.0 and cuDNN 7.0 for compatibility.
## Installation guide
We recommend creating a new Anaconda (https://www.anaconda.com/) environment with the dependencies above.
This repository can be cloned directly through:
```
git clone https://github.com/keiserlab/plaquebox-paper.git
```## Demo
Notebook [2.2) CNN Models - Test Cases](https://github.com/keiserlab/plaquebox-paper/blob/master/2.2%29%20CNN%20Models%20-%20Test%20Cases.ipynb) is a demo that shows how to apply the trained CNN model on unseen dataset. Simply download the tiles from Zenodo repository and unzip it to the /data folder, then the notebook can be run through Jupyter.
## Instructions for use
This repository contains 11 notebooks to reproduce the results from the linked paper. Each notebook includes details relevant to a portion of the described pipeline, with detailed descriptions at the top of each notebook. For results reproduction, these files are presented in sequential order and depend on the previous notebook.
### Data Download
Before running the code, it is necessary to download the raw datafiles from the corresponding Zenodo repository above and unzip the files to the /data folder.
### Modifying Filepaths
The filepaths must be specified as indicated in each notebook to specify the location of the downloaded data.
### 1. Preprocessing Steps
**Notebooks 1.1-1.3** describe necessary preprocessing steps, including: color normalization, whole slide image tiling, blob detection, and dataset splitting.
### 2. Model Training and Development
**Notebooks 2.1 and 2.2** detail model development, training, and testing.
### 3. Visualizing Predictions
**Notebook 3** describes prediction confidence heatmaps.
### 4. Saliency Maps
**Notebooks 4.1 and 4.2** describe feature interpretation studies, including feature occlusion and guided-grad cam studies.
### 5. CERAD-like Scoring on Whole Slide Images
**Notebooks 5.1-5.3** describe whole slide scoring.