Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/drorata/rossmann_store_sales
Playing around with the Rossmann data set from kaggle
https://github.com/drorata/rossmann_store_sales
Last synced: 5 days ago
JSON representation
Playing around with the Rossmann data set from kaggle
- Host: GitHub
- URL: https://github.com/drorata/rossmann_store_sales
- Owner: drorata
- License: mit
- Created: 2018-04-16T10:16:10.000Z (almost 7 years ago)
- Default Branch: master
- Last Pushed: 2018-04-16T19:04:35.000Z (almost 7 years ago)
- Last Synced: 2024-11-15T11:42:41.493Z (2 months ago)
- Language: Jupyter Notebook
- Size: 7.62 MB
- Stars: 0
- Watchers: 1
- Forks: 0
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
README
# Rossmann store sales - kaggle
Some playing around with kaggle competition dealing with predictions on the sales of [Rossmann shops](https://www.kaggle.com/c/rossmann-store-sales/data).
This project is derived from [ds-cookiecutter](https://github.com/drorata/ds-cookiecutter).## Running docker container
You can experiment with this analysis using the provided docker image.
This can be done in two steps, first build the image:```bash
docker build -t rossmann .
```You can replace the name of the image as per your convenience.
This step can take couple of minutes, depending on the machine you're using.
Next, you can run the built image:```bash
docker run -p 8899:8888 -it --rm rossmann
```and visit on `localhost:8899`.
The port `8899` can be replaced upon running the image.## Project Organization
├── LICENSE
├── README.md <- The top-level README (this file)
├── data
│ ├── external <- Data from third party sources.
│ ├── interim <- Intermediate data that has been transformed.
│ ├── processed <- The final, canonical data sets for modeling.
│ └── raw <- The original, immutable data dump.
│
├── notebooks <- Main analysis is happening in the notebook available here
│
├── environment.yml <- Definition of the computation environment when using `conda`
│
└── src <- Note used. Additional code could come in here.
├── __init__.py <- Makes src a Python module
│
├── data <- Scripts to download or generate data
│ └── make_dataset.py
│
├── features <- Scripts to turn raw data into features for modeling
│ └── build_features.py
│
├── models <- Scripts to train models and then use trained models to make
│ │ predictions
│ ├── predict_model.py
│ └── train_model.py
│
└── visualization <- Scripts to create exploratory and results oriented visualizations
└── visualize.py