Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/bryanyang0528/docker-cdh-spark
cdh with spark 2.2
https://github.com/bryanyang0528/docker-cdh-spark
cdh cloudera docker spark
Last synced: 28 days ago
JSON representation
cdh with spark 2.2
- Host: GitHub
- URL: https://github.com/bryanyang0528/docker-cdh-spark
- Owner: bryanyang0528
- License: apache-2.0
- Created: 2018-01-04T14:14:52.000Z (about 7 years ago)
- Default Branch: master
- Last Pushed: 2018-07-09T08:41:48.000Z (over 6 years ago)
- Last Synced: 2024-11-13T13:18:20.350Z (3 months ago)
- Topics: cdh, cloudera, docker, spark
- Language: Shell
- Homepage:
- Size: 14.6 KB
- Stars: 3
- Watchers: 2
- Forks: 0
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
README
# docker-cdh-spark
The version of Spark in the original Cloudera image is still 1.6.1.
This image upgrade it to 2.2 by hacking methods, so I CAN NOT promise 100% compatible with the original CDH.
## Prerequisite
### install docker
OSX: [https://docs.docker.com/docker-for-mac/install/](https://docs.docker.com/docker-for-mac/install/)
Others: [https://docs.docker.com/engine/installation/](https://docs.docker.com/engine/installation/)
### install docker-compose
[https://docs.docker.com/compose/install/#install-compose](https://docs.docker.com/compose/install/#install-compose)
## Usage
### clone this repo
`git clone https://github.com/bryanyang0528/docker-cdh-spark`
`cd docker-cdh-spark`
### Option 1: docker
`docker build -t docker-cdh-spark:latest .`
`docker run -p 8080:8080 --hostname=quickstart.cloudera --privileged=true -ti --rm docker-cdh-spark:latest /bin/bash`
### Option 2: docker compose (*recommend*)
`docker-compose up -d`
`docker-compose exec docker-cdh-spark /bin/bash`
### Start ipython notebook
`ipython notebook --ip 0.0.0.0 --port 8080 --allow-root --NotebookApp.token=''`
## Colab (If you do not have docker environment.)
* !apt-get install openjdk-8-jdk-headless -qq > /dev/null* !pip install pyspark