Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/apache/incubator-livy
Apache Livy is an open source REST interface for interacting with Apache Spark from anywhere.
https://github.com/apache/incubator-livy
apachelivy bigdata livy spark
Last synced: 3 days ago
JSON representation
Apache Livy is an open source REST interface for interacting with Apache Spark from anywhere.
- Host: GitHub
- URL: https://github.com/apache/incubator-livy
- Owner: apache
- License: apache-2.0
- Created: 2017-06-25T07:00:09.000Z (over 7 years ago)
- Default Branch: master
- Last Pushed: 2024-11-12T13:33:33.000Z (3 months ago)
- Last Synced: 2025-01-21T06:01:51.613Z (10 days ago)
- Topics: apachelivy, bigdata, livy, spark
- Language: Scala
- Homepage: https://livy.apache.org/
- Size: 3.54 MB
- Stars: 899
- Watchers: 59
- Forks: 604
- Open Issues: 111
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
README
# Apache Livy
[![Build Status](https://travis-ci.org/apache/incubator-livy.svg?branch=master)](https://travis-ci.org/apache/incubator-livy)
Apache Livy is an open source REST interface for interacting with
[Apache Spark](http://spark.apache.org) from anywhere. It supports executing snippets of code or
programs in a Spark context that runs locally or in
[Apache Hadoop YARN](http://hadoop.apache.org/docs/current/hadoop-yarn/hadoop-yarn-site/YARN.html).* Interactive Scala, Python and R shells
* Batch submissions in Scala, Java, Python
* Multiple users can share the same server (impersonation support)
* Can be used for submitting jobs from anywhere with REST
* Does not require any code change to your programs[Pull requests](https://github.com/apache/incubator-livy/pulls) are welcomed! But before you begin,
please check out the [Contributing](http://livy.incubator.apache.org/community#Contributing)
section on the [Community](http://livy.incubator.apache.org/community) page of our website.## Online Documentation
Guides and documentation on getting started using Livy, example code snippets, and Livy API
documentation can be found at [livy.incubator.apache.org](http://livy.incubator.apache.org).## Before Building Livy
To build Livy, you will need:
Debian/Ubuntu:
* mvn (from ``maven`` package or maven3 tarball)
* openjdk-8-jdk (or Oracle JDK 8)
* Python 2.7+
* R 3.xRedhat/CentOS:
* mvn (from ``maven`` package or maven3 tarball)
* java-1.8.0-openjdk (or Oracle JDK 8)
* Python 2.7+
* R 3.xMacOS:
* Xcode command line tools
* Oracle's JDK 1.8
* Maven (Homebrew)
* Python 2.7+
* R 3.xRequired python packages for building Livy:
* cloudpickle
* requests
* requests-kerberos
* flake8
* flaky
* pytestTo run Livy, you will also need a Spark installation. You can get Spark releases at
https://spark.apache.org/downloads.html.Livy requires Spark 2.4+. You can switch to a different version of Spark by setting the
``SPARK_HOME`` environment variable in the Livy server process, without needing to rebuild Livy.## Building Livy
Livy is built using [Apache Maven](http://maven.apache.org). To check out and build Livy, run:
```
git clone https://github.com/apache/incubator-livy.git
cd incubator-livy
mvn package
```You can also use the provided [Dockerfile](./dev/docker/livy-dev-base/Dockerfile):
```
git clone https://github.com/apache/incubator-livy.git
cd incubator-livy
docker build -t livy-ci dev/docker/livy-dev-base/
docker run --rm -it -v $(pwd):/workspace -v $HOME/.m2:/root/.m2 livy-ci mvn package
```> **Note**: The `docker run` command maps the maven repository to your host machine's maven cache so subsequent runs will not need to download dependencies.
By default Livy is built against Apache Spark 2.4.5, but the version of Spark used when running
Livy does not need to match the version used to build Livy. Livy internally handles the differences
between different Spark versions.The Livy package itself does not contain a Spark distribution. It will work with any supported
version of Spark without needing to rebuild.### Build Profiles
| Flag | Purpose |
|--------------|--------------------------------------------------------------------|
| -Phadoop2 | Choose Hadoop2 based build dependencies (default configuration) |
| -Pspark2 | Choose Spark 2.x based build dependencies (default configuration) |
| -Pspark3 | Choose Spark 3.x based build dependencies |
| -Pscala-2.11 | Choose Scala 2.11 based build dependencies (default configuration) |
| -Pscala-2.12 | Choose scala 2.12 based build dependencies |