Ecosyste.ms: Awesome

An open API service indexing awesome lists of open source software.

https://github.com/Stratio/sparta

Real Time Analytics and Data Pipelines based on Spark Streaming
https://github.com/Stratio/sparta

analytics hdfs kafka lambda olap real-time scala spark spark-streaming sparksql sparta stratio stratio-sparta streaming streaming-data triggers workflow

Last synced: 2 months ago
JSON representation

Real Time Analytics and Data Pipelines based on Spark Streaming

Lists

README

        

Discontinued
============
After around two years of development, we have decided to discontinue this project due to a major refactor in its structure and in a near future we will launch Sparta 2.0.

We would like to thank all the open source community for their contribution.
Needless to say that you can continue using this repository as a basis for your developments as it contains the latest stable version as of today and minor issues will be attended.

If you are interested in the new Sparta 2.0 with pipelines and workflows, please contact with us in the email [email protected]

About Stratio Sparta
============

At Stratio, we have implemented several real-time analytics projects based on Apache Spark, Kafka, Flume, Cassandra, ElasticSearch or MongoDB.
These technologies were always a perfect fit, but soon we found ourselves writing the same pieces of integration code over and over again.
Stratio Sparta is the easiest way to make use of the Apache Spark Streaming technology and all its ecosystem.
Choose your input, operations and outputs, and start extracting insights out of your data in real-time.

Strata Twitter Analytics with Kibana

Main Features
============

- Pure Spark
- No need of coding, only declarative analytical workflows
- Data continuously streamed in & processed in near real-time
- Ready to use out-of-the-box
- Plug & play: flexible workflows (inputs, outputs, transformations, etc…)
- High performance and Fault Tolerance
- Scalable and High Availability
- Big Data OLAP on real-time to small data
- ETLs
- Triggers over streaming data
- Spark SQL language with streaming and batch data
- Kerberos and CAS compatible

Main Features

Architecture
============

Send one workflow as a JSON to Sparta API and execute in one Spark Cluster your own real-time plugins
![Architecture](./images/architecture.jpg)

Sparta as a Job Manager
------------

Send more than one Streaming Job in the Spark Cluster and manage them with a simple UI

Job Manager

Run workflows over Mesos, Yarn or SparkStandAlone

Job Manager Architecture

Sparta as a SDK
------------

Modular components extensible with simple SDK
- You can extend several points of the platform to fulfill your needs, such as adding new inputs, outputs, operators, transformations.
- Add new functions to Kite SDK in order to extend the data cleaning, enrichment and normalization capabilities.
![Architecture Detail](./images/architectureDetail.jpg)

Components
========

On each workflow multiple components can be defined, but now all have the following architecture
![workflow](./images/workflow.jpg)
![Components](./images/components.jpg)

Core components
------------

Several plugins are been implemented by Stratio Sparta team
![Main plugins](./images/plugins.jpg)

Trigger component
------------

With Sparta is possible to execute queries over the streaming data, execute ETL, aggregations and Simple Event
Processing mixing streaming data with batch data on the trigger process.
![triggers](./images/triggers.jpg)

Aggregation component
------------

The aggregation process in Sparta is very powerful because is possible to generate efficient OLAP processes with
streaming data
![OLAP](./images/OLAPintegration.jpg)

Advanced feature are been implemented in order to optimize the stateful operations over Spark Streaming
![Aggregations](./images/aggregation.jpg)

Inputs
------------

- Twitter
- Kafka
- Flume
- RabbitMQ
- Socket
- WebSocket
- HDFS/S3

Outputs
------------

- MongoDB
- Cassandra
- ElasticSearch
- Redis
- JDBC
- CSV
- Parquet
- Http
- Kafka
- HDFS/S3
- Http Rest
- Avro
- Logger

![Outputs](./images/outputs.png)

Key technologies
========

- [Spark Streaming & Spark] (http://spark.apache.org)
- [SparkSQL] (https://spark.apache.org/sql)
- [Akka] (http://akka.io)
- [MongoDB] (http://www.mongodb.org/)
- [Apache Cassandra] (http://cassandra.apache.org)
- [ElasticSearch] (https://www.elastic.co)
- [Redis] (http://redis.io)
- [Apache Parquet] (http://parquet.apache.org/)
- [HDFS] (http://hadoop.apache.org/docs/r1.2.1/hdfs_design.html)
- [Apache Kafka] (http://kafka.apache.org)
- [Apache Flume] (https://flume.apache.org/)
- [RabbitMQ] (https://www.rabbitmq.com/)
- [Spray] (http://spray.io/)
- [KiteSDK (morphlines)] (http://kitesdk.org/docs/current)
- [Apache Avro] (https://avro.apache.org/)

Advantages
========

Sparta provide several advantages to final Users
![Advantages](./images/features.jpg)

Build
========

You can generate rpm and deb packages by running:

`mvn clean package -Ppackage`

**Note:** you need to have installed the following programs in order to build these packages:

In a debian distribution:

- fakeroot
- dpkg-dev
- rpm
- jq

In a centOS distribution:

- fakeroot
- dpkg-dev
- rpmdevtools
- jq

In all distributions:

- Java 8
- Maven 3

License
========

Licensed to STRATIO (C) under one or more contributor license agreements.
See the NOTICE file distributed with this work for additional information
regarding copyright ownership. The STRATIO (C) licenses this file
to you under the Apache License, Version 2.0 (the
"License"); you may not use this file except in compliance
with the License. You may obtain a copy of the License at

http://www.apache.org/licenses/LICENSE-2.0

Unless required by applicable law or agreed to in writing,
software distributed under the License is distributed on an
"AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
KIND, either express or implied. See the License for the
specific language governing permissions and limitations
under the License.