Ecosyste.ms: Awesome

An open API service indexing awesome lists of open source software.

Awesome Lists | Featured Topics | Projects

https://github.com/mesos/kafka

Apache Kafka on Apache Mesos
https://github.com/mesos/kafka

Last synced: 6 days ago
JSON representation

Apache Kafka on Apache Mesos

Awesome Lists containing this project

README

        

Kafka Mesos Framework [![Build Status](https://travis-ci.org/mesos/kafka.svg?branch=master)](https://travis-ci.org/mesos/kafka)
======================

**Join us on Slack at #kafka-mesos https://mesos-slackin.herokuapp.com/**

For issues https://github.com/mesos/kafka/issues

[Installation](#installation)
* [Environment Configuration](#environment-configuration)
* [Scheduler Configuration](#scheduler-configuration)
* [Run the scheduler](#run-the-scheduler)
* [Starting and using 1 broker](#starting-and-using-1-broker)

[Typical Operations](#typical-operations)
* [Run the scheduler with Docker](https://github.com/mesos/kafka/tree/master/src/docker#intro)
* [Run the scheduler on Marathon](https://github.com/mesos/kafka/tree/master/src/docker#running-image-in-marathon)
* [Changing the location where data is stored](#changing-the-location-where-data-is-stored)
* [Starting 3 brokers](#starting-3-brokers)
* [View broker log](#view-broker-log)
* [High Availability Scheduler State](#high-availability-scheduler-state)
* [Failed Broker Recovery](#failed-broker-recovery)
* [Passing multiple options](#passing-multiple-options)
* [Broker metrics](#broker-metrics)
* [Rolling restart](#rolling-restart)

[Navigating the CLI](#navigating-the-cli)
* [Adding brokers to the cluster](#adding-brokers-to-the-cluster)
* [Updating broker configurations](#updating-broker-configurations)
* [Starting brokers](#starting-brokers-in-the-cluster)
* [Stopping brokers](#stopping-brokers-in-the-cluster)
* [Restarting brokers](#restarting-brokers-in-the-cluster)
* [Removing brokers](#removing-brokers-from-the-cluster)
* [Cloning brokers](#cloning-brokers)
* [Retrieving broker log](#retrieving-broker-log)
* [Rebalancing brokers in the cluster](#rebalancing-topics)
* [Realigning brokers in the cluster](#realigning-topics)
* [Listing topics](#listing-topics)
* [Adding topic](#adding-topic)
* [Updating topic](#updating-topic)
* [Listing topic partitions details](#listing-topic-partition-details)

[Using the REST API](#using-the-rest-api)

[Project Goals](#project-goals)

Installation
-------------

Install OpenJDK 7 (or higher) http://openjdk.java.net/install/

Install gradle http://gradle.org/installation

Clone and build the project

# git clone https://github.com/mesos/kafka
# cd kafka
# ./gradlew jar downloadKafka

Environment Configuration
--------------------------

Before running `./kafka-mesos.sh`, set the location of libmesos:

# export MESOS_NATIVE_JAVA_LIBRARY=/usr/local/lib/libmesos.so

If the host running scheduler has several IP addresses you may also need to

# export LIBPROCESS_IP=

Scheduler Configuration
----------------------

The scheduler is configured through the command line or `kafka-mesos.properties` file.

The following options are available:
```
# ./kafka-mesos.sh help scheduler
Start scheduler
Usage: scheduler [options] [config.properties]

Option Description
------ -----------
--api Api url. Example: http://master:7000
--bind-address Scheduler bind address (master, 0.0.0.0, 192.168.50.*, if:eth1). Default - all
--debug Debug mode. Default - false
--framework-name Framework name. Default - kafka
--framework-role Framework role. Default - *
--framework-timeout Framework timeout (30s, 1m, 1h). Default - 30d
--jre JRE zip-file (jre-7-openjdk.zip). Default - none.
--log Log file to use. Default - stdout.
--master Master connection settings. Examples:
- master:5050
- master:5050,master2:5050
- zk://master:2181/mesos
- zk://username:password@master:2181
- zk://master:2181,master2:2181/mesos
--principal Principal (username) used to register framework. Default - none
--secret Secret (password) used to register framework. Default - none
--storage Storage for cluster state. Examples:
- file:kafka-mesos.json
- zk:/kafka-mesos
Default - file:kafka-mesos.json
--user Mesos user to run tasks. Default - none
--zk Kafka zookeeper.connect. Examples:
- master:2181
- master:2181,master2:2181
```

Additionally you can create `kafka-mesos.properties` containing values for CLI options of scheduler.

Example of `kafka-mesos.properties`:
```
storage=file:kafka-mesos.json
master=zk://master:2181/mesos
zk=master:2181
api=http://master:7000
```

Now if running scheduler via `./kafka-mesos.sh scheduler` (no options specified) the scheduler will read values for options
from the above file. You could also specify alternative config file by using `config` argument of the scheduler.

Another example of `kafka-mesos.properties`:
```
storage=zk:/kafka-mesos
master=zk://master:2181/mesos
zk=zookeep1:2181,zookeep2:2181,zookeep3:2181/KafkaCluster
api=http://master:7000
```

The second example creates /KafkaCluster chroot in zookeeper quorum (consisting of zookeep1,zookeep2,zookeep3 servers). Cluster data (--storage) is saved in
zookeeper quorum, /KafkaCluster/kafka-mesos. Example creating topics using tools that come with kafka would be
```
kafka-topics.sh --zookeeper zookeep1:2181,zookeep2:2181,zookeep3:2181/KafkaCluster --create --config min.insync.replicas=3 --config max.message.bytes=47185920 --config unclean.leader.election.enable=false --topic tw_spam --partitions 15 --replication-factor 3
```

Run the scheduler
-----------------

Start the Kafka scheduler using this command:

# ./kafka-mesos.sh scheduler

Note: you can also use Marathon to launch the scheduler process so it gets restarted if it crashes.

Starting and using 1 broker
---------------------------

First let's start up and use 1 broker with the default settings. Further in the readme you can see how to change these from the defaults.

```
# ./kafka-mesos.sh broker add 0
broker added:
id: 0
active: false
state: stopped
resources: cpus:1.00, mem:2048, heap:1024, port:auto
failover: delay:1m, max-delay:10m
stickiness: period:10m
```

You now have a cluster with 1 broker that is not started.

```
# ./kafka-mesos.sh broker list
broker:
id: 0
active: false
state: stopped
resources: cpus:1.00, mem:2048, heap:1024, port:auto
failover: delay:1m, max-delay:10m
stickiness: period:10m
```
Now let's start the broker.

```
# ./kafka-mesos.sh broker start 0
broker started:
id: 0
active: true
state: running
resources: cpus:1.00, mem:2048, heap:1024, port:auto
failover: delay:1m, max-delay:10m
stickiness: period:10m, hostname:slave0
task:
id: broker-0-d2d94520-2f3e-4779-b276-771b4843043c
running: true
endpoint: 172.16.25.62:31000
attributes: rack=r1
```

Great! Now let's produce and consume from the cluster. Let's use [kafkacat](https://github.com/edenhill/kafkacat), a nice third party c library command line tool for Kafka.

```
# echo "test"|kafkacat -P -b "172.16.25.62:31000" -t testTopic -p 0
```

And let's read it back.

```
# kafkacat -C -b "172.16.25.62:31000" -t testTopic -p 0 -e
test
```

This is a beta version.

Typical Operations
===================

Changing the location where data is stored
------------------------------------------

```
# ./kafka-mesos.sh broker stop 0
broker stopped:
id: 0
active: false
state: stopped
resources: cpus:1.00, mem:2048, heap:1024, port:auto
failover: delay:1m, max-delay:10m
stickiness: period:10m, hostname:slave0, expires:2015-07-10 15:51:43+03

# ./kafka-mesos.sh broker update 0 --options log.dirs=/mnt/array1/broker0
broker updated:
id: 0
active: false
state: stopped
resources: cpus:1.00, mem:2048, heap:1024, port:auto
options: log.dirs=/mnt/array1/broker0
failover: delay:1m, max-delay:10m
stickiness: period:10m, hostname:slave0, expires:2015-07-10 15:51:43+03

# ./kafka-mesos.sh broker start 0
broker started:
id: 0
active: true
state: running
resources: cpus:1.00, mem:2048, heap:1024, port:auto
failover: delay:1m, max-delay:10m
stickiness: period:10m, hostname:slave0
task:
id: broker-0-d2d94520-2f3e-4779-b276-771b4843043c
running: true
endpoint: 172.16.25.62:31000
attributes: rack=r1
```

Starting 3 brokers
-------------------------

```
#./kafka-mesos.sh broker add 0..2 --heap 1024 --mem 2048
brokers added:
id: 0
active: false
state: stopped
resources: cpus:1.00, mem:2048, heap:1024, port:auto
failover: delay:1m, max-delay:10m
stickiness: period:10m

id: 1
active: false
state: stopped
resources: cpus:1.00, mem:2048, heap:1024, port:auto
failover: delay:1m, max-delay:10m
stickiness: period:10m

id: 2
active: false
state: stopped
resources: cpus:1.00, mem:2048, heap:1024, port:auto
failover: delay:1m, max-delay:10m
stickiness: period:10m

#./kafka-mesos.sh broker start 0..2
brokers started:
id: 0
active: true
state: running
resources: cpus:1.00, mem:2048, heap:1024, port:auto
failover: delay:1m, max-delay:10m
stickiness: period:10m, hostname:slave0
task:
id: broker-0-d2d94520-2f3e-4779-b276-771b4843043c
running: true
endpoint: 172.16.25.62:31000
attributes: rack=r1

id: 1
active: true
state: running
...
```

View broker log
---------------

Strings are always being read from the end of a file.
Get last 100 lines from `stdout` file of broker 0,
```
./kafka-mesos.sh broker log 0
```

or from `stderr`

```
./kafka-mesos.sh broker log 0 --name stderr
```

or any file in kafka-*/log/, for example: `server.log`

```
./kafka-mesos.sh broker log 0 --name server.log
```

or maybe more lines

```
./kafka-mesos.sh broker log 0 --name server.log --lines 200
```

current limit is 100Kb no matter how many lines being requested.

High Availability Scheduler State
-------------------------
The scheduler supports storing the cluster state in Zookeeper. It currently shares a znode within the mesos ensemble. To turn this on in properties

```
clusterStorage=zk:/kafka-mesos
```

Failed Broker Recovery
------------------------
When a broker fails, kafka mesos scheduler assumes that the failure is recoverable. The scheduler will try
to restart the broker after waiting failover-delay (i.e. 30s, 2m). The initial waiting delay is equal to failover-delay setting.
After each consecutive failure this delay is doubled until it reaches failover-max-delay value.

If failover-max-tries is defined and the consecutive failure count exceeds it, the broker will be deactivated.

The following failover settings exists:
```
--failover-delay - initial failover delay to wait after failure, required
--failover-max-delay - max failover delay, required
--failover-max-tries - max failover tries to deactivate broker, optional
```

Broker Placement Stickiness
---------------------------
If a broker is started within a stickiness-period interval from it's stop time, the scheduler will place it on the same node
it was on during the last successful start. This applies both to failover and manual restarts.

The following stickiness settings exists:
```
--stickiness-period - period of time during which broker would be restarted on the same node
```

Passing multiple options
-----------------------
A common use case is to supply multiple `log.dirs`, or provide other options. To do this you may use comma escaping like this:

```
./kafka-mesos.sh broker update 0 --options log.dirs=/mnt/array1/broker0\\,/mnt/array2/broker0,num.io.threads=16
broker updated:
id: 0
active: false
state: stopped
resources: cpus:1.00, mem:2048, heap:1024, port:auto
options: log.dirs=/mnt/array1/broker0\,/mnt/array2/broker0,num.io.threads=16
failover: delay:1m, max-delay:10m
stickiness: period:10m, hostname:slave0, expires:2015-07-29 11:54:39Z
```

Broker metrics
--------------
Executor sends broker metrics to scheduler every 30 seconds, such as:
- under-replicated-partitions is number of under replicated partitions (| ISR | < | all replicas |)
- offline-partitions-count is number of partitions that don't have an active leader and are hence not writable or readable
- is-active-controller is controller active on broker, only one broker in cluster should have value "1"

```
./kafka-mesos.sh broker list
brokers:
id: 0
active: true
state: running
resources: cpus:1.00, mem:2048, heap:1024, port:auto
failover: delay:1m, max-delay:10m
stickiness: period:10m, hostname:slave0
task:
id: broker-0-826e8075-5dd3-49ab-b86e-6432fa03ef66
state: running
endpoint: slave0:9092 (slave0)
metrics:
collected: 2016-01-18 11:53:36Z
under-replicated-partitions: 0
offline-partitions-count: 0
is-active-controller: 1

id: 1
active: true
state: running
resources: cpus:1.00, mem:2048, heap:1024, port:auto
failover: delay:1m, max-delay:10m
stickiness: period:10m, hostname:slave1
task:
id: broker-1-4648130a-9aa0-4c7e-b8af-761c03aa111c
state: running
endpoint: slave1:9092 (slave1)
metrics:
collected: 2016-01-18 11:53:36Z
under-replicated-partitions: 0
offline-partitions-count: 0
is-active-controller: 0
```

Rolling restart
---------------

For example there are 3 running brokers and you want to add new log dir /mnt/array2/broker$id alongside
already used /mnt/array1/broker$id

```
# ./kafka-mesos.sh broker list
brokers:
id: 0
active: true
state: running
resources: cpus:0.50, mem:2048, heap:1024, port:9092
bind-address: slave0
constraints: hostname=like:slave0
options: log.dirs=/mnt/array1/broker$id
...

id: 1
active: true
state: running
resources: cpus:0.50, mem:2048, heap:1024, port:9092
bind-address: slave1
constraints: hostname=like:slave1
options: log.dirs=/mnt/array1/broker$id
...

id: 2
active: true
state: running
resources: cpus:0.50, mem:2048, heap:1024, port:9092
bind-address: slave2
constraints: hostname=like:slave2
options: log.dirs=/mnt/array1/broker$id
...
```

Update brokers by adding new log dir (they were already using dir /mnt/array1/broker$id).

NOTE: you could update brokers in running state, however configuration updated only in
scheduler, thus you need to restart brokers to apply changes. Whenever you update running
broker `broker list` will show you notification `(modified, needs restart)` right in state description,
notice will disappear once broker restarted.

```
# ./kafka-mesos.sh broker update 1..3 --options log.dirs=/mnt/array1/broker\$id\\,/mnt/array2/broker\$id
brokers updated:
id: 0
active: true
state: running (modified, needs restart)
resources: cpus:0.50, mem:2048, heap:1024, port:9092
bind-address: slave0
constraints: hostname=like:slave0
options: log.dirs=/mnt/array1/broker$id\,/mnt/array2/broker$id
...

id: 1
active: true
state: running (modified, needs restart)
resources: cpus:0.50, mem:2048, heap:1024, port:9092
bind-address: slave1
constraints: hostname=like:slave1
options: log.dirs=/mnt/array1/broker$id\,/mnt/array2/broker$id
...

id: 2
active: true
state: running (modified, needs restart)
resources: cpus:0.50, mem:2048, heap:1024, port:9092
bind-address: slave2
constraints: hostname=like:slave2
options: log.dirs=/mnt/array1/broker$id\,/mnt/array2/broker$id
...
```

Rolling restart (sequentially stop then start each broker) gives you ability to combine
`stop` then `start` for each broker in single action. See [restart CLI options](#restarting-brokers-in-the-cluster).

```
# ./kafka-mesos.sh broker restart 1..3 --timeout 5m
brokers restarted:
id: 0
active: true
state: running
resources: cpus:0.50, mem:2048, heap:1024, port:9092
bind-address: slave0
constraints: hostname=like:slave0
options: log.dirs=/mnt/array1/broker$id\,/mnt/array2/broker$id
...

id: 1
active: true
state: running
resources: cpus:0.50, mem:2048, heap:1024, port:9092
bind-address: slave1
constraints: hostname=like:slave1
options: log.dirs=/mnt/array1/broker$id\,/mnt/array2/broker$id
...

id: 2
active: true
state: running
resources: cpus:0.50, mem:2048, heap:1024, port:9092
bind-address: slave2
constraints: hostname=like:slave2
options: log.dirs=/mnt/array1/broker$id\,/mnt/array2/broker$id
...
```

It's possible that some broker could timeout on `stop` or `start`, in such case `restart` halts with notice:

```
# ./kafka-mesos.sh broker restart 1..3 --timeout 5m
Error: broker 1 timeout on stop
```

or for `start`

```
# ./kafka-mesos.sh broker restart 1..3 --timeout 5m
Error: broker 1 timeout on start
```

Navigating the CLI
==================

Adding brokers to the cluster
-------------------------------

```
# ./kafka-mesos.sh help broker add
Add broker
Usage: broker add [options]

Option Description
------ -----------
--bind-address broker bind address (broker0, 192.168.50.*, if:eth1). Default - auto
--constraints constraints (hostname=like:master,rack=like:1.*). See below.
--cpus cpu amount (0.5, 1, 2)
--failover-delay failover delay (10s, 5m, 3h)
--failover-max-delay max failover delay. See failoverDelay.
--failover-max-tries max failover tries. Default - none
--heap heap amount in Mb
--jvm-options jvm options string (-Xms128m -XX:PermSize=48m)
--log4j-options log4j options or file. Examples:
log4j.logger.kafka=DEBUG\, kafkaAppender
file:log4j.properties
--mem mem amount in Mb
--options options or file. Examples:
log.dirs=/tmp/kafka/$id,num.io.threads=16
file:server.properties
--port port or range (31092, 31090..31100). Default - auto
--stickiness-period stickiness period to preserve same node for broker (5m, 10m, 1h)
--volume pre-reserved persistent volume id

Generic Options
Option Description
------ -----------
--api Api url. Example: http://master:7000

broker-expr examples:
0 - broker 0
0,1 - brokers 0,1
0..2 - brokers 0,1,2
0,1..2 - brokers 0,1,2
* - any broker
attribute filtering:
*[rack=r1] - any broker having rack=r1
*[hostname=slave*] - any broker on host with name starting with 'slave'
0..4[rack=r1,dc=dc1] - any broker having rack=r1 and dc=dc1

constraint examples:
like:master - value equals 'master'
unlike:master - value not equals 'master'
like:slave.* - value starts with 'slave'
unique - all values are unique
cluster - all values are the same
cluster:master - value equals 'master'
groupBy - all values are the same
groupBy:3 - all values are within 3 different groups
```

Updating broker configurations
-----------------------------------

```
# ./kafka-mesos.sh help broker update
Update broker
Usage: broker update [options]

Option Description
------ -----------
--bind-address broker bind address (broker0, 192.168.50.*, if:eth1). Default - auto
--constraints constraints (hostname=like:master,rack=like:1.*). See below.
--cpus cpu amount (0.5, 1, 2)
--failover-delay failover delay (10s, 5m, 3h)
--failover-max-delay max failover delay. See failoverDelay.
--failover-max-tries max failover tries. Default - none
--heap heap amount in Mb
--jvm-options jvm options string (-Xms128m -XX:PermSize=48m)
--log4j-options log4j options or file. Examples:
log4j.logger.kafka=DEBUG\, kafkaAppender
file:log4j.properties
--mem mem amount in Mb
--options options or file. Examples:
log.dirs=/tmp/kafka/$id,num.io.threads=16
file:server.properties
--port port or range (31092, 31090..31100). Default - auto
--stickiness-period stickiness period to preserve same node for broker (5m, 10m, 1h)
--volume pre-reserved persistent volume id

Generic Options
Option Description
------ -----------
--api Api url. Example: http://master:7000

broker-expr examples:
0 - broker 0
0,1 - brokers 0,1
0..2 - brokers 0,1,2
0,1..2 - brokers 0,1,2
* - any broker
attribute filtering:
*[rack=r1] - any broker having rack=r1
*[hostname=slave*] - any broker on host with name starting with 'slave'
0..4[rack=r1,dc=dc1] - any broker having rack=r1 and dc=dc1

constraint examples:
like:master - value equals 'master'
unlike:master - value not equals 'master'
like:slave.* - value starts with 'slave'
unique - all values are unique
cluster - all values are the same
cluster:master - value equals 'master'
groupBy - all values are the same
groupBy:3 - all values are within 3 different groups

Note: use "" arg to unset an option
```

Starting brokers in the cluster
-------------------------------

```
# ./kafka-mesos.sh help broker start
Start broker
Usage: broker start [options]

Option Description
------ -----------
--timeout timeout (30s, 1m, 1h). 0s - no timeout

Generic Options
Option Description
------ -----------
--api Api url. Example: http://master:7000

broker-expr examples:
0 - broker 0
0,1 - brokers 0,1
0..2 - brokers 0,1,2
0,1..2 - brokers 0,1,2
* - any broker
attribute filtering:
*[rack=r1] - any broker having rack=r1
*[hostname=slave*] - any broker on host with name starting with 'slave'
0..4[rack=r1,dc=dc1] - any broker having rack=r1 and dc=dc1
```

Stopping brokers in the cluster
-------------------------------

```
# ./kafka-mesos.sh help broker stop
Stop broker
Usage: broker stop [options]

Option Description
------ -----------
--force forcibly stop
--timeout timeout (30s, 1m, 1h). 0s - no timeout

Generic Options
Option Description
------ -----------
--api Api url. Example: http://master:7000

broker-expr examples:
0 - broker 0
0,1 - brokers 0,1
0..2 - brokers 0,1,2
0,1..2 - brokers 0,1,2
* - any broker
attribute filtering:
*[rack=r1] - any broker having rack=r1
*[hostname=slave*] - any broker on host with name starting with 'slave'
0..4[rack=r1,dc=dc1] - any broker having rack=r1 and dc=dc1
```

Restarting brokers in the cluster
---------------------------------

```
# ./kafka-mesos.sh help broker restart
Restart broker
Usage: broker restart [options]

Option Description
------ -----------
--timeout time to wait until broker restarts (30s, 1m, 1h). Default - 2m

Generic Options
Option Description
------ -----------
--api Api url. Example: http://master:7000

broker-expr examples:
0 - broker 0
0,1 - brokers 0,1
0..2 - brokers 0,1,2
0,1..2 - brokers 0,1,2
* - any broker
attribute filtering:
*[rack=r1] - any broker having rack=r1
*[hostname=slave*] - any broker on host with name starting with 'slave'
0..4[rack=r1,dc=dc1] - any broker having rack=r1 and dc=dc1
```

Removing brokers from the cluster
----------------------------------

```
# ./kafka-mesos.sh help broker remove
Remove broker
Usage: broker remove [options]

Generic Options
Option Description
------ -----------
--api Api url. Example: http://master:7000

broker-expr examples:
0 - broker 0
0,1 - brokers 0,1
0..2 - brokers 0,1,2
0,1..2 - brokers 0,1,2
* - any broker
attribute filtering:
*[rack=r1] - any broker having rack=r1
*[hostname=slave*] - any broker on host with name starting with 'slave'
0..4[rack=r1,dc=dc1] - any broker having rack=r1 and dc=dc1
```

Cloning Brokers
---------------

```
# ./kafka-mesos.sh help broker clone
Clone broker
Usage: broker clone [options]

Option Description
------ -----------
--source source broker id to copy settings from

Generic Options
Option Description
------ -----------
--api Api url. Example: http://master:7000

broker-expr examples:
0 - broker 0
0,1 - brokers 0,1
0..2 - brokers 0,1,2
0,1..2 - brokers 0,1,2
* - any broker
attribute filtering:
*[rack=r1] - any broker having rack=r1
*[hostname=slave*] - any broker on host with name starting with 'slave'
0..4[rack=r1,dc=dc1] - any broker having rack=r1 and dc=dc1
```

Retrieving broker log
---------------------

```
Retrieve broker log
Usage: broker log [options]

Option Description
------ -----------
--lines maximum number of lines to read from the end of file.
Default - 100
--name name of log file (stdout, stderr, server.log). Default
- stdout
--timeout timeout (30s, 1m, 1h). Default - 30s

Generic Options
Option Description
------ -----------
--api Api url. Example: http://master:7000
```

Listing Topics
--------------
```
#./kafka-mesos.sh help topic list
List topics
Usage: topic list []

Generic Options
Option Description
------ -----------
--api Api url. Example: http://master:7000

topic-expr examples:
t0 - topic t0
t0,t1 - topics t0, t1
* - any topic
t* - topics starting with 't'
```

Adding Topic
------------
```
#./kafka-mesos.sh help topic add
Add topic
Usage: topic add [options]

Option Description
------ -----------
--broker . Default - *. See below.
--options topic options. Example: flush.ms=60000,retention.ms=6000000
--partitions partitions count. Default - 1
--replicas replicas count. Default - 1

Generic Options
Option Description
------ -----------
--api Api url. Example: http://master:7000

topic-expr examples:
t0 - topic t0
t0,t1 - topics t0, t1
* - any topic
t* - topics starting with 't'

broker-expr examples:
0 - broker 0
0,1 - brokers 0,1
0..2 - brokers 0,1,2
0,1..2 - brokers 0,1,2
* - any broker
attribute filtering:
*[rack=r1] - any broker having rack=r1
*[hostname=slave*] - any broker on host with name starting with 'slave'
0..4[rack=r1,dc=dc1] - any broker having rack=r1 and dc=dc1
```

Updating Topic
--------------
```
#./kafka-mesos.sh help topic update
Update topic
Usage: topic update [options]

Option Description
------ -----------
--options topic options. Example: flush.ms=60000,retention.ms=6000000

Generic Options
Option Description
------ -----------
--api Api url. Example: http://master:7000

topic-expr examples:
t0 - topic t0
t0,t1 - topics t0, t1
* - any topic
t* - topics starting with 't'
```

Deleting Topics
--------------
Note: needs to be supported and enabled in the broker configuration (i.e.: `delete.topic.enable=true`).
```
#./kafka-mesos.sh help topic delete
Delete topics
Usage: topic delete []

Generic Options
Option Description
------ -----------
--api Api url. Example: http://master:7000

topic-expr examples:
t0 - topic t0
t0,t1 - topics t0, t1
* - any topic
t* - topics starting with 't'
```

Listing topic partition details
-------------------------------
```
#./kafka-mesos.sh help topic partitions
Picked up _JAVA_OPTIONS: -Djava.net.preferIPv4Stack=true
List partitions
Usage: topic partition []

Generic Options
Option Description
------ -----------
--api Api url. Example: http://master:7000

topic-expr examples:
t0 - topic t0
t0,t1 - topics t0, t1
* - any topic
t* - topics starting with 't'
```

Rebalancing topics
----------------------------------
```
#./kafka-mesos.sh help topic rebalance
Rebalance topics
Usage: topic rebalance |status [options]

Option Description
------ -----------
--broker . Default - *. See below.
--replicas replicas count. Default - 1
--timeout timeout (30s, 1m, 1h). 0s - no timeout

Generic Options
Option Description
------ -----------
--api Api url. Example: http://master:7000

topic-expr examples:
t0 - topic t0
t0,t1 - topics t0, t1
* - any topic
t* - topics starting with 't'

broker-expr examples:
0 - broker 0
0,1 - brokers 0,1
0..2 - brokers 0,1,2
0,1..2 - brokers 0,1,2
* - any broker
attribute filtering:
*[rack=r1] - any broker having rack=r1
*[hostname=slave*] - any broker on host with name starting with 'slave'
0..4[rack=r1,dc=dc1] - any broker having rack=r1 and dc=dc1
```

Realigning topics
----------------------------------
```
#./kafka-mesos.sh help topic realign
Realign topics
Usage: topic realign |status [options]

Option Description
------ -----------
--broker . Default - *. See below.
--fixedStartIndex index into the broker set to start assigning partitions at. Default - -1
(random)
--replicas replicas count. Default - -1 (no change)
--timeout timeout (30s, 1m, 1h). 0s - no timeout

Generic Options
Option Description
------ -----------
--api Api url. Example: http://master:7000

topic-expr examples:
t0 - topic t0
t0,t1 - topics t0, t1
* - any topic
t* - topics starting with 't'

broker-expr examples:
0 - broker 0
0,1 - brokers 0,1
0..2 - brokers 0,1,2
0,1..2 - brokers 0,1,2
* - any broker
attribute filtering:
*[rack=r1] - any broker having rack=r1
*[hostname=slave*] - any broker on host with name starting with 'slave'
0..4[rack=r1,dc=dc1] - any broker having rack=r1 and dc=dc1
```

Using the REST API
========================

The scheduler REST API fully exposes all of the features of the CLI with the following request format:
```
/api/broker//broker={broker-expr}&=
/api/topic//topic={topic-expr}&=
```

Listing brokers

```
# curl "http://localhost:7000/api/broker/list"
{"brokers" : [{"id" : "0", "mem" : 128, "cpus" : 0.1, "heap" : 128, "failover" : {"delay" : "10s", "maxDelay" : "60s", "failures" : 5, "failureTime" : 1426651240585}, "active" : true}, {"id" : "5", "mem" : 128, "cpus" : 0.5, "heap" : 128, "failover" : {"delay" : "10s", "maxDelay" : "60s"}, "active" : false}, {"id" : "8", "mem" : 43008, "cpus" : 8.0, "heap" : 128, "failover" : {"delay" : "10s", "maxDelay" : "60s"}, "active" : true}]}
```

Adding a broker

```
# curl -X POST "http://localhost:7000/api/broker/add?broker=0&cpus=8&mem=43008"
{"brokers" : [{"id" : "0", "mem" : 43008, "cpus" : 8.0, "heap" : 128, "failover" : {"delay" : "10s", "maxDelay" : "60s"}, "active" : false}]}
```

Starting a broker

```
# curl -X POST "http://localhost:7000/api/broker/start?broker=0"
{"success" : true, "ids" : "0"}
```

Stopping a broker

```
# curl -X POST "http://localhost:7000/api/broker/stop?broker=0"
{"success" : true, "ids" : "0"}
```

Restarting a broker

```
# curl -X POST "http://localhost:7000/api/broker/restart?broker=0"
{"status" : "restarted", "brokers" : [{"task" : {"hostname" : "slave0", "state" : "running", "slaveId" : "fd935975-5db0-4732-bfa4-3063b534972d-S3", "executorId" : "broker-0-a8e0d084-b890-4482-800e-12e72ed7f9ed", "attributes" : {}, "id" : "broker-0-ff11db36-206a-4019-9cd2-6993376831eb", "endpoint" : "slave0:9092"}, "stickiness" : {"period" : "10m", "hostname" : "slave0"}, "bindAddress" : "slave0", "options" : "log.dirs=\/tmp\/kafka\/$id", "id" : "2", "port" : "9092", "constraints" : "hostname=like:slave0", "mem" : 1024, "cpus" : 0.5, "metrics" : {"underReplicatedPartitions" : 0, "offlinePartitionsCount" : 0, "activeControllerCount" : 0, "timestamp" : 1455557472857}, "heap" : 1024, "failover" : {"delay" : "1m", "maxDelay" : "14m"}, "active" : true}]}
```
Removing a broker

```
# curl -X POST "http://localhost:7000/api/broker/remove?broker=0"
{"ids" : "0"}
```

Listing topics
```
# curl "http://localhost:7000/api/topic/list"
{"topics" : [{"name" : "t", "partitions" : {"0" : "0, 1"}, "options" : {"flush.ms": "1000"}}]}
```

Adding topic
```
# curl -X POST "http://localhost:7000/api/topic/add?topic=t"
{"topic" : {"name" : "t", "partitions" : {"0" : "1"}, "options" : {}}}
```

Updating topic
```
# curl -X POST "http://localhost:7000/api/topic/update?topic=t&options=flush.ms%3D1000"
{"topic" : {"name" : "t", "partitions" : {"0" : "0, 1"}, "options" : {"flush.ms" : "1000"}}}
```

Project Goals
==============

* smart broker.id assignment.

* preservation of broker placement (through constraints/stickiness and/or new features).

* ability to-do configuration changes.

* scaling the cluster up and down with automatic, programmatic and manual options.

* smart partition assignment via constraints, roles, resources and attributes.