Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/camathieu/flume-ng-kafka-sink
flume-ng kafka sink
https://github.com/camathieu/flume-ng-kafka-sink
Last synced: 3 months ago
JSON representation
flume-ng kafka sink
- Host: GitHub
- URL: https://github.com/camathieu/flume-ng-kafka-sink
- Owner: camathieu
- License: apache-2.0
- Created: 2014-01-15T21:32:13.000Z (almost 11 years ago)
- Default Branch: master
- Last Pushed: 2014-08-26T15:34:13.000Z (over 10 years ago)
- Last Synced: 2024-04-16T19:18:17.826Z (9 months ago)
- Language: Java
- Size: 204 KB
- Stars: 6
- Watchers: 2
- Forks: 5
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
README
flume-ng-kafka-sink
===================Apache FlumeNG sink to push events to Kafka 0.8+
topic selection :
- statically via sink configuration
- dynamically via event header "topic"partition key :
- from event header "key"By default ( no "key" header ) event are spread
in a round robin fashion across topic's partitions.With the default kafka partitioner event with the same key
will always go to the same partition and therefore to the same consumer.
you can override this behaviour by providing a partitioner.class to
the kafka server configurationBuild/Install
-------------Build flume-ng-kafka-sink
mvn package
Configuration
-------------agent.sources = httpSource
agent.channels = memoryChannel
agent.sinks = kafkaSink# Source configuration
agent.sources.httpSource.type = http
agent.sources.httpSource.channels = memoryChannel
agent.sources.httpSource.port = 8026
agent.sources.httpSource.handler = org.apache.flume.source.http.JSONHandler# Sink configuration
agent.sinks.kafkaSink.type = org.apache.flume.sink.kafka.KafkaSink
agent.sinks.kafkaSink.channel = memoryChannel
# If omitted topic must be provided by event header topic
agent.sinks.kafkaSink.topic = flume
agent.sinks.kafkaSink.batch.size = 100
# Producer configuration
agent.sinks.kafkaSink.metadata.broker.list = localhost:9092
agent.sinks.kafkaSink.serializer.class = kafka.serializer.DefaultEncoder
agent.sinks.kafkaSink.producer.type = async
agent.sinks.kafkaSink.batch.num.messages = 100
agent.sinks.kafkaSink.queue.buffering.max.ms = 100Test
----
curl -X POST -d "[{"headers":{"topic":"test","key":"key"},"body":'{"foo":"bar"}'}]" http://localhost:8026Reporting
---------This sink follow the standard flume reporting interface
To enable json reporting you may start flume with the following options
-Dflume.monitoring.type=http -Dflume.monitoring.port=34545Just get /metrics to get them
curl http://localhost:34545/metrics 2> /dev/null | python -mjson.tool
{
"CHANNEL.memoryChannel": {
"ChannelCapacity": "10000",
"ChannelFillPercentage": "0.0",
"ChannelSize": "0",
"EventPutAttemptCount": "5783",
"EventPutSuccessCount": "5783",
"EventTakeAttemptCount": "5805",
"EventTakeSuccessCount": "5783",
"StartTime": "1404126896721",
"StopTime": "0",
"Type": "CHANNEL"
},
"SINK.kafkaSink": {
"BatchCompleteCount": "57",
"BatchEmptyCount": "21",
"BatchUnderflowCount": "1",
"ConnectionClosedCount": "0",
"ConnectionCreatedCount": "0",
"ConnectionFailedCount": "0",
"EventDrainAttemptCount": "5783",
"EventDrainSuccessCount": "5783",
"StartTime": "1404126896722",
"StopTime": "0",
"Type": "SINK"
},
"SOURCE.httpSource": {
"AppendAcceptedCount": "0",
"AppendBatchAcceptedCount": "5783",
"AppendBatchReceivedCount": "5783",
"AppendReceivedCount": "0",
"EventAcceptedCount": "5783",
"EventReceivedCount": "5783",
"OpenConnectionCount": "0",
"StartTime": "1404126896808",
"StopTime": "0",
"Type": "SOURCE"
}
}