An open API service indexing awesome lists of open source software.

https://github.com/cookpad/streamy

Basic toolset for hooking into event stream
https://github.com/cookpad/streamy

Last synced: 5 months ago
JSON representation

Basic toolset for hooking into event stream

Awesome Lists containing this project

README

        

# Streamy

[![CI](https://github.com/cookpad/streamy/actions/workflows/ci.yml/badge.svg)](https://github.com/cookpad/streamy/actions/workflows/ci.yml)

## Installation

Add this line to your application's Gemfile:

```ruby
gem "streamy"
```

## Usage

### Broadcasting events

Streamy includes support for two different types of event encoding (JSON and [Avro](https://avro.apache.org/docs/current/spec.html)).

#### Events with JSON encoding

Add this to config/initializer/streamy.rb

```ruby
require "streamy/message_buses/kafka_message_bus"
Streamy.message_bus = Streamy::MessageBuses::KafkaMessageBus.new(
client_id: "streamy",
seed_brokers: "broker.remote:9092",
ssl_ca_certs_from_system: true
)
```

Create an event:

```ruby
module Events
class ReceivedPayment < Streamy::JsonEvent
def topic
"payments.transactions"
end

def body
{
amount: 200
}
end

def event_time
Time.now
end
end
end
```

Publish it:

```ruby
Events::ReceivedPayment.publish
```

#### Events with Avro encoding

Add this to config/initializer/streamy.rb

```ruby
require "streamy/message_buses/kafka_message_bus"
Streamy.message_bus = Streamy::MessageBuses::KafkaMessageBus.new(
client_id: "streamy",
seed_brokers: "broker.remote:9092",
ssl_ca_certs_from_system: true,
)

Streamy.configure do |config|
config.avro_schema_registry_url = "http://registry.example.com",
config.avro_schemas_path = "app/schemas"
end
```

*Default schemas path is "app/schemas"*
*Schema Registry Url is required for encoding with Avro*

Create an event:

```ruby
module Events
class ReceivedPayment < Streamy::AvroEvent
def topic
"payments.transactions"
end

def body
{
amount: 200
}
end

def event_time
Time.now
end
end
end
```

Create Avro schema (`received_payment.asvc`) for event in schema path above:

```json
{
"type": "record",
"name": "received_payment",
"fields": [
{
"name": "type",
"type": "string"
},
{
"name": "event_time",
"type": {
"type": "long",
"logicalType": "timestamp-micros"
}
},
{
"name": "body",
"type": {
"type": "record",
"name": "body",
"fields": [
{
"name": "amount",
"type": ["null", "int"],
"default": null
}
]
}
}
]
}

```

Publish event:

```ruby
Events::ReceivedPayment.publish
```

---

### Consuming events

We use [karafka](https://github.com/karafka/karafka) to handle the bulk of the consumer logic. You can also use [karafka/avro](https://github.com/karafka/avro) to consume Avro encoded events.

Configure karafka consumer:

```rb
class ApplicationConsumer < Karafka::BaseConsumer
def consume
params_batch.each do |message|
Streamy::MessageProcessor.new(message).run
end
end
end
```

Add event handler(s):

```ruby
# app/handlers/received_payment_handler.rb
class ReceivedPaymentHandler
def initialize(body)
@body = message
end

def process
PaymentCounter.increment(body[:amount])
end

private

attr_reader :body
end
```

### Deserialization

Streamy provides an avro deserializer that is used with the schema registry as detailed above. More information on serialization can be found [here](https://github.com/karafka/karafka/wiki/Serialization)

Put the following line in your karafka routes file:

```ruby
deserializer Streamy::Deserializers::AvroDeserializer.new
```

---

## Advanced options

### Event Priority

You can choose a priority for your events. This is done by overriding the `priority` method on your event:

* `:low` - The event will be sent to Kafka by a background thread, events are buffered until `delivery_threshold` messages are waiting or until `delivery_interval` seconds have passed since the last delivery. Calling publish on a low priority event is non blocking, and no errors should be thrown, unless the buffer is full.
* `:standard` (default) - The event will be sent to Kafka by a background thread, but the thread is signaled to send any buffered events as soon as possible. The call to publish is non blocking, and should not throw errors, unless the buffer is full.
* `:essential` - The event will be sent to Kafka immediately. The call to publish is blocking, and may throw errors.
* `:batched` - The event will be queued to send to Kafka using a synchronous producer, but no events are sent until `batched_message_limit` is reached (which is set to `max_buffer_size - 1`), or the synchronous producer in the specific thread has `deliver_messages` called by another service. This allows efficient event batching, when creating many events, e.g. in batch jobs. When a batch of events is being delivered the call to publish will block, and may throw errors.

Please read the `ruby-kafka` notes here on [buffering and error handling](https://github.com/zendesk/ruby-kafka#buffering-and-error-handling)

### Shutdown

To ensure that all `:low` `:batched` or `:standard` priority events are published `Streamy.shutdown` should be called before your process exits to avoid losing any events.
Streamy automatically adds an `at_exit` hook to initiate this, but if you are doing something unusual you might need to be aware of this.

## Testing

Streamy provides a few helpers to make testing a breeze:

### RSpec

```ruby
it "does publish an received payment" do
ReceivedPayment.publish

expect_event(
type: "received_payment",
topic: "payments.transactions",
body: {
amount: 200
}
)
end

it "does not publish an received payment" do
ReceivedPayment.publish

expect_no_event(type: "received_payment")
end
```

### Minitest and TestUnit

```ruby
def test_publish_received_payment
ReceivedPayment.publish

assert_event(
type: "received_payment",
topic: "payments.transactions",
body: {
amount: 200
}
)
end
```

### Testing Avro schemas

Streamy will test your Avro messages against your Avro schemas located in your host application for type errors and schema composition errors. To do this you will need to set up your schema path and registry_url in your specs, and stub any requests to `FakeConfluentSchemaRegistryServer`. Again for example with Webmock and RSpec:

```ruby
RSpec.configure do |config|
config.before(:each) do
require "avro_turf/test/fake_confluent_schema_registry_server"
Streamy.configuration.avro_schema_registry_url = "http://registry.example.com"
Streamy.configuration.avro_schemas_path = "app/events/schemas"
stub_request(:any, /^#{Streamy.configuration.avro_schema_registry_url}/).to_rack(FakeConfluentSchemaRegistryServer)
end
end
```
---

## License

The gem is available as open source under the terms of the [MIT License](http://opensource.org/licenses/MIT).