Add this line to your application's Gemfile:
gem "streamy"Streamy includes support for two different types of event encoding (JSON and Avro).
Add this to config/initializer/streamy.rb
require "streamy/message_buses/kafka_message_bus"
Streamy.message_bus = Streamy::MessageBuses::KafkaMessageBus.new(
client_id: "streamy",
seed_brokers: "broker.remote:9092",
ssl_ca_certs_from_system: true
)Create an event:
module Events
class ReceivedPayment < Streamy::JsonEvent
def topic
"payments.transactions"
end
def body
{
amount: 200
}
end
def event_time
Time.now
end
end
endPublish it:
Events::ReceivedPayment.publishAdd this to config/initializer/streamy.rb
require "streamy/message_buses/kafka_message_bus"
Streamy.message_bus = Streamy::MessageBuses::KafkaMessageBus.new(
client_id: "streamy",
seed_brokers: "broker.remote:9092",
ssl_ca_certs_from_system: true,
)
Streamy.configure do |config|
config.avro_schema_registry_url = "http://registry.example.com",
config.avro_schemas_path = "app/schemas"
end Default schemas path is "app/schemas" Schema Registry Url is required for encoding with Avro
Create an event:
module Events
class ReceivedPayment < Streamy::AvroEvent
def topic
"payments.transactions"
end
def body
{
amount: 200
}
end
def event_time
Time.now
end
end
endCreate Avro schema (received_payment.asvc) for event in schema path above:
{
"type": "record",
"name": "received_payment",
"fields": [
{
"name": "type",
"type": "string"
},
{
"name": "event_time",
"type": {
"type": "long",
"logicalType": "timestamp-micros"
}
},
{
"name": "body",
"type": {
"type": "record",
"name": "body",
"fields": [
{
"name": "amount",
"type": ["null", "int"],
"default": null
}
]
}
}
]
}
Publish event:
Events::ReceivedPayment.publishWe use karafka to handle the bulk of the consumer logic. You can also use karafka/avro to consume Avro encoded events.
Configure karafka consumer:
class ApplicationConsumer < Karafka::BaseConsumer
def consume
params_batch.each do |message|
Streamy::MessageProcessor.new(message).run
end
end
endAdd event handler(s):
# app/handlers/received_payment_handler.rb
class ReceivedPaymentHandler
def initialize(body)
@body = message
end
def process
PaymentCounter.increment(body[:amount])
end
private
attr_reader :body
endStreamy provides an avro deserializer that is used with the schema registry as detailed above. More information on serialization can be found here
Put the following line in your karafka routes file:
deserializer Streamy::Deserializers::AvroDeserializer.newYou can choose a priority for your events. This is done by overriding the priority method on your event:
:low- The event will be sent to Kafka by a background thread, events are buffered untildelivery_thresholdmessages are waiting or untildelivery_intervalseconds have passed since the last delivery. Calling publish on a low priority event is non blocking, and no errors should be thrown, unless the buffer is full.:standard(default) - The event will be sent to Kafka by a background thread, but the thread is signaled to send any buffered events as soon as possible. The call to publish is non blocking, and should not throw errors, unless the buffer is full.:essential- The event will be sent to Kafka immediately. The call to publish is blocking, and may throw errors.:batched- The event will be queued to send to Kafka using a synchronous producer, but no events are sent untilbatched_message_limitis reached (which is set tomax_buffer_size - 1), or the synchronous producer in the specific thread hasdeliver_messagescalled by another service. This allows efficient event batching, when creating many events, e.g. in batch jobs. When a batch of events is being delivered the call to publish will block, and may throw errors.
Please read the ruby-kafka notes here on buffering and error handling
To ensure that all :low :batched or :standard priority events are published Streamy.shutdown should be called before your process exits to avoid losing any events.
Streamy automatically adds an at_exit hook to initiate this, but if you are doing something unusual you might need to be aware of this.
Streamy provides a few helpers to make testing a breeze:
it "does publish an received payment" do
ReceivedPayment.publish
expect_event(
type: "received_payment",
topic: "payments.transactions",
body: {
amount: 200
}
)
end
it "does not publish an received payment" do
ReceivedPayment.publish
expect_no_event(type: "received_payment")
enddef test_publish_received_payment
ReceivedPayment.publish
assert_event(
type: "received_payment",
topic: "payments.transactions",
body: {
amount: 200
}
)
endStreamy will test your Avro messages against your Avro schemas located in your host application for type errors and schema composition errors. To do this you will need to set up your schema path and registry_url in your specs, and stub any requests to FakeConfluentSchemaRegistryServer. Again for example with Webmock and RSpec:
RSpec.configure do |config|
config.before(:each) do
require "avro_turf/test/fake_confluent_schema_registry_server"
Streamy.configuration.avro_schema_registry_url = "http://registry.example.com"
Streamy.configuration.avro_schemas_path = "app/events/schemas"
stub_request(:any, /^#{Streamy.configuration.avro_schema_registry_url}/).to_rack(FakeConfluentSchemaRegistryServer)
end
endThe gem is available as open source under the terms of the MIT License.