Ziggurat is a framework built to simplify Stream processing on Kafka. It can be used to create a full-fledged Clojure app that reads and processes messages from Kafka. Ziggurat is built with the intent to abstract out the following features
- reading messages from Kafka
- retrying failed messages via RabbitMQ
- setting up an HTTP server
Refer concepts to understand the concepts referred to in this document.
(For mac users only)
Install Clojure: brew install clojure
Install leiningen: brew install leiningen
Run docker-compose: docker-compose up
. This starts
Run tests: make test
make setup-cluster
This clears up the volume and starts
make test-cluster
config.test.cluster.edn
instead of config.test.edn
Add this to your project.clj
[tech.gojek/ziggurat "4.8.0"]
Please refer clojars for the latest stable version
To start a stream (a thread that reads messages from Kafka), add this to your core namespace.
(require '[ziggurat.init :as ziggurat])
(defn start-fn []
;; your logic that runs at startup goes here
)
(defn stop-fn []
;; your logic that runs at shutdown goes here
)
(defn main-fn
[{:keys [message metadata] :as message-payload}]
(println message)
:success)
(def handler-fn
(-> main-fn
(middleware/protobuf->hash ProtoClass :stream-id)))
;; Here ProtoClass refers to the fully qualified name of the Java class which the code is used to de-serialize the message.
(ziggurat/main start-fn stop-fn {:stream-id {:handler-fn handler-fn}})
NOTE: this example assumes that the message is serialized in Protobuf format
Please refer the Middleware section for understanding handler-fn
here.
:stream-id
is a unique identifier per stream. All configs, queues and metrics will be namespaced under this id.(ziggurat/main start-fn stop-fn {:stream-id-1 {:handler-fn handler-fn-1}
:stream-id-2 {:handler-fn handler-fn-2}})
(require '[ziggurat.init :as ziggurat])
(defn start-fn []
;; your logic that runs at startup goes here
)
(defn stop-fn []
;; your logic that runs at shutdown goes here
)
(defn api-handler [_request]
{:status 200
:headers {"Content-Type" "application/json"}
:body (get-resource)})
(def routes [["v1/resources" {:get api-handler}]])
(defn main-fn
[{:keys [message metadata] :as message-payload}]
(println message)
:success)
(def handler-fn
(-> main-fn
(middleware/protobuf->hash ProtoClass :stream-id)))
(ziggurat/main start-fn stop-fn {:stream-id {:handler-fn handler-fn}} routes)
NOTE: this example assumes that the message is serialized in Protobuf format
Ziggurat also sets up a HTTP server by default and you can pass in your own routes that it will serve. The above example demonstrates how you can pass in your own route.
or
(ziggurat/main {:start-fn start-fn
:stop-fn stop-fn
:stream-routes {:stream-id {:handler-fn main-fn}}
:actor-routes routes
:modes [:api-server :stream-worker]})
This will start both api-server and stream-worker modes
There are four modes supported by ziggurat
:api-server - Mode by which only server will be started with actor routes and management routes(Dead set management)
:stream-worker - Only start the server plus rabbitmq for only producing the messages for retry and channels
:worker - Starts the rabbitmq consumer for retry and channel
:management-api - Servers only routes which used for deadset management
You can pass in multiple modes and it will start accordingly If nothing passed to modes then it will start all the modes.
Feature implementation of issue #56. Stop and start streams on a running process using nREPL. A nREPL server starts at port 7011
(default) when an actor using ziggurat starts. Check ZIGGURAT_NREPL_SERVER_PORT
in your config.
Connect to the shell using
lein repl :connect <host>:<port>
The functions can be accessed via the following commands to stop and start streams using their topic-entity
> (ziggurat.streams/stop-stream :booking)
> (ziggurat.streams/start-stream :booking)
where booking
is the topic-entity
Version 3.0.0 of Ziggurat introduces the support of Middleware. Old versions of Ziggurat (< 3.0) assumed that the messages read from kafka were serialized in proto-format and thus it deserialized them and passed a clojure map to the mapper-fn. We have now pulled the deserialization function into a middleware and users have the freedom to use this function to deserialize their messages or define their custom middlewares. This enables ziggurat to process messages serialized in any format.
The default middleware default/protobuf->hash
assumes that the message is serialized in proto format.
(require '[ziggurat.init :as ziggurat])
(defn start-fn []
;; your logic that runs at startup goes here
)
(defn stop-fn []
;; your logic that runs at shutdown goes here
)
(defn main-fn
[{:keys [message metadata] :as message-payload}]
(println message)
:success)
(defn wrap-middleware-fn
[mapper-fn :stream-id]
(fn [message]
(println "processing message for stream: " :stream-id)
(mapper-fn (deserialize-message message))))
(def handler-fn
(-> main-fn
(wrap-middleware-fn :stream-id)))
(ziggurat/main start-fn stop-fn {:stream-id {:handler-fn handler-fn}})
The handler-fn gets a serialized message from kafka and thus we need a deserialize-message function. We have provided default deserializers in Ziggurat
Ziggurat 3.1.0 provides a middleware to deserialize JSON messages, along with proto. It can be used like this.
(def message-handler-fn
(-> actual-message-handler-function
(parse-json :stream-route-config)))
Here, message-handler-fn
calls parse-json
with a message handler function
actual-message-handler-function
as the first argument and the key of a stream-route
config (as defined in config.edn
) as the second argument.
To enable publishing data to kafka, Ziggurat provides producing support through ziggurat.producer namespace. This namespace defines methods for publishing data to Kafka topics. The methods defined here are essentially wrapper around variants of send
methods defined in org.apache.kafka.clients.producer.KafkaProducer
.
At the time of initialization, an instance of org.apache.kafka.clients.producer.KafkaProducer
is constructed using config values provided in resources/config.edn
. A producer can be configured for each of the stream-routes in config.edn. Please see the example below.
At present, only a few configurations are supported for constructing KafkaProducer. These have been explained here. Please see Producer configs for a complete list of all producer configs available in Kafka.
Ziggurat.producer namespace defines a multi-arity send
function which is a thin wrapper around KafkaProducer#send. This method publishes data to a Kafka topic through a Kafka producer
defined in the stream router configuration. See configuration section below.
E.g.
For publishing data using a producer which is defined for the stream router config with key :default
, use send like this:
(send :default "test-topic" "key" "value")
(send :default "test-topic" 1 "key" "value")
With Ziggurat version 3.5.1, both Kafka Streams API and Kafka Consumer API can be used to consume the messages in real time. Kafka Consumer API is an efficient way to consume messages from high throughput Kafka topics.
With Kafka Streams API, one message is processed at a time. But, with Kafka Consumer API integration in Ziggurat, a user can consume messages in bulk and can control how many messages it wants to consume at a time. This batch size can be configured using max-poll-records config https://docs.confluent.io/current/installation/configuration/consumer-configs.html#max.poll.records.
Like Streams, Ziggurat also provides the facility to specify multiple batch routes.
:batch-routes {:restaurants-updates-to-non-personalized-es
{:consumer-group-id "restaurants-updates-consumer"
:bootstrap-servers "g-gojek-id-mainstream.golabs.io:6668"
:origin-topic "restaurant-updates-stream"}}
A full list of supported configs is given below. These configs can be added to config.edn
as per the requirements.
(defn -main [& args]
(init/main {:start-fn start
:stop-fn stop
:stream-routes {:booking {:handler-fn (stream-deserializer/protobuf->hash
stream-handler
BookingLogMessage
:booking)}}
:batch-routes {:batch-consumer-1 {:handler-fn (batch-deserialzer/deserialize-batch-of-proto-messages
batch-handler
BookingLogKey
BookingLogMessage
:batch-consumer-1)}}
:actor-routes [["v1/hello" {:get get-hello}]]}))
(defn- single-message-handler
[message]
(log/info "Batch Message: " message))
(defn batch-handler
[batch]
(log/infof "Received a batch of %d messages" (count batch))
(doseq [single-message batch]
(single-message-handler single-message))
(if (retry?)
(do (log/info "Retrying the batch..")
{:retry batch :skip []})
{:retry [] :skip []}))
Ziggurat Config | Default Value | Description | Mandatory? |
---|---|---|---|
:bootstrap-servers | NA | https://kafka.apache.org/documentation/#bootstrap.servers | Yes |
:consumer-group-id | NA | https://kafka.apache.org/documentation/#group.id | Yes |
:origin-topic | NA | Kafka Topic to read data from | Yes |
:max-poll-records | 500 | https://kafka.apache.org/documentation/#max.poll.records | No |
:session-timeout-ms-config | 60000 | https://kafka.apache.org/documentation/#session.timeout.ms | No |
:key-deserializer-class-config | "org.apache.kafka.common.serialization.ByteArrayDeserializer" | https://kafka.apache.org/documentation/#key.deserializer | No |
:value-deserializer-class-config | "org.apache.kafka.common.serialization.ByteArrayDeserializer" | https://kafka.apache.org/documentation/#value.deserializer | No |
:poll-timeout-ms-config | 1000 | Timeout value used for polling with a Kafka Consumer | No |
:thread-count | 2 | Number of Kafka Consumer threads for each batch-route | No |
:default-api-timeout-ms | 60000 | https://cwiki.apache.org/confluence/display/KAFKA/KIP-266%3A+Fix+consumer+indefinite+blocking+behavior | No |
config.edn
{:ziggurat {:messaging {:constructor "ziggurat.messaging.rabbitmq-cluster-wrapper/->RabbitMQMessaging"
:rabbit-mq-connection {:hosts "g-lambda-lambda-rabbitmq-a-01,g-lambda-lambda-rabbitmq-a-02,g-lambda-lambda-rabbitmq-a-03"
:port [5672 :int]
:prefetch-count [3 :int]
:username "guest"
:password "guest"
:channel-timeout [2000 :int]
:address-resolver [:dns :keyword] ;;possible values [:dns :ip-list]. Default is :dns
:channel-pool {:max-wait-ms [5000 :int]
:min-idle [10 :int]
:max-idle [20 :int]}
:publish-retry {:back-off-ms 5000
:non-recoverable-exception {:enabled true
:back-off-ms 1000
:count 5}}}}}}
:hosts
is a comma separated values of RabbitMQ hostnames (dns-names OR IPs).:port
specifies the port number on which the RabbitMQ nodes are running.:channel-pool
specifies the properties for the RabbitMQ channel pool used for publishing:address-resolver
specifies the strategy to figure out RabbitMQ hosts IP addresses. :dns
is the default and shoud
be used when :hosts
specifies a DNS address. :ip-list
should be used when comma separated IPs are provided.:publish-retry
defines the config for recoverable and non-recoverable exceptions.
:back-off-ms
- defines the time period after which a retry should happen:enabled
- defines whether retries should happen:back-off-ms
- defines the time period after which a retry should happen:count
- defines the number of retriesIn addition to linear retries, Ziggurat users can now use exponential backoff strategy for retries. This means that the message
timeouts after every retry increase by a factor of 2. So, if your configured timeout is 100ms the backoffs will have timeouts as
200, 300, 700, 1500 ..
. These timeouts are calculated using the formula (queue-timeout-ms * ((2**exponent) - 1))
where exponent
falls in this range [1,(min 25, configured-retry-count)]
.
The number of retries possible in this case are capped at 25.
The number of queues created in the RabbitMQ are equal to the configured-retry-count or 25, whichever is smaller.
Exponential retries can be configured as described below.
:ziggurat {:stream-router {:default {:application-id "application_name"...}}}
:retry {:type [:exponential :keyword]
:count [10 :int]
:enable [true :bool]}
Exponential retries can be configured for channels too. Additionally, a user can specify a custom queue-timeout-ms
value per channel.
Timeouts for exponential backoffs are calculated using queue-timeout-ms
. This implies that each channel can have separate count of retries
and different timeout values.
:ziggurat {:stream-router {:default {:application-id "application_name"...
:channels {:channel-1 .....
:retry {:type [:exponential :keyword]
:count [10 :int]
:queue-timeout-ms 2000
:enable [true :bool]}}}}}
Open Tracing enables to identify the amount of time spent in various stages of the work flow.
Currently, the execution of the handler function is traced. If the message consumed has the corresponding tracing headers, then the E2E life time of the message from the time of production till the time of consumption can be traced.
Tracing has been added to the following flows:
By default, tracing is done via Jaeger based on the env configs. Please refer Jaeger Configuration
and Jaeger Architecture to set the respective env variables.
To enable custom tracer, a custom tracer provider function name can be set in :custom-provider
. The corresponding function will be executed in runtime to create a tracer. In the event of any errors while executing the custom tracer provider, a Noop tracer will be created.
To enable tracing, the following config needs to be added to the config.edn
under :ziggurat
key.
:tracer {:enabled [true :bool]
:custom-provider ""}
Example Jaeger Env Config:
JAEGER_SERVICE_NAME: "service-name"
JAEGER_AGENT_HOST: "localhost"
JAEGER_AGENT_PORT: 6831
As of Ziggurat version 3.13.0, all the official Kafka configs Kafka configurations for Streams API, Consumer API and Producer API are supported.
All Ziggurat configs should be in your clonfig
config.edn
under the :ziggurat
key.
{:ziggurat {:app-name "application_name"
:nrepl-server {:port [7011 :int]}
:stream-router {:stream-id {:application-id "kafka_consumer_id"
:bootstrap-servers "kafka-broker-1:6667,Kafka-broker-2:6667"
:stream-threads-count [1 :int]
:origin-topic "kafka-topic-*"
:oldest-processed-message-in-s [604800 :int]
:changelog-topic-replication-factor [3 :int]
:stream-thread-exception-response [:shutdown-client :keyword]
:producer {:bootstrap-servers "localhost:9092"
:acks "all"
:retries-config 5
:max-in-flight-requests-per-connection 5
:enable-idempotence false
:value-serializer "org.apache.kafka.common.serialization.StringSerializer"
:key-serializer "org.apache.kafka.common.serialization.StringSerializer"}}}
:batch-routes {:restaurants-updates-to-non-personalized-es
{:consumer-group-id "restaurants-updates-consumer"
:bootstrap-servers "g-gojek-id-mainstream.golabs.io:6668"
:origin-topic "restaurant-updates-stream"}}
:ssl {:enabled true
:ssl-keystore-location "/location/to/keystore"
:ssl-keystore-password "some-password"
{:jaas {:username "username"
:password "password"
:mechanism "SCRAM_SHA-512"}}}
:default-api-timeout-ms-config [600000 :int]
:statsd {:host "localhost"
:port [8125 :int]
:enabled [false :bool]}
:statsd {:host "localhost"
:port [8125 :int]
:enabled [false :bool]}
:sentry {:enabled [false :bool]
:dsn "dummy"
:worker-count [5 :int]
:queue-size [5 :int]
:thread-termination-wait-s [1 :int]}
:rabbit-mq-connection {:host "localhost"
:port [5672 :int]
:prefetch-count [3 :int]
:username "guest"
:password "guest"
:channel-timeout [2000 :int]}
:rabbit-mq {:delay {:queue-name "application_name_delay_queue"
:exchange-name "application_name_delay_exchange"
:dead-letter-exchange "application_name_instant_exchange"
:queue-timeout-ms [5000 :int]}
:instant {:queue-name "application_name_instant_queue"
:exchange-name "application_name_instant_exchange"}
:dead-letter {:queue-name "application_name_dead_letter_queue"
:exchange-name "application_name_dead_letter_exchange"}}
:retry {:count [5 :int]
:enabled [false :bool]}
:jobs {:instant {:worker-count [4 :int]
:prefetch-count [4 :int]}}
:http-server {:port [8010 :int]
:graceful-shutdown-timeout-ms [30000 :int]
:new-relic {:report-errors [false :bool]}}}}
app-name - Refers to the name of the application. Used to namespace queues and metrics.
nrepl-server - Port on which the repl server will be hosted
default-api-timeout-ms-config - Specifies the timeout (in milliseconds) for client APIs. This configuration is used as the default timeout for all client operations that do not specify a timeout parameter. The recommended value for Ziggurat based apps is 600000 ms (10 minutes).
stream-router - Configs related to all the Kafka streams the application is reading from
batch-routes - This has been explained in the Batch Routes section above. All the properties provided with Kafka Consumer Config are accepted as kebab case keywords
ssl - All Kafka SSL configs and SASL configs can be provided as kebab case keywords. These configs are automatically applied to all kafka stream, kafka producer and kafka consumer objects created in Ziggurat.
statsd - Formerly known as datadog, The statsd host and port that metrics should be sent to.
sentry - Whenever a :failure keyword is returned from the mapper-function or an exception is raised while executing the mapper-function, an event is sent to sentry. You can skip this flow by disabling it.
rabbit-mq-connection - The details required to make a connection to rabbitmq. We use rabbitmq for the retry mechanism.
rabbit-mq - The queues that are part of the retry mechanism
retry - The number of times the message should be retried and if retry flow should be enabled or not
jobs - The number of consumers that should be reading from the retry queues and the prefetch count of each consumer
http-server - Ziggurat starts an http server by default and gives apis for ping health-check and deadset management. This defines the port and the number of threads of the http server. It also controls the graceful shutdown timeout of the HTTP server. Default is 30000ms
new-relic - If report-errors is true, whenever a :failure keyword is returned from the mapper-function or an exception is raised while executing it, an error is reported to new-relic. You can skip this flow by disabling it.
Copyright 2018, GO-JEK Tech <http://gojek.tech>
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
Can you improve this documentation? These fine people already did:
Anirudh, prateek.khatri, Kartik Gupta, Michael Angelo Calimlim, shubhang.balkundi, Rooba Limsa, mjayprateek, anmol1vw13, Sandilya Jandhyala, Anirudh Vyas, Saptanto Sindu, Wito Chandra, Prateek Khatri, Shubhang Balkundi, Lakshya Gupta, Anmol Vijaywargiya, Gurupratap, WickedBrat, williamChang80, led & s7saxenaEdit on GitHub
cljdoc is a website building & hosting documentation for Clojure/Script libraries
× close