μ/log is a micro-logging library that logs events and data, not words!
μ, mu (Pronunciation: /mjuː/)
The twelfth letter of the Greek alphabet (Μ, μ), often used as a prefix for mirco- which is 10-6 in the SI (System of Unis). Lowercase letter "u
" is often substituted for "μ
" when the Greek character is not typographically available.
Here some features and key design decisions that make μ/log special:
Available publishers:
Existing logging libraries are based on a design from the 80s and early 90s. Most of the systems at the time where developed in standalone servers where logging messages to console or file was the predominant thing to do. Logging was mostly providing debbugging information and system behavioural introspection.
Most of modern systems are distributed in virtualized machines that live in the cloud. These machines could disappear any time. In this context logging on the local file system isn't useful as logs are easily lost if virtual machines are destroyed. Therefore it is common practice to use log collectors and centralized log processors. The ELK stack it has been predominant in this space for years, but there are a multitude of other commercial and open-source products.
Most of these systems have to deal with non structured data represented as formatted strings in files. The process of extracting information out of these strings is very tedious, error prone, and definitely not fun. But the question is: why did we encode these as strings in the first place? This is just because existing log frameworks, which have been redesigned in various decades follow the same structure as when systems lived on the same single server for decades.
I believe we need the break free of these anachronistic design and use event loggers, not message loggers, which are designed for dynamic distributed systems living in cloud and using centralized log aggregators. So here is μ/log designed for this very purpose.
In order to use the library add the dependency to your project.clj
;; Leiningen project
[com.brunobonacci/mulog "0.1.6"]
;; deps.edn format
{:deps { com.brunobonacci/mulog "0.1.6" }}
Then require the namespace:
(ns your-ns
(:require [com.brunobonacci.mulog :as μ]))
;; or for the more ASCII traditionalists
(ns your-ns
(:require [com.brunobonacci.mulog :as u]))
Check the online documentation
The instrument your code with the log you deem useful. Then general structure is
(μ/log event-name, key1 value1, key2 value2, ... keyN valueN)
For example:
;; good to use namespaced keywords for the event-name
(μ/log ::hello :to "New World!")
However you will NOT be able to see any events until you add a publisher which it will take your events and send to a distributed logger of your local console (if you are developing).
You can add as many key-value pairs as you deem useful to express the event in your system
(μ/start-publisher! {:type :console})
At this point you should be able to see the previous event in your REPL terminal and it will look as follow:
{:mulog/timestamp 1572707670555, :mulog/event-name :your-ns/hello, :mulog/namespace "your-ns", :to "New World!"}
Here some example of events you might want to log.
(μ/log ::system-started :version "0.1.0" :init-time 32)
(μ/log ::user-logged :user-id "1234567" :remote-ip "1.2.3.4" :auth-method :password-login)
(μ/log ::http-request :path "/orders", :method :post, :remote-ip "1.2.3.4", :http-status 201, :request-time 129)
(μ/log ::invalid-request :exception x, :user-id "123456789", :items-requested 47)
(μ/log ::position-updated :poi "1234567" :location {:lat 51.4978128, :lng -0.1767122 )
All above are examples of events you might want to track, collect and aggregate on it in a specialized timeseries database.
Adding events which are rich in attributes and dimensions is extremely useful, however it is not easy to have all the attributes and dimensions at your disposal everywhere in the code. To get around this problem μ/log supports the use of context.
There are two levels of context, a global level and a local one.
The global context allows you to define properties and values which will be added to all the events logged afterwards.
For example:
(μ/log ::system-started :init-time 32)
;; {:mulog/timestamp 1572709206048, :mulog/event-name :your-ns/system-started, :mulog/namespace "your-ns", :init-time 32}
;; set global context
(μ/set-global-context! {:app-name "mulog-demo", :version "0.1.0", :env "local"})
(μ/log ::system-started :init-time 32)
;; {:mulog/timestamp 1572709332340,
;; :mulog/event-name :your-ns/system-started,
;; :mulog/namespace "your-ns",
;; :app-name "mulog-demo",
;; :version "0.1.0",
;; :env "local",
;; :init-time 32}
Typically, you will set the global context once in your main function
at the starting of your application with properties which are valid
for all events emitted by the process. Use set-global-context!
to
specify a given value, or update-global-context!
with a update
function to change some of the values. Examples of properties you
should consider adding in the global context are app-name
,
version
, environment
, process-id
, host-ip
, os-type
,
jvm-version
etc etc
The second type of context is the (thread) local context. It can be used to inject information about the current processing and all the events withing the scope of the context will inherit the properties and their values.
For example the following line will contain all the properties of the global context, all the properties of the local context and all inline properties.
(μ/with-context {:order "abc123"}
(μ/log ::item-processed :item-id "sku-123" :qt 2))
;; {:mulog/timestamp 1572711123826,
;; :mulog/event-name :your-ns/item-processed,
;; :mulog/namespace "your-ns",
;; :app-name "mulog-demo",
;; :version "0.1.0",
;; :env "local",
;; :order "abc123",
;; :item-id "sku-123",
;; :qt 2}
The local context can be nested:
(μ/with-context {:transaction-id "tx-098765"}
(μ/with-context {:order "abc123"}
(μ/log ::item-processed :item-id "sku-123" :qt 2)))
;; {:mulog/timestamp 1572711123826,
;; :mulog/event-name :your-ns/item-processed,
;; :mulog/namespace "your-ns",
;; :app-name "mulog-demo",
;; :version "0.1.0",
;; :env "local",
;; :transaction-id "tx-098765",
;; :order "abc123",
;; :item-id "sku-123",
;; :qt 2}
Local context works across function boundaries:
(defn process-item [sku quantity]
;; ... do something
(u/log ::item-processed :item-id "sku-123" :qt quantity)
;; ... do something
)
(μ/with-context {:order "abc123"}
(process-item "sku-123" 2))
;; {:mulog/timestamp 1572711818791,
;; :mulog/event-name :your-ns/item-processed,
;; :mulog/namespace "your-ns",
;; :app-name "mulog-demo",
;; :version "0.1.0",
;; :env "local",
;; :order "abc123",
;; :item-id "sku-123",
;; :qt 2}
Here some best practices to follow while logging events:
event-name
Publishers allow to send the events to external system where they can be stored, indexed, transformed or visualized.
It outputs the events into the standard output in EDN format, mostly intended for local development.
The available configuration options:
{:type :console
;; a function to apply to the sequence of events before publishing.
;; This transformation function can be used to filter, tranform,
;; anonymise events before they are published to a external system.
;; by defatult there is no transformation. (since v0.1.8)
:transform identity
}
How to use it:
(μ/start-publisher! {:type :console})
It sends the output of each log into a file in EDN format.
The available configuration options:
{:type :simple-file
;; the name of the file, including the path, where the logs will be written
;; If the directory doesn't exists, it will try to create them, same for the file.
;; If the file already exists, it will append the new events.
:filename "/tmp/mulog/events.log"
;; a function to apply to the sequence of events before publishing.
;; This transformation function can be used to filter, tranform,
;; anonymise events before they are published to a external system.
;; by defatult there is no transformation. (since v0.1.8)
:transform identity
}
How to use it:
(μ/start-publisher! {:type :simple-file :filename "/tmp/mulog/events.log"})
The multi publisher allows you to define multiple publishers
configuration all in one place. It is equivalent to calling
μ/start-publisher!
on all the individual configurations, it is just
provided for ease of use.
;; it will initialize all the configured publishers
(μ/start-publisher!
{:type :multi
:publishers
[{:type :console}
{:type :simple-file :filename "/tmp/disk1/mulog/events1.log"}
{:type :simple-file :filename "/tmp/disk2/mulog/events2.log"}]}))
It will initialize all the configured publishers and return a function with no arguments which when called will stop all the publishers.
The events must be serializeable in JSON format (Cheshire)
The available configuration options:
{:type :elasticsearch
;; ElasticSearch endpoint (REQUIRED)
:url "http://localhost:9200/"
;; the maximum number of events which can be sent in a single
;; batch request to ElasticSearch
:max-items 5000
;; Interval in milliseconds between publish requests.
;; μ/log will try to send the records to ElasticSearch
;; with the interval specified.
:publish-delay 5000
;; The index pattern to use for the events
:index-pattern "'mulog-'yyyy.MM.dd"
;; Whether or not to change the attribute names
;; to facilitate queries and avoid type clashing
:name-mangling true
;; a function to apply to the sequence of events before publishing.
;; This transformation function can be used to filter, tranform,
;; anonymise events before they are published to a external system.
;; by defatult there is no transformation. (since v0.1.8)
:transform identity
}
How to use it:
(μ/start-publisher!
{:type :elasticsearch
:url "http://localhost:9200/"})
Supported versions: 6.7+
, 7.x
Read more on Elasticsearch name mangling here.
The events must be serializeable in JSON format (Cheshire)
The available configuration options:
{:type :kafka
;; kafka configuration
:kafka {;; the comma-separated list of brokers (REQUIRED)
:bootstrap.servers "localhost:9092"
;; you can add more kafka connection properties here
}
;; the name of the kafka topic where events will be sent
;; :topic "mulog"
;; maximum number of events in a single batch
;; :max-items 1000
;; how often it will send the events Kafka (in millis)
;; :publish-delay 1000
;; the format of the events to send into the topic
;; can be one of: :json, :edn (default :json)
;; :format :json
;; The name of the field which it will be used as partition key
;; the :puid is the process unique identifier which can be injected
;; as global context
;; :key-field :puid
;; a function to apply to the sequence of events before publishing.
;; This transformation function can be used to filter, tranform,
;; anonymise events before they are published to a external system.
;; by defatult there is no transformation. (since v0.1.8)
:transform identity
}
How to use it:
(μ/start-publisher!
{:type :kafka
:kafka {:bootstrap.servers "localhost:9092"}})
To use your own publisher create a function which take a configuration
and return an instance of PPublisher
protocol and then use the
:custom
dynamic loader. Ensure that the jar is added to the
classpath and then just add the fully qualified function name:
(μ/start-publisher!
{:type :custom
:fqn-function "my-namespace.publisher/my-custom-publisher"
;; add here additional configuration options which will be passed
;; to the custom publisher.
})
For more information about how to implement custom publisher see: How to write custom publishers
Coming soon:
PRs are welcome ;-)
Copyright © 2019-2020 Bruno Bonacci - Distributed under the Apache License v2.0
Can you improve this documentation?Edit on GitHub
cljdoc is a website building & hosting documentation for Clojure/Script libraries
× close