Skip to content

ETCDEVTeam/geth-ELK

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

48 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Geth + Elasticsearch

Please note that this document that the geth mlog APIs referenced herein are still (as of 24 Nov 2017) in an experimental stage and should not be relied on for production purposes.

If you're running Geth classic you're likely to see some console logs something like this:

$ geth --vmodule="eth/fetcher/*=6,eth/downloader/*=6,eth/peer.go=6,eth/handler.go=5,core/database_util.go=6,core/blockchain.go=6" --log-dir="mainnet/log" --log-status="sync=10" --rpc --ws

geth-logs

These are logs whose style and intent trace back to, well, traces, debugging, developer feedback, and event updates. While useful for developers to track and understand program decision making on a micro level, when you see thousands of lines like these flying across your screen, it can be not a little tricky to make sense of the "big picture" and general performance of your node.

One option for a simpler and more easily grokable output would be to use a few flags. Here we can enable sync status logs to print every 20 seconds, and turn down normal event reporting to just show errors (1).

$ geth --log-status="sync=20" --verbosity 1

geth-logs-status

While in some cases this might fit the bill, it also has it's own shortcomings: hard to debug, single-dimensioned, exclusively macro-level, and any performance analysis will require a calculator.

This is where geth's mlog API comes in. MLog stands for "machine log," and it's job is to make logs that computers can read and readily analyse. Sometimes these are also called structured logs.

Once you tell geth to start churning out mlogs, you can then fire them over to the Elastic stack, and open the door to something like this:

Blockchain & Sync:

dash-sync

Network & Peers:

dash-p2p

Miners & Mining:

dash-mining


Big picture

Once we get geth's mlogs up and running, we're going to point Filebeat at them to watch, read, and send off to Logstash. Once Logstash hears about some incoming logs from Filebeat, Logstash will parse and transform those raw logs into a structured data format for Elasticsearch. Elasticsearch then does a bunch of indexing and stuff and fires up a port to be able to talk to Kibana about that data. Kibana is the application that provides a nice UI to interact with Elasticsearch data; build visualizations, make searches and filters... etc.

It's kind of a long ass pipeline, but the team at Elastic have done a great job of making modular software that, while it seems kind of long and complicated, offers a lot of flexibility because of the modularity; you can run Filebeat on one server next to your geth, then pipe those logs along to Logstash and Elasticsearch on another server (or two); separate your concerns, user privileges, and public APIs. Weird but great.

ℹ️ We're going to run each of these tools in their own terminal console. You're going to have 5 terminals open by the end of this. There are probably ways to use services, screens, backgrounding, etc. to keep your tab count smaller, but that's for another time.

Install all the things

OSX:

$ brew install ethereumproject/classic/geth
$ brew install filebeat logstash elasticsearch kibana

If you're on Linux or Windows, you're on your own for the (F)ELK stack for now. You can install geth by checking out our Releases Page and just downloading the (archived) binary, un-archiving the executable, and moving it to your $PATH.

Geth

Terminal 1

So far for geth mlog is disabled by default; you'll need to turn it on as you fire up your geth instance.

$ geth --mlog=json

Mlog has 3 output modes (or structures): plain, kv (key-value), and json.

Example kv line:

2017-08-24T17:17:39Z [discover] PING HANDLE FROM from.udp_address=$STRING from.id=$STRING ping.bytes_transferred=$INT

Example json line:

{"event":"ping.handle.from","from.id":"string","from.udp_address":"string","ping.bytes_transferred":0,"ts":"2017-10-18T09:01:06.54592753-07:00"}

In this walkthrough I'm only going to use the json format for mlogs. You can also use kv, or plain. I like JSON because Logstash likes JSON, so it saves a lot of groking.

You can also use --mlog-dir to pick where you want to keep these logs.

ℹ️ Be advised that these logs aren't tiny. Geth will generate about 200MB of machine logs every 12 hours. If you're only interested in mlogs for certain components, you can also use the --mlog-components flag. As usual, check out geth help for a little more information there.

You get out what you put in.

Elasticsearch is going to ingest and store the data it gets from geth. This means that Elasticsearch will only have the data that you mlog from geth until you run it with mlog tomorrow... and so on. If you've only used --mlog with geth today, that's all the data Elasticsearch is going to get. It won't automatically have lifetime of data for your geth client or somehow magically know about the whole blockchain.

Filebeat

Terminal 2

Turn Filebeat on:

$ filebeat -e -path.config=. -c filebeat.geth.mlog.json.yml -d "publish"

Filebeat config: ./filebeat.geth.mlog.json.yml

filebeat.prospectors:
- input_type: log
  paths:
    - /Users/ia/Library/EthereumClassic/mainnet/mlogs/geth*
  symlinks: true
  json.message_key: event
  json.keys_under_root: true
output.logstash:
  hosts: ["localhost:5043"]

You want to use this config file. Filebeat's default config dir is /usr/local/etc/filebeat and it comes with a default config file, so you can either toss this config in there or point to it.

Note that we're using wildcard globbing to watch all mlog* files. This is convenient because geth will create a new mlog file each time it fires up, and automatically roll the log file when it reaches a max size. If you use mlogs consistently, your mlogs/ directory may end up looking like this:

-rw-r--r-- 1 ia staff 3.8M Nov 21 11:43 geth.mh.ia.mlog.20171121-114144.51125
-rw-r--r-- 1 ia staff  53M Nov 21 16:54 geth.mh.ia.mlog.20171121-114910.52716
-rw-r--r-- 1 ia staff  40M Nov 21 20:35 geth.mh.ia.mlog.20171121-165414.8245
-rw-r--r-- 1 ia staff 112M Nov 22 10:43 geth.mh.ia.mlog.20171121-223130.69466
-rw-r--r-- 1 ia staff 183M Nov 23 10:55 geth.mh.ia.mlog.20171122-213235.3464
-rw-r--r-- 1 ia staff 3.5M Nov 23 10:55 geth.mh.ia.mlog.20171123-105520.37157
-rw-r--r-- 1 ia staff 6.2M Nov 23 11:06 geth.mh.ia.mlog.20171123-105957.38013
-rw-r--r-- 1 ia staff  14M Nov 23 12:01 geth.mh.ia.mlog.20171123-110659.39536
-rw-r--r-- 1 ia staff  18M Nov 23 13:47 geth.mh.ia.mlog.20171123-120155.50707
-rw-r--r-- 1 ia staff 3.7M Nov 23 13:48 geth.mh.ia.mlog.20171123-134731.70141
-rw-r--r-- 1 ia staff  40M Nov 23 19:27 geth.mh.ia.mlog.20171123-135124.71278

Filebeat is now publishing your logs on port 5043.

Logstash

Terminal 3

Turn Logstash on:

$ logstash -f ls-pipeline-json.conf --config.reload.automatic

Logstash config: ./ls-pipeline-json.conf

input {
    beats {
        port => "5043"
    }
}
filter {
    grok {
      match => { "to.udp_address" => "%{IP:to.host}:%{NUMBER:port}" }
      match => { "from.udp_address" => "%{IP:from.host}:%{NUMBER:port}" }
    }
    date {
        match => [ "block.time","UNIX" ]
        target => "block.TimeFromUnix"
    }
    geoip {
        source => "to.host"
        target => "geoip"
        add_field => [ "[geoip][coordinates]", "%{[geoip][longitude]}" ]
        add_field => [ "[geoip][coordinates]", "%{[geoip][latitude]}" ]
    }
    geoip {
        source => "from.host"
        target => "geoip"
        add_field => [ "[geoip][coordinates]", "%{[geoip][longitude]}" ]
        add_field => [ "[geoip][coordinates]", "%{[geoip][latitude]}" ]
    }
    mutate {
        convert => [ "[geoip][coordinates]", "float" ]
    }
    ruby {
        code => "
            begin
                block_time = DateTime.parse(Time.at(event.get('block.time')).to_s).to_time
                block_received_at = DateTime.parse(event.get('block.received_at')).to_time

                event.set('block.time_vs_received_elapsed', block_received_at - block_time)
            rescue Exception => e
                event.set('logstash_ruby_exception', 'elapsed: ' + e.message)
            end
        "
    }
}
output {
    stdout { }
    elasticsearch {
        hosts => "localhost:9200"
    }
}

There are a couple of filters going on here:

  • the grok filters are helping geoip parse the IP's so we can make the fancy where-in-the-world map
  • the date filter is using the Unix default to establish block.time as a time intead of just a bunch of seconds
  • in the ruby filter we're actually creating a new field by subtracting block.received_at from block.time fields to create block.time_vs_received_elapsed which is an interesting network latency/propagation measure

These filters are the tip of the iceberg, too. You can use other cool elastic packages like aggregate and a bunch more I don't know off the top of my head to clean, manipulate, and measure your raw log data before passing it to elasticsearch.

Logstash is now grabbing your logs from Filebeat on 5043, load-extract-transforming them, and making them available for Elasticsearch on 9200.

Elasticsearch and Kibana

Terminal 4

Turn elastic search on.

$ elasticsearch
Terminal 5

Turn Kibana on.

$ kibana

Kibana connects with Elasticsearch on 9200, and publishes on 5601. Finally, open localhost:5601 in your browser of choice.

ℹ️ Elasticsearch will only be able to create indexes for events and fields it has ingested.

This means that if you're just starting out and don't have tons of mlogs (yet!), then you won't be able to index or visualize using those as-yet-unseen values. For example, if you just fired everything up but geth hasn't seen a chain reorg yet, you won't be able to index on blockchain.reorg.blocks or use that value in a visualization.

As you accrue mlogs and ingest them in elasticsearch, you may find it useful to refresh your field list:

kibana-refresh-indexes


Hit the ground running

If you'd like to import the indexes, dashboards, and visualizations used to make the dashboards shown above, they're all available here in ./kibana-export.

kibana_import

Releases

No releases published

Packages

No packages published

Languages