You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
.Sample log files
I tried to parse small log file (having 100logs) - 1 emit count got increased
I tried to parse big log file (having 200logs)- suddenly emit got increased with large number
we were trying to find a metrics/ratio/or formula to identify failure of fluentd destinations such as (opensearch,kafka). While browsing we found flush error ratio can be used which is an error count to an emit count. But we found emit count drastically increased whereas an error increased by just 1. and hence ratio was way low to raise an alarm.
Below are the metrics that we used to calculate flush error ratio:
fluentd_output_status_emit_count -> To calculate emit count
fluentd_output_status_num_errors -> To calculate no of errors
Below are the configuration for output destination
I tried to limit out the no of retries to 0 so to check does an error count reports to an every emit happened since output destination (kafka, opensearch) is unreachable.
So Can you tell how emit count is calculated.
The text was updated successfully, but these errors were encountered:
.Sample log files
I tried to parse small log file (having 100logs) - 1 emit count got increased
I tried to parse big log file (having 200logs)- suddenly emit got increased with large number
we were trying to find a metrics/ratio/or formula to identify failure of fluentd destinations such as (opensearch,kafka). While browsing we found flush error ratio can be used which is an error count to an emit count. But we found emit count drastically increased whereas an error increased by just 1. and hence ratio was way low to raise an alarm.
Below are the metrics that we used to calculate flush error ratio:
Below are the configuration for output destination
For kafka2 plugin
<match>
@type kafka2
brokers kf-train-headless:9092
default_topic your_topic
output_data_type json
@log_level debug
<buffer>
@type file
chunk_limit_size 8MB
path /var/log/fluent/is-fluentd-buffer/kafka.logging.all
total_limit_size 1024m
flush_interval 3s
retry_max_times 0
</buffer>
<format>
@type json
</format>
</match>
For opensearch
<match>
@type opensearch
@log_level info
include_tag_key true
host indexsearch
port 9200
logstash_format true
logstash_prefix log-${tag[2]}
reload_connections false
reconnect_on_error true
reload_on_failure true
request_timeout 10s
<buffer>
@type file
chunk_limit_size 8MB
path /var/log/fluent/is-fluentd-buffer/os.logging.all
total_limit_size 1024m
retry_max_times 0
</buffer>
</match>
I tried to limit out the no of retries to 0 so to check does an error count reports to an every emit happened since output destination (kafka, opensearch) is unreachable.
So Can you tell how emit count is calculated.
The text was updated successfully, but these errors were encountered: