Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

admin: backpressure: updating for style and consistency #1511

Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
169 changes: 110 additions & 59 deletions administration/backpressure.md
Original file line number Diff line number Diff line change
Expand Up @@ -2,68 +2,119 @@

<img referrerpolicy="no-referrer-when-downgrade" src="https://static.scarf.sh/a.png?x-pxid=63e37cfe-9ce3-4a18-933a-76b9198958c1" />

Under certain scenarios it is possible for logs or data to be ingested or created faster than the ability to flush it to some destinations. One such common scenario is when reading from big log files, especially with a large backlog, and dispatching the logs to a backend over the network, which takes time to respond. This generates backpressure leading to high memory consumption in the service.

In order to avoid backpressure, Fluent Bit implements a mechanism in the engine that restricts the amount of data that an input plugin can ingest, this is done through the configuration parameters **Mem\_Buf\_Limit** and **storage.Max\_Chunks\_Up**.

As described in the [Buffering](../concepts/buffering.md) concepts section, Fluent Bit offers two modes for data handling: in-memory only (default) and in-memory + filesystem \(optional\).

The default `storage.type memory` buffer can be restricted with **Mem\_Buf\_Limit**. If memory reaches this limit and you reach a backpressure scenario, you will not be able to ingest more data until the data chunks that are in memory can be flushed. The input will be paused and Fluent Bit will [emit](https://github.com/fluent/fluent-bit/blob/v2.0.0/src/flb_input_chunk.c#L1334) a `[warn] [input] {input name or alias} paused (mem buf overlimit)` log message. Depending on the input plugin in use, this might lead to discard incoming data \(e.g: TCP input plugin\). The tail plugin can handle pause without data loss; it will store its current file offset and resume reading later. When buffer memory is available, the input will resume collecting/accepting logs and Fluent Bit will [emit](https://github.com/fluent/fluent-bit/blob/v2.0.0/src/flb_input_chunk.c#L1277) a `[info] [input] {input name or alias} resume (mem buf overlimit)` message.

This risk of data loss can be mitigated by configuring secondary storage on the filesystem using the `storage.type` of `filesystem` \(as described in [Buffering & Storage](buffering-and-storage.md)\). Initially, logs will be buffered to *both* memory and filesystem. When the `storage.max_chunks_up` limit is reached, all the new data will be stored safely only in the filesystem. Fluent Bit will stop enqueueing new data in memory and will only buffer to the filesystem. Please note that when `storage.type filesystem` is set, the `Mem_Buf_Limit` setting no longer has any effect, instead, the `[SERVICE]` level `storage.max_chunks_up` setting controls the size of the memory buffer.

## Mem\_Buf\_Limit

This option is disabled by default and can be applied to all input plugins. Please note that `Mem_Buf_Limit` only applies with the default `storage.type memory`. Let's explain its behavior using the following scenario:

* Mem\_Buf\_Limit is set to 1MB \(one megabyte\)
* input plugin tries to append 700KB
* engine route the data to an output plugin
* output plugin backend \(HTTP Server\) is down
* engine scheduler will retry the flush after 10 seconds
* input plugin tries to append 500KB

At this exact point, the engine will **allow** appending those 500KB of data into the memory; in total it will have 1.2MB of data buffered. The limit is permissive and will allow a single write past the limit, but once the limit is **exceeded** the following actions are taken:

* block local buffers for the input plugin \(cannot append more data\)
* notify the input plugin invoking a **pause** callback

The engine will protect itself and will not append more data coming from the input plugin in question; note that it is the responsibility of the plugin to keep state and decide what to do in that _paused_ state.

After some time, usually measured in seconds, if the scheduler was able to flush the initial 700KB of data or it has given up after retrying, that amount of memory is released and the following actions will occur:

* Upon data buffer release \(700KB\), the internal counters get updated
* Counters now are set at 500KB
* Since 500KB is &lt; 1MB it checks the input plugin state
* If the plugin is paused, it invokes a **resume** callback
* input plugin can continue appending more data

## storage.max\_chunks\_up

Please note that when `storage.type filesystem` is set, the `Mem_Buf_Limit` setting no longer has any effect, instead, the `[SERVICE]` level `storage.max_chunks_up` setting controls the size of the memory buffer.

The setting behaves similarly to the above scenario with `Mem_Buf_Limit` when the non-default `storage.pause_on_chunks_overlimit` is enabled.

When (default) `storage.pause_on_chunks_overlimit` is disabled, the input will not pause when the memory limit is reached. Instead, it will switch to only buffering logs in the filesystem. The disk spaced used for filesystem buffering can be limited with `storage.total_limit_size`.

See the [Buffering & Storage](buffering-and-storage.md) docs for more information.

## About pause and resume Callbacks

Each plugin is independent and not all of them implements the **pause** and **resume** callbacks. As said, these callbacks are just a notification mechanism for the plugin.

One example of a plugin that implements these callbacks and keeps state correctly is the [Tail Input](../pipeline/inputs/tail.md) plugin. When the **pause** callback is triggered, it pauses its collectors and stops appending data. Upon **resume**, it resumes the collectors and continues ingesting data. Tail will track the current file offset when it pauses and resume at the same position. If the file has not been deleted or moved, it can still be read.

With the default `storage.type memory` and `Mem_Buf_Limit`, the following log messages will be emitted for pause and resume:

```
It's possible for logs or data to be ingested or created faster than the ability to
flush it to some destinations. A common scenario is when reading from big log files,
especially with a large backlog, and dispatching the logs to a backend over the
network, which takes time to respond. This generates _backpressure_, leading to high
memory consumption in the service.

To avoid backpressure, Fluent Bit implements a mechanism in the engine that restricts
the amount of data an input plugin can ingest. Restriction is done through the
configuration parameters `Mem_Buf_Limit` and `storage.Max_Chunks_Up`.

As described in the [Buffering](../concepts/buffering.md) concepts section, Fluent
Bit offers two modes for data handling: in-memory only (default) and in-memory and
filesystem (optional).

The default `storage.type memory` buffer can be restricted with `Mem_Buf_Limit`. If
memory reaches this limit and you reach a backpressure scenario, you won't be able
to ingest more data until the data chunks that are in memory can be flushed. The
input pauses and Fluent Bit
[emits](https://github.com/fluent/fluent-bit/blob/v2.0.0/src/flb_input_chunk.c#L1334)
a `[warn] [input] {input name or alias} paused (mem buf overlimit)` log message.

Depending on the input plugin in use, this might cause incoming data to be discarded
(for example, TCP input plugin). The tail plugin can handle pauses without data
ingloss, storing its current file offset and resuming reading later. When buffer
memory is available, the input resumes accepting logs. Fluent Bit
[emits](https://github.com/fluent/fluent-bit/blob/v2.0.0/src/flb_input_chunk.c#L1277)
a `[info] [input] {input name or alias} resume (mem buf overlimit)` message.

Mitigate the risk of data loss by configuring secondary storage on the filesystem
using the `storage.type` of `filesystem` (as described in [Buffering &
Storage](buffering-and-storage.md)). Initially, logs will be buffered to both memory
and the filesystem. When the `storage.max_chunks_up` limit is reached, all new data
will be stored in the filesystem. Fluent Bit stops queueing new data in memory and
buffers only to the filesystem. When `storage.type filesystem` is set, the
`Mem_Buf_Limit` setting no longer has any effect. Instead, the `[SERVICE]` level
`storage.max_chunks_up` setting controls the size of the memory buffer.

## `Mem_Buf_Limit`

`Mem_Buf_Limit` applies only with the default `storage.type memory`. This option is
disabled by default and can be applied to all input plugins.

As an example situation:

- `Mem_Buf_Limit` is set to `1MB`.
- The input plugin tries to append 700&nbsp;KB.
- The engine routes the data to an output plugin.
- The output plugin backend (HTTP Server) is down.
- Engine scheduler retries the flush after 10 seconds.
- The input plugin tries to append 500&nbsp;KB.

In this situation, the engine allows appending those 500&nbsp;KB of data into the memory,
with a total of 1.2&nbsp;MB of data buffered. The limit is permissive and will
allow a single write past the limit. When the limit is exceeded, the following
actions are taken:

- Block local buffers for the input plugin (can't append more data).
- Notify the input plugin, invoking a `pause` callback.

The engine protects itself and won't append more data coming from the input plugin in
question. It's the responsibility of the plugin to keep state and decide what to do
in a `paused` state.

In a few seconds, if the scheduler was able to flush the initial 700&nbsp;KB of data or it
has given up after retrying, that amount of memory is released and the following
actions occur:

- Upon data buffer release (700&nbsp;KB), the internal counters get updated.
- Counters now are set at 500&nbsp;KB.
- Because 500&nbsp;KB isless than 1&nbsp;MB, it checks the input plugin state.
- If the plugin is paused, it invokes a `resume` callback.
- The input plugin can continue appending more data.

## `storage.max_chunks_up`

The `[SERVICE]` level `storage.max_chunks_up` setting controls the size of the memory
buffer. When `storage.type filesystem` is set, the `Mem_Buf_Limit` setting no longer
has an effect.

The setting behaves similar to the `Mem_Buf_Limit` scenario when the non-default
`storage.pause_on_chunks_overlimit` is enabled.

When (default) `storage.pause_on_chunks_overlimit` is disabled, the input won't pause
when the memory limit is reached. Instead, it switches to buffering logs only in
the filesystem. Limit the disk spaced used for filesystem buffering with
`storage.total_limit_size`.

See [Buffering & Storage](buffering-and-storage.md) docs for more information.

## About pause and resume callbacks

Each plugin is independent and not all of them implement `pause` and `resume`
callbacks. These callbacks are a notification mechanism for the plugin.

One example of a plugin that implements these callbacks and keeps state correctly is
the [Tail Input](../pipeline/inputs/tail.md) plugin. When the `pause` callback
triggers, it pauses its collectors and stops appending data. Upon `resume`, it
resumes the collectors and continues ingesting data. Tail tracks the current file
offset when it pauses, and resumes at the same position. If the file hasn't been
deleted or moved, it can still be read.

With the default `storage.type memory` and `Mem_Buf_Limit`, the following log
messages emit for `pause` and `resume`:

```text
[warn] [input] {input name or alias} paused (mem buf overlimit)
[info] [input] {input name or alias} resume (mem buf overlimit)
```

With `storage.type filesystem` and `storage.max_chunks_up`, the following log messages will be emitted for pause and resume:
With `storage.type filesystem` and `storage.max_chunks_up`, the following log
messages emit for `pause` and `resume`:

```
[input] {input name or alias} paused (storage buf overlimit
[input] {input name or alias} resume (storage buf overlimit
```text
[input] {input name or alias} paused (storage buf overlimit)
[input] {input name or alias} resume (storage buf overlimit)
```