-
Notifications
You must be signed in to change notification settings - Fork 1.7k
could not enqueue records into the ring buffer #9906
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
please attach your full Fluent Bit log file |
@edsiper do you want me to enable a higher debug level than info and re-run this ? |
I'm seeing same issue in fluent-bit 3.2.1 It happens with multiple inputs, see log entries below for TCP and prometheus_remote_write inputs. If I change the inputs to threaded: false, the error goes away. However, that puts all inputs in the main thread, and not good for performance. I have also seen below issue which mentions same error, and the root cause was supposed to be fixed in an earlier version. [2025/02/27 08:11:13] [error] [input:tcp:tcp.1] could not enqueue records into the ring buffer [2025/02/27 08:55:23] [error] [input:prometheus_remote_write:prometheus_remote_write.2] could not enqueue records into the ring buffer |
Same issue in v3.1.7 here |
Issue is still happening on 4.0.1 Turning off
|
Uh oh!
There was an error while loading. Please reload this page.
Bug Report
Describe the bug
We've got a fluent-bit aggregator VM running 3.2.5.
The node is behaving in a strange manner, throughput used to be about 80MB/s, 40 in and 40 out. However, now it's doing < 10MB in, and 0 out. The only time there is output is when I restarted the process. Also resources are readily available.
It was working fine, but all of a sudden the traffic came crashing down and output went to 0. Almost like it hit a race condition.
After a processes reboot, network traffic will spike up to 70-80mb/s then come down to 10mb/s.
We're seeing the following errors pop up once every second:
[2025/02/02 03:13:53] [error] [input:forward:forward.0] could not enqueue records into the ring buffer [2025/02/02 03:13:54] [error] [input:forward:forward.0] could not enqueue records into the ring buffer [2025/02/02 03:13:55] [error] [input:forward:forward.0] could not enqueue records into the ring buffer [2025/02/02 03:13:56] [error] [input:forward:forward.0] could not enqueue records into the ring buffer [2025/02/02 03:13:57] [error] [input:forward:forward.0] could not enqueue records into the ring buffer [2025/02/02 03:13:58] [error] [input:forward:forward.0] could not enqueue records into the ring buffer [2025/02/02 03:13:59] [error] [input:forward:forward.0] could not enqueue records into the ring buffer [2025/02/02 03:14:00] [error] [input:forward:forward.0] could not enqueue records into the ring buffer [2025/02/02 03:14:01] [error] [input:forward:forward.0] could not enqueue records into the ring buffer
To Reproduce
Install fluent-bit 3.2.5, with config:
We have 5 VMs acting as forwarders, each sending a 2.2M lined CSV file, which the aggregator then ingests writes out to disk and sends to redpanda (kafka).
There is NO compression on the forwarders, and NO compression going to redpanda (kafka).
Update
So I realized when fluent-bit's network performance decreases, it also seems like all cores stop working and only 1 core is working primarily on the forwarder input.
This can best be seen in the screens above, something puts it into this odd single core state, even though as you can see from the config it should be threaded for input and has multiple workers for output.
Expected behavior
Having run a dozen tests we expect throughput to be 80MB/s with 40 in and 40 hour, and the disk maintaining 40mb/s writes.
Your Environment
VMware ESXi, 7.0.3, 21424296
Hardware is a Dell PowerEdge R720XD, 512GB of ram, Intel(R) Xeon(R) CPU E5-2667 v2 @ 3.30GHz with NVMe drives and 10G networking.
RHEL 9.5 (Plow)
The VM for fluent-bit aggregator is 16vCPU, 16GB of Mem, 300GB on NVMe, 78% free disk space.
The text was updated successfully, but these errors were encountered: