-
Notifications
You must be signed in to change notification settings - Fork 1.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fluent-bit input plugin tail doesn't process all logs: scan_blog add(): dismissed: #4155
Comments
@rguske I also have this problem, how did you solve it? |
Hi @zy286938348, I have no clue at the moment but I will follow some of the made suggestions here: https://fluent-all.slack.com/archives/C0CTQGHKJ/p1637005220493300 |
Hi @zy286938348, |
@rguske @zy286938348 I am also seeing this issue using containerd as my runtime. |
I think this is the RCA for me: #3873 When I removed Please tag me if you are able to try this and confirm it works for you too. |
Hi @kc-dot-io, |
I also have this issue. Upgrading Fargate 1.3(using Docker as runtime) to 1.4 (using containerd) did not help. |
Circling back to this and I'm no longer certain that The issue still appears to occur from time to time and while it's not limited to a single node usually, I have observed the circumstance where one node will work while another isn't. I'm now testing increasing the resources that are being allocated to the fluebt bit replicas since the issue appears to come and go. If anyone has any more ideas, I'm happy to troubleshoot it with my setup. Please do ping me in follow ups. |
Is there any updates on this issue? I am facing the same on K8s 1.21 deployment with Fluent Bit v1.8.12 and elasticsearch. |
@srideok2 I don't think there is a permanent solution yet but I've talked to a few other people who had issues and we've had luck with reconfiguring our fluentbits to get to a place where things were stable. From what I can see the issue mostly happens when buffer sizes get exceeded due to large log lines. The solution I used and worked for at least one other person was to turn on the option for skipping long lines and then jack up the buffer size on the inputs and filters. IRC there is another open issue that speaks to the issue at a lower level and I think there are multiple different symptoms of them same root cause here, which is basically that when the memory buffer is exceed, fluent-bit just silently stops working until it can be reset. Ping me on Slack community if you have any questions. I hope this helps. |
Hi all, |
Has this been resolved? |
@edsiper can you please give us some pointers here to debug this issue. As I have understood fluent-bit/plugins/in_tail/tail_file.c Line 861 in cd92cf3
|
There is a bug in v1.8.13, we have just published the new container image for v1.8.14, please upgrade:
|
I'm using version 1.9.0 and I'm getting dismissed logs. This doesn't really seems fixed. |
Same issue here. Openshift 4 CoreOS and cri-o. Any input regarding this would be great. |
This got fixed when we started using 1.9.0. @danielserrao |
@ramatudaan i switched my fluent-bit image to 1.9.0 and my issue now seems to be security/permission related or missconfiguration of templates to match Openshift 4 CoreOS/cri-o. I created and issue here on it: here |
Hello guys, is there any ETA on this @edsiper? |
Same issue guys, tail logging not works |
Have same issue. Tried both output as http to logstash and es plugin to ES |
fluent-bit版本:1.9.3 目前暂时没有遇到tail停止采集的问题 |
Has anyone tried turning |
I'm having this issue as well on 1.9.6. The problem is that I NEED the long lines to be processed. |
@Jonathan-w6d yes this helped for me. Also increasing the max buffer size. That said you'll still run into an issue with a memory leak and eventually need to restart the process. I have to do this every 90-120 days. @zulrang Try increasing your max buffer size to something very large - you'll probably still have to restart the processes every so often, but in theory if they buffer is large enough it should be able to process long lines until it can't again. |
I have all buffers set to 2M and it has never successfully flushed chunks from one specific node - the one that is logging stack traces (about 8k long). The other 13 work fine. |
@leonardo-albertovich i can give you the last values.yaml containing the confs that I used before switching to another solution if you'd like ? |
Here you go, that's actually pretty straightforward, just adding k8s metadata and modifying tag on some logs so I can differentiate them, then output to opensearch directly.
|
Thank you @Jonathan-w6d, we'll try to have someone from the team reproduce the issue. |
We are encountering this in v2.1.2 in EKS 1.22. |
@msolters That's indeed a good clue to follow, hopefully this can bring some lights for debugging this @leonardo-albertovich fluent-bit has difficulty detecting new files? |
Fluent-bit shouldn't have issues detecting new files, at least I never had that happen. Could you please share more information about your context @msolters? I'm interested in knowing which platform is it, if you are using inotify, how many files are there in the paths where this is happening (or how many glob matchs), how does your configuation file look like and a copy of your log if possible. Feel free to share things with me privately in slack if you don't feel comfortable doing it in public. |
This is some information about the cluster, including the yaml of the Fluent Bit daemonset, the configuration file of Fluent Bit, the version information of k8s and docker, the OS version, and kernel parameters. I hope this helps to reproduce the problem, If you need other information please let me know. PS: The k8s version is |
@leonardo-albertovich @edsiper @agup006 @patrick-stephens we hope this can help to narrow down the investigation scope and to locate the root cause. Let us know what else you need |
Same issue using latest version, has this been resolved? |
Looks like 'scan_blog add(): dismissed' is generic message, when the root cause can be on different places like in parser etc. It would be great to add more details in debug mode, which can help to user uncover the issue for example error message coming from parser etc. Would it be possible? In general .. where to find the debug information related to parsers processing? |
@leonardo-albertovich what do you think? Is it possible to add more debug messages to help narrow the scope? |
@studyhuang1996 fluentbit v2.1.6 no longer skips any logs, issue resolved in your env that has this issue before? |
FYI team, issue is still present on |
I am facing the same issue, one additional observation. The |
I tried with latest version of fluentbit (2.1.10) Still facing the same error. Do we have any news on this ? [2023/10/10 09:28:34] [debug] [input:tail:tail.0] scan_blog add(): dismissed: Please see the input config used below. inputs: | |
这个问题一直存在,当然也不影响使用,只是不好排查问题,有一定的误解性,但是现在遇到一些问题会出现数据丢失,字段丢失还有一些多行匹配场景不满足,目前打算使用其他工具替换
fluent-bit
Benjamin Huo ***@***.***> 于2023年8月31日周四 14:13写道:
… [image: image]
<https://user-images.githubusercontent.com/27665214/264525540-e69a6936-8321-4e39-a5ee-757ca975ecc9.png>
fluent-bit version 2.1.6 has this been resolved?
@studyhuang1996 <https://github.com/studyhuang1996> fluentbit v2.1.6 no
longer skips any logs, issue resolved in your env that has this issue
before?
—
Reply to this email directly, view it on GitHub
<#4155 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AGTCGPUBFCRO36HI2CXGA3LXYATRPANCNFSM5FHYQJOA>
.
You are receiving this because you were mentioned.Message ID:
***@***.***>
|
I have the same issue fluent/helm-charts#415:
|
same here with fluent/fluent-bit helm chart 0.42.0, and 2.2.1 image version |
This issue is stale because it has been open 90 days with no activity. Remove stale label or comment or this will be closed in 5 days. Maintainers can add the |
This issue was closed because it has been stalled for 5 days with no activity. |
@kc-dot-io reopen |
Hello. If you have this issue, please build your binary using this: https://github.com/lecaros/fluent-bit/tree/master |
I'll gladly reopen the ticket if you can provide the requested indo. |
Hi, we had a few scenarios where some logs weren't delivered (using Fluent Bit 2.2.2, "tail" input plugin and "splunk" output plugin). Looking at Fluent Bit logs I've noticed a
Basically after FluentBit starts monitoring a file it will always print `dismissed" because it doesn't need to register it again (although TBH the message is misleading, it makes you think that the file is not monitored anymore). TL,DR: if you came here because you've lost some events and you're seeing Thanks! |
Can we reopen this issue? We see that this message is misleading, instead of the proper root cause issue. |
Still seeing this issue even for @leonardo-albertovich Logs should paint an appropriate picture of what's wrong. An abrupt event out of nowhere contemplating switching to a different tool OR use the forked version |
Bug Report
Describe the bug
Fluent Bit is not processing all logs located in
/var/log/containers/
.To Reproduce
The following messages are displayed:
Configuration details:
ServiceAccount:
ConfigMap:
DaemonSet:
Expected behavior
All logs in
/var/log/containers/
should be processed.Your Environment
See: https://github.com/vmware-samples/vcenter-event-broker-appliance
@include input-kubernetes.conf
@include input-systemd.conf
@include input-kube-apiserver.conf
@include input-auditd.conf
@include filter-kubernetes.conf
@include filter-record.conf
@include output-syslog.conf
Additional context
Running
tail -f
manually from within the system on a specific pod log, which is writing tostdout
, works.Logs which e.g. aren't processed:
I've also tried running the DS in
Similar issues I found but which doesn't provide the solution for this issue:
#3857
#4014
Your help would be much appreciated. Thanks
The text was updated successfully, but these errors were encountered: