question-mark
Stuck on an issue?

Lightrun Answers was designed to reduce the constant googling that comes with debugging 3rd party libraries. It collects links to all the places you might be looking at while hunting down a tough bug.

And, if you’re still stuck at the end, we’re happy to hop on a call to see how we can help out.

Splunk-connect consuming lot of API Server resources | 98% of the API connections are consumed by Fluentd

See original GitHub issue

Hello Team,

We were using splunk-connect 1.0.1 and it was fine. Now we have upgraded to splunk-connect 1.4.0 and it exponentially increased number of API connections. We see 2 issues in our clusters due to this:

  1. API Server is very busy processing these API calls and its consuming lot of resources even with less workload in the cluster.

  2. We have enabled audit logging which is filled up very quickly and 98% of the events are from Splunk pods.

We need to see why fluentd consuming these many connections and resources. We understand it uses watch API but how we can minimize the consumption ?

What happened: As Above consuming CPU and Memory resources along with Disk space for audit logs.

What you expected to happen: We expected to happen this smooth without much stress on API server

How to reproduce it (as minimally and precisely as possible): Install splunk-connect 1.4.0 with Openshift 3.11.x / Docker enterprise v1.14 version.

Anything else we need to know?:

Environment:

  • Kubernetes version (use kubectl version):v1.14.8
  • Ruby version (use ruby --version):
  • OS (e.g: cat /etc/os-release): CentOS 7.7
  • Splunk version: 7.x
  • Others:

Issue Analytics

  • State:closed
  • Created 3 years ago
  • Comments:17 (1 by maintainers)

github_iconTop GitHub Comments

1reaction
crdnbcommented, May 7, 2020

Thanks @matthewmodestino - it’s working very well now!

1reaction
mghermancommented, Mar 25, 2020

We are also seeing this occur, every few days the CPU utilisation on the control plane nodes (api server) inceases to 100%. Restarting the splunk-connect logging pods resolves this issue for another few days.

During the period of hi api-server usage the audit logs indicate a large number of repeated queries to /api/v1/watch/pods?fieldSelector=spec.nodeName%3Dus-prod-kubewrk-001.atl01.example.org&resourceVersion=78464987 for each worker node (6). These requests appear to be being made at a rate of approximately 1600 r/min on each api node (3).

Read more comments on GitHub >

github_iconTop Results From Across the Web

i am seeing this error when i connected to Splunk
when i tried to configure splunk dbconnect app i am seeing below error 2020 -07-24 ... -0400 [ dw-55 - POST /api/connections/status ]...
Read more >
Amazon Elastic Container Service - Best Practices Guide
API Gateway functions which use a VPC link that allows the AWS managed service to connect to hosts inside the private subnet of...
Read more >
Pumped-up logging with Fluent Bit and Splunk - Medium
If you are not familiar with Fluent Bit, it is an open-source application written in C to help ship data from many sources...
Read more >
Security perspective on container-based applications in safety ...
One of the most important ones is the API server. It informs all other components within the cluster via JSON based REST-API. This....
Read more >
Search Results - CVE
0 and earlier is vulnerable to a JDBC Deserialisation attack if the attacker is able to control the JDBC URL used or cause...
Read more >

github_iconTop Related Medium Post

No results found

github_iconTop Related StackOverflow Question

No results found

github_iconTroubleshoot Live Code

Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free

github_iconTop Related Reddit Thread

No results found

github_iconTop Related Hackernoon Post

No results found

github_iconTop Related Tweet

No results found

github_iconTop Related Dev.to Post

No results found

github_iconTop Related Hashnode Post

No results found