Introduction Log pipeline backlogs cause delayed or lost log data, impacting incident investigation, compliance, and security monitoring. When the pipeline cannot keep up with log volume, logs accumulate in buffers or are dropped.
Symptoms - Log search shows gaps in recent data - Buffer utilization increasing continuously - Error: "buffer queue is full" or "chunk flush failed" - Log shipping latency increasing - Disk space consumed by log buffers
Common Causes - Log volume spike exceeding pipeline capacity - Output destination slow or unreachable (Elasticsearch down) - Buffer configuration too small - Parsing errors causing pipeline stalls - Network bandwidth bottleneck
Step-by-Step Fix 1. **Check buffer status**: ```bash # Fluentd curl http://localhost:24220/api/plugins.json | jq '.[] | select(.type == "buffer")' # Filebeat filebeat test output ```
- 1.Increase buffer capacity:
- 2.```yaml
- 3.# Fluentd
- 4.<buffer>
- 5.@type file
- 6.path /var/log/fluentd-buffers
- 7.chunk_limit_size 8m
- 8.queue_limit_length 64
- 9.flush_interval 5s
- 10.</buffer>
- 11.
` - 12.Check output destination health:
- 13.```bash
- 14.curl -s http://elasticsearch:9200/_cluster/health
- 15.
`