Indexer Pipeline Fails Under Heavy Traffic — Log Stream Stuck (v0.8.2) #5818
Replies: 2 comments
-
In your description you mention an issue with heavy load, but it seems that you have identified that the problem is actually a document or a batch of documents that systematically crash the indexing pipeline. Am I correct?
Not through the API, but you can modify the index_metadata directly in the metastore to move the checkpoint forward (which obviously creates data loss).
In general, the indexing pipeline should not fail because of bad documents. Do you have any panic or error message prior to the indexing pipeline failure that would suggest the reason? |
Beta Was this translation helpful? Give feedback.
-
Hi @rdettai, really appreciate it your response
Suppose to be yes, because in the daily normal operation, even if we have some heavy traffics, the logs are normal, and can be ingested to QuickWit, but sometimes, when there's a scanning from the outside, it makes the quickwit indexing pipelines are crashed
I think it will be fine as long as the indexing can back running properly, may i know for this thing, do i need to update the data on database side or is there an api for updating the index metadata in the metastore ?
Sure here's the panic logs
|
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Hello everyone,
Is anyone else experiencing the same issue with Quickwit running on version 0.8.2?
Error log:
I’m using Kafka as the source. Under normal traffic conditions, everything works fine. However, during heavy traffic or when there’s an external security scan, the Quickwit indexer pipeline encounters errors and fails to consume data from Kafka, which causes the log stream to become stuck.
I’ve tried skipping offsets from the Kafka consumer group, but it didn’t work. From what I’ve read in other GitHub issues, it seems there’s no built-in way to skip Kafka offsets in Quickwit.
Questions:
Is it true that there’s currently no way to skip Kafka offsets in Quickwit?
Is there any recommended approach or best practice to prevent the logs from getting stuck in this scenario?
Any advice or workaround would be greatly appreciated.
Thank you!
Beta Was this translation helpful? Give feedback.
All reactions