Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Make the river more resistant to bulk import failures #570

Open
PeterBackman opened this issue Nov 12, 2015 · 1 comment
Open

Make the river more resistant to bulk import failures #570

PeterBackman opened this issue Nov 12, 2015 · 1 comment

Comments

@PeterBackman
Copy link

Hi,
in our system we can occasionally get documents that can not be imported into ES. We do not have full control of the input so sometimes we hit the 32k limit in Lucene which will prevent the document to be inserted into ES. The bulk import will fail and the river is stopped.

Locally I made a patch to make afterBulk(long executionId, BulkRequest request, BulkResponse response) öog an error and continue without stopping the river. Seems to work fine.

Is there a reason the river must be stopped or would the above change be interesting?

@ankon
Copy link
Contributor

ankon commented Jan 4, 2016

I guess the main question here is whether someone actually reads the logs :)

In our case stopping the river is preferable, because we can monitor that easily, and then call a human to investigate the failure. There should never be ignorable import failures for us. Other situations might differ, so even if the change might not ever be applied to avoid accidental foot-shooting, it could still be useful for other people that can tolerate some data loss.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants