-
Notifications
You must be signed in to change notification settings - Fork 103
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Is it possible to set document_id to avoid duplication of events? #156
Comments
I'm looking for this exact same feature. If I use action => "update" I get a I've been trying to use doc_as_upsert as true but it still duplicates my data This is my config
|
Any updates on this issue? |
I am also currently blocked by this. |
I am trying to deduplicate too! my way around this is to delete the index and recreating it again :( |
I ended up not using this plugin and using the elasticsearch output plugin like so:
|
How did you set up the ELASTIC_USERNAME and password? |
After being quite confused by this thread (and still questioning whether we were ever doing upserts correctly), I realized that document_id does actually work for de-duplication, at least in my case.
|
Yeah document_id does work for de-duplication. For the username and password I believe we're using Amazon Cognito Authentication. |
I would like to deduplicate events coming from logstash. This can be done by providing an unique id. In normal logstash es output this can be achieved by setting the option
document_id
, see https://www.elastic.co/guide/en/beats/filebeat/current/filebeat-deduplication.html#ls-doc-idBut I see no such configuration in logstash-output-amazon_es plugin. Is there a way to pass the id to es somehow? Or is it possible to achieve deduplication in another way?
The text was updated successfully, but these errors were encountered: