As you see,I use primary key ID in my table as the document_id and "tracking_column => ID" to avoid dumplicate data, but now, the problem is that I found my data loss, about 15%.I think it happens when
logstash's query timeout or failed but the sql_last_value still updated. Is there any way to resolve this problem?
hi, i'm using logstash 2.3 and logstash-input-jdbc plugin,below is my config file:
input {
}
filter { mutate { rename => ["updation_date", "creation_date"] } }
output { elasticsearch { hosts => ["10.1.3.1"] user => "user" password => "pass2" workers => 5 index => "caslog-%{+YYYY.MM.dd}" template_overwrite => true document_id => "%{log_id}" } }
As you see,I use primary key ID in my table as the document_id and "tracking_column => ID" to avoid dumplicate data, but now, the problem is that I found my data loss, about 15%.I think it happens when logstash's query timeout or failed but the sql_last_value still updated. Is there any way to resolve this problem?