Skip to content

Bug: Duplicate records can be introduced when failure before commits happen #233

@lyuboxa

Description

@lyuboxa

Bug description

The source connector batch acker, will commit offsets to Kafka at every n or 1000 (as const).
This can result in failure introducing between 1 < duplicates < n to the destination.
When the consumer restarts, it will assume the consumer group contains the latest positioning data
ask'd by Conduit, however the discrepancy described above can be observed.

Steps to reproduce

  1. Create kafka pipeline with conduit
  2. Produce 10 records
  3. Kill pipeline (kill command not teardown, as this will force a flush)
  4. Start pipeline, same records will be produced

Version

latest

Metadata

Metadata

Assignees

No one assigned

    Labels

    bugSomething isn't working

    Type

    No type

    Projects

    Status

    Todo

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions