Kafka source plugin skips, rather than fails, on messages that are too large

Description

Kafka source plugin seems to skip, rather than fail, on messages that are too large.
If we send messages larger than the default value (1 MB) then the pipeline will be successful.
Kafka batch source statistics show that we have not read any messages (0/0), but the offset will be updated in hbase table (as result next run will skip these messages).
Instead, the pipeline should throw error message and fail. I tried to read the same messages using Flume agent. It throws exceptions saying that the message size is too large. After adding special property to consumer config (consumer.max.partition.fetch.bytes = e.g. 5 MB) everything worked properly.

Release Notes

None

Activity

Show:

Yaojie FengApril 24, 2018 at 5:19 PM

this is fixed by https://github.com/data-integrations/kafka-plugins/pull/23, user can provide consumer config to get large messages.

Fixed
Pinned fields
Click on the next to a field label to start pinning.

Details

Assignee

Reporter

Labels

Components

Fix versions

Priority

Created September 8, 2017 at 5:44 PM
Updated July 17, 2020 at 12:27 AM
Resolved April 24, 2018 at 5:19 PM