-
Notifications
You must be signed in to change notification settings - Fork 11
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Merge pull request #388 from TikhomirovSergey/0.24.7-ALPHA
0.24.7-ALPHA: improved serialization of ConsumerRecord
- Loading branch information
Showing
10 changed files
with
376 additions
and
33 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -6,7 +6,7 @@ plugins { | |
} | ||
|
||
ext { | ||
globalVersion = '0.24.6-ALPHA' | ||
globalVersion = '0.24.7-ALPHA' | ||
} | ||
|
||
repositories { | ||
|
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
48 changes: 48 additions & 0 deletions
48
...c/main/java/ru/tinkoff/qa/neptune/kafka/jackson/desrializer/ConsumerRecordSerializer.java
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,48 @@ | ||
package ru.tinkoff.qa.neptune.kafka.jackson.desrializer; | ||
|
||
import com.fasterxml.jackson.core.JsonGenerator; | ||
import com.fasterxml.jackson.databind.JsonSerializer; | ||
import com.fasterxml.jackson.databind.ObjectMapper; | ||
import com.fasterxml.jackson.databind.SerializerProvider; | ||
import org.apache.kafka.clients.consumer.ConsumerRecord; | ||
import org.apache.kafka.common.header.Headers; | ||
|
||
import java.io.IOException; | ||
import java.util.function.Function; | ||
|
||
import static java.util.Objects.nonNull; | ||
|
||
@SuppressWarnings("rawtypes") | ||
public class ConsumerRecordSerializer extends JsonSerializer<ConsumerRecord> { | ||
|
||
private void writeSerializedProperty(String property, | ||
JsonGenerator gen, | ||
ConsumerRecord<?, ?> consumerRecord, | ||
Function<ConsumerRecord<?, ?>, Object> getPropertyValue) throws IOException { | ||
|
||
var propertyValue = getPropertyValue.apply(consumerRecord); | ||
if (nonNull(propertyValue)) { | ||
gen.writeStringField(property, new ObjectMapper().writerWithDefaultPrettyPrinter().writeValueAsString(propertyValue)); | ||
} else { | ||
gen.writeObjectField(property, null); | ||
} | ||
} | ||
|
||
@Override | ||
public void serialize(ConsumerRecord value, JsonGenerator gen, SerializerProvider serializers) throws IOException { | ||
gen.writeStartObject(); | ||
gen.writeStringField("topic", value.topic()); | ||
gen.writeNumberField("partition", value.partition()); | ||
gen.writeObjectField("leaderEpoch", value.leaderEpoch().orElse(null)); | ||
gen.writeNumberField("offset", value.offset()); | ||
if (nonNull(value.timestampType())) { | ||
gen.writeObjectField(value.timestampType().toString(), value.timestamp()); | ||
} | ||
serializers.findValueSerializer(Headers.class).serialize(value.headers(), gen, serializers); | ||
|
||
writeSerializedProperty("key", gen, value, ConsumerRecord::key); | ||
writeSerializedProperty("value", gen, value, ConsumerRecord::value); | ||
|
||
gen.writeEndObject(); | ||
} | ||
} |
31 changes: 31 additions & 0 deletions
31
...a/src/main/java/ru/tinkoff/qa/neptune/kafka/jackson/desrializer/HeaderJsonSerializer.java
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,31 @@ | ||
package ru.tinkoff.qa.neptune.kafka.jackson.desrializer; | ||
|
||
import com.fasterxml.jackson.core.JsonGenerator; | ||
import com.fasterxml.jackson.databind.JsonSerializer; | ||
import com.fasterxml.jackson.databind.SerializerProvider; | ||
import org.apache.kafka.common.header.Headers; | ||
|
||
import java.io.IOException; | ||
import java.util.LinkedHashMap; | ||
import java.util.LinkedHashSet; | ||
import java.util.Set; | ||
|
||
import static java.util.Objects.isNull; | ||
|
||
public class HeaderJsonSerializer extends JsonSerializer<Headers> { | ||
|
||
@Override | ||
public void serialize(Headers value, JsonGenerator gen, SerializerProvider serializers) throws IOException { | ||
|
||
if (isNull(value)) { | ||
return; | ||
} | ||
|
||
var headerMap = new LinkedHashMap<String, Set<String>>(); | ||
value.forEach(header -> { | ||
var set = headerMap.computeIfAbsent(header.key(), s -> new LinkedHashSet<>()); | ||
set.add(new String(header.value())); | ||
}); | ||
gen.writeObjectField("headers", headerMap); | ||
} | ||
} |
13 changes: 13 additions & 0 deletions
13
kafka/src/main/java/ru/tinkoff/qa/neptune/kafka/jackson/desrializer/KafkaJacksonModule.java
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,13 @@ | ||
package ru.tinkoff.qa.neptune.kafka.jackson.desrializer; | ||
|
||
import com.fasterxml.jackson.databind.module.SimpleModule; | ||
import org.apache.kafka.clients.consumer.ConsumerRecord; | ||
import org.apache.kafka.common.header.Headers; | ||
|
||
public class KafkaJacksonModule extends SimpleModule { | ||
|
||
public KafkaJacksonModule() { | ||
addSerializer(Headers.class, new HeaderJsonSerializer()); | ||
addSerializer(ConsumerRecord.class, new ConsumerRecordSerializer()); | ||
} | ||
} |
105 changes: 105 additions & 0 deletions
105
kafka/src/test/java/ru/tinkoff/qa/neptune/kafka/ConsumerRecordDeserializationTest.java
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,105 @@ | ||
package ru.tinkoff.qa.neptune.kafka; | ||
|
||
import com.fasterxml.jackson.databind.ObjectMapper; | ||
import org.apache.kafka.clients.consumer.ConsumerRecord; | ||
import org.apache.kafka.common.header.internals.RecordHeader; | ||
import org.apache.kafka.common.header.internals.RecordHeaders; | ||
import org.testng.annotations.DataProvider; | ||
import org.testng.annotations.Test; | ||
import ru.tinkoff.qa.neptune.kafka.jackson.desrializer.KafkaJacksonModule; | ||
|
||
import java.util.Date; | ||
import java.util.Optional; | ||
|
||
import static java.nio.ByteBuffer.allocateDirect; | ||
import static org.apache.kafka.common.record.TimestampType.LOG_APPEND_TIME; | ||
import static org.hamcrest.MatcherAssert.assertThat; | ||
import static org.hamcrest.Matchers.is; | ||
|
||
public class ConsumerRecordDeserializationTest { | ||
|
||
@DataProvider | ||
public static Object[][] data() { | ||
var date = new Date(); | ||
return new Object[][]{ | ||
{new ConsumerRecord<>("testTopic", 1, | ||
1L, | ||
date.getTime(), | ||
LOG_APPEND_TIME, | ||
0, | ||
0, | ||
new DraftDto().setName("Some Key"), | ||
new DraftDto().setName("Some Value"), | ||
new RecordHeaders() | ||
.add(new RecordHeader("header1", "value1".getBytes())) | ||
.add(new RecordHeader("header1", "value2".getBytes())) | ||
.add(new RecordHeader("header2", "value1".getBytes())), | ||
Optional.of(5)), | ||
"{\n" + | ||
" \"topic\" : \"testTopic\",\n" + | ||
" \"partition\" : 1,\n" + | ||
" \"leaderEpoch\" : 5,\n" + | ||
" \"offset\" : 1,\n" + | ||
" \"LogAppendTime\" : " + date.getTime() + ",\n" + | ||
" \"headers\" : {\n" + | ||
" \"header1\" : [ \"value1\", \"value2\" ],\n" + | ||
" \"header2\" : [ \"value1\" ]\n" + | ||
" },\n" + | ||
" \"key\" : \"{\\n \\\"name\\\" : \\\"Some Key\\\"\\n}\",\n" + | ||
" \"value\" : \"{\\n \\\"name\\\" : \\\"Some Value\\\"\\n}\"\n" + | ||
"}"}, | ||
|
||
{new ConsumerRecord<>("testTopic", 1, | ||
1L, | ||
new Date().getTime(), | ||
null, | ||
0, | ||
0, | ||
null, | ||
null, | ||
new RecordHeaders(), | ||
Optional.empty()), | ||
"{\n" + | ||
" \"topic\" : \"testTopic\",\n" + | ||
" \"partition\" : 1,\n" + | ||
" \"leaderEpoch\" : null,\n" + | ||
" \"offset\" : 1,\n" + | ||
" \"headers\" : { },\n" + | ||
" \"key\" : null,\n" + | ||
" \"value\" : null\n" + | ||
"}" | ||
}, | ||
|
||
{new ConsumerRecord<>("testTopic", 1, | ||
1L, | ||
new Date().getTime(), | ||
null, | ||
0, | ||
0, | ||
allocateDirect(5), | ||
allocateDirect(5), | ||
new RecordHeaders(), | ||
Optional.empty()), | ||
"{\n" + | ||
" \"topic\" : \"testTopic\",\n" + | ||
" \"partition\" : 1,\n" + | ||
" \"leaderEpoch\" : null,\n" + | ||
" \"offset\" : 1,\n" + | ||
" \"headers\" : { },\n" + | ||
" \"key\" : \"\\\"AAAAAAA=\\\"\",\n" + | ||
" \"value\" : \"\\\"AAAAAAA=\\\"\"\n" + | ||
"}" | ||
}, | ||
}; | ||
} | ||
|
||
@Test(dataProvider = "data") | ||
public void deserializationTest(ConsumerRecord<?, ?> consumerRecord, String expected) throws Exception { | ||
var serialized = new ObjectMapper() | ||
.registerModule(new KafkaJacksonModule()) | ||
.writerWithDefaultPrettyPrinter() | ||
.writeValueAsString(consumerRecord); | ||
|
||
assertThat(serialized, is(expected)); | ||
} | ||
} |
Oops, something went wrong.