use of io.smallrye.reactive.messaging.kafka.KafkaConnectorIncomingConfiguration in project smallrye-reactive-messaging by smallrye.
the class KafkaCommitHandlerTest method testSourceWithAutoCommitDisabled.
@Test
public void testSourceWithAutoCommitDisabled() throws ExecutionException, InterruptedException, TimeoutException {
MapBasedConfig config = newCommonConfigForSource().with("group.id", "test-source-with-auto-commit-disabled").with("value.deserializer", IntegerDeserializer.class.getName()).with("commit-strategy", "latest");
KafkaConnectorIncomingConfiguration ic = new KafkaConnectorIncomingConfiguration(config);
source = new KafkaSource<>(vertx, "test-source-with-auto-commit-disabled", ic, UnsatisfiedInstance.instance(), CountKafkaCdiEvents.noCdiEvents, UnsatisfiedInstance.instance(), -1);
List<Message<?>> messages = Collections.synchronizedList(new ArrayList<>());
source.getStream().subscribe().with(messages::add);
companion.produceIntegers().usingGenerator(i -> new ProducerRecord<>(topic, i), 10);
await().atMost(2, TimeUnit.MINUTES).until(() -> messages.size() >= 10);
assertThat(messages.stream().map(m -> ((KafkaRecord<String, Integer>) m).getPayload()).collect(Collectors.toList())).containsExactly(0, 1, 2, 3, 4, 5, 6, 7, 8, 9);
Message<?> last = messages.get(messages.size() - 1);
last.ack().toCompletableFuture().get(2, TimeUnit.MINUTES);
await().ignoreExceptions().untilAsserted(() -> {
TopicPartition topicPartition = new TopicPartition(topic, 0);
OffsetAndMetadata offset = companion.consumerGroups().offsets("test-source-with-auto-commit-disabled", topicPartition);
assertNotNull(offset);
assertEquals(10L, offset.offset());
});
}
use of io.smallrye.reactive.messaging.kafka.KafkaConnectorIncomingConfiguration in project smallrye-reactive-messaging by smallrye.
the class KafkaCommitHandlerTest method testSourceWithThrottledLatestProcessedCommitEnabled.
@Test
public void testSourceWithThrottledLatestProcessedCommitEnabled() {
MapBasedConfig config = newCommonConfigForSource().with("client.id", UUID.randomUUID().toString()).with("group.id", "test-source-with-throttled-latest-processed-commit").with("auto.offset.reset", "earliest").with("value.deserializer", IntegerDeserializer.class.getName()).with("commit-strategy", "throttled").with("throttled.unprocessed-record-max-age.ms", 100);
KafkaConnectorIncomingConfiguration ic = new KafkaConnectorIncomingConfiguration(config);
source = new KafkaSource<>(vertx, "test-source-with-throttled-latest-processed-commit", ic, UnsatisfiedInstance.instance(), CountKafkaCdiEvents.noCdiEvents, UnsatisfiedInstance.instance(), -1);
List<Message<?>> messages = Collections.synchronizedList(new ArrayList<>());
source.getStream().subscribe().with(messages::add);
companion.produceIntegers().usingGenerator(i -> new ProducerRecord<>(topic, i), 10);
await().atMost(2, TimeUnit.MINUTES).until(() -> messages.size() >= 10);
assertThat(messages.stream().map(m -> ((KafkaRecord<String, Integer>) m).getPayload()).collect(Collectors.toList())).containsExactly(0, 1, 2, 3, 4, 5, 6, 7, 8, 9);
await().atMost(2, TimeUnit.MINUTES).ignoreExceptions().untilAsserted(() -> {
// we must keep acking to eventually induce a commit
messages.forEach(Message::ack);
TopicPartition topicPartition = new TopicPartition(topic, 0);
OffsetAndMetadata offset = companion.consumerGroups().offsets("test-source-with-throttled-latest-processed-commit", topicPartition);
assertNotNull(offset);
assertEquals(10L, offset.offset());
});
await().atMost(2, TimeUnit.MINUTES).untilAsserted(() -> {
HealthReport.HealthReportBuilder healthReportBuilder = HealthReport.builder();
source.isAlive(healthReportBuilder);
assertTrue(healthReportBuilder.build().isOk());
});
}
use of io.smallrye.reactive.messaging.kafka.KafkaConnectorIncomingConfiguration in project smallrye-reactive-messaging by smallrye.
the class KafkaSourceWithCloudEventsTest method testReceivingStructuredCloudEventsWithByteArrayDeserializer.
@SuppressWarnings("unchecked")
@Test
public void testReceivingStructuredCloudEventsWithByteArrayDeserializer() {
KafkaMapBasedConfig config = newCommonConfig();
config.put("topic", topic);
config.put("value.deserializer", ByteArrayDeserializer.class.getName());
config.put("channel-name", topic);
KafkaConnectorIncomingConfiguration ic = new KafkaConnectorIncomingConfiguration(config);
source = new KafkaSource<>(vertx, UUID.randomUUID().toString(), ic, UnsatisfiedInstance.instance(), CountKafkaCdiEvents.noCdiEvents, UnsatisfiedInstance.instance(), -1);
List<Message<?>> messages = new ArrayList<>();
source.getStream().subscribe().with(messages::add);
companion.produce(String.class, JsonObject.class).fromRecords(new ProducerRecord<>(topic, null, null, "key", new JsonObject().put("specversion", CloudEventMetadata.CE_VERSION_1_0).put("type", "type").put("id", "id").put("source", "test://test").put("subject", "foo").put("data", new JsonObject().put("name", "neo")), Collections.singletonList(new RecordHeader("content-type", "application/cloudevents+json; charset=utf-8".getBytes()))));
await().atMost(2, TimeUnit.MINUTES).until(() -> messages.size() >= 1);
Message<?> message = messages.get(0);
IncomingKafkaCloudEventMetadata<String, JsonObject> metadata = message.getMetadata(IncomingKafkaCloudEventMetadata.class).orElse(null);
assertThat(metadata).isNotNull();
assertThat(metadata.getSpecVersion()).isEqualTo(CloudEventMetadata.CE_VERSION_1_0);
assertThat(metadata.getType()).isEqualTo("type");
assertThat(metadata.getId()).isEqualTo("id");
assertThat(metadata.getSource()).isEqualTo(URI.create("test://test"));
assertThat(metadata.getSubject()).hasValue("foo");
assertThat(metadata.getData().getString("name")).isEqualTo("neo");
// Extensions
assertThat(metadata.getKey()).isEqualTo("key");
// Rule 3.1 - partitionkey attribute
assertThat(metadata.<String>getExtension("partitionkey")).hasValue("key");
assertThat(metadata.getTopic()).isEqualTo(topic);
assertThat(message.getPayload()).isInstanceOf(JsonObject.class);
assertThat(((JsonObject) message.getPayload()).getString("name")).isEqualTo("neo");
}
use of io.smallrye.reactive.messaging.kafka.KafkaConnectorIncomingConfiguration in project smallrye-reactive-messaging by smallrye.
the class ClientTestBase method createSourceSeekToBeginning.
public KafkaSource<Integer, String> createSourceSeekToBeginning() {
String groupId = UUID.randomUUID().toString();
MapBasedConfig config = createConsumerConfig(groupId).put("topic", topic);
SingletonInstance<KafkaConsumerRebalanceListener> listeners = new SingletonInstance<>(groupId, getKafkaConsumerRebalanceListenerAwaitingAssignationAndSeekToBeginning());
source = new KafkaSource<>(vertx, groupId, new KafkaConnectorIncomingConfiguration(config), listeners, CountKafkaCdiEvents.noCdiEvents, UnsatisfiedInstance.instance(), 0);
return source;
}
use of io.smallrye.reactive.messaging.kafka.KafkaConnectorIncomingConfiguration in project smallrye-reactive-messaging by smallrye.
the class HighLatencyTest method testHighLatency.
@Test
public void testHighLatency() throws InterruptedException, IOException {
MapBasedConfig config = newCommonConfigForSource().with("value.deserializer", IntegerDeserializer.class.getName()).with(ConsumerConfig.SESSION_TIMEOUT_MS_CONFIG, 6000).with(ConsumerConfig.HEARTBEAT_INTERVAL_MS_CONFIG, 100).with("retry", true).with("retry-attempts", 100).with("retry-max-wait", 30);
KafkaConnectorIncomingConfiguration ic = new KafkaConnectorIncomingConfiguration(config);
source = new KafkaSource<>(vertx, UUID.randomUUID().toString(), ic, UnsatisfiedInstance.instance(), CountKafkaCdiEvents.noCdiEvents, UnsatisfiedInstance.instance(), -1);
List<KafkaRecord<?, ?>> messages1 = new ArrayList<>();
source.getStream().subscribe().with(messages1::add);
companion.produceIntegers().usingGenerator(i -> new ProducerRecord<>(topic, i), 10);
await().atMost(2, TimeUnit.MINUTES).until(() -> messages1.size() >= 10);
toxics().latency("latency", ToxicDirection.UPSTREAM, 6000 + 1000);
// session timeout + a bit more just in case.
Thread.sleep(6000 + 2000);
toxics().get("latency").remove();
companion.produceIntegers().usingGenerator(i -> new ProducerRecord<>(topic, 10 + i), 10);
await().atMost(2, TimeUnit.MINUTES).until(() -> messages1.size() >= 20);
assertThat(messages1.size()).isGreaterThanOrEqualTo(20);
}
Aggregations