use of io.pravega.client.SynchronizerClientFactory in project pravega by pravega.
the class WatermarkWorkflowTest method testWatermarkClient.
@Test(timeout = 10000L)
public void testWatermarkClient() {
Stream stream = new StreamImpl("scope", "stream");
SynchronizerClientFactory clientFactory = spy(SynchronizerClientFactory.class);
@Cleanup MockRevisionedStreamClient revisionedClient = new MockRevisionedStreamClient();
doAnswer(x -> revisionedClient).when(clientFactory).createRevisionedStreamClient(anyString(), any(), any());
@Cleanup PeriodicWatermarking.WatermarkClient client = new PeriodicWatermarking.WatermarkClient(stream, clientFactory);
// iteration 1 ==> null -> w1
client.reinitialize();
// There is no watermark in the stream. All values should be null and all writers active and participating.
assertEquals(revisionedClient.getMark(), MockRevision.EMPTY);
assertTrue(revisionedClient.watermarks.isEmpty());
assertEquals(client.getPreviousWatermark(), Watermark.EMPTY);
Map.Entry<String, WriterMark> entry0 = new AbstractMap.SimpleEntry<>("writerId", new WriterMark(0L, ImmutableMap.of()));
Map.Entry<String, WriterMark> entry1 = new AbstractMap.SimpleEntry<>("writerId", new WriterMark(1L, ImmutableMap.of()));
Map.Entry<String, WriterMark> entry2 = new AbstractMap.SimpleEntry<>("writerId", new WriterMark(2L, ImmutableMap.of()));
Map.Entry<String, WriterMark> entry3 = new AbstractMap.SimpleEntry<>("writerId", new WriterMark(3L, ImmutableMap.of()));
Map.Entry<String, WriterMark> entry4 = new AbstractMap.SimpleEntry<>("writerId", new WriterMark(4L, ImmutableMap.of()));
Map.Entry<String, WriterMark> entry5 = new AbstractMap.SimpleEntry<>("writerId", new WriterMark(5L, ImmutableMap.of()));
assertTrue(client.isWriterActive(entry0, 0L));
assertTrue(client.isWriterParticipating(0L));
Watermark first = new Watermark(1L, 2L, ImmutableMap.of());
client.completeIteration(first);
// iteration 2 : do not emit ==> w1 -> w1
client.reinitialize();
// There is one watermark. All writers should be active and writers greater than last watermark should be participating
assertEquals(revisionedClient.getMark(), MockRevision.EMPTY);
assertEquals(revisionedClient.watermarks.size(), 1);
assertEquals(client.getPreviousWatermark(), first);
assertTrue(client.isWriterActive(entry2, 0L));
assertFalse(client.isWriterActive(entry1, 0L));
assertTrue(client.isWriterTracked(entry1.getKey()));
assertFalse(client.isWriterParticipating(1L));
assertTrue(client.isWriterParticipating(2L));
// dont emit a watermark. Everything stays same as before.
client.completeIteration(null);
// iteration 3 : emit ==> w1 -> w1 w2
client.reinitialize();
// There is one watermark. All writers should be active and writers greater than last watermark should be participating
assertEquals(revisionedClient.getMark(), MockRevision.EMPTY);
assertEquals(revisionedClient.watermarks.size(), 1);
assertEquals(client.getPreviousWatermark(), first);
assertTrue(client.isWriterActive(entry2, 0L));
assertFalse(client.isWriterParticipating(1L));
assertTrue(client.isWriterParticipating(2L));
// emit second watermark
Watermark second = new Watermark(2L, 3L, ImmutableMap.of());
client.completeIteration(second);
// iteration 4: do not emit ==> w1 w2 -> w1 w2
client.reinitialize();
assertEquals(revisionedClient.getMark(), revisionedClient.watermarks.get(0).getKey());
assertEquals(2, revisionedClient.watermarks.size());
assertEquals(client.getPreviousWatermark(), second);
assertFalse(client.isWriterActive(entry2, 0L));
assertTrue(client.isWriterTracked(entry2.getKey()));
assertTrue(client.isWriterActive(entry3, 0L));
assertFalse(client.isWriterParticipating(2L));
assertTrue(client.isWriterParticipating(3L));
assertTrue(client.isWriterActive(entry0, 1000L));
assertTrue(client.isWriterTracked(entry0.getKey()));
// dont emit a watermark but complete this iteration.
client.completeIteration(null);
// iteration 6: emit ==> w1 w2 -> w1 w2 w3
client.reinitialize();
assertEquals(revisionedClient.getMark(), revisionedClient.watermarks.get(0).getKey());
assertEquals(2, revisionedClient.watermarks.size());
assertEquals(client.getPreviousWatermark(), second);
assertTrue(client.isWriterActive(entry3, 0L));
assertFalse(client.isWriterTracked(entry3.getKey()));
assertFalse(client.isWriterParticipating(2L));
assertTrue(client.isWriterParticipating(3L));
// emit third watermark
Watermark third = new Watermark(3L, 4L, ImmutableMap.of());
client.completeIteration(third);
// iteration 7: do not emit ==> w1 w2 w3 -> w1 w2 w3
client.reinitialize();
// active writers should be ahead of first watermark. participating writers should be ahead of second watermark
assertEquals(revisionedClient.getMark(), revisionedClient.watermarks.get(1).getKey());
assertEquals(3, revisionedClient.watermarks.size());
assertEquals(client.getPreviousWatermark(), third);
assertFalse(client.isWriterActive(entry3, 0L));
assertTrue(client.isWriterActive(entry4, 0L));
assertFalse(client.isWriterParticipating(3L));
assertTrue(client.isWriterParticipating(4L));
client.completeIteration(null);
// iteration 8 : emit ==> w2 w3 -> w2 w3 w4
client.reinitialize();
assertEquals(revisionedClient.getMark(), revisionedClient.watermarks.get(1).getKey());
// window = w2 w3
assertEquals(revisionedClient.watermarks.size(), 3);
assertEquals(client.getPreviousWatermark(), third);
assertFalse(client.isWriterActive(entry3, 0L));
assertTrue(client.isWriterActive(entry4, 0L));
assertFalse(client.isWriterParticipating(3L));
assertTrue(client.isWriterParticipating(4L));
// emit fourth watermark
Watermark fourth = new Watermark(4L, 5L, ImmutableMap.of());
client.completeIteration(fourth);
// iteration 9: do not emit ==> w1 w2 w3 w4 -> w1 w2 w3 w4.. check writer timeout
client.reinitialize();
assertEquals(revisionedClient.getMark(), revisionedClient.watermarks.get(2).getKey());
assertEquals(revisionedClient.watermarks.size(), 4);
assertEquals(client.getPreviousWatermark(), fourth);
assertFalse(client.isWriterActive(entry3, 0L));
assertTrue(client.isWriterTracked(entry4.getKey()));
assertFalse(client.isWriterParticipating(4L));
assertTrue(client.isWriterParticipating(5L));
// verify that writer is active if we specify a higher timeout
assertTrue(client.isWriterActive(entry1, 1000L));
assertTrue(client.isWriterTracked(entry1.getKey()));
// now that the writer is being tracked
assertFalse(Futures.delayedTask(() -> client.isWriterActive(entry1, 1L), Duration.ofSeconds(1), executor).join());
assertTrue(client.isWriterTracked(entry1.getKey()));
// dont emit a watermark but complete this iteration. This should shrink the window again.
client.completeIteration(null);
// iteration 10
client.reinitialize();
assertEquals(revisionedClient.getMark(), revisionedClient.watermarks.get(2).getKey());
assertEquals(revisionedClient.watermarks.size(), 4);
assertEquals(client.getPreviousWatermark(), fourth);
assertFalse(client.isWriterActive(entry4, 0L));
assertTrue(client.isWriterActive(entry5, 0L));
assertFalse(client.isWriterParticipating(4L));
assertTrue(client.isWriterParticipating(5L));
}
use of io.pravega.client.SynchronizerClientFactory in project pravega by pravega.
the class WatermarkWorkflowTest method testWatermarkingWorkflow.
@Test(timeout = 30000L)
public void testWatermarkingWorkflow() {
SynchronizerClientFactory clientFactory = spy(SynchronizerClientFactory.class);
ConcurrentHashMap<String, MockRevisionedStreamClient> revisionedStreamClientMap = new ConcurrentHashMap<>();
doAnswer(x -> {
String streamName = x.getArgument(0);
return revisionedStreamClientMap.compute(streamName, (s, rsc) -> {
if (rsc != null) {
return rsc;
} else {
return new MockRevisionedStreamClient();
}
});
}).when(clientFactory).createRevisionedStreamClient(anyString(), any(), any());
@Cleanup PeriodicWatermarking periodicWatermarking = new PeriodicWatermarking(streamMetadataStore, bucketStore, sp -> clientFactory, executor, new RequestTracker(false));
String streamName = "stream";
String scope = "scope";
streamMetadataStore.createScope(scope, null, executor).join();
streamMetadataStore.createStream(scope, streamName, StreamConfiguration.builder().scalingPolicy(ScalingPolicy.fixed(3)).timestampAggregationTimeout(10000L).build(), System.currentTimeMillis(), null, executor).join();
streamMetadataStore.setState(scope, streamName, State.ACTIVE, null, executor).join();
// set minimum number of segments to 1
StreamConfiguration config = StreamConfiguration.builder().scalingPolicy(ScalingPolicy.fixed(1)).timestampAggregationTimeout(10000L).build();
streamMetadataStore.startUpdateConfiguration(scope, streamName, config, null, executor).join();
VersionedMetadata<StreamConfigurationRecord> configRecord = streamMetadataStore.getConfigurationRecord(scope, streamName, null, executor).join();
streamMetadataStore.completeUpdateConfiguration(scope, streamName, configRecord, null, executor).join();
// 2. note writer1, writer2, writer3 marks
// writer 1 reports segments 0, 1.
// writer 2 reports segments 1, 2,
// writer 3 reports segment 0, 2
String writer1 = "writer1";
Map<Long, Long> map1 = ImmutableMap.of(0L, 100L, 1L, 200L);
streamMetadataStore.noteWriterMark(scope, streamName, writer1, 100L, map1, null, executor).join();
String writer2 = "writer2";
Map<Long, Long> map2 = ImmutableMap.of(1L, 100L, 2L, 200L);
streamMetadataStore.noteWriterMark(scope, streamName, writer2, 101L, map2, null, executor).join();
String writer3 = "writer3";
Map<Long, Long> map3 = ImmutableMap.of(2L, 100L, 0L, 200L);
streamMetadataStore.noteWriterMark(scope, streamName, writer3, 102L, map3, null, executor).join();
// 3. run watermarking workflow.
StreamImpl stream = new StreamImpl(scope, streamName);
periodicWatermarking.watermark(stream).join();
// verify that a watermark has been emitted.
// this should emit a watermark that contains all three segments with offsets = 200L
// and timestamp = 100L
MockRevisionedStreamClient revisionedClient = revisionedStreamClientMap.get(NameUtils.getMarkStreamForStream(streamName));
assertEquals(revisionedClient.watermarks.size(), 1);
Watermark watermark = revisionedClient.watermarks.get(0).getValue();
assertEquals(watermark.getLowerTimeBound(), 100L);
assertEquals(watermark.getStreamCut().size(), 3);
assertEquals(getSegmentOffset(watermark, 0L), 200L);
assertEquals(getSegmentOffset(watermark, 1L), 200L);
assertEquals(getSegmentOffset(watermark, 2L), 200L);
// send positions only on segment 1 and segment 2. nothing on segment 0.
map1 = ImmutableMap.of(1L, 300L);
streamMetadataStore.noteWriterMark(scope, streamName, writer1, 200L, map1, null, executor).join();
map2 = ImmutableMap.of(1L, 100L, 2L, 300L);
streamMetadataStore.noteWriterMark(scope, streamName, writer2, 201L, map2, null, executor).join();
map3 = ImmutableMap.of(2L, 300L);
streamMetadataStore.noteWriterMark(scope, streamName, writer3, 202L, map3, null, executor).join();
// run watermark workflow. this will emit a watermark with time = 200L and streamcut = 0 -> 200L, 1 -> 300L, 2 -> 300L
periodicWatermarking.watermark(stream).join();
assertEquals(revisionedClient.watermarks.size(), 2);
watermark = revisionedClient.watermarks.get(1).getValue();
assertEquals(watermark.getLowerTimeBound(), 200L);
assertEquals(watermark.getStreamCut().size(), 3);
assertEquals(getSegmentOffset(watermark, 0L), 200L);
assertEquals(getSegmentOffset(watermark, 1L), 300L);
assertEquals(getSegmentOffset(watermark, 2L), 300L);
// scale stream 0, 1, 2 -> 3, 4
scaleStream(streamName, scope);
// writer 1 reports segments 0, 1.
// writer 2 reports segments 1, 2
// writer 3 reports segment 3
map1 = ImmutableMap.of(0L, 300L, 1L, 400L);
streamMetadataStore.noteWriterMark(scope, streamName, writer1, 302L, map1, null, executor).join();
map2 = ImmutableMap.of(1L, 100L, 2L, 400L);
streamMetadataStore.noteWriterMark(scope, streamName, writer2, 301L, map2, null, executor).join();
long segment3 = NameUtils.computeSegmentId(3, 1);
long segment4 = NameUtils.computeSegmentId(4, 1);
map3 = ImmutableMap.of(segment3, 100L);
// writer 3 has lowest reported time.
streamMetadataStore.noteWriterMark(scope, streamName, writer3, 300L, map3, null, executor).join();
// run watermark workflow. this will emit a watermark with time = 300L and streamcut = 3 -> 100L, 4 -> 0L
periodicWatermarking.watermark(stream).join();
assertEquals(revisionedClient.watermarks.size(), 3);
watermark = revisionedClient.watermarks.get(2).getValue();
assertEquals(watermark.getLowerTimeBound(), 300L);
assertEquals(watermark.getStreamCut().size(), 2);
assertEquals(getSegmentOffset(watermark, segment3), 100L);
assertEquals(getSegmentOffset(watermark, segment4), 0L);
// report complete positions from writers.
// writer 1 reports 0, 1, 2
// writer 2 reports 0, 1, 2
// writer 3 doesnt report.
map1 = ImmutableMap.of(0L, 400L, 1L, 400L);
streamMetadataStore.noteWriterMark(scope, streamName, writer1, 400L, map1, null, executor).join();
map2 = ImmutableMap.of(1L, 100L, 2L, 400L);
streamMetadataStore.noteWriterMark(scope, streamName, writer2, 401L, map2, null, executor).join();
// run watermark workflow. there shouldn't be a watermark emitted because writer 3 is active and has not reported a time.
periodicWatermarking.watermark(stream).join();
assertEquals(revisionedClient.watermarks.size(), 3);
// even though writer3 is excluded from computation, its mark is still not removed because it is still active
WriterMark writer3Mark = streamMetadataStore.getWriterMark(scope, streamName, writer3, null, executor).join();
assertTrue(writer3Mark.isAlive());
assertEquals(writer3Mark.getTimestamp(), 300L);
// report shutdown of writer 3
streamMetadataStore.shutdownWriter(scope, streamName, writer3, null, executor).join();
writer3Mark = streamMetadataStore.getWriterMark(scope, streamName, writer3, null, executor).join();
assertFalse(writer3Mark.isAlive());
assertEquals(writer3Mark.getTimestamp(), 300L);
// now a watermark should be generated. Time should be advanced. But watermark's stream cut is already ahead of writer's
// positions so stream cut should not advance.
// Also writer 3 being inactive and shutdown, should be removed.
periodicWatermarking.watermark(stream).join();
assertEquals(revisionedClient.watermarks.size(), 4);
watermark = revisionedClient.watermarks.get(3).getValue();
assertEquals(watermark.getLowerTimeBound(), 400L);
assertEquals(watermark.getStreamCut().size(), 2);
assertEquals(getSegmentOffset(watermark, segment3), 100L);
assertEquals(getSegmentOffset(watermark, segment4), 0L);
AssertExtensions.assertFutureThrows("Writer 3 should have been removed from store", streamMetadataStore.getWriterMark(scope, streamName, writer3, null, executor), e -> Exceptions.unwrap(e) instanceof StoreException.DataNotFoundException);
// writer 1, 2 and 3 report marks. With writer 3 reporting mark on segment 4. Writer3 will get added again
map1 = ImmutableMap.of(0L, 500L, 1L, 500L);
streamMetadataStore.noteWriterMark(scope, streamName, writer1, 500L, map1, null, executor).join();
map2 = ImmutableMap.of(1L, 100L, 2L, 500L);
streamMetadataStore.noteWriterMark(scope, streamName, writer2, 501L, map2, null, executor).join();
map3 = ImmutableMap.of(segment4, 500L);
streamMetadataStore.noteWriterMark(scope, streamName, writer3, 502L, map3, null, executor).join();
// run watermarking workflow. It should generate watermark that includes segments 3 -> 100L and 4 -> 500L with time 500L
periodicWatermarking.watermark(stream).join();
assertEquals(revisionedClient.watermarks.size(), 5);
watermark = revisionedClient.watermarks.get(4).getValue();
assertEquals(watermark.getLowerTimeBound(), 500L);
assertEquals(watermark.getStreamCut().size(), 2);
assertEquals(getSegmentOffset(watermark, segment3), 100L);
assertEquals(getSegmentOffset(watermark, segment4), 500L);
}
use of io.pravega.client.SynchronizerClientFactory in project pravega by pravega.
the class RestoreBackUpDataRecoveryTest method getWatermarks.
/**
* Gets watermarks used while writing the events
*/
private LinkedBlockingQueue<Watermark> getWatermarks(PravegaRunner pravegaRunner, AtomicBoolean stopFlag, CompletableFuture<Void> writerFuture) throws Exception {
@Cleanup SynchronizerClientFactory syncClientFactory = SynchronizerClientFactory.withScope(SCOPE, ClientConfig.builder().controllerURI(pravegaRunner.controllerRunner.controllerURI).build());
String markStream = NameUtils.getMarkStreamForStream(STREAM1);
RevisionedStreamClient<Watermark> watermarkReader = syncClientFactory.createRevisionedStreamClient(markStream, new WatermarkSerializer(), SynchronizerConfig.builder().build());
LinkedBlockingQueue<Watermark> watermarks = new LinkedBlockingQueue<>();
CompletableFuture<Void> fetchWaterMarksFuture = fetchWatermarks(watermarkReader, watermarks, stopFlag);
AssertExtensions.assertEventuallyEquals(true, () -> watermarks.size() >= 2, 100000);
stopFlag.set(true);
fetchWaterMarksFuture.join();
writerFuture.join();
return watermarks;
}
use of io.pravega.client.SynchronizerClientFactory in project pravega by pravega.
the class WatermarkingTest method progressingWatermarkWithWriterTimeouts.
@Test(timeout = 60000)
public void progressingWatermarkWithWriterTimeouts() throws Exception {
String scope = "Timeout";
String streamName = "progressingWatermarkWithWriterTimeouts";
int numSegments = 1;
ClientConfig clientConfig = ClientConfig.builder().controllerURI(PRAVEGA.getControllerURI()).build();
@Cleanup StreamManager streamManager = StreamManager.create(clientConfig);
assertNotNull(streamManager);
streamManager.createScope(scope);
streamManager.createStream(scope, streamName, StreamConfiguration.builder().scalingPolicy(ScalingPolicy.fixed(numSegments)).build());
@Cleanup EventStreamClientFactory clientFactory = EventStreamClientFactory.withScope(scope, clientConfig);
@Cleanup SynchronizerClientFactory syncClientFactory = SynchronizerClientFactory.withScope(scope, clientConfig);
String markStream = NameUtils.getMarkStreamForStream(streamName);
@Cleanup RevisionedStreamClient<Watermark> watermarkReader = syncClientFactory.createRevisionedStreamClient(markStream, new WatermarkSerializer(), SynchronizerConfig.builder().build());
LinkedBlockingQueue<Watermark> watermarks = new LinkedBlockingQueue<>();
AtomicBoolean stopFlag = new AtomicBoolean(false);
fetchWatermarks(watermarkReader, watermarks, stopFlag);
// create two writers and write two sevent and call note time for each writer.
@Cleanup EventStreamWriter<String> writer1 = clientFactory.createEventWriter(streamName, new JavaSerializer<>(), EventWriterConfig.builder().build());
writer1.writeEvent("1").get();
writer1.noteTime(100L);
@Cleanup EventStreamWriter<String> writer2 = clientFactory.createEventWriter(streamName, new JavaSerializer<>(), EventWriterConfig.builder().build());
writer2.writeEvent("2").get();
writer2.noteTime(102L);
// writer0 should timeout. writer1 and writer2 should result in two more watermarks with following times:
// 1: 100L-101L 2: 101-101
// then first writer should timeout and be discarded. But second writer should continue to be active as its time
// is higher than first watermark. This should result in a second watermark to be emitted.
AssertExtensions.assertEventuallyEquals(true, () -> watermarks.size() == 2, 100000);
Watermark watermark1 = watermarks.poll();
Watermark watermark2 = watermarks.poll();
assertEquals(100L, watermark1.getLowerTimeBound());
assertEquals(102L, watermark1.getUpperTimeBound());
assertEquals(102L, watermark2.getLowerTimeBound());
assertEquals(102L, watermark2.getUpperTimeBound());
// stream cut should be same
assertTrue(watermark2.getStreamCut().entrySet().stream().allMatch(x -> watermark1.getStreamCut().get(x.getKey()).equals(x.getValue())));
// bring back writer1 and post an event with note time smaller than current watermark
writer1.writeEvent("3").get();
writer1.noteTime(101L);
// no watermark should be emitted.
Watermark nullMark = watermarks.poll(10, TimeUnit.SECONDS);
assertNull(nullMark);
}
use of io.pravega.client.SynchronizerClientFactory in project pravega by pravega.
the class RevisionedStreamClientTest method testSegmentTruncation.
@Test
public void testSegmentTruncation() {
String scope = "scope";
String stream = "stream";
PravegaNodeUri endpoint = new PravegaNodeUri("localhost", SERVICE_PORT);
@Cleanup MockConnectionFactoryImpl connectionFactory = new MockConnectionFactoryImpl();
@Cleanup MockController controller = new MockController(endpoint.getEndpoint(), endpoint.getPort(), connectionFactory, false);
createScopeAndStream(scope, stream, controller);
MockSegmentStreamFactory streamFactory = new MockSegmentStreamFactory();
@Cleanup SynchronizerClientFactory clientFactory = new ClientFactoryImpl(scope, controller, connectionFactory, streamFactory, streamFactory, streamFactory, streamFactory);
SynchronizerConfig config = SynchronizerConfig.builder().build();
@Cleanup RevisionedStreamClient<String> client = clientFactory.createRevisionedStreamClient(stream, new JavaSerializer<>(), config);
Revision r0 = client.fetchLatestRevision();
client.writeUnconditionally("a");
Revision ra = client.fetchLatestRevision();
client.writeUnconditionally("b");
Revision rb = client.fetchLatestRevision();
client.writeUnconditionally("c");
Revision rc = client.fetchLatestRevision();
assertEquals(r0, client.fetchOldestRevision());
client.truncateToRevision(r0);
assertEquals(r0, client.fetchOldestRevision());
client.truncateToRevision(ra);
assertEquals(ra, client.fetchOldestRevision());
client.truncateToRevision(r0);
assertEquals(ra, client.fetchOldestRevision());
assertThrows(TruncatedDataException.class, () -> client.readFrom(r0));
Iterator<Entry<Revision, String>> iterA = client.readFrom(ra);
assertTrue(iterA.hasNext());
Iterator<Entry<Revision, String>> iterB = client.readFrom(ra);
assertTrue(iterB.hasNext());
assertEquals("b", iterA.next().getValue());
assertEquals("b", iterB.next().getValue());
client.truncateToRevision(rb);
assertTrue(iterA.hasNext());
assertEquals("c", iterA.next().getValue());
client.truncateToRevision(rc);
assertFalse(iterA.hasNext());
assertTrue(iterB.hasNext());
assertThrows(TruncatedDataException.class, () -> iterB.next());
}
Aggregations