use of io.pravega.segmentstore.contracts.SegmentProperties in project pravega by pravega.
the class GetInfoOperation method call.
@Override
public SegmentProperties call() throws IOException {
String segmentName = getTarget();
long traceId = LoggerHelpers.traceEnter(log, "getStreamSegmentInfo", segmentName);
List<FileDescriptor> allFiles = findAll(segmentName, true);
SegmentProperties result = null;
int attemptCount = 0;
do {
val last = allFiles.get(allFiles.size() - 1);
long length = last.getOffset() + last.getLength();
boolean isSealed;
try {
isSealed = isSealed(last);
} catch (FileNotFoundException fnf) {
// be deleted in that case so we need to try our luck again (in which case we need to refresh the file list).
if (++attemptCount < MAX_ATTEMPT_COUNT) {
allFiles = findAll(segmentName, true);
continue;
}
throw fnf;
}
result = StreamSegmentInformation.builder().name(segmentName).length(length).sealed(isSealed).build();
} while (result == null);
LoggerHelpers.traceLeave(log, "getStreamSegmentInfo", traceId, segmentName, result);
return result;
}
use of io.pravega.segmentstore.contracts.SegmentProperties in project pravega by pravega.
the class AppendProcessorTest method testEventNumbers.
@Test
public void testEventNumbers() {
String streamSegmentName = "testAppendSegment";
UUID clientId = UUID.randomUUID();
byte[] data = new byte[] { 1, 2, 3, 4, 6, 7, 8, 9 };
StreamSegmentStore store = mock(StreamSegmentStore.class);
ServerConnection connection = mock(ServerConnection.class);
AppendProcessor processor = new AppendProcessor(store, connection, new FailingRequestProcessor(), null);
CompletableFuture<SegmentProperties> propsFuture = CompletableFuture.completedFuture(StreamSegmentInformation.builder().name(streamSegmentName).build());
when(store.getStreamSegmentInfo(streamSegmentName, true, AppendProcessor.TIMEOUT)).thenReturn(propsFuture);
processor.setupAppend(new SetupAppend(1, clientId, streamSegmentName, ""));
verify(store).getStreamSegmentInfo(streamSegmentName, true, AppendProcessor.TIMEOUT);
CompletableFuture<Void> result = CompletableFuture.completedFuture(null);
int eventCount = 100;
when(store.append(streamSegmentName, data, updateEventNumber(clientId, 100, SegmentMetadata.NULL_ATTRIBUTE_VALUE, eventCount), AppendProcessor.TIMEOUT)).thenReturn(result);
processor.append(new Append(streamSegmentName, clientId, 100, eventCount, Unpooled.wrappedBuffer(data), null));
verify(store).append(streamSegmentName, data, updateEventNumber(clientId, 100, SegmentMetadata.NULL_ATTRIBUTE_VALUE, eventCount), AppendProcessor.TIMEOUT);
Map<UUID, Long> map = new HashMap<>();
map.put(clientId, 100L);
map.put(EVENT_COUNT, 100L);
propsFuture = CompletableFuture.completedFuture(StreamSegmentInformation.builder().name(streamSegmentName).attributes(map).build());
when(store.append(streamSegmentName, data, updateEventNumber(clientId, 200, 100, eventCount), AppendProcessor.TIMEOUT)).thenReturn(result);
processor.append(new Append(streamSegmentName, clientId, 200, eventCount, Unpooled.wrappedBuffer(data), null));
verify(store).append(streamSegmentName, data, updateEventNumber(clientId, 200, 100, eventCount), AppendProcessor.TIMEOUT);
verifyNoMoreInteractions(store);
}
use of io.pravega.segmentstore.contracts.SegmentProperties in project pravega by pravega.
the class AppendProcessorTest method testEventNumbersOldClient.
@Test
public void testEventNumbersOldClient() {
String streamSegmentName = "testAppendSegment";
UUID clientId = UUID.randomUUID();
byte[] data = new byte[] { 1, 2, 3, 4, 6, 7, 8, 9 };
StreamSegmentStore store = mock(StreamSegmentStore.class);
ServerConnection connection = mock(ServerConnection.class);
AppendProcessor processor = new AppendProcessor(store, connection, new FailingRequestProcessor(), null);
CompletableFuture<SegmentProperties> propsFuture = CompletableFuture.completedFuture(StreamSegmentInformation.builder().name(streamSegmentName).attributes(Collections.singletonMap(clientId, 100L)).build());
when(store.getStreamSegmentInfo(streamSegmentName, true, AppendProcessor.TIMEOUT)).thenReturn(propsFuture);
processor.setupAppend(new SetupAppend(1, clientId, streamSegmentName, ""));
verify(store).getStreamSegmentInfo(streamSegmentName, true, AppendProcessor.TIMEOUT);
int eventCount = 10;
CompletableFuture<Void> result = CompletableFuture.completedFuture(null);
when(store.append(streamSegmentName, data, updateEventNumber(clientId, 200, 100, eventCount), AppendProcessor.TIMEOUT)).thenReturn(result);
processor.append(new Append(streamSegmentName, clientId, 200, eventCount, Unpooled.wrappedBuffer(data), null));
verify(store).append(streamSegmentName, data, updateEventNumber(clientId, 200, 100, eventCount), AppendProcessor.TIMEOUT);
when(store.append(streamSegmentName, data, updateEventNumber(clientId, 300, 200, eventCount), AppendProcessor.TIMEOUT)).thenReturn(result);
processor.append(new Append(streamSegmentName, clientId, 300, eventCount, Unpooled.wrappedBuffer(data), null));
verify(store).append(streamSegmentName, data, updateEventNumber(clientId, 300, 200, eventCount), AppendProcessor.TIMEOUT);
verifyNoMoreInteractions(store);
}
use of io.pravega.segmentstore.contracts.SegmentProperties in project pravega by pravega.
the class AppendProcessorTest method testDelayedDataAppended.
/**
* Test to ensure newer appends are processed only after successfully sending the DataAppended acknowledgement
* back to client. This test tests the following:
* - If sending first DataAppended is blocked, ensure future appends are not written to store.
* - Once the first DataAppended is sent ensure the remaining appends are written to store and DataAppended ack'ed
* back.
*/
@Test(timeout = 15 * 1000)
public void testDelayedDataAppended() throws Exception {
ReusableLatch firstStoreAppendInvoked = new ReusableLatch();
ReusableLatch completeFirstDataAppendedAck = new ReusableLatch();
ReusableLatch secondStoreAppendInvoked = new ReusableLatch();
@Cleanup("shutdownNow") ScheduledExecutorService nettyExecutor = ExecutorServiceHelpers.newScheduledThreadPool(1, "Netty-threadPool");
String streamSegmentName = "testDelayedAppend";
UUID clientId = UUID.randomUUID();
byte[] data = new byte[] { 1, 2, 3, 4, 6, 7, 8, 9 };
StreamSegmentStore store = mock(StreamSegmentStore.class);
ServerConnection connection = mock(ServerConnection.class);
// Ensure the first DataAppended is hung/delayed.
doAnswer(invocation -> {
firstStoreAppendInvoked.release();
// wait, simulating a hung/delayed dataAppended acknowledgement.
completeFirstDataAppendedAck.await();
return null;
}).doAnswer(invocation -> {
secondStoreAppendInvoked.release();
return null;
}).when(connection).send(any(DataAppended.class));
AppendProcessor processor = new AppendProcessor(store, connection, new FailingRequestProcessor(), null);
CompletableFuture<SegmentProperties> propsFuture = CompletableFuture.completedFuture(StreamSegmentInformation.builder().name(streamSegmentName).build());
when(store.getStreamSegmentInfo(streamSegmentName, true, AppendProcessor.TIMEOUT)).thenReturn(propsFuture);
processor.setupAppend(new SetupAppend(1, clientId, streamSegmentName, ""));
verify(store).getStreamSegmentInfo(streamSegmentName, true, AppendProcessor.TIMEOUT);
CompletableFuture<Void> result = CompletableFuture.completedFuture(null);
int eventCount = 100;
when(store.append(streamSegmentName, data, updateEventNumber(clientId, 100, SegmentMetadata.NULL_ATTRIBUTE_VALUE, eventCount), AppendProcessor.TIMEOUT)).thenReturn(result);
// Trigger the first append, here the sending of DataAppended ack will be delayed/hung.
nettyExecutor.submit(() -> processor.append(new Append(streamSegmentName, clientId, 100, eventCount, Unpooled.wrappedBuffer(data), null)));
firstStoreAppendInvoked.await();
verify(store).append(streamSegmentName, data, updateEventNumber(clientId, 100, SegmentMetadata.NULL_ATTRIBUTE_VALUE, eventCount), AppendProcessor.TIMEOUT);
/* Trigger the next append. This should be completed immediately and should not cause a store.append to be
invoked as the previous DataAppended ack is still not sent. */
processor.append(new Append(streamSegmentName, clientId, 200, eventCount, Unpooled.wrappedBuffer(data), null));
// Since the first Ack was never sent the next append should not be written to the store.
verifyNoMoreInteractions(store);
// Setup mock for check behaviour after the delayed/hung dataAppended completes.
when(store.append(streamSegmentName, data, updateEventNumber(clientId, 200, 100, eventCount), AppendProcessor.TIMEOUT)).thenReturn(result);
// Now ensure the dataAppended sent
completeFirstDataAppendedAck.release();
// wait until the next store append is invoked.
secondStoreAppendInvoked.await();
// Verify that the next store append invoked.
verify(store).append(streamSegmentName, data, updateEventNumber(clientId, 200, 100, eventCount), AppendProcessor.TIMEOUT);
// Verify two DataAppended acks are sent out.
verify(connection, times(2)).send(any(DataAppended.class));
verify(connection).send(new DataAppended(clientId, 100, Long.MIN_VALUE));
verify(connection).send(new DataAppended(clientId, 200, 100));
}
use of io.pravega.segmentstore.contracts.SegmentProperties in project pravega by pravega.
the class SegmentStatsRecorderTest method setup.
@Before
public void setup() {
AutoScaleProcessor processor = mock(AutoScaleProcessor.class);
StreamSegmentStore store = mock(StreamSegmentStore.class);
CompletableFuture<SegmentProperties> toBeReturned = CompletableFuture.completedFuture(new SegmentProperties() {
@Override
public String getName() {
return STREAM_SEGMENT_NAME;
}
@Override
public boolean isSealed() {
return false;
}
@Override
public boolean isDeleted() {
return false;
}
@Override
public ImmutableDate getLastModified() {
return null;
}
@Override
public long getStartOffset() {
return 0;
}
@Override
public long getLength() {
return 0;
}
@Override
public Map<UUID, Long> getAttributes() {
Map<UUID, Long> map = new HashMap<>();
map.put(Attributes.SCALE_POLICY_TYPE, 0L);
map.put(Attributes.SCALE_POLICY_RATE, 10L);
latch.complete(null);
return map;
}
});
when(store.getStreamSegmentInfo(STREAM_SEGMENT_NAME, false, Duration.ofMinutes(1))).thenReturn(toBeReturned);
statsRecorder = new SegmentStatsRecorderImpl(processor, store, 10000, 2, TimeUnit.SECONDS, executor, maintenanceExecutor);
}
Aggregations