use of com.palantir.atlasdb.table.description.UniformRowNamePartitioner in project atlasdb by palantir.
the class Scrubber method runBackgroundScrubTask.
@VisibleForTesting
void runBackgroundScrubTask(final TransactionManager txManager) {
log.debug("Starting scrub task");
// Warning: Let T be the hard delete transaction that triggered a scrub, and let S be its
// start timestamp. If the locks for T happen to time out right after T checks that its
// locks are held but right before T writes its commit timestamp (extremely rare case), AND
// the unreadable timestamp is greater than S, then the scrub task could actually roll back
// the hard delete transaction (forcing it to abort or retry). Note that this doesn't affect
// correctness, but could be an annoying edge cause that causes hard delete to take longer
// than it otherwise would have.
Long immutableTimestamp = immutableTimestampSupplier.get();
Long unreadableTimestamp = unreadableTimestampSupplier.get();
final long maxScrubTimestamp = aggressiveScrub ? immutableTimestamp : Math.min(unreadableTimestamp, immutableTimestamp);
log.debug("Scrub task immutableTimestamp: {}, unreadableTimestamp: {}, maxScrubTimestamp: {}", immutableTimestamp, unreadableTimestamp, maxScrubTimestamp);
final int batchSize = (int) Math.ceil(batchSizeSupplier.get() * ((double) threadCount / readThreadCount));
List<byte[]> rangeBoundaries = Lists.newArrayList();
rangeBoundaries.add(PtBytes.EMPTY_BYTE_ARRAY);
if (readThreadCount > 1) {
// This will actually partition into the closest higher power of 2 number of ranges.
rangeBoundaries.addAll(Ordering.from(UnsignedBytes.lexicographicalComparator()).sortedCopy(new UniformRowNamePartitioner(ValueType.BLOB).getPartitions(readThreadCount - 1)));
}
rangeBoundaries.add(PtBytes.EMPTY_BYTE_ARRAY);
List<Future<Void>> readerFutures = Lists.newArrayList();
final AtomicInteger totalCellsRead = new AtomicInteger(0);
for (int i = 0; i < rangeBoundaries.size() - 1; i++) {
final byte[] startRow = rangeBoundaries.get(i);
final byte[] endRow = rangeBoundaries.get(i + 1);
readerFutures.add(readerExec.submit(() -> {
BatchingVisitable<SortedMap<Long, Multimap<TableReference, Cell>>> scrubQueue = scrubberStore.getBatchingVisitableScrubQueue(maxScrubTimestamp, startRow, endRow);
scrubQueue.batchAccept(batchSize, batch -> {
for (SortedMap<Long, Multimap<TableReference, Cell>> cells : batch) {
// We may actually get more cells than the batch size. The batch size is used
// for pulling off the scrub queue, and a single entry in the scrub queue may
// match multiple tables. These will get broken down into smaller batches later
// on when we actually do deletes.
int numCellsRead = scrubSomeCells(cells, txManager, maxScrubTimestamp);
int totalRead = totalCellsRead.addAndGet(numCellsRead);
log.debug("Scrub task processed {} cells in a batch, total {} processed so far.", numCellsRead, totalRead);
if (!isScrubEnabled.get()) {
log.debug("Stopping scrub for banned hours.");
break;
}
}
return isScrubEnabled.get();
});
return null;
}));
}
for (Future<Void> readerFuture : readerFutures) {
Futures.getUnchecked(readerFuture);
}
log.debug("Scrub background task running at timestamp {} processed a total of {} cells", maxScrubTimestamp, totalCellsRead.get());
log.debug("Finished scrub task");
}
Aggregations