use of org.apache.cassandra.locator.AbstractReplicationStrategy in project cassandra by apache.
the class ActiveRepairServiceTest method testGetNeighborsTimesTwo.
@Test
public void testGetNeighborsTimesTwo() throws Throwable {
TokenMetadata tmd = StorageService.instance.getTokenMetadata();
// generate rf*2 nodes, and ensure that only neighbors specified by the ARS are returned
addTokens(2 * Keyspace.open(KEYSPACE5).getReplicationStrategy().getReplicationFactor());
AbstractReplicationStrategy ars = Keyspace.open(KEYSPACE5).getReplicationStrategy();
Set<InetAddress> expected = new HashSet<>();
for (Range<Token> replicaRange : ars.getAddressRanges().get(FBUtilities.getBroadcastAddress())) {
expected.addAll(ars.getRangeAddresses(tmd.cloneOnlyTokenMap()).get(replicaRange));
}
expected.remove(FBUtilities.getBroadcastAddress());
Collection<Range<Token>> ranges = StorageService.instance.getLocalRanges(KEYSPACE5);
Set<InetAddress> neighbors = new HashSet<>();
for (Range<Token> range : ranges) {
neighbors.addAll(ActiveRepairService.getNeighbors(KEYSPACE5, ranges, range, null, null));
}
assertEquals(expected, neighbors);
}
use of org.apache.cassandra.locator.AbstractReplicationStrategy in project cassandra by apache.
the class MoveTest method testSimultaneousMove.
/*
* Test ranges and write endpoints when multiple nodes are on the move simultaneously
*/
@Test
public void testSimultaneousMove() throws UnknownHostException {
StorageService ss = StorageService.instance;
final int RING_SIZE = 10;
TokenMetadata tmd = ss.getTokenMetadata();
IPartitioner partitioner = RandomPartitioner.instance;
VersionedValue.VersionedValueFactory valueFactory = new VersionedValue.VersionedValueFactory(partitioner);
ArrayList<Token> endpointTokens = new ArrayList<Token>();
ArrayList<Token> keyTokens = new ArrayList<Token>();
List<InetAddress> hosts = new ArrayList<InetAddress>();
List<UUID> hostIds = new ArrayList<UUID>();
// create a ring or 10 nodes
Util.createInitialRing(ss, partitioner, endpointTokens, keyTokens, hosts, hostIds, RING_SIZE);
// nodes 6, 8 and 9 leave
final int[] MOVING = new int[] { 6, 8, 9 };
Map<Integer, Token> newTokens = new HashMap<Integer, Token>();
for (int movingIndex : MOVING) {
Token newToken = positionToken(movingIndex);
ss.onChange(hosts.get(movingIndex), ApplicationState.STATUS, valueFactory.moving(newToken));
// storing token associated with a node index
newTokens.put(movingIndex, newToken);
}
Collection<InetAddress> endpoints;
tmd = tmd.cloneAfterAllSettled();
ss.setTokenMetadataUnsafe(tmd);
// boot two new nodes with keyTokens.get(5) and keyTokens.get(7)
InetAddress boot1 = InetAddress.getByName("127.0.1.1");
Gossiper.instance.initializeNodeUnsafe(boot1, UUID.randomUUID(), 1);
Gossiper.instance.injectApplicationState(boot1, ApplicationState.TOKENS, valueFactory.tokens(Collections.singleton(keyTokens.get(5))));
ss.onChange(boot1, ApplicationState.STATUS, valueFactory.bootstrapping(Collections.<Token>singleton(keyTokens.get(5))));
PendingRangeCalculatorService.instance.blockUntilFinished();
InetAddress boot2 = InetAddress.getByName("127.0.1.2");
Gossiper.instance.initializeNodeUnsafe(boot2, UUID.randomUUID(), 1);
Gossiper.instance.injectApplicationState(boot2, ApplicationState.TOKENS, valueFactory.tokens(Collections.singleton(keyTokens.get(7))));
ss.onChange(boot2, ApplicationState.STATUS, valueFactory.bootstrapping(Collections.<Token>singleton(keyTokens.get(7))));
PendingRangeCalculatorService.instance.blockUntilFinished();
// don't require test update every time a new keyspace is added to test/conf/cassandra.yaml
Map<String, AbstractReplicationStrategy> keyspaceStrategyMap = new HashMap<String, AbstractReplicationStrategy>();
for (int i = 1; i <= 4; i++) {
keyspaceStrategyMap.put("MoveTestKeyspace" + i, getStrategy("MoveTestKeyspace" + i, tmd));
}
/**
* Keyspace1 & Keyspace2 RF=1
* {
* /127.0.0.1=[(97,0]],
* /127.0.0.2=[(0,10]],
* /127.0.0.3=[(10,20]],
* /127.0.0.4=[(20,30]],
* /127.0.0.5=[(30,40]],
* /127.0.0.6=[(40,50]],
* /127.0.0.7=[(50,67]],
* /127.0.0.8=[(67,70]],
* /127.0.0.9=[(70,87]],
* /127.0.0.10=[(87,97]]
* }
*/
Multimap<InetAddress, Range<Token>> keyspace1ranges = keyspaceStrategyMap.get(Simple_RF1_KeyspaceName).getAddressRanges();
Collection<Range<Token>> ranges1 = keyspace1ranges.get(InetAddress.getByName("127.0.0.1"));
assertEquals(1, collectionSize(ranges1));
assertEquals(generateRange(97, 0), ranges1.iterator().next());
Collection<Range<Token>> ranges2 = keyspace1ranges.get(InetAddress.getByName("127.0.0.2"));
assertEquals(1, collectionSize(ranges2));
assertEquals(generateRange(0, 10), ranges2.iterator().next());
Collection<Range<Token>> ranges3 = keyspace1ranges.get(InetAddress.getByName("127.0.0.3"));
assertEquals(1, collectionSize(ranges3));
assertEquals(generateRange(10, 20), ranges3.iterator().next());
Collection<Range<Token>> ranges4 = keyspace1ranges.get(InetAddress.getByName("127.0.0.4"));
assertEquals(1, collectionSize(ranges4));
assertEquals(generateRange(20, 30), ranges4.iterator().next());
Collection<Range<Token>> ranges5 = keyspace1ranges.get(InetAddress.getByName("127.0.0.5"));
assertEquals(1, collectionSize(ranges5));
assertEquals(generateRange(30, 40), ranges5.iterator().next());
Collection<Range<Token>> ranges6 = keyspace1ranges.get(InetAddress.getByName("127.0.0.6"));
assertEquals(1, collectionSize(ranges6));
assertEquals(generateRange(40, 50), ranges6.iterator().next());
Collection<Range<Token>> ranges7 = keyspace1ranges.get(InetAddress.getByName("127.0.0.7"));
assertEquals(1, collectionSize(ranges7));
assertEquals(generateRange(50, 67), ranges7.iterator().next());
Collection<Range<Token>> ranges8 = keyspace1ranges.get(InetAddress.getByName("127.0.0.8"));
assertEquals(1, collectionSize(ranges8));
assertEquals(generateRange(67, 70), ranges8.iterator().next());
Collection<Range<Token>> ranges9 = keyspace1ranges.get(InetAddress.getByName("127.0.0.9"));
assertEquals(1, collectionSize(ranges9));
assertEquals(generateRange(70, 87), ranges9.iterator().next());
Collection<Range<Token>> ranges10 = keyspace1ranges.get(InetAddress.getByName("127.0.0.10"));
assertEquals(1, collectionSize(ranges10));
assertEquals(generateRange(87, 97), ranges10.iterator().next());
/**
* Keyspace3 RF=5
* {
* /127.0.0.1=[(97,0], (70,87], (50,67], (87,97], (67,70]],
* /127.0.0.2=[(97,0], (70,87], (87,97], (0,10], (67,70]],
* /127.0.0.3=[(97,0], (70,87], (87,97], (0,10], (10,20]],
* /127.0.0.4=[(97,0], (20,30], (87,97], (0,10], (10,20]],
* /127.0.0.5=[(97,0], (30,40], (20,30], (0,10], (10,20]],
* /127.0.0.6=[(40,50], (30,40], (20,30], (0,10], (10,20]],
* /127.0.0.7=[(40,50], (30,40], (50,67], (20,30], (10,20]],
* /127.0.0.8=[(40,50], (30,40], (50,67], (20,30], (67,70]],
* /127.0.0.9=[(40,50], (70,87], (30,40], (50,67], (67,70]],
* /127.0.0.10=[(40,50], (70,87], (50,67], (87,97], (67,70]]
* }
*/
Multimap<InetAddress, Range<Token>> keyspace3ranges = keyspaceStrategyMap.get(KEYSPACE3).getAddressRanges();
ranges1 = keyspace3ranges.get(InetAddress.getByName("127.0.0.1"));
assertEquals(collectionSize(ranges1), 5);
assertTrue(ranges1.equals(generateRanges(97, 0, 70, 87, 50, 67, 87, 97, 67, 70)));
ranges2 = keyspace3ranges.get(InetAddress.getByName("127.0.0.2"));
assertEquals(collectionSize(ranges2), 5);
assertTrue(ranges2.equals(generateRanges(97, 0, 70, 87, 87, 97, 0, 10, 67, 70)));
ranges3 = keyspace3ranges.get(InetAddress.getByName("127.0.0.3"));
assertEquals(collectionSize(ranges3), 5);
assertTrue(ranges3.equals(generateRanges(97, 0, 70, 87, 87, 97, 0, 10, 10, 20)));
ranges4 = keyspace3ranges.get(InetAddress.getByName("127.0.0.4"));
assertEquals(collectionSize(ranges4), 5);
assertTrue(ranges4.equals(generateRanges(97, 0, 20, 30, 87, 97, 0, 10, 10, 20)));
ranges5 = keyspace3ranges.get(InetAddress.getByName("127.0.0.5"));
assertEquals(collectionSize(ranges5), 5);
assertTrue(ranges5.equals(generateRanges(97, 0, 30, 40, 20, 30, 0, 10, 10, 20)));
ranges6 = keyspace3ranges.get(InetAddress.getByName("127.0.0.6"));
assertEquals(collectionSize(ranges6), 5);
assertTrue(ranges6.equals(generateRanges(40, 50, 30, 40, 20, 30, 0, 10, 10, 20)));
ranges7 = keyspace3ranges.get(InetAddress.getByName("127.0.0.7"));
assertEquals(collectionSize(ranges7), 5);
assertTrue(ranges7.equals(generateRanges(40, 50, 30, 40, 50, 67, 20, 30, 10, 20)));
ranges8 = keyspace3ranges.get(InetAddress.getByName("127.0.0.8"));
assertEquals(collectionSize(ranges8), 5);
assertTrue(ranges8.equals(generateRanges(40, 50, 30, 40, 50, 67, 20, 30, 67, 70)));
ranges9 = keyspace3ranges.get(InetAddress.getByName("127.0.0.9"));
assertEquals(collectionSize(ranges9), 5);
assertTrue(ranges9.equals(generateRanges(40, 50, 70, 87, 30, 40, 50, 67, 67, 70)));
ranges10 = keyspace3ranges.get(InetAddress.getByName("127.0.0.10"));
assertEquals(collectionSize(ranges10), 5);
assertTrue(ranges10.equals(generateRanges(40, 50, 70, 87, 50, 67, 87, 97, 67, 70)));
/**
* Keyspace4 RF=3
* {
* /127.0.0.1=[(97,0], (70,87], (87,97]],
* /127.0.0.2=[(97,0], (87,97], (0,10]],
* /127.0.0.3=[(97,0], (0,10], (10,20]],
* /127.0.0.4=[(20,30], (0,10], (10,20]],
* /127.0.0.5=[(30,40], (20,30], (10,20]],
* /127.0.0.6=[(40,50], (30,40], (20,30]],
* /127.0.0.7=[(40,50], (30,40], (50,67]],
* /127.0.0.8=[(40,50], (50,67], (67,70]],
* /127.0.0.9=[(70,87], (50,67], (67,70]],
* /127.0.0.10=[(70,87], (87,97], (67,70]]
* }
*/
Multimap<InetAddress, Range<Token>> keyspace4ranges = keyspaceStrategyMap.get(Simple_RF3_KeyspaceName).getAddressRanges();
ranges1 = keyspace4ranges.get(InetAddress.getByName("127.0.0.1"));
assertEquals(collectionSize(ranges1), 3);
assertTrue(ranges1.equals(generateRanges(97, 0, 70, 87, 87, 97)));
ranges2 = keyspace4ranges.get(InetAddress.getByName("127.0.0.2"));
assertEquals(collectionSize(ranges2), 3);
assertTrue(ranges2.equals(generateRanges(97, 0, 87, 97, 0, 10)));
ranges3 = keyspace4ranges.get(InetAddress.getByName("127.0.0.3"));
assertEquals(collectionSize(ranges3), 3);
assertTrue(ranges3.equals(generateRanges(97, 0, 0, 10, 10, 20)));
ranges4 = keyspace4ranges.get(InetAddress.getByName("127.0.0.4"));
assertEquals(collectionSize(ranges4), 3);
assertTrue(ranges4.equals(generateRanges(20, 30, 0, 10, 10, 20)));
ranges5 = keyspace4ranges.get(InetAddress.getByName("127.0.0.5"));
assertEquals(collectionSize(ranges5), 3);
assertTrue(ranges5.equals(generateRanges(30, 40, 20, 30, 10, 20)));
ranges6 = keyspace4ranges.get(InetAddress.getByName("127.0.0.6"));
assertEquals(collectionSize(ranges6), 3);
assertTrue(ranges6.equals(generateRanges(40, 50, 30, 40, 20, 30)));
ranges7 = keyspace4ranges.get(InetAddress.getByName("127.0.0.7"));
assertEquals(collectionSize(ranges7), 3);
assertTrue(ranges7.equals(generateRanges(40, 50, 30, 40, 50, 67)));
ranges8 = keyspace4ranges.get(InetAddress.getByName("127.0.0.8"));
assertEquals(collectionSize(ranges8), 3);
assertTrue(ranges8.equals(generateRanges(40, 50, 50, 67, 67, 70)));
ranges9 = keyspace4ranges.get(InetAddress.getByName("127.0.0.9"));
assertEquals(collectionSize(ranges9), 3);
assertTrue(ranges9.equals(generateRanges(70, 87, 50, 67, 67, 70)));
ranges10 = keyspace4ranges.get(InetAddress.getByName("127.0.0.10"));
assertEquals(collectionSize(ranges10), 3);
assertTrue(ranges10.equals(generateRanges(70, 87, 87, 97, 67, 70)));
// pre-calculate the results.
Map<String, Multimap<Token, InetAddress>> expectedEndpoints = new HashMap<String, Multimap<Token, InetAddress>>();
expectedEndpoints.put(Simple_RF1_KeyspaceName, HashMultimap.<Token, InetAddress>create());
expectedEndpoints.get(Simple_RF1_KeyspaceName).putAll(new BigIntegerToken("5"), makeAddrs("127.0.0.2"));
expectedEndpoints.get(Simple_RF1_KeyspaceName).putAll(new BigIntegerToken("15"), makeAddrs("127.0.0.3"));
expectedEndpoints.get(Simple_RF1_KeyspaceName).putAll(new BigIntegerToken("25"), makeAddrs("127.0.0.4"));
expectedEndpoints.get(Simple_RF1_KeyspaceName).putAll(new BigIntegerToken("35"), makeAddrs("127.0.0.5"));
expectedEndpoints.get(Simple_RF1_KeyspaceName).putAll(new BigIntegerToken("45"), makeAddrs("127.0.0.6"));
expectedEndpoints.get(Simple_RF1_KeyspaceName).putAll(new BigIntegerToken("55"), makeAddrs("127.0.0.7", "127.0.1.1"));
expectedEndpoints.get(Simple_RF1_KeyspaceName).putAll(new BigIntegerToken("65"), makeAddrs("127.0.0.7"));
expectedEndpoints.get(Simple_RF1_KeyspaceName).putAll(new BigIntegerToken("75"), makeAddrs("127.0.0.9", "127.0.1.2"));
expectedEndpoints.get(Simple_RF1_KeyspaceName).putAll(new BigIntegerToken("85"), makeAddrs("127.0.0.9"));
expectedEndpoints.get(Simple_RF1_KeyspaceName).putAll(new BigIntegerToken("95"), makeAddrs("127.0.0.10"));
expectedEndpoints.put(KEYSPACE2, HashMultimap.<Token, InetAddress>create());
expectedEndpoints.get(KEYSPACE2).putAll(new BigIntegerToken("5"), makeAddrs("127.0.0.2"));
expectedEndpoints.get(KEYSPACE2).putAll(new BigIntegerToken("15"), makeAddrs("127.0.0.3"));
expectedEndpoints.get(KEYSPACE2).putAll(new BigIntegerToken("25"), makeAddrs("127.0.0.4"));
expectedEndpoints.get(KEYSPACE2).putAll(new BigIntegerToken("35"), makeAddrs("127.0.0.5"));
expectedEndpoints.get(KEYSPACE2).putAll(new BigIntegerToken("45"), makeAddrs("127.0.0.6"));
expectedEndpoints.get(KEYSPACE2).putAll(new BigIntegerToken("55"), makeAddrs("127.0.0.7", "127.0.1.1"));
expectedEndpoints.get(KEYSPACE2).putAll(new BigIntegerToken("65"), makeAddrs("127.0.0.7"));
expectedEndpoints.get(KEYSPACE2).putAll(new BigIntegerToken("75"), makeAddrs("127.0.0.9", "127.0.1.2"));
expectedEndpoints.get(KEYSPACE2).putAll(new BigIntegerToken("85"), makeAddrs("127.0.0.9"));
expectedEndpoints.get(KEYSPACE2).putAll(new BigIntegerToken("95"), makeAddrs("127.0.0.10"));
expectedEndpoints.put(KEYSPACE3, HashMultimap.<Token, InetAddress>create());
expectedEndpoints.get(KEYSPACE3).putAll(new BigIntegerToken("5"), makeAddrs("127.0.0.2", "127.0.0.3", "127.0.0.4", "127.0.0.5", "127.0.0.6"));
expectedEndpoints.get(KEYSPACE3).putAll(new BigIntegerToken("15"), makeAddrs("127.0.0.3", "127.0.0.4", "127.0.0.5", "127.0.0.6", "127.0.0.7", "127.0.1.1"));
expectedEndpoints.get(KEYSPACE3).putAll(new BigIntegerToken("25"), makeAddrs("127.0.0.4", "127.0.0.5", "127.0.0.6", "127.0.0.7", "127.0.0.8", "127.0.1.1"));
expectedEndpoints.get(KEYSPACE3).putAll(new BigIntegerToken("35"), makeAddrs("127.0.0.5", "127.0.0.6", "127.0.0.7", "127.0.0.8", "127.0.0.9", "127.0.1.1", "127.0.1.2"));
expectedEndpoints.get(KEYSPACE3).putAll(new BigIntegerToken("45"), makeAddrs("127.0.0.6", "127.0.0.7", "127.0.0.8", "127.0.0.9", "127.0.0.10", "127.0.1.1", "127.0.1.2"));
expectedEndpoints.get(KEYSPACE3).putAll(new BigIntegerToken("55"), makeAddrs("127.0.0.7", "127.0.0.8", "127.0.0.9", "127.0.0.10", "127.0.0.1", "127.0.1.1", "127.0.1.2"));
expectedEndpoints.get(KEYSPACE3).putAll(new BigIntegerToken("65"), makeAddrs("127.0.0.7", "127.0.0.8", "127.0.0.9", "127.0.0.10", "127.0.0.1", "127.0.1.2"));
expectedEndpoints.get(KEYSPACE3).putAll(new BigIntegerToken("75"), makeAddrs("127.0.0.9", "127.0.0.10", "127.0.0.1", "127.0.0.2", "127.0.0.3", "127.0.1.2"));
expectedEndpoints.get(KEYSPACE3).putAll(new BigIntegerToken("85"), makeAddrs("127.0.0.9", "127.0.0.10", "127.0.0.1", "127.0.0.2", "127.0.0.3"));
expectedEndpoints.get(KEYSPACE3).putAll(new BigIntegerToken("95"), makeAddrs("127.0.0.10", "127.0.0.1", "127.0.0.2", "127.0.0.3", "127.0.0.4"));
expectedEndpoints.put(Simple_RF3_KeyspaceName, HashMultimap.<Token, InetAddress>create());
expectedEndpoints.get(Simple_RF3_KeyspaceName).putAll(new BigIntegerToken("5"), makeAddrs("127.0.0.2", "127.0.0.3", "127.0.0.4"));
expectedEndpoints.get(Simple_RF3_KeyspaceName).putAll(new BigIntegerToken("15"), makeAddrs("127.0.0.3", "127.0.0.4", "127.0.0.5"));
expectedEndpoints.get(Simple_RF3_KeyspaceName).putAll(new BigIntegerToken("25"), makeAddrs("127.0.0.4", "127.0.0.5", "127.0.0.6"));
expectedEndpoints.get(Simple_RF3_KeyspaceName).putAll(new BigIntegerToken("35"), makeAddrs("127.0.0.5", "127.0.0.6", "127.0.0.7", "127.0.1.1"));
expectedEndpoints.get(Simple_RF3_KeyspaceName).putAll(new BigIntegerToken("45"), makeAddrs("127.0.0.6", "127.0.0.7", "127.0.0.8", "127.0.1.1"));
expectedEndpoints.get(Simple_RF3_KeyspaceName).putAll(new BigIntegerToken("55"), makeAddrs("127.0.0.7", "127.0.0.8", "127.0.0.9", "127.0.1.1", "127.0.1.2"));
expectedEndpoints.get(Simple_RF3_KeyspaceName).putAll(new BigIntegerToken("65"), makeAddrs("127.0.0.7", "127.0.0.8", "127.0.0.9", "127.0.1.2"));
expectedEndpoints.get(Simple_RF3_KeyspaceName).putAll(new BigIntegerToken("75"), makeAddrs("127.0.0.9", "127.0.0.10", "127.0.0.1", "127.0.1.2"));
expectedEndpoints.get(Simple_RF3_KeyspaceName).putAll(new BigIntegerToken("85"), makeAddrs("127.0.0.9", "127.0.0.10", "127.0.0.1"));
expectedEndpoints.get(Simple_RF3_KeyspaceName).putAll(new BigIntegerToken("95"), makeAddrs("127.0.0.10", "127.0.0.1", "127.0.0.2"));
for (Map.Entry<String, AbstractReplicationStrategy> keyspaceStrategy : keyspaceStrategyMap.entrySet()) {
String keyspaceName = keyspaceStrategy.getKey();
AbstractReplicationStrategy strategy = keyspaceStrategy.getValue();
for (Token token : keyTokens) {
endpoints = tmd.getWriteEndpoints(token, keyspaceName, strategy.getNaturalEndpoints(token));
assertEquals(expectedEndpoints.get(keyspaceName).get(token).size(), endpoints.size());
assertTrue(expectedEndpoints.get(keyspaceName).get(token).containsAll(endpoints));
}
// just to be sure that things still work according to the old tests, run them:
if (strategy.getReplicationFactor() != 3)
continue;
// tokens 5, 15 and 25 should go three nodes
for (int i = 0; i < 3; i++) {
endpoints = tmd.getWriteEndpoints(keyTokens.get(i), keyspaceName, strategy.getNaturalEndpoints(keyTokens.get(i)));
assertEquals(3, endpoints.size());
assertTrue(endpoints.contains(hosts.get(i + 1)));
assertTrue(endpoints.contains(hosts.get(i + 2)));
assertTrue(endpoints.contains(hosts.get(i + 3)));
}
// token 35 should go to nodes 4, 5, 6 and boot1
endpoints = tmd.getWriteEndpoints(keyTokens.get(3), keyspaceName, strategy.getNaturalEndpoints(keyTokens.get(3)));
assertEquals(4, endpoints.size());
assertTrue(endpoints.contains(hosts.get(4)));
assertTrue(endpoints.contains(hosts.get(5)));
assertTrue(endpoints.contains(hosts.get(6)));
assertTrue(endpoints.contains(boot1));
// token 45 should go to nodes 5, 6, 7 boot1
endpoints = tmd.getWriteEndpoints(keyTokens.get(4), keyspaceName, strategy.getNaturalEndpoints(keyTokens.get(4)));
assertEquals(4, endpoints.size());
assertTrue(endpoints.contains(hosts.get(5)));
assertTrue(endpoints.contains(hosts.get(6)));
assertTrue(endpoints.contains(hosts.get(7)));
assertTrue(endpoints.contains(boot1));
// token 55 should go to nodes 6, 7, 8 boot1 and boot2
endpoints = tmd.getWriteEndpoints(keyTokens.get(5), keyspaceName, strategy.getNaturalEndpoints(keyTokens.get(5)));
assertEquals(5, endpoints.size());
assertTrue(endpoints.contains(hosts.get(6)));
assertTrue(endpoints.contains(hosts.get(7)));
assertTrue(endpoints.contains(hosts.get(8)));
assertTrue(endpoints.contains(boot1));
assertTrue(endpoints.contains(boot2));
// token 65 should go to nodes 6, 7, 8 and boot2
endpoints = tmd.getWriteEndpoints(keyTokens.get(6), keyspaceName, strategy.getNaturalEndpoints(keyTokens.get(6)));
assertEquals(4, endpoints.size());
assertTrue(endpoints.contains(hosts.get(6)));
assertTrue(endpoints.contains(hosts.get(7)));
assertTrue(endpoints.contains(hosts.get(8)));
assertTrue(endpoints.contains(boot2));
// token 75 should to go nodes 8, 9, 0 and boot2
endpoints = tmd.getWriteEndpoints(keyTokens.get(7), keyspaceName, strategy.getNaturalEndpoints(keyTokens.get(7)));
assertEquals(4, endpoints.size());
assertTrue(endpoints.contains(hosts.get(8)));
assertTrue(endpoints.contains(hosts.get(9)));
assertTrue(endpoints.contains(hosts.get(0)));
assertTrue(endpoints.contains(boot2));
// token 85 should go to nodes 8, 9 and 0
endpoints = tmd.getWriteEndpoints(keyTokens.get(8), keyspaceName, strategy.getNaturalEndpoints(keyTokens.get(8)));
assertEquals(3, endpoints.size());
assertTrue(endpoints.contains(hosts.get(8)));
assertTrue(endpoints.contains(hosts.get(9)));
assertTrue(endpoints.contains(hosts.get(0)));
// token 95 should go to nodes 9, 0 and 1
endpoints = tmd.getWriteEndpoints(keyTokens.get(9), keyspaceName, strategy.getNaturalEndpoints(keyTokens.get(9)));
assertEquals(3, endpoints.size());
assertTrue(endpoints.contains(hosts.get(9)));
assertTrue(endpoints.contains(hosts.get(0)));
assertTrue(endpoints.contains(hosts.get(1)));
}
// all moving nodes are back to the normal state
for (Integer movingIndex : MOVING) {
ss.onChange(hosts.get(movingIndex), ApplicationState.STATUS, valueFactory.normal(Collections.singleton(newTokens.get(movingIndex))));
}
}
use of org.apache.cassandra.locator.AbstractReplicationStrategy in project eiger by wlloyd.
the class BootStrapper method getRangesWithSources.
/** get potential sources for each range, ordered by proximity (as determined by EndpointSnitch) */
Multimap<Range<Token>, InetAddress> getRangesWithSources(String table) {
assert tokenMetadata.sortedTokens().size() > 0;
final AbstractReplicationStrategy strat = Table.open(table).getReplicationStrategy();
Collection<Range<Token>> myRanges = strat.getPendingAddressRanges(tokenMetadata, token, address);
Multimap<Range<Token>, InetAddress> myRangeAddresses = ArrayListMultimap.create();
Multimap<Range<Token>, InetAddress> rangeAddresses = strat.getRangeAddresses(tokenMetadata);
for (Range<Token> myRange : myRanges) {
for (Range<Token> range : rangeAddresses.keySet()) {
if (range.contains(myRange)) {
List<InetAddress> preferred = DatabaseDescriptor.getEndpointSnitch().getSortedListByProximity(address, rangeAddresses.get(range));
myRangeAddresses.putAll(myRange, preferred);
break;
}
}
assert myRangeAddresses.keySet().contains(myRange);
}
return myRangeAddresses;
}
use of org.apache.cassandra.locator.AbstractReplicationStrategy in project eiger by wlloyd.
the class StorageProxy method mutateCounter.
/**
* Handle counter mutation on the coordinator host.
*
* A counter mutation needs to first be applied to a replica (that we'll call the leader for the mutation) before being
* replicated to the other endpoint. To achieve so, there is two case:
* 1) the coordinator host is a replica: we proceed to applying the update locally and replicate throug
* applyCounterMutationOnCoordinator
* 2) the coordinator is not a replica: we forward the (counter)mutation to a chosen replica (that will proceed through
* applyCounterMutationOnLeader upon receive) and wait for its acknowledgment.
*
* Implementation note: We check if we can fulfill the CL on the coordinator host even if he is not a replica to allow
* quicker response and because the WriteResponseHandlers don't make it easy to send back an error. We also always gather
* the write latencies at the coordinator node to make gathering point similar to the case of standard writes.
*/
public static IWriteResponseHandler mutateCounter(CounterMutation cm, String localDataCenter) throws UnavailableException, TimeoutException, IOException {
InetAddress endpoint = findSuitableEndpoint(cm.getTable(), cm.key(), localDataCenter);
if (endpoint.equals(FBUtilities.getBroadcastAddress())) {
return applyCounterMutationOnCoordinator(cm, localDataCenter);
} else {
assert false : "In COPS2 an add == mutateCounter should always go to the coordinator";
// Exit now if we can't fulfill the CL here instead of forwarding to the leader replica
String table = cm.getTable();
AbstractReplicationStrategy rs = Table.open(table).getReplicationStrategy();
Collection<InetAddress> writeEndpoints = getWriteEndpoints(table, cm.key());
rs.getWriteResponseHandler(writeEndpoints, cm.consistency()).assureSufficientLiveNodes();
// Forward the actual update to the chosen leader replica
IWriteResponseHandler responseHandler = WriteResponseHandler.create(endpoint);
Message message = cm.makeMutationMessage(Gossiper.instance.getVersion(endpoint));
if (logger.isDebugEnabled())
logger.debug("forwarding counter update of key " + ByteBufferUtil.bytesToHex(cm.key()) + " to " + endpoint);
MessagingService.instance().sendRR(message, endpoint, responseHandler);
return responseHandler;
}
}
use of org.apache.cassandra.locator.AbstractReplicationStrategy in project eiger by wlloyd.
the class StorageService method move.
/**
* move the node to new token or find a new token to boot to according to load
*
* @param newToken new token to boot to, or if null, find balanced token to boot to
*
* @throws IOException on any I/O operation error
*/
private void move(Token newToken) throws IOException {
if (newToken == null)
throw new IOException("Can't move to the undefined (null) token.");
if (tokenMetadata_.sortedTokens().contains(newToken))
throw new IOException("target token " + newToken + " is already owned by another node.");
// address of the current node
InetAddress localAddress = FBUtilities.getBroadcastAddress();
List<String> tablesToProcess = Schema.instance.getNonSystemTables();
// checking if data is moving to this node
for (String table : tablesToProcess) {
if (tokenMetadata_.getPendingRanges(table, localAddress).size() > 0)
throw new UnsupportedOperationException("data is currently moving to this node; unable to leave the ring");
}
// setting 'moving' application state
Gossiper.instance.addLocalApplicationState(ApplicationState.STATUS, valueFactory.moving(newToken));
logger_.info(String.format("Moving %s from %s to %s.", localAddress, getLocalToken(), newToken));
IEndpointSnitch snitch = DatabaseDescriptor.getEndpointSnitch();
Map<String, Multimap<InetAddress, Range<Token>>> rangesToFetch = new HashMap<String, Multimap<InetAddress, Range<Token>>>();
Map<String, Multimap<Range<Token>, InetAddress>> rangesToStreamByTable = new HashMap<String, Multimap<Range<Token>, InetAddress>>();
TokenMetadata tokenMetaClone = tokenMetadata_.cloneAfterAllSettled();
// which current node will handle after move to the new token
for (String table : tablesToProcess) {
// replication strategy of the current keyspace (aka table)
AbstractReplicationStrategy strategy = Table.open(table).getReplicationStrategy();
// getting collection of the currently used ranges by this keyspace
Collection<Range<Token>> currentRanges = getRangesForEndpoint(table, localAddress);
// collection of ranges which this node will serve after move to the new token
Collection<Range<Token>> updatedRanges = strategy.getPendingAddressRanges(tokenMetadata_, newToken, localAddress);
// ring ranges and endpoints associated with them
// this used to determine what nodes should we ping about range data
Multimap<Range<Token>, InetAddress> rangeAddresses = strategy.getRangeAddresses(tokenMetadata_);
// calculated parts of the ranges to request/stream from/to nodes in the ring
Pair<Set<Range<Token>>, Set<Range<Token>>> rangesPerTable = calculateStreamAndFetchRanges(currentRanges, updatedRanges);
/**
* In this loop we are going through all ranges "to fetch" and determining
* nodes in the ring responsible for data we are interested in
*/
Multimap<Range<Token>, InetAddress> rangesToFetchWithPreferredEndpoints = ArrayListMultimap.create();
for (Range<Token> toFetch : rangesPerTable.right) {
for (Range<Token> range : rangeAddresses.keySet()) {
if (range.contains(toFetch)) {
List<InetAddress> endpoints = snitch.getSortedListByProximity(localAddress, rangeAddresses.get(range));
// storing range and preferred endpoint set
rangesToFetchWithPreferredEndpoints.putAll(toFetch, endpoints);
}
}
}
// calculating endpoints to stream current ranges to if needed
// in some situations node will handle current ranges as part of the new ranges
Multimap<Range<Token>, InetAddress> rangeWithEndpoints = HashMultimap.create();
for (Range<Token> toStream : rangesPerTable.left) {
Set<InetAddress> currentEndpoints = ImmutableSet.copyOf(strategy.calculateNaturalEndpoints(toStream.right, tokenMetadata_));
Set<InetAddress> newEndpoints = ImmutableSet.copyOf(strategy.calculateNaturalEndpoints(toStream.right, tokenMetaClone));
rangeWithEndpoints.putAll(toStream, Sets.difference(newEndpoints, currentEndpoints));
}
// associating table with range-to-endpoints map
rangesToStreamByTable.put(table, rangeWithEndpoints);
Multimap<InetAddress, Range<Token>> workMap = BootStrapper.getWorkMap(rangesToFetchWithPreferredEndpoints);
rangesToFetch.put(table, workMap);
if (logger_.isDebugEnabled())
logger_.debug("Table {}: work map {}.", table, workMap);
}
if (!rangesToStreamByTable.isEmpty() || !rangesToFetch.isEmpty()) {
logger_.info("Sleeping {} ms before start streaming/fetching ranges.", RING_DELAY);
try {
Thread.sleep(RING_DELAY);
} catch (InterruptedException e) {
throw new RuntimeException("Sleep interrupted " + e.getMessage());
}
setMode(Mode.MOVING, "fetching new ranges and streaming old ranges", true);
if (logger_.isDebugEnabled())
logger_.debug("[Move->STREAMING] Work Map: " + rangesToStreamByTable);
CountDownLatch streamLatch = streamRanges(rangesToStreamByTable);
if (logger_.isDebugEnabled())
logger_.debug("[Move->FETCHING] Work Map: " + rangesToFetch);
CountDownLatch fetchLatch = requestRanges(rangesToFetch);
try {
streamLatch.await();
fetchLatch.await();
} catch (InterruptedException e) {
throw new RuntimeException("Interrupted latch while waiting for stream/fetch ranges to finish: " + e.getMessage());
}
}
// setting new token as we have everything settled
setToken(newToken);
if (logger_.isDebugEnabled())
logger_.debug("Successfully moved to new token {}", getLocalToken());
}
Aggregations