use of com.hazelcast.internal.serialization.SerializationService in project hazelcast by hazelcast.
the class CacheBackupTest method entrySuccessfullyRetrievedFromBackup.
private void entrySuccessfullyRetrievedFromBackup(int backupCount, boolean sync) {
final String KEY = "key";
final String VALUE = "value";
final int nodeCount = backupCount + 1;
final TestHazelcastInstanceFactory instanceFactory = createHazelcastInstanceFactory(nodeCount);
final HazelcastInstance[] instances = new HazelcastInstance[nodeCount];
for (int i = 0; i < instances.length; i++) {
instances[i] = instanceFactory.newHazelcastInstance();
}
final HazelcastInstance hz = instances[0];
final CachingProvider cachingProvider = createServerCachingProvider(hz);
final CacheManager cacheManager = cachingProvider.getCacheManager();
final String cacheName = randomName();
final CacheConfig cacheConfig = new CacheConfig().setName(cacheName);
if (sync) {
cacheConfig.setBackupCount(backupCount);
} else {
cacheConfig.setAsyncBackupCount(backupCount);
}
final Cache cache = cacheManager.createCache(cacheName, cacheConfig);
warmUpPartitions(instances);
waitAllForSafeState(instances);
cache.put(KEY, VALUE);
final Node node = getNode(hz);
final InternalPartitionService partitionService = node.getPartitionService();
final int keyPartitionId = partitionService.getPartitionId(KEY);
for (int i = 1; i <= backupCount; i++) {
final Node backupNode = getNode(instances[i]);
final SerializationService serializationService = backupNode.getSerializationService();
final ICacheService cacheService = backupNode.getNodeEngine().getService(ICacheService.SERVICE_NAME);
if (sync) {
checkSavedRecordOnBackup(KEY, VALUE, cacheName, keyPartitionId, serializationService, cacheService);
} else {
assertTrueEventually(new AssertTask() {
@Override
public void run() throws Exception {
checkSavedRecordOnBackup(KEY, VALUE, cacheName, keyPartitionId, serializationService, cacheService);
}
});
}
}
}
use of com.hazelcast.internal.serialization.SerializationService in project hazelcast by hazelcast.
the class CacheLoadAllTest method testLoadAll.
@Test
public void testLoadAll() throws InterruptedException {
ICache<String, String> cache = createCache();
String cacheName = cache.getName();
Map<String, String> entries = createAndFillEntries();
final CountDownLatch latch = new CountDownLatch(1);
cache.loadAll(entries.keySet(), true, new CompletionListener() {
@Override
public void onCompletion() {
latch.countDown();
}
@Override
public void onException(Exception e) {
latch.countDown();
}
});
assertTrue(latch.await(60, TimeUnit.SECONDS));
// Verify that load-all works
for (Map.Entry<String, String> entry : entries.entrySet()) {
String key = entry.getKey();
String expectedValue = entries.get(key);
String actualValue = cache.get(key);
assertEquals(expectedValue, actualValue);
}
Node node = getNode(hazelcastInstance);
InternalPartitionService partitionService = node.getPartitionService();
SerializationService serializationService = node.getSerializationService();
// Verify that backup of load-all works
for (Map.Entry<String, String> entry : entries.entrySet()) {
String key = entry.getKey();
String expectedValue = entries.get(key);
Data keyData = serializationService.toData(key);
int keyPartitionId = partitionService.getPartitionId(keyData);
for (int i = 0; i < INSTANCE_COUNT; i++) {
Node n = getNode(hazelcastInstances[i]);
ICacheService cacheService = n.getNodeEngine().getService(ICacheService.SERVICE_NAME);
ICacheRecordStore recordStore = cacheService.getRecordStore("/hz/" + cacheName, keyPartitionId);
assertNotNull(recordStore);
String actualValue = serializationService.toObject(recordStore.get(keyData, null));
assertEquals(expectedValue, actualValue);
}
}
}
use of com.hazelcast.internal.serialization.SerializationService in project hazelcast by hazelcast.
the class CacheBasicAbstractTest method testInitableIterator.
@Test
public void testInitableIterator() {
int testSize = 3007;
SerializationService serializationService = new DefaultSerializationServiceBuilder().build();
for (int fetchSize = 1; fetchSize < 102; fetchSize++) {
SampleableConcurrentHashMap<Data, String> map = new SampleableConcurrentHashMap<>(1000);
for (int i = 0; i < testSize; i++) {
Integer key = i;
Data data = serializationService.toData(key);
String value1 = "value" + i;
map.put(data, value1);
}
IterationPointer[] pointers = { new IterationPointer(Integer.MAX_VALUE, -1) };
int total = 0;
int remaining = testSize;
while (remaining > 0 && pointers[pointers.length - 1].getIndex() > 0) {
int size = (Math.min(remaining, fetchSize));
List<Data> keys = new ArrayList<>(size);
pointers = map.fetchKeys(pointers, size, keys);
remaining -= keys.size();
total += keys.size();
}
assertEquals(testSize, total);
}
}
use of com.hazelcast.internal.serialization.SerializationService in project hazelcast by hazelcast.
the class ClusterWideConfigurationService method cloneConfig.
private IdentifiedDataSerializable cloneConfig(IdentifiedDataSerializable config) {
SerializationService serializationService = nodeEngine.getSerializationService();
Data data = serializationService.toData(config);
return serializationService.toObject(data);
}
use of com.hazelcast.internal.serialization.SerializationService in project hazelcast by hazelcast.
the class WriteMapP method init.
@Override
public void init(@Nonnull Outbox outbox, @Nonnull Context context) {
map = instance().getMap(mapName);
boolean hasCustomSerializers = serializationService instanceof DelegatingSerializationService && ((DelegatingSerializationService) serializationService).hasAddedSerializers();
boolean hasNearCache = map instanceof NearCachedMapProxyImpl;
if (hasNearCache && hasCustomSerializers) {
// See https://github.com/hazelcast/hazelcast-jet/issues/3046
throw new JetException("Writing into IMap with both near cache and custom serializers not supported");
}
if (!hasCustomSerializers) {
addToBuffer = item -> buffer.add(new SimpleEntry<>(key(item), value(item)));
} else if (map instanceof MapProxyImpl) {
PartitioningStrategy<?> partitionStrategy = ((MapProxyImpl<K, V>) map).getPartitionStrategy();
addToBuffer = item -> {
Data key = serializationService.toData(key(item), partitionStrategy);
Data value = serializationService.toData(value(item));
buffer.add(new SimpleEntry<>(key, value));
};
} else if (map instanceof ClientMapProxy) {
// TODO: add strategy/unify after https://github.com/hazelcast/hazelcast/issues/13950 is fixed
addToBuffer = item -> {
Data key = serializationService.toData(key(item));
Data value = serializationService.toData(value(item));
buffer.add(new SimpleEntry<>(key, value));
};
} else {
throw new RuntimeException("Unexpected map class: " + map.getClass().getName());
}
}
Aggregations