use of net.dempsy.transport.NodeAddress in project Dempsy by Dempsy.
the class OutgoingDispatcher method dispatch.
@Override
public void dispatch(final KeyedMessageWithType messageParam, final MessageResourceManager disposer) throws InterruptedException {
final boolean traceEnabled = LOGGER.isTraceEnabled();
if (messageParam == null)
throw new NullPointerException("Attempt to dispatch a null message.");
final Object messageKey = messageParam.key;
if (messageKey == null)
throw new NullPointerException("Message " + SafeString.objectDescription(messageParam) + " has a null key.");
boolean messageSentSomewhere = false;
try (ResourceManagerClosable x = new ResourceManagerClosable(disposer, messageParam)) {
final KeyedMessageWithType message = x.toUse;
ApplicationState tmp = outbounds.get();
// if we're in the midst of an update then we really want to wait for the new state.
while (tmp == null) {
if (!isRunning.get()) {
LOGGER.debug("[{}] Router dispatch called while stopped.", thisNodeId);
return;
}
if (// however, if we never were ready then we're not in the midst
!isReady.get())
// of an update.
throw new IllegalStateException("Dispatch used before Router is ready.");
// let the other threads do their thing. Maybe we'll be updated sooner.
Thread.yield();
// are we updated yet?
tmp = outbounds.get();
}
final ApplicationState cur = tmp;
final Map<String, RoutingStrategy.Router[]> outboundsByMessageType = cur.outboundsByMessageType;
// =================================================================================
// For each message type, determine the set of Routers. The goal of this loop is to set
// 'containerByNodeAddress'
final Map<NodeAddress, ContainerAddress> containerByNodeAddress = new HashMap<>();
for (final String mt : message.messageTypes) {
final RoutingStrategy.Router[] routers = outboundsByMessageType.get(mt);
if (routers == null) {
if (traceEnabled)
LOGGER.trace("[{}] No cluster that handles messages of type {}", thisNodeId, mt);
} else {
// the set of ContainerAddresses that this message will be sent to.
for (int i = 0; i < routers.length; i++) {
final ContainerAddress ca = routers[i].selectDestinationForMessage(message);
// it's possible 'ca' is null when we don't know where to send the message.
if (ca == null) {
if (LOGGER.isDebugEnabled())
LOGGER.debug("[{}] No way to send the message {} to specific cluster for the time being", thisNodeId, message.message);
} else {
// When the message will be sent to 2 different clusters, but both clusters
// are hosted in the same node, then we send 1 message to 1 ContainerAddress
// where the 'clusters' field contains both container ids.
final ContainerAddress already = containerByNodeAddress.get(ca.node);
if (already != null) {
final int[] ia = new int[already.clusters.length + ca.clusters.length];
System.arraycopy(already.clusters, 0, ia, 0, already.clusters.length);
System.arraycopy(ca.clusters, 0, ia, already.clusters.length, ca.clusters.length);
containerByNodeAddress.put(ca.node, new ContainerAddress(ca.node, ia));
} else
containerByNodeAddress.put(ca.node, ca);
}
}
}
}
if (containerByNodeAddress.size() == 0) {
if (traceEnabled)
LOGGER.trace("[{}] There appears to be no valid destination addresses for the message {}", thisNodeId, SafeString.objectDescription(message.message));
}
for (final Map.Entry<NodeAddress, ContainerAddress> e : containerByNodeAddress.entrySet()) {
final NodeAddress curNode = e.getKey();
final ContainerAddress curAddr = e.getValue();
// If we're local then just send this message directly back to our own node.
if (curNode.equals(thisNode)) {
if (traceEnabled)
LOGGER.trace("Sending local {}", message);
// if the message is a resource then the disposer will be used to dispose of the message
// but it needs an additional replicate. See propogateMessageToNode javadoc.
nodeReciever.propogateMessageToNode(new RoutedMessage(curAddr.clusters, messageKey, disposer == null ? message.message : disposer.replicate(message.message)), // this shouldn't count since Router is an OUTGOING class
false, disposer);
messageSentSomewhere = true;
} else {
if (traceEnabled)
LOGGER.trace("Sending {} to {}", message, curNode);
final Sender sender = cur.getSender(curNode);
if (sender == null) {
// router update is probably behind the routing strategy update
if (isRunning.get())
LOGGER.error("[{}] Couldn't send message to " + curNode + " from " + thisNodeId + " because there's no " + Sender.class.getSimpleName(), thisNodeId);
} else {
sender.send(new RoutedMessage(curAddr.clusters, messageKey, sender.considerMessageOwnsershipTransfered() ? (disposer == null ? message.message : disposer.replicate(message.message)) : message.message));
messageSentSomewhere = true;
}
}
}
} finally {
if (!messageSentSomewhere) {
if (traceEnabled)
LOGGER.trace("Message not sent.");
statsCollector.messageNotSent();
}
}
}
use of net.dempsy.transport.NodeAddress in project Dempsy by Dempsy.
the class NodeManager method start.
public NodeManager start() throws DempsyException {
validate();
nodeStatsCollector = tr.track((NodeStatsCollector) node.getNodeStatsCollector());
// TODO: cleaner?
statsCollectorFactory = tr.track(new Manager<>(ClusterStatsCollectorFactory.class).getAssociatedInstance(node.getClusterStatsCollectorFactoryId()));
// =====================================
// set the dispatcher on adaptors and create containers for mp clusters
final AtomicReference<String> firstAdaptorClusterName = new AtomicReference<>(null);
node.getClusters().forEach(c -> {
if (c.isAdaptor()) {
if (firstAdaptorClusterName.get() == null)
firstAdaptorClusterName.set(c.getClusterId().clusterName);
final Adaptor adaptor = c.getAdaptor();
adaptors.put(c.getClusterId(), adaptor);
if (c.getRoutingStrategyId() != null && !"".equals(c.getRoutingStrategyId().trim()) && !" ".equals(c.getRoutingStrategyId().trim()))
LOGGER.warn("The cluster " + c.getClusterId() + " contains an adaptor but also has the routingStrategy set. The routingStrategy will be ignored.");
if (c.getOutputScheduler() != null)
LOGGER.warn("The cluster " + c.getClusterId() + " contains an adaptor but also has an output executor set. The output executor will never be used.");
} else {
String containerTypeId = c.getContainerTypeId();
if (containerTypeId == null)
// can't be null
containerTypeId = node.getContainerTypeId();
final Container con = makeContainer(containerTypeId).setMessageProcessor(c.getMessageProcessor()).setClusterId(c.getClusterId()).setMaxPendingMessagesPerContainer(c.getMaxPendingMessagesPerContainer());
// TODO: This is a hack for now.
final Manager<RoutingStrategy.Inbound> inboundManager = new RoutingInboundManager();
final RoutingStrategy.Inbound is = inboundManager.getAssociatedInstance(c.getRoutingStrategyId());
final boolean outputSupported = c.getMessageProcessor().isOutputSupported();
final Object outputScheduler = c.getOutputScheduler();
// if there mp handles output but there's no output scheduler then we should warn.
if (outputSupported && outputScheduler == null)
LOGGER.warn("The cluster " + c.getClusterId() + " contains a message processor that supports an output cycle but there's no executor set so it will never be invoked.");
if (!outputSupported && outputScheduler != null)
LOGGER.warn("The cluster " + c.getClusterId() + " contains a message processor that doesn't support an output cycle but there's an output cycle executor set. The output cycle executor will never be used.");
final OutputScheduler os = (outputSupported && outputScheduler != null) ? (OutputScheduler) outputScheduler : null;
containers.add(new PerContainer(con, is, c, os));
}
});
// it we're all adaptor then don't bother to get the receiver.
if (containers.size() == 0) {
// here there's no point in a receiver since there's nothing to receive.
if (firstAdaptorClusterName.get() == null)
throw new IllegalStateException("There seems to be no clusters or adaptors defined for this node \"" + node.toString() + "\"");
} else {
receiver = (Receiver) node.getReceiver();
if (// otherwise we're all adaptor
receiver != null)
nodeAddress = receiver.getAddress(this);
else if (firstAdaptorClusterName.get() == null)
throw new IllegalStateException("There seems to be no clusters or adaptors defined for this node \"" + node.toString() + "\"");
}
nodeId = Optional.ofNullable(nodeAddress).map(n -> n.getGuid()).orElse(firstAdaptorClusterName.get());
if (nodeStatsCollector == null) {
LOGGER.warn("There is no {} set for the the application '{}'", StatsCollector.class.getSimpleName(), node.application);
nodeStatsCollector = new DummyNodeStatsCollector();
}
nodeStatsCollector.setNodeId(nodeId);
if (nodeAddress == null && node.getReceiver() != null)
LOGGER.warn("The node at " + nodeId + " contains no message processors but has a Reciever set. The receiver will never be started.");
if (nodeAddress == null && node.getDefaultRoutingStrategyId() != null)
LOGGER.warn("The node at " + nodeId + " contains no message processors but has a defaultRoutingStrategyId set. The routingStrategyId will never be used.");
if (threading == null)
threading = tr.track(new DefaultThreadingModel(nodeId)).configure(node.getConfiguration()).start(nodeId);
else if (!threading.isStarted())
threading.start(nodeId);
nodeStatsCollector.setMessagesPendingGauge(() -> threading.getNumberLimitedPending());
final NodeReceiver nodeReciever = receiver == null ? null : tr.track(new NodeReceiver(containers.stream().map(pc -> pc.container).collect(Collectors.toList()), threading, nodeStatsCollector));
final Map<ClusterId, ClusterInformation> messageTypesByClusterId = new HashMap<>();
containers.stream().map(pc -> pc.clusterDefinition).forEach(c -> {
messageTypesByClusterId.put(c.getClusterId(), new ClusterInformation(c.getRoutingStrategyId(), c.getClusterId(), c.getMessageProcessor().messagesTypesHandled()));
});
final NodeInformation nodeInfo = nodeAddress != null ? new NodeInformation(receiver.transportTypeId(), nodeAddress, messageTypesByClusterId) : null;
// Then actually register the Node
if (nodeInfo != null) {
keepNodeRegstered = new PersistentTask(LOGGER, isRunning, persistenceScheduler, RETRY_PERIOND_MILLIS) {
@Override
public boolean execute() {
try {
session.recursiveMkdir(rootPaths.clustersDir, null, DirMode.PERSISTENT, DirMode.PERSISTENT);
session.recursiveMkdir(rootPaths.nodesDir, null, DirMode.PERSISTENT, DirMode.PERSISTENT);
final String nodePath = rootPaths.nodesDir + "/" + nodeId;
session.mkdir(nodePath, nodeInfo, DirMode.EPHEMERAL);
final NodeInformation reread = (NodeInformation) session.getData(nodePath, this);
final boolean ret = nodeInfo.equals(reread);
if (ret == true)
ptaskReady.set(true);
return ret;
} catch (final ClusterInfoException e) {
final String logmessage = "Failed to register the node. Retrying in " + RETRY_PERIOND_MILLIS + " milliseconds.";
if (LOGGER.isDebugEnabled())
LOGGER.info(logmessage, e);
else
LOGGER.info(logmessage, e);
}
return false;
}
@Override
public String toString() {
return "register node information";
}
};
}
// =====================================
// The layering works this way.
//
// Receiver -> NodeReceiver -> adaptor -> container -> OutgoingDispatcher -> RoutingStrategyOB -> Transport
//
// starting needs to happen in reverse.
// =====================================
isRunning.set(true);
this.tManager = tr.start(new TransportManager(), this);
this.rsManager = tr.start(new RoutingStrategyManager(), this);
// create the router but don't start it yet.
this.router = new OutgoingDispatcher(rsManager, nodeAddress, nodeId, nodeReciever, tManager, nodeStatsCollector);
// set up containers
containers.forEach(pc -> pc.container.setDispatcher(router).setEvictionCycle(pc.clusterDefinition.getEvictionFrequency().evictionFrequency, pc.clusterDefinition.getEvictionFrequency().evictionTimeUnit));
// IB routing strategy
final int numContainers = containers.size();
for (int i = 0; i < numContainers; i++) {
final PerContainer c = containers.get(i);
c.inboundStrategy.setContainerDetails(c.clusterDefinition.getClusterId(), new ContainerAddress(nodeAddress, i), c.container);
}
// setup the output executors by passing the containers
containers.stream().filter(pc -> pc.outputScheduler != null).forEach(pc -> pc.outputScheduler.setOutputInvoker(pc.container));
// set up adaptors
adaptors.values().forEach(a -> a.setDispatcher(router));
// start containers after setting inbound
containers.forEach(pc -> tr.start(pc.container.setInbound(pc.inboundStrategy), this));
// start the output schedulers now that the containers have been started.
containers.stream().map(pc -> pc.outputScheduler).filter(os -> os != null).forEach(os -> tr.start(os, this));
// start IB routing strategy
containers.forEach(pc -> tr.start(pc.inboundStrategy, this));
// start router
tr.start(this.router, this);
final PersistentTask startAdaptorAfterRouterIsRunning = new PersistentTask(LOGGER, isRunning, this.persistenceScheduler, 500) {
@Override
public boolean execute() {
if (!router.isReady())
return false;
adaptors.entrySet().forEach(e -> threading.runDaemon(() -> tr.track(e.getValue()).start(), "Adaptor-" + e.getKey().clusterName));
return true;
}
};
// make sure the router is running. Once it is, start the adaptor(s)
startAdaptorAfterRouterIsRunning.process();
if (receiver != null)
tr.track(receiver).start(nodeReciever, this);
// close this session when we're done.
tr.track(session);
// make it known we're here and ready
if (keepNodeRegstered != null)
keepNodeRegstered.process();
else
ptaskReady.set(true);
return this;
}
use of net.dempsy.transport.NodeAddress in project Dempsy by Dempsy.
the class ApplicationState method update.
public ApplicationState.Update update(final Set<NodeInformation> newState, final NodeAddress thisNodeX, final String thisNodeId) {
final Set<NodeInformation> toAdd = new HashSet<>();
final Set<NodeAddress> toDelete = new HashSet<>();
final Set<NodeAddress> knownAddr = new HashSet<>(current.keySet());
final Set<NodeInformation> leaveAlone = new HashSet<>();
NodeAddress oursOnTheList = null;
for (final NodeInformation cur : newState) {
final NodeInformation known = current.get(cur.nodeAddress);
if (cur.nodeAddress.equals(thisNodeX))
oursOnTheList = cur.nodeAddress;
if (// then we don't know about this one yet.
known == null)
// we need to add this one
toAdd.add(cur);
else {
if (!known.equals(cur)) {
// known but changed ... we need to add and delete it
toAdd.add(cur);
toDelete.add(known.nodeAddress);
} else
leaveAlone.add(known);
// remove it from the known ones. Whatever is leftover will
// end up needing to be deleted.
knownAddr.remove(known.nodeAddress);
}
}
if (// we don't seem to have our own address registered with the collaborator.
oursOnTheList == null && thisNodeX != null)
// this condition is actually okay since the Router is started before the
// node is registered with the collaborator.
LOGGER_SESSION.trace("Router at {} doesn't seem to have its own address registered with the collaborator yet", thisNodeId);
// dump the remaining knownAddrs on the toDelete list
toDelete.addAll(knownAddr);
return new Update(leaveAlone, toAdd, toDelete);
}
use of net.dempsy.transport.NodeAddress in project Dempsy by Dempsy.
the class NioSenderFactory method getSender.
@Override
public NioSender getSender(final NodeAddress destination) throws MessageTransportException {
final TcpAddress tcpaddr = (TcpAddress) destination;
final NioSender ret;
if (isRunning.get()) {
ret = senders.computeIfAbsent(tcpaddr, a -> new NioSender(a, this));
} else
throw new MessageTransportException(nodeId + " sender had getSender called while stopped.");
try {
ret.connect(false);
} catch (final IOException e) {
throw new MessageTransportException(nodeId + " sender failed to connect to " + destination, e);
}
return ret;
}
use of net.dempsy.transport.NodeAddress in project Dempsy by Dempsy.
the class TestGroupRoutingStrategy method testInboundResillience.
@Test
public void testInboundResillience() throws Exception {
final int numShardsToExpect = Integer.parseInt(Utils.DEFAULT_TOTAL_SHARDS);
final String groupName = "testInboundResillience";
final Manager<RoutingStrategy.Inbound> manager = new RoutingInboundManager();
try (final RoutingStrategy.Inbound ib = manager.getAssociatedInstance(ClusterGroupInbound.class.getPackage().getName() + ":" + groupName)) {
final ClusterId clusterId = super.setTestName("testInboundResillience");
final NodeAddress na = new DummyNodeAddress("theOnlyNode");
final ContainerAddress ca = new ContainerAddress(na, 0);
final GroupDetails gd = new GroupDetails(groupName, na);
final Infrastructure infra = makeInfra(session, sched);
final Utils<GroupDetails> msutils = new Utils<>(infra, groupName, gd);
ib.setContainerDetails(clusterId, ca, (l, m) -> {
});
ib.start(infra);
checkForShardDistribution(session, msutils, numShardsToExpect, 1);
disruptor.accept(session);
checkForShardDistribution(session, msutils, numShardsToExpect, 1);
}
}
Aggregations