use of com.google.common.collect.ImmutableMultimap in project cassandra by apache.
the class NetworkTopologyStrategy method calculateNaturalReplicas.
/**
* calculate endpoints in one pass through the tokens by tracking our progress in each DC.
*/
@Override
public EndpointsForRange calculateNaturalReplicas(Token searchToken, TokenMetadata tokenMetadata) {
// we want to preserve insertion order so that the first added endpoint becomes primary
ArrayList<Token> sortedTokens = tokenMetadata.sortedTokens();
Token replicaEnd = TokenMetadata.firstToken(sortedTokens, searchToken);
Token replicaStart = tokenMetadata.getPredecessor(replicaEnd);
Range<Token> replicatedRange = new Range<>(replicaStart, replicaEnd);
EndpointsForRange.Builder builder = new EndpointsForRange.Builder(replicatedRange);
Set<Pair<String, String>> seenRacks = new HashSet<>();
Topology topology = tokenMetadata.getTopology();
// all endpoints in each DC, so we can check when we have exhausted all the members of a DC
Multimap<String, InetAddressAndPort> allEndpoints = topology.getDatacenterEndpoints();
// all racks in a DC so we can check when we have exhausted all racks in a DC
Map<String, ImmutableMultimap<String, InetAddressAndPort>> racks = topology.getDatacenterRacks();
assert !allEndpoints.isEmpty() && !racks.isEmpty() : "not aware of any cluster members";
int dcsToFill = 0;
Map<String, DatacenterEndpoints> dcs = new HashMap<>(datacenters.size() * 2);
// Create a DatacenterEndpoints object for each non-empty DC.
for (Map.Entry<String, ReplicationFactor> en : datacenters.entrySet()) {
String dc = en.getKey();
ReplicationFactor rf = en.getValue();
int nodeCount = sizeOrZero(allEndpoints.get(dc));
if (rf.allReplicas <= 0 || nodeCount <= 0)
continue;
DatacenterEndpoints dcEndpoints = new DatacenterEndpoints(rf, sizeOrZero(racks.get(dc)), nodeCount, builder, seenRacks);
dcs.put(dc, dcEndpoints);
++dcsToFill;
}
Iterator<Token> tokenIter = TokenMetadata.ringIterator(sortedTokens, searchToken, false);
while (dcsToFill > 0 && tokenIter.hasNext()) {
Token next = tokenIter.next();
InetAddressAndPort ep = tokenMetadata.getEndpoint(next);
Pair<String, String> location = topology.getLocation(ep);
DatacenterEndpoints dcEndpoints = dcs.get(location.left);
if (dcEndpoints != null && dcEndpoints.addEndpointAndCheckIfDone(ep, location, replicatedRange))
--dcsToFill;
}
return builder.build();
}
use of com.google.common.collect.ImmutableMultimap in project calcite by apache.
the class ScalarFunctionImpl method createAll.
/**
* Creates {@link org.apache.calcite.schema.ScalarFunction} for each method in
* a given class.
*/
public static ImmutableMultimap<String, ScalarFunction> createAll(Class<?> clazz) {
final ImmutableMultimap.Builder<String, ScalarFunction> builder = ImmutableMultimap.builder();
for (Method method : clazz.getMethods()) {
if (method.getDeclaringClass() == Object.class) {
continue;
}
if (!Modifier.isStatic(method.getModifiers()) && !classHasPublicZeroArgsConstructor(clazz)) {
continue;
}
final ScalarFunction function = create(method);
builder.put(method.getName(), function);
}
return builder.build();
}
use of com.google.common.collect.ImmutableMultimap in project intellij by bazelbuild.
the class BlazeSyncTask method doSyncProject.
/**
* @return true if sync successfully completed
*/
private SyncResult doSyncProject(BlazeContext context, @Nullable BlazeProjectData oldBlazeProjectData) {
long syncStartTime = System.currentTimeMillis();
if (!FileOperationProvider.getInstance().exists(workspaceRoot.directory())) {
IssueOutput.error(String.format("Workspace '%s' doesn't exist.", workspaceRoot.directory())).submit(context);
return SyncResult.FAILURE;
}
BlazeVcsHandler vcsHandler = BlazeVcsHandler.vcsHandlerForProject(project);
if (vcsHandler == null) {
IssueOutput.error("Could not find a VCS handler").submit(context);
return SyncResult.FAILURE;
}
ListeningExecutorService executor = BlazeExecutor.getInstance().getExecutor();
WorkspacePathResolverAndProjectView workspacePathResolverAndProjectView = computeWorkspacePathResolverAndProjectView(context, vcsHandler, executor);
if (workspacePathResolverAndProjectView == null) {
return SyncResult.FAILURE;
}
ProjectViewSet projectViewSet = workspacePathResolverAndProjectView.projectViewSet;
List<String> syncFlags = BlazeFlags.blazeFlags(project, projectViewSet, BlazeCommandName.INFO, BlazeInvocationContext.Sync);
syncStats.setSyncFlags(syncFlags);
ListenableFuture<BlazeInfo> blazeInfoFuture = BlazeInfoRunner.getInstance().runBlazeInfo(context, importSettings.getBuildSystem(), Blaze.getBuildSystemProvider(project).getSyncBinaryPath(), workspaceRoot, syncFlags);
ListenableFuture<WorkingSet> workingSetFuture = vcsHandler.getWorkingSet(project, context, workspaceRoot, executor);
BlazeInfo blazeInfo = FutureUtil.waitForFuture(context, blazeInfoFuture).timed(Blaze.buildSystemName(project) + "Info", EventType.BlazeInvocation).withProgressMessage(String.format("Running %s info...", Blaze.buildSystemName(project))).onError(String.format("Could not run %s info", Blaze.buildSystemName(project))).run().result();
if (blazeInfo == null) {
return SyncResult.FAILURE;
}
BlazeVersionData blazeVersionData = BlazeVersionData.build(importSettings.getBuildSystem(), workspaceRoot, blazeInfo);
if (!BuildSystemVersionChecker.verifyVersionSupported(context, blazeVersionData)) {
return SyncResult.FAILURE;
}
WorkspacePathResolver workspacePathResolver = workspacePathResolverAndProjectView.workspacePathResolver;
ArtifactLocationDecoder artifactLocationDecoder = new ArtifactLocationDecoderImpl(blazeInfo, workspacePathResolver);
WorkspaceLanguageSettings workspaceLanguageSettings = LanguageSupport.createWorkspaceLanguageSettings(projectViewSet);
syncStats.setLanguagesActive(new ArrayList<>(workspaceLanguageSettings.activeLanguages));
syncStats.setWorkspaceType(workspaceLanguageSettings.getWorkspaceType());
for (BlazeSyncPlugin syncPlugin : BlazeSyncPlugin.EP_NAME.getExtensions()) {
syncPlugin.installSdks(context);
}
if (!ProjectViewVerifier.verifyProjectView(project, context, workspacePathResolver, projectViewSet, workspaceLanguageSettings)) {
return SyncResult.FAILURE;
}
final BlazeProjectData newBlazeProjectData;
WorkingSet workingSet = FutureUtil.waitForFuture(context, workingSetFuture).timed("WorkingSet", EventType.Other).withProgressMessage("Computing VCS working set...").onError("Could not compute working set").run().result();
if (context.isCancelled()) {
return SyncResult.CANCELLED;
}
if (context.hasErrors()) {
return SyncResult.FAILURE;
}
if (workingSet != null) {
printWorkingSet(context, workingSet);
}
SyncState.Builder syncStateBuilder = new SyncState.Builder();
SyncState previousSyncState = oldBlazeProjectData != null ? oldBlazeProjectData.syncState : null;
List<TargetExpression> targets = Lists.newArrayList();
if (syncParams.addProjectViewTargets) {
Collection<TargetExpression> projectViewTargets = projectViewSet.listItems(TargetSection.KEY);
if (!projectViewTargets.isEmpty()) {
syncStats.setBlazeProjectTargets(new ArrayList<>(projectViewTargets));
targets.addAll(projectViewTargets);
printTargets(context, "project view", projectViewTargets);
}
}
if (syncParams.addWorkingSet && workingSet != null) {
Collection<? extends TargetExpression> workingSetTargets = getWorkingSetTargets(projectViewSet, workingSet);
if (!workingSetTargets.isEmpty()) {
targets.addAll(workingSetTargets);
syncStats.setWorkingSetTargets(new ArrayList<>(workingSetTargets));
printTargets(context, "working set", workingSetTargets);
}
}
if (!syncParams.targetExpressions.isEmpty()) {
targets.addAll(syncParams.targetExpressions);
printTargets(context, syncParams.title, syncParams.targetExpressions);
}
ShardedTargetsResult shardedTargetsResult = BlazeBuildTargetSharder.expandAndShardTargets(project, context, workspaceRoot, projectViewSet, workspacePathResolver, targets);
if (shardedTargetsResult.buildResult.status == BuildResult.Status.FATAL_ERROR) {
return SyncResult.FAILURE;
}
ShardedTargetList shardedTargets = shardedTargetsResult.shardedTargets;
syncStats.setSyncSharded(shardedTargets.shardedTargets.size() > 1);
BlazeConfigurationHandler configHandler = new BlazeConfigurationHandler(blazeInfo);
boolean mergeWithOldState = !syncParams.addProjectViewTargets;
BlazeIdeInterface.IdeResult ideQueryResult = getIdeQueryResult(project, context, projectViewSet, blazeVersionData, configHandler, shardedTargets, workspaceLanguageSettings, artifactLocationDecoder, syncStateBuilder, previousSyncState, mergeWithOldState);
if (context.isCancelled()) {
return SyncResult.CANCELLED;
}
context.output(PrintOutput.log("ide-query result: " + ideQueryResult.buildResult.status));
if (ideQueryResult.targetMap == null || ideQueryResult.buildResult.status == BuildResult.Status.FATAL_ERROR) {
context.setHasError();
if (ideQueryResult.buildResult.outOfMemory()) {
SuggestBuildShardingNotification.syncOutOfMemoryError(project, context);
}
return SyncResult.FAILURE;
}
TargetMap targetMap = ideQueryResult.targetMap;
context.output(PrintOutput.log("Target map size: " + ideQueryResult.targetMap.targets().size()));
BuildResult ideInfoResult = ideQueryResult.buildResult;
ListenableFuture<ImmutableMultimap<TargetKey, TargetKey>> reverseDependenciesFuture = BlazeExecutor.getInstance().submit(() -> ReverseDependencyMap.createRdepsMap(targetMap));
BuildResult ideResolveResult = resolveIdeArtifacts(project, context, workspaceRoot, projectViewSet, blazeVersionData, workspaceLanguageSettings, shardedTargets);
if (ideResolveResult.status == BuildResult.Status.FATAL_ERROR) {
context.setHasError();
if (ideResolveResult.outOfMemory()) {
SuggestBuildShardingNotification.syncOutOfMemoryError(project, context);
}
return SyncResult.FAILURE;
}
if (context.isCancelled()) {
return SyncResult.CANCELLED;
}
Scope.push(context, (childContext) -> {
childContext.push(new TimingScope("UpdateSyncState", EventType.Other));
for (BlazeSyncPlugin syncPlugin : BlazeSyncPlugin.EP_NAME.getExtensions()) {
syncPlugin.updateSyncState(project, childContext, workspaceRoot, projectViewSet, workspaceLanguageSettings, blazeInfo, workingSet, workspacePathResolver, artifactLocationDecoder, targetMap, syncStateBuilder, previousSyncState);
}
});
ImmutableMultimap<TargetKey, TargetKey> reverseDependencies = FutureUtil.waitForFuture(context, reverseDependenciesFuture).timed("ReverseDependencies", EventType.Other).onError("Failed to compute reverse dependency map").run().result();
if (reverseDependencies == null) {
return SyncResult.FAILURE;
}
newBlazeProjectData = new BlazeProjectData(syncStartTime, targetMap, blazeInfo, blazeVersionData, workspacePathResolver, artifactLocationDecoder, workspaceLanguageSettings, syncStateBuilder.build(), reverseDependencies);
FileCaches.onSync(project, context, projectViewSet, newBlazeProjectData, syncParams.syncMode);
ListenableFuture<?> prefetch = PrefetchService.getInstance().prefetchProjectFiles(project, projectViewSet, newBlazeProjectData);
FutureUtil.waitForFuture(context, prefetch).withProgressMessage("Prefetching files...").timed("PrefetchFiles", EventType.Prefetching).onError("Prefetch failed").run();
ListenableFuture<DirectoryStructure> directoryStructureFuture = DirectoryStructure.getRootDirectoryStructure(project, workspaceRoot, projectViewSet);
refreshVirtualFileSystem(context, newBlazeProjectData);
DirectoryStructure directoryStructure = FutureUtil.waitForFuture(context, directoryStructureFuture).withProgressMessage("Computing directory structure...").timed("DirectoryStructure", EventType.Other).onError("Directory structure computation failed").run().result();
if (directoryStructure == null) {
return SyncResult.FAILURE;
}
boolean success = updateProject(context, projectViewSet, blazeVersionData, directoryStructure, oldBlazeProjectData, newBlazeProjectData);
if (!success) {
return SyncResult.FAILURE;
}
SyncResult syncResult = SyncResult.SUCCESS;
if (ideInfoResult.status == BuildResult.Status.BUILD_ERROR || ideResolveResult.status == BuildResult.Status.BUILD_ERROR) {
final String errorType = ideInfoResult.status == BuildResult.Status.BUILD_ERROR ? "BUILD file errors" : "compilation errors";
String message = String.format("Sync was successful, but there were %s. " + "The project may not be fully updated or resolve until fixed. " + "If the errors are from your working set, please uncheck " + "'Blaze > Sync > Expand Sync to Working Set' and try again.", errorType);
context.output(PrintOutput.error(message));
IssueOutput.warn(message).submit(context);
syncResult = SyncResult.PARTIAL_SUCCESS;
}
return syncResult;
}
use of com.google.common.collect.ImmutableMultimap in project graylog2-server by Graylog2.
the class EntityDependencyPermissionChecker method check.
/**
* Runs permission checks for the given dependencies for every selected grantee and returns the entities that
* grantees cannot access.
*
* @param sharingUser the sharing user
* @param dependencies the dependencies to check
* @param selectedGrantees the selected grantees
* @return dependencies that grantees cannot access, grouped by grantee
*/
public ImmutableMultimap<GRN, EntityDescriptor> check(GRN sharingUser, ImmutableSet<EntityDescriptor> dependencies, Set<GRN> selectedGrantees) {
final ImmutableMultimap.Builder<GRN, EntityDescriptor> deniedDependencies = ImmutableMultimap.builder();
final GranteeAuthorizer sharerAuthorizer = granteeAuthorizerFactory.create(sharingUser);
for (final GRN grantee : selectedGrantees) {
// We only check for existing grants for the actual grantee. If the grantee is a team, we only check if
// the team has a grant, not if any users in the team can access the dependency via other grants.
// The same for the "everyone" grantee, we only check if the "everyone" grantee has access to a dependency.
final GranteeAuthorizer granteeAuthorizer = granteeAuthorizerFactory.create(grantee);
for (final EntityDescriptor dependency : dependencies) {
// leaking information to the sharing user.
if (cannotView(sharerAuthorizer, dependency)) {
continue;
}
if (cannotView(granteeAuthorizer, dependency)) {
deniedDependencies.put(grantee, dependency);
}
}
}
return deniedDependencies.build();
}
use of com.google.common.collect.ImmutableMultimap in project beam by apache.
the class ScalarFunctionImpl method createAll.
/**
* Creates {@link org.apache.beam.vendor.calcite.v1_28_0.org.apache.calcite.schema.Function} for
* each method in a given class.
*/
public static ImmutableMultimap<String, Function> createAll(Class<?> clazz) {
final ImmutableMultimap.Builder<String, Function> builder = ImmutableMultimap.builder();
for (Method method : clazz.getMethods()) {
if (method.getDeclaringClass() == Object.class) {
continue;
}
if (!Modifier.isStatic(method.getModifiers()) && !classHasPublicZeroArgsConstructor(clazz)) {
continue;
}
final Function function = create(method);
builder.put(method.getName(), function);
}
return builder.build();
}
Aggregations