use of io.dingodb.exec.base.Output in project dingo by dingodb.
the class DingoJobVisitor method visit.
@Override
public Collection<Output> visit(@Nonnull DingoSort rel) {
Collection<Output> inputs = dingo(rel.getInput()).accept(this);
List<Output> outputs = new LinkedList<>();
for (Output input : inputs) {
Operator operator = new SortOperator(rel.getCollation().getFieldCollations().stream().map(c -> new SortCollation(c.getFieldIndex(), c.direction, c.nullDirection)).collect(Collectors.toList()), rel.fetch == null ? -1 : RexLiteral.intValue(rel.fetch), rel.offset == null ? 0 : RexLiteral.intValue(rel.offset));
Task task = input.getTask();
operator.setId(idGenerator.get());
task.putOperator(operator);
input.setLink(operator.getInput(0));
operator.getSoleOutput().copyHint(input);
outputs.addAll(operator.getOutputs());
}
return outputs;
}
use of io.dingodb.exec.base.Output in project dingo by dingodb.
the class DingoJobVisitor method createJob.
@Nonnull
public static Job createJob(RelNode input, boolean addRoot) {
IdGenerator idGenerator = new IdGenerator();
DingoJobVisitor visitor = new DingoJobVisitor(idGenerator);
Collection<Output> outputs = dingo(input).accept(visitor);
if (addRoot) {
if (outputs.size() == 1) {
Output output = sole(outputs);
Task task = output.getTask();
RootOperator root = new RootOperator(TupleSchema.fromRelDataType(input.getRowType()));
root.setId(idGenerator.get());
task.putOperator(root);
output.setLink(root.getInput(0));
} else if (!outputs.isEmpty()) {
throw new IllegalStateException("There must be zero or one output to job root.");
}
}
Job job = visitor.getJob();
log.info("job = {}", job);
return job;
}
use of io.dingodb.exec.base.Output in project dingo by dingodb.
the class DingoJobVisitor method visit.
@Override
public Collection<Output> visit(@Nonnull DingoPartition rel) {
Collection<Output> inputs = dingo(rel.getInput()).accept(this);
String tableName = getSimpleName(rel.getTable());
List<Output> outputs = new LinkedList<>();
final Map<String, Location> partLocations = Services.META.getPartLocations(tableName);
final TableDefinition td = Services.META.getTableDefinition(tableName);
final PartitionStrategy ps = new SimpleHashStrategy(partLocations.size());
for (Output input : inputs) {
Task task = input.getTask();
PartitionOperator operator = new PartitionOperator(ps, td.getKeyMapping());
operator.setId(idGenerator.get());
operator.createOutputs(tableName, partLocations);
task.putOperator(operator);
input.setLink(operator.getInput(0));
outputs.addAll(operator.getOutputs());
}
return outputs;
}
use of io.dingodb.exec.base.Output in project dingo by dingodb.
the class DingoJobVisitor method visit.
@Override
public Collection<Output> visit(@Nonnull DingoPartScan rel) {
String tableName = getSimpleName(rel.getTable());
TableDefinition td = Services.META.getTableDefinition(tableName);
Map<String, Location> parts = Services.META.getPartLocations(tableName);
List<Output> outputs = new ArrayList<>(parts.size());
TableId tableId = new TableId(Services.META.getTableKey(tableName));
String filterStr = null;
if (rel.getFilter() != null) {
filterStr = RexConverter.convert(rel.getFilter()).toString();
}
for (Map.Entry<String, Location> entry : parts.entrySet()) {
final Object partId = entry.getKey();
PartScanOperator operator = new PartScanOperator(tableId, entry.getKey(), td.getTupleSchema(), td.getKeyMapping(), filterStr, rel.getSelection());
operator.setId(idGenerator.get());
Task task = job.getOrCreate(entry.getValue());
task.putOperator(operator);
operator.getSoleOutput().setHint(OutputHint.of(tableName, partId));
outputs.addAll(operator.getOutputs());
}
return outputs;
}
use of io.dingodb.exec.base.Output in project dingo by dingodb.
the class DingoJobVisitor method visit.
@Override
public Collection<Output> visit(@Nonnull DingoGetByKeys rel) {
String tableName = getSimpleName(rel.getTable());
final Map<String, Location> partLocations = Services.META.getPartLocations(tableName);
final TableDefinition td = Services.META.getTableDefinition(tableName);
final PartitionStrategy ps = new SimpleHashStrategy(partLocations.size());
final TableId tableId = new TableId(Services.META.getTableKey(tableName));
Map<String, List<Object[]>> partMap = ps.partKeyTuples(rel.getKeyTuples());
List<Output> outputs = new LinkedList<>();
for (Map.Entry<String, List<Object[]>> entry : partMap.entrySet()) {
final Object partId = entry.getKey();
GetByKeysOperator operator = new GetByKeysOperator(tableId, partId, td.getTupleSchema(), td.getKeyMapping(), entry.getValue(), rel.getSelection());
operator.setId(idGenerator.get());
Task task = job.getOrCreate(partLocations.get(entry.getKey()));
task.putOperator(operator);
operator.getSoleOutput().setHint(OutputHint.of(tableName, partId));
outputs.addAll(operator.getOutputs());
}
return outputs;
}
Aggregations