Search in sources :

Example 1 with KeepInputAsIsDistribution

use of org.apache.flink.table.planner.plan.nodes.exec.InputProperty.KeepInputAsIsDistribution in project flink by apache.

the class BatchExecExchange method translateToPlanInternal.

@SuppressWarnings("unchecked")
@Override
protected Transformation<RowData> translateToPlanInternal(PlannerBase planner, ExecNodeConfig config) {
    final ExecEdge inputEdge = getInputEdges().get(0);
    final Transformation<RowData> inputTransform = (Transformation<RowData>) inputEdge.translateToPlan(planner);
    final RowType inputType = (RowType) inputEdge.getOutputType();
    boolean requireUndefinedExchangeMode = false;
    final StreamPartitioner<RowData> partitioner;
    final int parallelism;
    final InputProperty inputProperty = getInputProperties().get(0);
    final RequiredDistribution requiredDistribution = inputProperty.getRequiredDistribution();
    final InputProperty.DistributionType distributionType = requiredDistribution.getType();
    switch(distributionType) {
        case ANY:
            partitioner = null;
            parallelism = ExecutionConfig.PARALLELISM_DEFAULT;
            break;
        case BROADCAST:
            partitioner = new BroadcastPartitioner<>();
            parallelism = ExecutionConfig.PARALLELISM_DEFAULT;
            break;
        case SINGLETON:
            partitioner = new GlobalPartitioner<>();
            parallelism = 1;
            break;
        case HASH:
            partitioner = createHashPartitioner(((HashDistribution) requiredDistribution), inputType, config);
            parallelism = ExecutionConfig.PARALLELISM_DEFAULT;
            break;
        case KEEP_INPUT_AS_IS:
            KeepInputAsIsDistribution keepInputAsIsDistribution = (KeepInputAsIsDistribution) requiredDistribution;
            if (keepInputAsIsDistribution.isStrict()) {
                // explicitly use ForwardPartitioner to guarantee the data distribution is
                // exactly the same as input
                partitioner = new ForwardPartitioner<>();
                requireUndefinedExchangeMode = true;
            } else {
                RequiredDistribution inputDistribution = ((KeepInputAsIsDistribution) requiredDistribution).getInputDistribution();
                checkArgument(inputDistribution instanceof HashDistribution, "Only HashDistribution is supported now");
                partitioner = new ForwardForConsecutiveHashPartitioner<>(createHashPartitioner(((HashDistribution) inputDistribution), inputType, config));
            }
            parallelism = inputTransform.getParallelism();
            break;
        default:
            throw new TableException(distributionType + "is not supported now!");
    }
    final StreamExchangeMode exchangeMode = requireUndefinedExchangeMode ? StreamExchangeMode.UNDEFINED : getBatchStreamExchangeMode(config, requiredExchangeMode);
    final Transformation<RowData> transformation = new PartitionTransformation<>(inputTransform, partitioner, exchangeMode);
    transformation.setParallelism(parallelism);
    transformation.setOutputType(InternalTypeInfo.of(getOutputType()));
    return transformation;
}
Also used : RequiredDistribution(org.apache.flink.table.planner.plan.nodes.exec.InputProperty.RequiredDistribution) PartitionTransformation(org.apache.flink.streaming.api.transformations.PartitionTransformation) Transformation(org.apache.flink.api.dag.Transformation) TableException(org.apache.flink.table.api.TableException) ExecEdge(org.apache.flink.table.planner.plan.nodes.exec.ExecEdge) InputProperty(org.apache.flink.table.planner.plan.nodes.exec.InputProperty) RowType(org.apache.flink.table.types.logical.RowType) PartitionTransformation(org.apache.flink.streaming.api.transformations.PartitionTransformation) HashDistribution(org.apache.flink.table.planner.plan.nodes.exec.InputProperty.HashDistribution) RowData(org.apache.flink.table.data.RowData) KeepInputAsIsDistribution(org.apache.flink.table.planner.plan.nodes.exec.InputProperty.KeepInputAsIsDistribution) StreamExchangeModeUtils.getBatchStreamExchangeMode(org.apache.flink.table.planner.utils.StreamExchangeModeUtils.getBatchStreamExchangeMode) StreamExchangeMode(org.apache.flink.streaming.api.transformations.StreamExchangeMode)

Example 2 with KeepInputAsIsDistribution

use of org.apache.flink.table.planner.plan.nodes.exec.InputProperty.KeepInputAsIsDistribution in project flink by apache.

the class BatchExecExchange method getDescription.

@Override
public String getDescription() {
    // make sure the description be consistent with before, update this once plan is stable
    RequiredDistribution requiredDistribution = getInputProperties().get(0).getRequiredDistribution();
    StringBuilder sb = new StringBuilder();
    String type = requiredDistribution.getType().name().toLowerCase();
    if (type.equals("singleton")) {
        type = "single";
    } else if (requiredDistribution instanceof KeepInputAsIsDistribution && ((KeepInputAsIsDistribution) requiredDistribution).isStrict()) {
        type = "forward";
    }
    sb.append("distribution=[").append(type);
    if (requiredDistribution instanceof HashDistribution) {
        sb.append(getHashDistributionDescription((HashDistribution) requiredDistribution));
    } else if (requiredDistribution instanceof KeepInputAsIsDistribution && !((KeepInputAsIsDistribution) requiredDistribution).isStrict()) {
        KeepInputAsIsDistribution distribution = (KeepInputAsIsDistribution) requiredDistribution;
        sb.append("[hash").append(getHashDistributionDescription((HashDistribution) distribution.getInputDistribution())).append("]");
    }
    sb.append("]");
    if (requiredExchangeMode == StreamExchangeMode.BATCH) {
        sb.append(", shuffle_mode=[BATCH]");
    }
    return String.format("Exchange(%s)", sb);
}
Also used : RequiredDistribution(org.apache.flink.table.planner.plan.nodes.exec.InputProperty.RequiredDistribution) KeepInputAsIsDistribution(org.apache.flink.table.planner.plan.nodes.exec.InputProperty.KeepInputAsIsDistribution) HashDistribution(org.apache.flink.table.planner.plan.nodes.exec.InputProperty.HashDistribution)

Aggregations

HashDistribution (org.apache.flink.table.planner.plan.nodes.exec.InputProperty.HashDistribution)2 KeepInputAsIsDistribution (org.apache.flink.table.planner.plan.nodes.exec.InputProperty.KeepInputAsIsDistribution)2 RequiredDistribution (org.apache.flink.table.planner.plan.nodes.exec.InputProperty.RequiredDistribution)2 Transformation (org.apache.flink.api.dag.Transformation)1 PartitionTransformation (org.apache.flink.streaming.api.transformations.PartitionTransformation)1 StreamExchangeMode (org.apache.flink.streaming.api.transformations.StreamExchangeMode)1 TableException (org.apache.flink.table.api.TableException)1 RowData (org.apache.flink.table.data.RowData)1 ExecEdge (org.apache.flink.table.planner.plan.nodes.exec.ExecEdge)1 InputProperty (org.apache.flink.table.planner.plan.nodes.exec.InputProperty)1 StreamExchangeModeUtils.getBatchStreamExchangeMode (org.apache.flink.table.planner.utils.StreamExchangeModeUtils.getBatchStreamExchangeMode)1 RowType (org.apache.flink.table.types.logical.RowType)1