use of org.apache.flink.api.java.tuple.Tuple3 in project flink by apache.
the class PartitionITCase method testHashPartitionByKeyFieldAndDifferentParallelism.
@Test
public void testHashPartitionByKeyFieldAndDifferentParallelism() throws Exception {
/*
* Test hash partition by key field and different parallelism
*/
final ExecutionEnvironment env = ExecutionEnvironment.getExecutionEnvironment();
env.setParallelism(3);
DataSet<Tuple3<Integer, Long, String>> ds = CollectionDataSets.get3TupleDataSet(env);
DataSet<Long> uniqLongs = ds.partitionByHash(1).setParallelism(4).mapPartition(new UniqueTupleLongMapper());
List<Long> result = uniqLongs.collect();
String expected = "1\n" + "2\n" + "3\n" + "4\n" + "5\n" + "6\n";
compareResultAsText(result, expected);
}
use of org.apache.flink.api.java.tuple.Tuple3 in project flink by apache.
the class SampleITCase method verifySamplerWithFraction.
private void verifySamplerWithFraction(boolean withReplacement, double fraction, long seed) throws Exception {
final ExecutionEnvironment env = ExecutionEnvironment.getExecutionEnvironment();
FlatMapOperator<Tuple3<Integer, Long, String>, String> ds = getSourceDataSet(env);
MapPartitionOperator<String, String> sampled = DataSetUtils.sample(ds, withReplacement, fraction, seed);
List<String> result = sampled.collect();
containsResultAsText(result, getSourceStrings());
}
use of org.apache.flink.api.java.tuple.Tuple3 in project flink by apache.
the class SampleITCase method verifySamplerWithFixedSize.
private void verifySamplerWithFixedSize(boolean withReplacement, int numSamples, long seed) throws Exception {
final ExecutionEnvironment env = ExecutionEnvironment.getExecutionEnvironment();
FlatMapOperator<Tuple3<Integer, Long, String>, String> ds = getSourceDataSet(env);
DataSet<String> sampled = DataSetUtils.sampleWithSize(ds, withReplacement, numSamples, seed);
List<String> result = sampled.collect();
assertEquals(numSamples, result.size());
containsResultAsText(result, getSourceStrings());
}
use of org.apache.flink.api.java.tuple.Tuple3 in project flink by apache.
the class SortPartitionITCase method testSortPartitionByFieldExpression.
@SuppressWarnings({ "rawtypes", "unchecked" })
@Test
public void testSortPartitionByFieldExpression() throws Exception {
/*
* Test sort partition on field expression
*/
final ExecutionEnvironment env = ExecutionEnvironment.getExecutionEnvironment();
env.setParallelism(4);
DataSet<Tuple3<Integer, Long, String>> ds = CollectionDataSets.get3TupleDataSet(env);
List<Tuple1<Boolean>> result = ds.map(new IdMapper()).setParallelism(// parallelize input
4).sortPartition("f1", Order.DESCENDING).mapPartition(new OrderCheckMapper<>(new Tuple3Checker())).distinct().collect();
String expected = "(true)\n";
compareResultAsText(result, expected);
}
use of org.apache.flink.api.java.tuple.Tuple3 in project flink by apache.
the class SumMinMaxITCase method testGroupedAggregate.
@Test
public void testGroupedAggregate() throws Exception {
/*
* Grouped Aggregate
*/
final ExecutionEnvironment env = ExecutionEnvironment.getExecutionEnvironment();
DataSet<Tuple3<Integer, Long, String>> ds = CollectionDataSets.get3TupleDataSet(env);
DataSet<Tuple2<Long, Integer>> aggregateDs = ds.groupBy(1).sum(0).project(1, 0);
List<Tuple2<Long, Integer>> result = aggregateDs.collect();
String expected = "1,1\n" + "2,5\n" + "3,15\n" + "4,34\n" + "5,65\n" + "6,111\n";
compareResultAsTuples(result, expected);
}
Aggregations