use of org.talend.hadoop.distribution.component.SparkBatchComponent in project tbd-studio-se by Talend.
the class AbstractDynamicDistributionTemplate method buildModuleGroupsTemplateMap.
protected Map<ComponentType, IDynamicModuleGroupTemplate> buildModuleGroupsTemplateMap() {
Map<ComponentType, IDynamicModuleGroupTemplate> moduleGroupsTemplateMap = new HashMap<>();
DynamicPluginAdapter pluginAdapter = getPluginAdapter();
if (this instanceof HDFSComponent) {
moduleGroupsTemplateMap.put(ComponentType.HDFS, new DynamicHDFSModuleGroupTemplate(pluginAdapter));
}
if (this instanceof HBaseComponent) {
moduleGroupsTemplateMap.put(ComponentType.HBASE, new DynamicHBaseModuleGroupTemplate(pluginAdapter));
}
if (this instanceof HCatalogComponent) {
moduleGroupsTemplateMap.put(ComponentType.HCATALOG, new DynamicHCatalogModuleGroupTemplate(pluginAdapter));
}
if (this instanceof HiveComponent) {
moduleGroupsTemplateMap.put(ComponentType.HIVE, new DynamicHiveModuleGroupTemplate(pluginAdapter));
}
if (this instanceof HiveOnSparkComponent) {
moduleGroupsTemplateMap.put(ComponentType.HIVEONSPARK, new DynamicHiveOnSparkModuleGroupTemplate(pluginAdapter));
}
if (this instanceof MapRDBComponent) {
moduleGroupsTemplateMap.put(ComponentType.MAPRDB, new DynamicMapRDBModuleGroupTemplate(pluginAdapter));
}
if (this instanceof MRComponent) {
moduleGroupsTemplateMap.put(ComponentType.MAPREDUCE, new DynamicMapReduceModuleGroupTemplate(pluginAdapter));
}
if (this instanceof SparkBatchComponent) {
moduleGroupsTemplateMap.put(ComponentType.SPARKBATCH, new DynamicSparkBatchModuleGroupTemplate(pluginAdapter));
}
if (this instanceof SparkStreamingComponent) {
moduleGroupsTemplateMap.put(ComponentType.SPARKSTREAMING, new DynamicSparkStreamingModuleGroupTemplate(pluginAdapter));
}
if (this instanceof SqoopComponent) {
moduleGroupsTemplateMap.put(ComponentType.SQOOP, new DynamicSqoopModuleGroupTemplate(pluginAdapter));
}
if (this instanceof ImpalaComponent) {
moduleGroupsTemplateMap.put(ComponentType.IMPALA, new DynamicImpalaModuleGroupTemplate(pluginAdapter));
}
return moduleGroupsTemplateMap;
}
use of org.talend.hadoop.distribution.component.SparkBatchComponent in project tbd-studio-se by Talend.
the class AbstractDynamicCDHDistributionTemplate method buildModuleGroupsTemplateMap.
@Override
protected Map<ComponentType, IDynamicModuleGroupTemplate> buildModuleGroupsTemplateMap() {
Map<ComponentType, IDynamicModuleGroupTemplate> groupTemplateMap = super.buildModuleGroupsTemplateMap();
DynamicPluginAdapter pluginAdapter = getPluginAdapter();
if (this instanceof HiveOnSparkComponent) {
groupTemplateMap.put(ComponentType.HIVEONSPARK, new DynamicCDHHiveOnSparkModuleGroupTemplate(pluginAdapter));
}
if (this instanceof MRComponent) {
groupTemplateMap.put(ComponentType.MAPREDUCE, new DynamicCDHMapReduceModuleGroupTemplate(pluginAdapter));
}
if (this instanceof SparkBatchComponent) {
groupTemplateMap.put(ComponentType.SPARKBATCH, new DynamicCDHSparkBatchModuleGroupTemplate(pluginAdapter));
}
if (this instanceof SparkStreamingComponent) {
groupTemplateMap.put(ComponentType.SPARKSTREAMING, new DynamicCDHSparkStreamingModuleGroupTemplate(pluginAdapter));
}
if (this instanceof SqoopComponent) {
groupTemplateMap.put(ComponentType.SQOOP, new DynamicCDHSqoopModuleGroupTemplate(pluginAdapter));
}
if (this instanceof ImpalaComponent) {
groupTemplateMap.put(ComponentType.IMPALA, new DynamicImpalaModuleGroupTemplate(pluginAdapter));
}
return groupTemplateMap;
}
use of org.talend.hadoop.distribution.component.SparkBatchComponent in project tbd-studio-se by Talend.
the class CustomDistributionTest method testCustomDistribution.
@Test
public void testCustomDistribution() throws Exception {
HadoopComponent distribution = new CustomDistribution();
assertNotNull(distribution.getDistributionName());
assertNull(distribution.getVersionName(null));
assertTrue(distribution.doSupportS3());
assertEquals(CustomDistribution.DISTRIBUTION_NAME, distribution.getDistribution());
assertNull(distribution.getVersion());
assertNull(distribution.getHadoopVersion());
assertTrue(distribution.doSupportKerberos());
assertTrue(distribution.doSupportUseDatanodeHostname());
assertFalse(distribution.doSupportGroup());
assertTrue(distribution.doSupportOldImportMode());
assertTrue(((HDFSComponent) distribution).doSupportSequenceFileShortType());
assertFalse(((MRComponent) distribution).isExecutedThroughWebHCat());
assertFalse(((MRComponent) distribution).doSupportCrossPlatformSubmission());
assertTrue(((MRComponent) distribution).doSupportImpersonation());
assertEquals(DEFAULT_YARN_APPLICATION_CLASSPATH, ((MRComponent) distribution).getYarnApplicationClasspath());
assertFalse(((HBaseComponent) distribution).doSupportNewHBaseAPI());
assertTrue(((SqoopComponent) distribution).doJavaAPISupportStorePasswordInFile());
assertFalse(((SqoopComponent) distribution).doJavaAPISqoopImportSupportDeleteTargetDir());
assertTrue(((SqoopComponent) distribution).doJavaAPISqoopImportAllTablesSupportExcludeTable());
assertTrue(((HiveComponent) distribution).doSupportEmbeddedMode());
assertTrue(((HiveComponent) distribution).doSupportStandaloneMode());
assertTrue(((HiveComponent) distribution).doSupportHive1());
assertTrue(((HiveComponent) distribution).doSupportHive2());
assertTrue(((HiveComponent) distribution).doSupportTezForHive());
assertTrue(((HiveComponent) distribution).doSupportHBaseForHive());
assertTrue(((HiveComponent) distribution).doSupportSSL());
assertTrue(((HiveComponent) distribution).doSupportORCFormat());
assertTrue(((HiveComponent) distribution).doSupportAvroFormat());
assertTrue(((HiveComponent) distribution).doSupportParquetFormat());
assertFalse(((SparkBatchComponent) distribution).getSparkVersions().contains(ESparkVersion.SPARK_2_0));
assertFalse(((SparkBatchComponent) distribution).getSparkVersions().contains(ESparkVersion.SPARK_1_6));
assertFalse(((SparkBatchComponent) distribution).getSparkVersions().contains(ESparkVersion.SPARK_1_5));
assertFalse(((SparkBatchComponent) distribution).getSparkVersions().contains(ESparkVersion.SPARK_1_4));
assertTrue(((SparkBatchComponent) distribution).getSparkVersions().contains(ESparkVersion.SPARK_1_3));
assertTrue(((SparkBatchComponent) distribution).doSupportDynamicMemoryAllocation());
assertFalse(((SparkBatchComponent) distribution).isExecutedThroughSparkJobServer());
assertTrue(((SparkBatchComponent) distribution).doSupportSparkStandaloneMode());
assertTrue(((SparkBatchComponent) distribution).doSupportSparkYarnClientMode());
assertFalse(((SparkStreamingComponent) distribution).getSparkVersions().contains(ESparkVersion.SPARK_2_0));
assertFalse(((SparkStreamingComponent) distribution).getSparkVersions().contains(ESparkVersion.SPARK_1_6));
assertFalse(((SparkStreamingComponent) distribution).getSparkVersions().contains(ESparkVersion.SPARK_1_5));
assertFalse(((SparkStreamingComponent) distribution).getSparkVersions().contains(ESparkVersion.SPARK_1_4));
assertTrue(((SparkStreamingComponent) distribution).getSparkVersions().contains(ESparkVersion.SPARK_1_3));
assertTrue(((SparkStreamingComponent) distribution).doSupportDynamicMemoryAllocation());
assertFalse(((SparkStreamingComponent) distribution).isExecutedThroughSparkJobServer());
assertTrue(((SparkStreamingComponent) distribution).doSupportCheckpointing());
assertTrue(((SparkStreamingComponent) distribution).doSupportSparkStandaloneMode());
assertTrue(((SparkStreamingComponent) distribution).doSupportSparkYarnClientMode());
assertTrue(((SparkStreamingComponent) distribution).doSupportBackpressure());
assertFalse(((HiveComponent) distribution).doSupportStoreAsParquet());
assertFalse(((HiveComponent) distribution).doSupportClouderaNavigator());
assertTrue(distribution instanceof HCatalogComponent);
assertTrue(distribution instanceof ImpalaComponent);
assertTrue(distribution.doSupportCreateServiceConnection());
assertTrue((distribution.getNecessaryServiceName() == null ? 0 : distribution.getNecessaryServiceName().size()) == 0);
assertFalse(distribution.doSupportAzureDataLakeStorage());
assertFalse(distribution.doSupportWebHDFS());
}
use of org.talend.hadoop.distribution.component.SparkBatchComponent in project tbd-studio-se by Talend.
the class DBR550DistributionTest method testDatabricksDistribution.
@Test
public void testDatabricksDistribution() throws Exception {
HadoopComponent distribution = new DBR550Distribution();
assertNotNull(distribution.getDistributionName());
assertNotNull(distribution.getVersionName(null));
assertTrue(distribution.doSupportS3());
assertEquals(DBR550Distribution.DISTRIBUTION_NAME, distribution.getDistribution());
assertEquals(DBR550Distribution.VERSION, distribution.getVersion());
assertEquals(EHadoopVersion.HADOOP_2, distribution.getHadoopVersion());
assertFalse(distribution.doSupportKerberos());
assertTrue(distribution.doSupportUseDatanodeHostname());
assertFalse(distribution.doSupportGroup());
assertTrue(((SparkBatchComponent) distribution).getSparkVersions().contains(ESparkVersion.SPARK_2_4_X));
assertFalse(((SparkBatchComponent) distribution).getSparkVersions().contains(ESparkVersion.SPARK_2_3));
assertFalse(((SparkBatchComponent) distribution).getSparkVersions().contains(ESparkVersion.SPARK_2_2));
assertFalse(((SparkBatchComponent) distribution).getSparkVersions().contains(ESparkVersion.SPARK_2_0));
assertFalse(((SparkBatchComponent) distribution).getSparkVersions().contains(ESparkVersion.SPARK_1_6));
assertFalse(((SparkBatchComponent) distribution).getSparkVersions().contains(ESparkVersion.SPARK_1_5));
assertFalse(((SparkBatchComponent) distribution).getSparkVersions().contains(ESparkVersion.SPARK_1_4));
assertFalse(((SparkBatchComponent) distribution).getSparkVersions().contains(ESparkVersion.SPARK_1_3));
assertTrue(((SparkBatchComponent) distribution).doSupportDynamicMemoryAllocation());
assertFalse(((SparkBatchComponent) distribution).isExecutedThroughSparkJobServer());
assertTrue(((SparkBatchComponent) distribution).doSupportSparkStandaloneMode());
assertFalse(((SparkBatchComponent) distribution).doSupportSparkYarnClientMode());
assertFalse(((SparkBatchComponent) distribution).doSupportSparkYarnClusterMode());
assertTrue(((SparkStreamingComponent) distribution).doSupportBackpressure());
assertTrue(((SparkStreamingComponent) distribution).doSupportCheckpointing());
assertTrue(distribution.doSupportCreateServiceConnection());
assertTrue((distribution.getNecessaryServiceName() == null ? 0 : distribution.getNecessaryServiceName().size()) == 0);
assertTrue(distribution.doSupportAzureDataLakeStorage());
}
use of org.talend.hadoop.distribution.component.SparkBatchComponent in project tbd-studio-se by Talend.
the class DBR640DistributionTest method testDatabricksDistribution.
@Test
public void testDatabricksDistribution() throws Exception {
HadoopComponent distribution = new DBR640Distribution();
assertNotNull(distribution.getDistributionName());
assertNotNull(distribution.getVersionName(null));
assertTrue(distribution.doSupportS3());
assertEquals(DBR640Distribution.DISTRIBUTION_NAME, distribution.getDistribution());
assertEquals(DBR640Distribution.VERSION, distribution.getVersion());
assertEquals(EHadoopVersion.HADOOP_2, distribution.getHadoopVersion());
assertFalse(distribution.doSupportKerberos());
assertTrue(distribution.doSupportUseDatanodeHostname());
assertFalse(distribution.doSupportGroup());
assertTrue(((SparkBatchComponent) distribution).getSparkVersions().contains(ESparkVersion.SPARK_2_4_X));
assertFalse(((SparkBatchComponent) distribution).getSparkVersions().contains(ESparkVersion.SPARK_2_3));
assertFalse(((SparkBatchComponent) distribution).getSparkVersions().contains(ESparkVersion.SPARK_2_2));
assertFalse(((SparkBatchComponent) distribution).getSparkVersions().contains(ESparkVersion.SPARK_2_0));
assertFalse(((SparkBatchComponent) distribution).getSparkVersions().contains(ESparkVersion.SPARK_1_6));
assertFalse(((SparkBatchComponent) distribution).getSparkVersions().contains(ESparkVersion.SPARK_1_5));
assertFalse(((SparkBatchComponent) distribution).getSparkVersions().contains(ESparkVersion.SPARK_1_4));
assertFalse(((SparkBatchComponent) distribution).getSparkVersions().contains(ESparkVersion.SPARK_1_3));
assertTrue(((SparkBatchComponent) distribution).doSupportDynamicMemoryAllocation());
assertFalse(((SparkBatchComponent) distribution).isExecutedThroughSparkJobServer());
assertTrue(((SparkBatchComponent) distribution).doSupportSparkStandaloneMode());
assertFalse(((SparkBatchComponent) distribution).doSupportSparkYarnClientMode());
assertTrue(((SparkStreamingComponent) distribution).doSupportBackpressure());
assertTrue(((SparkStreamingComponent) distribution).doSupportCheckpointing());
assertTrue(distribution.doSupportCreateServiceConnection());
assertTrue((distribution.getNecessaryServiceName() == null ? 0 : distribution.getNecessaryServiceName().size()) == 0);
assertTrue(distribution.doSupportAzureDataLakeStorage());
}
Aggregations