Search in sources :

Example 1 with Config

use of org.apache.hudi.utilities.HDFSParquetImporter.Config in project hudi by apache.

the class SparkMain method compact.

private static int compact(JavaSparkContext jsc, String basePath, String tableName, String compactionInstant, int parallelism, String schemaFile, int retry, String mode, String propsFilePath, List<String> configs) {
    HoodieCompactor.Config cfg = new HoodieCompactor.Config();
    cfg.basePath = basePath;
    cfg.tableName = tableName;
    cfg.compactionInstantTime = compactionInstant;
    // TODO: Make this configurable along with strategy specific config - For now, this is a generic enough strategy
    cfg.strategyClassName = UnBoundedCompactionStrategy.class.getCanonicalName();
    cfg.parallelism = parallelism;
    cfg.schemaFile = schemaFile;
    cfg.runningMode = mode;
    cfg.propsFilePath = propsFilePath;
    cfg.configs = configs;
    return new HoodieCompactor(jsc, cfg).compact(retry);
}
Also used : HoodieWriteConfig(org.apache.hudi.config.HoodieWriteConfig) HoodieIndexConfig(org.apache.hudi.config.HoodieIndexConfig) Config(org.apache.hudi.utilities.HDFSParquetImporter.Config) HoodieBootstrapConfig(org.apache.hudi.config.HoodieBootstrapConfig) HoodieCompactor(org.apache.hudi.utilities.HoodieCompactor) UnBoundedCompactionStrategy(org.apache.hudi.table.action.compact.strategy.UnBoundedCompactionStrategy)

Example 2 with Config

use of org.apache.hudi.utilities.HDFSParquetImporter.Config in project hudi by apache.

the class SparkMain method dataLoad.

private static int dataLoad(JavaSparkContext jsc, String command, String srcPath, String targetPath, String tableName, String tableType, String rowKey, String partitionKey, int parallelism, String schemaFile, int retry, String propsFilePath, List<String> configs) {
    Config cfg = new Config();
    cfg.command = command;
    cfg.srcPath = srcPath;
    cfg.targetPath = targetPath;
    cfg.tableName = tableName;
    cfg.tableType = tableType;
    cfg.rowKey = rowKey;
    cfg.partitionKey = partitionKey;
    cfg.parallelism = parallelism;
    cfg.schemaFile = schemaFile;
    cfg.propsFilePath = propsFilePath;
    cfg.configs = configs;
    return new HDFSParquetImporter(cfg).dataImport(jsc, retry);
}
Also used : HoodieWriteConfig(org.apache.hudi.config.HoodieWriteConfig) HoodieIndexConfig(org.apache.hudi.config.HoodieIndexConfig) Config(org.apache.hudi.utilities.HDFSParquetImporter.Config) HoodieBootstrapConfig(org.apache.hudi.config.HoodieBootstrapConfig) HDFSParquetImporter(org.apache.hudi.utilities.HDFSParquetImporter)

Aggregations

HoodieBootstrapConfig (org.apache.hudi.config.HoodieBootstrapConfig)2 HoodieIndexConfig (org.apache.hudi.config.HoodieIndexConfig)2 HoodieWriteConfig (org.apache.hudi.config.HoodieWriteConfig)2 Config (org.apache.hudi.utilities.HDFSParquetImporter.Config)2 UnBoundedCompactionStrategy (org.apache.hudi.table.action.compact.strategy.UnBoundedCompactionStrategy)1 HDFSParquetImporter (org.apache.hudi.utilities.HDFSParquetImporter)1 HoodieCompactor (org.apache.hudi.utilities.HoodieCompactor)1