use of org.pentaho.di.repository.Repository in project pentaho-kettle by pentaho.
the class RestMetaTest method testStepChecks.
@Test
public void testStepChecks() {
RestMeta meta = new RestMeta();
List<CheckResultInterface> remarks = new ArrayList<CheckResultInterface>();
TransMeta transMeta = new TransMeta();
StepMeta step = new StepMeta();
RowMetaInterface prev = new RowMeta();
RowMetaInterface info = new RowMeta();
String[] input = new String[0];
String[] output = new String[0];
VariableSpace variables = new Variables();
Repository repo = null;
IMetaStore metaStore = null;
// In a default configuration, it's expected that some errors will occur.
// For this, we'll grab a baseline count of the number of errors
// as the error count should decrease as we change configuration settings to proper values.
remarks.clear();
meta.check(remarks, transMeta, step, prev, input, output, info, variables, repo, metaStore);
final int errorsDefault = getCheckResultErrorCount(remarks);
assertTrue(errorsDefault > 0);
// Setting the step to read the URL from a field should fix one of the check() errors
meta.setUrlInField(true);
meta.setUrlField("urlField");
prev.addValueMeta(new ValueMetaString("urlField"));
remarks.clear();
meta.check(remarks, transMeta, step, prev, input, output, info, variables, repo, metaStore);
int errorsCurrent = getCheckResultErrorCount(remarks);
assertTrue(errorsDefault > errorsCurrent);
}
use of org.pentaho.di.repository.Repository in project pentaho-kettle by pentaho.
the class Import method main.
public static void main(String[] a) throws KettleException {
BufferedReader reader = new BufferedReader(new InputStreamReader(System.in));
KettleEnvironment.init();
Props.init(Props.TYPE_PROPERTIES_SPOON);
List<String> args = pickupCmdArguments(a);
StringBuilder optionRepname, optionUsername, optionPassword, optionDirname;
StringBuilder optionLimitDir, optionFilename, optionRules, optionComment;
StringBuilder optionReplace, optionContinueOnError, optionVersion, optionFileDir, optionNoRules;
CommandLineOption[] options = new CommandLineOption[] { //
createOption("rep", "Import.CmdLine.RepName", optionRepname = new StringBuilder()), createOption("user", "Import.CmdLine.RepUsername", optionUsername = new StringBuilder()), createOption("pass", "Import.CmdLine.RepPassword", optionPassword = new StringBuilder()), createOption("dir", "Import.CmdLine.RepDir", optionDirname = new StringBuilder()), createOption("limitdir", "Import.CmdLine.LimitDir", optionLimitDir = new StringBuilder()), createOption("file", "Import.CmdLine.File", optionFilename = new StringBuilder()), createOption("filedir", "Import.CmdLine.FileDir", optionFileDir = new StringBuilder()), createOption("rules", "Import.CmdLine.RulesFile", optionRules = new StringBuilder()), createOption("norules", "Import.CmdLine.NoRules", optionNoRules = new StringBuilder(), true, false), createOption("comment", "Import.CmdLine.Comment", optionComment = new StringBuilder(), false, false), createOption("replace", "Import.CmdLine.Replace", optionReplace = new StringBuilder(), true, false), createOption("coe", "Import.CmdLine.ContinueOnError", optionContinueOnError = new StringBuilder(), true, false), createOption("version", "Import.CmdLine.Version", optionVersion = new StringBuilder(), true, false), new CommandLineOption("", BaseMessages.getString(PKG, "Import.CmdLine.ExtraFiles"), new StringBuilder(), false, true, true) };
if (args.isEmpty()) {
CommandLineOption.printUsage(options);
exitJVM(9);
}
final LogChannelInterface log = new LogChannel(STRING_IMPORT);
CommandLineOption.parseArguments(args, options, log);
// The arguments that are still left in args are in fact filenames that need to be imported.
// This list is otherwise empty.
// To that we add the normal filename option
//
List<String> filenames = new ArrayList<String>(args);
if (!Utils.isEmpty(optionFilename)) {
filenames.add(optionFilename.toString());
}
String kettleRepname = Const.getEnvironmentVariable("KETTLE_REPOSITORY", null);
String kettleUsername = Const.getEnvironmentVariable("KETTLE_USER", null);
String kettlePassword = Const.getEnvironmentVariable("KETTLE_PASSWORD", null);
if (!Utils.isEmpty(kettleRepname)) {
optionRepname = new StringBuilder(kettleRepname);
}
if (!Utils.isEmpty(kettleUsername)) {
optionUsername = new StringBuilder(kettleUsername);
}
if (!Utils.isEmpty(kettlePassword)) {
optionPassword = new StringBuilder(kettlePassword);
}
if (!Utils.isEmpty(optionVersion)) {
BuildVersion buildVersion = BuildVersion.getInstance();
log.logBasic(BaseMessages.getString(PKG, "Import.Log.KettleVersion", buildVersion.getVersion(), buildVersion.getRevision(), buildVersion.getBuildDate()));
if (a.length == 1) {
exitJVM(6);
}
}
//
if (Utils.isEmpty(optionRepname)) {
log.logError(BaseMessages.getString(PKG, "Import.Error.NoRepProvided"));
exitJVM(1);
}
if (Utils.isEmpty(filenames)) {
log.logError(BaseMessages.getString(PKG, "Import.Error.NoExportFileProvided"));
exitJVM(1);
}
if (Utils.isEmpty(optionDirname)) {
log.logError(BaseMessages.getString(PKG, "Import.Error.NoRepositoryDirectoryProvided"));
exitJVM(1);
}
if (Utils.isEmpty(optionRules) && Utils.isEmpty(optionNoRules) && !"Y".equalsIgnoreCase(optionNoRules.toString())) {
log.logError(BaseMessages.getString(PKG, "Import.Error.NoRulesFileProvided"));
exitJVM(1);
}
// Load the rules file!
//
ImportRules importRules = new ImportRules();
String rulesFile = optionRules.toString();
if (!Utils.isEmpty(rulesFile)) {
try {
Document document = XMLHandler.loadXMLFile(rulesFile);
Node rulesNode = XMLHandler.getSubNode(document, ImportRules.XML_TAG);
importRules.loadXML(rulesNode);
log.logMinimal(BaseMessages.getString(PKG, "Import.Log.RulesLoaded", rulesFile, Integer.toString(importRules.getRules().size())));
for (ImportRuleInterface rule : importRules.getRules()) {
log.logBasic(" - " + rule.toString());
}
} catch (KettleException e) {
log.logError(BaseMessages.getString(PKG, "Import.Log.ExceptionLoadingRules", rulesFile), e);
exitJVM(7);
}
}
// Get the list of limiting source directories
//
List<String> limitDirs;
if (!Utils.isEmpty(optionLimitDir)) {
String[] directories = optionLimitDir.toString().split(",");
limitDirs = Arrays.asList(directories);
} else {
limitDirs = Collections.emptyList();
}
// Find the repository metadata...
//
RepositoriesMeta repsinfo = new RepositoriesMeta();
repsinfo.getLog().setLogLevel(log.getLogLevel());
try {
repsinfo.readData();
} catch (Exception e) {
log.logError(BaseMessages.getString(PKG, "Import.Error.UnableToLoadRepositoryInformation"), e);
exitJVM(1);
}
RepositoryMeta repositoryMeta = repsinfo.findRepository(optionRepname.toString());
if (repositoryMeta == null) {
log.logError(BaseMessages.getString(PKG, "Import.Error.RepositoryCouldNotBeFound", optionRepname.toString()));
exitJVM(1);
}
if (Utils.isEmpty(optionRepname)) {
log.logError(BaseMessages.getString(PKG, "Import.Error.NoRepProvided"));
exitJVM(1);
}
// Load the repository object as a plugin...
//
Repository repository = null;
try {
repository = PluginRegistry.getInstance().loadClass(RepositoryPluginType.class, repositoryMeta, Repository.class);
repository.init(repositoryMeta);
repository.getLog().setLogLevel(log.getLogLevel());
} catch (Exception e) {
log.logError(BaseMessages.getString(PKG, "Import.Error.UnableToLoadOrInitializeRepository"));
exitJVM(1);
}
try {
repository.connect(optionUsername != null ? optionUsername.toString() : null, optionPassword != null ? optionPassword.toString() : null);
} catch (KettleException ke) {
log.logError(ke.getMessage());
exitJVM(1);
} catch (Exception e) {
log.logError(BaseMessages.getString(PKG, "Import.Error.UnableToConnectToRepository"));
exitJVM(1);
}
final boolean replace = Utils.isEmpty(optionReplace) ? false : ValueMetaString.convertStringToBoolean(optionReplace.toString());
final boolean continueOnError = Utils.isEmpty(optionContinueOnError) ? false : ValueMetaString.convertStringToBoolean(optionContinueOnError.toString());
// Start the import!
//
log.logMinimal(BaseMessages.getString(PKG, "Import.Log.Starting"));
Date start, stop;
SimpleDateFormat df = new SimpleDateFormat("yyyy/MM/dd HH:mm:ss.SSS");
start = new Date();
int returnCode = 0;
try {
RepositoryDirectoryInterface tree = repository.loadRepositoryDirectoryTree();
RepositoryDirectoryInterface targetDirectory = tree.findDirectory(optionDirname.toString());
if (targetDirectory == null) {
log.logError(BaseMessages.getString(PKG, "Import.Error.UnableToFindTargetDirectoryInRepository", optionDirname.toString()));
exitJVM(1);
}
// Perform the actual import
IRepositoryImporter importer = repository.getImporter();
importer.setImportRules(importRules);
if (!limitDirs.isEmpty()) {
if (importer instanceof CanLimitDirs) {
((CanLimitDirs) importer).setLimitDirs(limitDirs);
} else {
throw new KettleException(BaseMessages.getString(PKG, "Import.CouldntLimitDirs", importer.getClass().getCanonicalName()));
}
}
RepositoryImportFeedbackInterface feedbackInterface = new ImportFeedback(log, continueOnError, replace, reader);
// Import files in a certain directory
importer.importAll(feedbackInterface, optionFileDir.toString(), filenames.toArray(new String[filenames.size()]), targetDirectory, replace, continueOnError, optionComment.toString());
// If the importer has exceptions, then our return code is 2
List<Exception> exceptions = importer.getExceptions();
if (exceptions != null && !exceptions.isEmpty()) {
log.logError(BaseMessages.getString(PKG, "Import.Error.UnexpectedErrorDuringImport"), exceptions.get(0));
returnCode = 2;
}
} catch (Exception e) {
log.logError(BaseMessages.getString(PKG, "Import.Error.UnexpectedErrorDuringImport"), e);
exitJVM(2);
}
log.logMinimal(BaseMessages.getString(PKG, "Import.Log.Finished"));
stop = new Date();
String begin = df.format(start);
String end = df.format(stop);
log.logMinimal(BaseMessages.getString(PKG, "Import.Log.StartStop", begin, end));
long seconds = (stop.getTime() - start.getTime()) / 1000;
if (seconds <= 60) {
log.logMinimal(BaseMessages.getString(PKG, "Import.Log.ProcessEndAfter", String.valueOf(seconds)));
} else if (seconds <= 60 * 60) {
int min = (int) (seconds / 60);
int rem = (int) (seconds % 60);
log.logMinimal(BaseMessages.getString(PKG, "Import.Log.ProcessEndAfterLong", String.valueOf(min), String.valueOf(rem), String.valueOf(seconds)));
} else if (seconds <= 60 * 60 * 24) {
int rem;
int hour = (int) (seconds / (60 * 60));
rem = (int) (seconds % (60 * 60));
int min = rem / 60;
rem = rem % 60;
log.logMinimal(BaseMessages.getString(PKG, "Import.Log.ProcessEndAfterLonger", String.valueOf(hour), String.valueOf(min), String.valueOf(rem), String.valueOf(seconds)));
} else {
int rem;
int days = (int) (seconds / (60 * 60 * 24));
rem = (int) (seconds % (60 * 60 * 24));
int hour = rem / (60 * 60);
rem = rem % (60 * 60);
int min = rem / 60;
rem = rem % 60;
log.logMinimal(BaseMessages.getString(PKG, "Import.Log.ProcessEndAfterLongest", String.valueOf(days), String.valueOf(hour), String.valueOf(min), String.valueOf(rem), String.valueOf(seconds)));
}
exitJVM(returnCode);
}
use of org.pentaho.di.repository.Repository in project pentaho-kettle by pentaho.
the class JobTest method testTwoJobsGetDifferentLogChannelIdWithDifferentCarteId.
/**
* This test demonstrates the fix for PDI-17398.
* Two schedules -> two Carte object Ids -> two log channel Ids
*/
@Test
public void testTwoJobsGetDifferentLogChannelIdWithDifferentCarteId() {
Repository repository = mock(Repository.class);
JobMeta meta1 = mock(JobMeta.class);
JobMeta meta2 = mock(JobMeta.class);
String carteId1 = UUID.randomUUID().toString();
String carteId2 = UUID.randomUUID().toString();
doReturn(carteId1).when(meta1).getContainerObjectId();
doReturn(carteId2).when(meta2).getContainerObjectId();
Job job1 = new Job(repository, meta1);
Job job2 = new Job(repository, meta2);
assertNotEquals(job1.getContainerObjectId(), job2.getContainerObjectId());
assertNotEquals(job1.getLogChannelId(), job2.getLogChannelId());
}
use of org.pentaho.di.repository.Repository in project pentaho-kettle by pentaho.
the class JobTest method testNewJobWithContainerObjectId.
@Test
public void testNewJobWithContainerObjectId() {
Repository repository = mock(Repository.class);
JobMeta meta = mock(JobMeta.class);
String carteId = UUID.randomUUID().toString();
doReturn(carteId).when(meta).getContainerObjectId();
Job job = new Job(repository, meta);
assertEquals(carteId, job.getContainerObjectId());
}
use of org.pentaho.di.repository.Repository in project pentaho-kettle by pentaho.
the class JobTest method testTwoJobsGetSameLogChannelId.
/**
* This test demonstrates the issue fixed in PDI-17398.
* When a job is scheduled twice, it gets the same log channel Id and both logs get merged
*/
@Test
public void testTwoJobsGetSameLogChannelId() {
Repository repository = mock(Repository.class);
JobMeta meta = mock(JobMeta.class);
Job job1 = new Job(repository, meta);
Job job2 = new Job(repository, meta);
assertEquals(job1.getLogChannelId(), job2.getLogChannelId());
}
Aggregations