use of org.apache.asterix.external.util.FileSystemWatcher in project asterixdb by apache.
the class RecordWithMetaTest method runTest.
@SuppressWarnings({ "unchecked", "rawtypes" })
public // @Test commented out due to ASTERIXDB-1881
void runTest() throws Exception {
File file = new File("target/beer.adm");
File expected = new File(getClass().getResource("/openbeerdb/beer.txt").toURI().getPath());
try {
FileUtils.deleteQuietly(file);
PrintStream printStream = new PrintStream(Files.newOutputStream(Paths.get(file.toURI())));
// create key type
IAType[] keyTypes = { BuiltinType.ASTRING };
String keyName = "id";
List<String> keyNameAsList = new ArrayList<>(1);
keyNameAsList.add(keyName);
// create record type
String[] recordFieldNames = {};
IAType[] recordFieldTypes = {};
recordType = new ARecordType("value", recordFieldNames, recordFieldTypes, true);
// create the meta type
String[] metaFieldNames = { keyName, "flags", "expiration", "cas", "rev", "vbid", "dtype" };
IAType[] metaFieldTypes = { BuiltinType.ASTRING, BuiltinType.AINT32, BuiltinType.AINT64, BuiltinType.AINT64, BuiltinType.AINT32, BuiltinType.AINT32, BuiltinType.AINT32 };
ARecordType metaType = new ARecordType("meta", metaFieldNames, metaFieldTypes, true);
int valueIndex = 4;
char delimiter = ',';
int numOfTupleFields = 3;
int[] pkIndexes = { 0 };
int[] pkIndicators = { 1 };
List<Path> paths = new ArrayList<>();
paths.add(Paths.get(getClass().getResource("/openbeerdb/beer.csv").toURI()));
FileSystemWatcher watcher = new FileSystemWatcher(paths, null, false);
// create input stream
LocalFSInputStream inputStream = new LocalFSInputStream(watcher);
// create reader record reader
Map<String, String> config = new HashMap<>();
config.put(ExternalDataConstants.KEY_HEADER, "true");
config.put(ExternalDataConstants.KEY_QUOTE, ExternalDataConstants.DEFAULT_QUOTE);
LineRecordReader lineReader = new LineRecordReader();
lineReader.configure(inputStream, config);
// create csv with json record reader
CSVToRecordWithMetadataAndPKConverter recordConverter = new CSVToRecordWithMetadataAndPKConverter(valueIndex, delimiter, metaType, recordType, pkIndicators, pkIndexes, keyTypes);
// create the value parser <ADM in this case>
ADMDataParser valueParser = new ADMDataParser(recordType, false);
// create parser.
RecordWithMetadataParser parser = new RecordWithMetadataParser(metaType, valueParser, recordConverter);
// create serializer deserializer and printer factories
ISerializerDeserializer[] serdes = new ISerializerDeserializer[keyTypes.length + 2];
IPrinterFactory[] printerFactories = new IPrinterFactory[keyTypes.length + 2];
for (int i = 0; i < keyTypes.length; i++) {
serdes[i + 2] = SerializerDeserializerProvider.INSTANCE.getSerializerDeserializer(keyTypes[i]);
printerFactories[i + 2] = ADMPrinterFactoryProvider.INSTANCE.getPrinterFactory(keyTypes[i]);
}
serdes[0] = SerializerDeserializerProvider.INSTANCE.getSerializerDeserializer(recordType);
serdes[1] = SerializerDeserializerProvider.INSTANCE.getSerializerDeserializer(metaType);
printerFactories[0] = ADMPrinterFactoryProvider.INSTANCE.getPrinterFactory(recordType);
printerFactories[1] = ADMPrinterFactoryProvider.INSTANCE.getPrinterFactory(metaType);
// create output descriptor
IPrinter[] printers = new IPrinter[printerFactories.length];
for (int i = 0; i < printerFactories.length; i++) {
printers[i] = printerFactories[i].createPrinter();
}
ArrayTupleBuilder tb = new ArrayTupleBuilder(numOfTupleFields);
while (lineReader.hasNext()) {
IRawRecord<char[]> record = lineReader.next();
tb.reset();
parser.parse(record, tb.getDataOutput());
tb.addFieldEndOffset();
parser.parseMeta(tb.getDataOutput());
tb.addFieldEndOffset();
parser.appendLastParsedPrimaryKeyToTuple(tb);
//print tuple
printTuple(tb, printers, printStream);
}
lineReader.close();
printStream.close();
Assert.assertTrue(FileUtils.contentEquals(file, expected));
} catch (Throwable th) {
System.err.println("TEST FAILED");
th.printStackTrace();
throw th;
} finally {
FileUtils.deleteQuietly(file);
}
System.err.println("TEST PASSED.");
}
Aggregations