use of org.apache.hadoop.io.VIntWritable in project SQLWindowing by hbutani.
the class PartitionedByteBasedSortedMapTest method testGeti.
@Test
public void testGeti() throws Exception {
int i = 0;
for (i = table.length - 1; i >= 0; i--) {
bm.put(new Text(table[i]), new VIntWritable(i));
}
Assert.assertEquals(table.length, bm.size());
for (i = 0; i < table.length; i++) {
bm.getKey(i, wObj);
Assert.assertEquals(table[i], wObj.toString());
bm.getValue(wObj, value);
Assert.assertEquals(value.get(), i);
}
}
use of org.apache.hadoop.io.VIntWritable in project SQLWindowing by hbutani.
the class PersistentByteBasedSortedMapTest method testEntryItr.
@Test
public void testEntryItr() throws Exception {
int i = 0;
for (i = table.length - 1; i >= 0; i--) {
bm.put(new Text(table[i]), new VIntWritable(i));
}
File f = File.createTempFile("wdw", null);
f.deleteOnExit();
PersistentByteBasedSortedMap.store(bm, f);
bm = null;
PersistentByteBasedSortedMap pm = new PersistentByteBasedSortedMap(f, comparator);
Assert.assertEquals(table.length, pm.size());
Iterator<ByteBasedSortedMap.WritableEntry> it = pm.entryIterator(wEntry);
i = 0;
while (it.hasNext()) {
ByteBasedSortedMap.WritableEntry e = it.next();
Assert.assertEquals(table[i], e.getKey().toString());
Assert.assertEquals(((VIntWritable) e.getValue()).get(), i);
i++;
}
}
use of org.apache.hadoop.io.VIntWritable in project SQLWindowing by hbutani.
the class PersistentByteBasedSortedMapTest method testGeti.
@Test
public void testGeti() throws Exception {
int i = 0;
for (i = table.length - 1; i >= 0; i--) {
bm.put(new Text(table[i]), new VIntWritable(i));
}
File f = File.createTempFile("wdw", null);
f.deleteOnExit();
PersistentByteBasedSortedMap.store(bm, f);
bm = null;
PersistentByteBasedSortedMap pm = new PersistentByteBasedSortedMap(f, comparator);
Assert.assertEquals(table.length, pm.size());
for (i = 0; i < table.length; i++) {
pm.getKey(i, wObj);
Assert.assertEquals(table[i], wObj.toString());
pm.getValue(wObj, value);
Assert.assertEquals(value.get(), i);
}
}
use of org.apache.hadoop.io.VIntWritable in project hive by apache.
the class ReaderWriter method readDatum.
public static Object readDatum(DataInput in) throws IOException {
byte type = in.readByte();
switch(type) {
case DataType.STRING:
byte[] buffer = new byte[in.readInt()];
in.readFully(buffer);
return new String(buffer, UTF8);
case DataType.INTEGER:
VIntWritable vint = new VIntWritable();
vint.readFields(in);
return vint.get();
case DataType.LONG:
VLongWritable vlong = new VLongWritable();
vlong.readFields(in);
return vlong.get();
case DataType.FLOAT:
return in.readFloat();
case DataType.DOUBLE:
return in.readDouble();
case DataType.BOOLEAN:
return in.readBoolean();
case DataType.BYTE:
return in.readByte();
case DataType.SHORT:
return in.readShort();
case DataType.NULL:
return null;
case DataType.BINARY:
int len = in.readInt();
byte[] ba = new byte[len];
in.readFully(ba);
return ba;
case DataType.MAP:
int size = in.readInt();
Map<Object, Object> m = new HashMap<Object, Object>(size);
for (int i = 0; i < size; i++) {
m.put(readDatum(in), readDatum(in));
}
return m;
case DataType.LIST:
int sz = in.readInt();
List<Object> list = new ArrayList<Object>(sz);
for (int i = 0; i < sz; i++) {
list.add(readDatum(in));
}
return list;
case DataType.CHAR:
HiveCharWritable hcw = new HiveCharWritable();
hcw.readFields(in);
return hcw.getHiveChar();
case DataType.VARCHAR:
HiveVarcharWritable hvw = new HiveVarcharWritable();
hvw.readFields(in);
return hvw.getHiveVarchar();
case DataType.DECIMAL:
HiveDecimalWritable hdw = new HiveDecimalWritable();
hdw.readFields(in);
return hdw.getHiveDecimal();
case DataType.DATE:
DateWritable dw = new DateWritable();
dw.readFields(in);
return dw.get();
case DataType.TIMESTAMP:
TimestampWritable tw = new TimestampWritable();
tw.readFields(in);
return tw.getTimestamp();
default:
throw new IOException("Unexpected data type " + type + " found in stream.");
}
}
Aggregations