use of org.apache.hadoop.hive.serde2.io.DateWritableV2 in project hive by apache.
the class TestGenericUDFDateAdd method testTimestampToDate.
@Test
public void testTimestampToDate() throws HiveException {
GenericUDFDateAdd udf = new GenericUDFDateAdd();
ObjectInspector valueOI1 = PrimitiveObjectInspectorFactory.writableTimestampObjectInspector;
ObjectInspector valueOI2 = PrimitiveObjectInspectorFactory.javaIntObjectInspector;
ObjectInspector[] arguments = { valueOI1, valueOI2 };
udf.initialize(arguments);
DeferredObject valueObj1 = new DeferredJavaObject(new TimestampWritableV2(Timestamp.valueOf(LocalDateTime.of(109, 06, 20, 4, 17, 52, 0).toString())));
DeferredObject valueObj2 = new DeferredJavaObject(Integer.valueOf("3"));
DeferredObject[] args = { valueObj1, valueObj2 };
DateWritableV2 output = (DateWritableV2) udf.evaluate(args);
assertEquals("date_add() test for TIMESTAMP failed ", "0109-06-23", output.toString());
// Test with null args
args = new DeferredObject[] { new DeferredJavaObject(null), valueObj2 };
assertNull("date_add() 1st arg null", udf.evaluate(args));
args = new DeferredObject[] { valueObj1, new DeferredJavaObject(null) };
assertNull("date_add() 2nd arg null", udf.evaluate(args));
args = new DeferredObject[] { new DeferredJavaObject(null), new DeferredJavaObject(null) };
assertNull("date_add() both args null", udf.evaluate(args));
}
use of org.apache.hadoop.hive.serde2.io.DateWritableV2 in project hive by apache.
the class TestGenericUDFDateAdd method testStringToDate.
@Test
public void testStringToDate() throws HiveException {
GenericUDFDateAdd udf = new GenericUDFDateAdd();
ObjectInspector valueOI1 = PrimitiveObjectInspectorFactory.javaStringObjectInspector;
ObjectInspector valueOI2 = PrimitiveObjectInspectorFactory.javaIntObjectInspector;
ObjectInspector[] arguments = { valueOI1, valueOI2 };
udf.initialize(arguments);
DeferredObject valueObj1 = new DeferredJavaObject(new Text("2009-07-20 04:17:52"));
DeferredObject valueObj2 = new DeferredJavaObject(Integer.valueOf("2"));
DeferredObject[] args = { valueObj1, valueObj2 };
DateWritableV2 output = (DateWritableV2) udf.evaluate(args);
assertEquals("date_add() test for STRING failed ", "2009-07-22", output.toString());
// Test with null args
args = new DeferredObject[] { new DeferredJavaObject(null), valueObj2 };
assertNull("date_add() 1st arg null", udf.evaluate(args));
args = new DeferredObject[] { valueObj1, new DeferredJavaObject(null) };
assertNull("date_add() 2nd arg null", udf.evaluate(args));
args = new DeferredObject[] { new DeferredJavaObject(null), new DeferredJavaObject(null) };
assertNull("date_add() both args null", udf.evaluate(args));
}
use of org.apache.hadoop.hive.serde2.io.DateWritableV2 in project hive by apache.
the class TestGenericUDFDateAdd method testShortDataTypeAsDays.
@Test
public void testShortDataTypeAsDays() throws HiveException {
GenericUDFDateAdd udf = new GenericUDFDateAdd();
ObjectInspector valueOI1 = PrimitiveObjectInspectorFactory.writableDateObjectInspector;
ObjectInspector valueOI2 = PrimitiveObjectInspectorFactory.javaShortObjectInspector;
ObjectInspector[] arguments = { valueOI1, valueOI2 };
udf.initialize(arguments);
DeferredObject valueObj1 = new DeferredJavaObject(new DateWritableV2(Date.of(109, 06, 20)));
DeferredObject valueObj2 = new DeferredJavaObject(Short.valueOf("4"));
DeferredObject[] args = { valueObj1, valueObj2 };
DateWritableV2 output = (DateWritableV2) udf.evaluate(args);
assertEquals("date_add() test for SHORT failed ", "0109-06-24", output.toString());
}
use of org.apache.hadoop.hive.serde2.io.DateWritableV2 in project hive by apache.
the class TestGenericUDFDateAdd method testByteDataTypeAsDays.
@Test
public void testByteDataTypeAsDays() throws HiveException {
GenericUDFDateAdd udf = new GenericUDFDateAdd();
ObjectInspector valueOI1 = PrimitiveObjectInspectorFactory.writableDateObjectInspector;
ObjectInspector valueOI2 = PrimitiveObjectInspectorFactory.javaByteObjectInspector;
ObjectInspector[] arguments = { valueOI1, valueOI2 };
udf.initialize(arguments);
DeferredObject valueObj1 = new DeferredJavaObject(new DateWritableV2(Date.of(109, 06, 20)));
DeferredObject valueObj2 = new DeferredJavaObject(Byte.valueOf("4"));
DeferredObject[] args = { valueObj1, valueObj2 };
DateWritableV2 output = (DateWritableV2) udf.evaluate(args);
assertEquals("date_add() test for BYTE failed ", "0109-06-24", output.toString());
}
use of org.apache.hadoop.hive.serde2.io.DateWritableV2 in project hive by apache.
the class VectorUDFDateDiffScalarCol method evaluate.
@Override
public void evaluate(VectorizedRowBatch batch) throws HiveException {
if (childExpressions != null) {
super.evaluateChildren(batch);
}
LongColumnVector outputColVector = (LongColumnVector) batch.cols[outputColumnNum];
ColumnVector inputCol = batch.cols[this.inputColumnNum[0]];
/* every line below this is identical for evaluateLong & evaluateString */
final int n = inputCol.isRepeating ? 1 : batch.size;
int[] sel = batch.selected;
final boolean selectedInUse = (inputCol.isRepeating == false) && batch.selectedInUse;
boolean[] outputIsNull = outputColVector.isNull;
if (batch.size == 0) {
/* n != batch.size when isRepeating */
return;
}
// We do not need to do a column reset since we are carefully changing the output.
outputColVector.isRepeating = false;
PrimitiveCategory primitiveCategory0 = ((PrimitiveTypeInfo) inputTypeInfos[0]).getPrimitiveCategory();
switch(primitiveCategory0) {
case DATE:
baseDate = (int) longValue;
break;
case TIMESTAMP:
date.setTime(timestampValue.getTime());
baseDate = DateWritableV2.dateToDays(date);
break;
case STRING:
case CHAR:
case VARCHAR:
try {
org.apache.hadoop.hive.common.type.Date hiveDate = org.apache.hadoop.hive.common.type.Date.valueOf(new String(stringValue, StandardCharsets.UTF_8));
baseDate = hiveDate.toEpochDay();
break;
} catch (Exception e) {
outputColVector.noNulls = false;
if (selectedInUse) {
for (int j = 0; j < n; j++) {
int i = sel[j];
outputColVector.isNull[i] = true;
}
} else {
for (int i = 0; i < n; i++) {
outputColVector.isNull[i] = true;
}
}
return;
}
default:
throw new Error("Unsupported input type " + primitiveCategory0.name());
}
PrimitiveCategory primitiveCategory1 = ((PrimitiveTypeInfo) inputTypeInfos[1]).getPrimitiveCategory();
switch(primitiveCategory1) {
case DATE:
if (inputCol.isRepeating) {
if (inputCol.noNulls || !inputCol.isNull[0]) {
outputColVector.isNull[0] = false;
outputColVector.vector[0] = evaluateDate(inputCol, 0);
} else {
outputColVector.isNull[0] = true;
outputColVector.noNulls = false;
}
outputColVector.isRepeating = true;
} else if (inputCol.noNulls) {
if (batch.selectedInUse) {
if (!outputColVector.noNulls) {
for (int j = 0; j != n; j++) {
final int i = sel[j];
// Set isNull before call in case it changes it mind.
outputIsNull[i] = false;
outputColVector.vector[i] = evaluateDate(inputCol, i);
}
} else {
for (int j = 0; j != n; j++) {
final int i = sel[j];
outputColVector.vector[i] = evaluateDate(inputCol, i);
}
}
} else {
if (!outputColVector.noNulls) {
// Assume it is almost always a performance win to fill all of isNull so we can
// safely reset noNulls.
Arrays.fill(outputIsNull, false);
outputColVector.noNulls = true;
}
for (int i = 0; i != n; i++) {
outputColVector.vector[i] = evaluateDate(inputCol, i);
}
}
} else /* there are NULLs in the inputColVector */
{
// Carefully handle NULLs..
// Handle case with nulls. Don't do function if the value is null, to save time,
// because calling the function can be expensive.
outputColVector.noNulls = false;
if (selectedInUse) {
for (int j = 0; j < n; j++) {
int i = sel[j];
outputColVector.isNull[i] = inputCol.isNull[i];
if (!inputCol.isNull[i]) {
outputColVector.vector[i] = evaluateDate(inputCol, i);
}
}
} else {
for (int i = 0; i < n; i++) {
outputColVector.isNull[i] = inputCol.isNull[i];
if (!inputCol.isNull[i]) {
outputColVector.vector[i] = evaluateDate(inputCol, i);
}
}
}
}
break;
case TIMESTAMP:
if (inputCol.isRepeating) {
if (inputCol.noNulls || !inputCol.isNull[0]) {
outputColVector.isNull[0] = false;
outputColVector.vector[0] = evaluateTimestamp(inputCol, 0);
} else {
outputColVector.isNull[0] = true;
outputColVector.noNulls = false;
}
outputColVector.isRepeating = true;
} else if (inputCol.noNulls) {
if (batch.selectedInUse) {
if (!outputColVector.noNulls) {
for (int j = 0; j != n; j++) {
final int i = sel[j];
// Set isNull before call in case it changes it mind.
outputIsNull[i] = false;
outputColVector.vector[i] = evaluateTimestamp(inputCol, i);
}
} else {
for (int j = 0; j != n; j++) {
final int i = sel[j];
outputColVector.vector[i] = evaluateTimestamp(inputCol, i);
}
}
} else {
if (!outputColVector.noNulls) {
// Assume it is almost always a performance win to fill all of isNull so we can
// safely reset noNulls.
Arrays.fill(outputIsNull, false);
outputColVector.noNulls = true;
}
for (int i = 0; i != n; i++) {
outputColVector.vector[i] = evaluateTimestamp(inputCol, i);
}
}
} else /* there are nulls in the inputColVector */
{
// Carefully handle NULLs..
// Handle case with nulls. Don't do function if the value is null, to save time,
// because calling the function can be expensive.
outputColVector.noNulls = false;
if (selectedInUse) {
for (int j = 0; j < n; j++) {
int i = sel[j];
outputColVector.isNull[i] = inputCol.isNull[i];
if (!inputCol.isNull[i]) {
outputColVector.vector[i] = evaluateTimestamp(inputCol, i);
}
}
} else {
for (int i = 0; i < n; i++) {
outputColVector.isNull[i] = inputCol.isNull[i];
if (!inputCol.isNull[i]) {
outputColVector.vector[i] = evaluateTimestamp(inputCol, i);
}
}
}
}
break;
case STRING:
case CHAR:
case VARCHAR:
if (inputCol.isRepeating) {
if (inputCol.noNulls || !inputCol.isNull[0]) {
outputColVector.isNull[0] = false;
evaluateString(inputCol, outputColVector, 0);
} else {
outputColVector.isNull[0] = true;
outputColVector.noNulls = false;
}
outputColVector.isRepeating = true;
} else if (inputCol.noNulls) {
if (batch.selectedInUse) {
if (!outputColVector.noNulls) {
for (int j = 0; j != n; j++) {
final int i = sel[j];
// Set isNull before call in case it changes it mind.
outputIsNull[i] = false;
evaluateString(inputCol, outputColVector, i);
}
} else {
for (int j = 0; j != n; j++) {
final int i = sel[j];
evaluateString(inputCol, outputColVector, i);
}
}
} else {
if (!outputColVector.noNulls) {
// Assume it is almost always a performance win to fill all of isNull so we can
// safely reset noNulls.
Arrays.fill(outputIsNull, false);
outputColVector.noNulls = true;
}
for (int i = 0; i != n; i++) {
evaluateString(inputCol, outputColVector, i);
}
}
} else /* there are nulls in the inputColVector */
{
// Carefully handle NULLs..
// Handle case with nulls. Don't do function if the value is null, to save time,
// because calling the function can be expensive.
outputColVector.noNulls = false;
if (selectedInUse) {
for (int j = 0; j < n; j++) {
int i = sel[j];
outputColVector.isNull[i] = inputCol.isNull[i];
if (!inputCol.isNull[i]) {
evaluateString(inputCol, outputColVector, i);
}
}
} else {
for (int i = 0; i < n; i++) {
outputColVector.isNull[i] = inputCol.isNull[i];
if (!inputCol.isNull[i]) {
evaluateString(inputCol, outputColVector, i);
}
}
}
}
break;
default:
throw new Error("Unsupported input type " + primitiveCategory1.name());
}
}
Aggregations