HadoopHudiColumnValue.java

// Licensed to the Apache Software Foundation (ASF) under one
// or more contributor license agreements.  See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership.  The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use this file except in compliance
// with the License.  You may obtain a copy of the License at
//
//   http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing,
// software distributed under the License is distributed on an
// "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
// KIND, either express or implied.  See the License for the
// specific language governing permissions and limitations
// under the License.

package org.apache.doris.hudi;

import org.apache.doris.common.jni.vec.ColumnType;
import org.apache.doris.common.jni.vec.ColumnValue;

import org.apache.hadoop.hive.common.type.HiveDecimal;
import org.apache.hadoop.hive.serde2.io.TimestampWritableV2;
import org.apache.hadoop.hive.serde2.objectinspector.ListObjectInspector;
import org.apache.hadoop.hive.serde2.objectinspector.MapObjectInspector;
import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector;
import org.apache.hadoop.hive.serde2.objectinspector.PrimitiveObjectInspector;
import org.apache.hadoop.hive.serde2.objectinspector.StructField;
import org.apache.hadoop.hive.serde2.objectinspector.StructObjectInspector;
import org.apache.hadoop.hive.serde2.objectinspector.primitive.DateObjectInspector;
import org.apache.hadoop.hive.serde2.objectinspector.primitive.TimestampObjectInspector;
import org.apache.hadoop.io.LongWritable;

import java.math.BigDecimal;
import java.math.BigInteger;
import java.sql.Timestamp;
import java.time.Instant;
import java.time.LocalDate;
import java.time.LocalDateTime;
import java.time.ZoneId;
import java.util.List;
import java.util.Map;

public class HadoopHudiColumnValue implements ColumnValue {
    private ColumnType dorisType;
    private ObjectInspector fieldInspector;
    private Object fieldData;
    private final ZoneId zoneId;

    public HadoopHudiColumnValue(ZoneId zoneId) {
        this.zoneId = zoneId;
    }

    public void setRow(Object record) {
        this.fieldData = record;
    }

    public void setField(ColumnType dorisType, ObjectInspector fieldInspector) {
        this.dorisType = dorisType;
        this.fieldInspector = fieldInspector;
    }

    private Object inspectObject() {
        return ((PrimitiveObjectInspector) fieldInspector).getPrimitiveJavaObject(fieldData);
    }

    @Override
    public boolean getBoolean() {
        return (boolean) inspectObject();
    }

    @Override
    public short getShort() {
        return (short) inspectObject();
    }

    @Override
    public int getInt() {
        return (int) inspectObject();
    }

    @Override
    public float getFloat() {
        return (float) inspectObject();
    }

    @Override
    public long getLong() {
        return (long) inspectObject();
    }

    @Override
    public double getDouble() {
        return (double) inspectObject();
    }

    @Override
    public String getString() {
        return inspectObject().toString();
    }

    @Override
    public byte[] getBytes() {
        return (byte[]) inspectObject();
    }


    @Override
    public byte getByte() {
        throw new UnsupportedOperationException("Hoodie type does not support tinyint");
    }

    @Override
    public BigDecimal getDecimal() {
        return ((HiveDecimal) inspectObject()).bigDecimalValue();
    }

    @Override
    public LocalDate getDate() {
        return LocalDate.ofEpochDay((((DateObjectInspector) fieldInspector).getPrimitiveJavaObject(fieldData))
                .toEpochDay());
    }

    @Override
    public LocalDateTime getDateTime() {
        if (fieldData instanceof Timestamp) {
            return ((Timestamp) fieldData).toLocalDateTime();
        } else if (fieldData instanceof TimestampWritableV2) {
            return LocalDateTime.ofInstant(Instant.ofEpochSecond((((TimestampObjectInspector) fieldInspector)
                    .getPrimitiveJavaObject(fieldData)).toEpochSecond()), zoneId);
        } else {
            long datetime = ((LongWritable) fieldData).get();
            long seconds;
            long nanoseconds;
            if (dorisType.getPrecision() == 3) {
                seconds = datetime / 1000;
                nanoseconds = (datetime % 1000) * 1000000;
            } else if (dorisType.getPrecision() == 6) {
                seconds = datetime / 1000000;
                nanoseconds = (datetime % 1000000) * 1000;
            } else {
                throw new RuntimeException("Hoodie timestamp only support milliseconds and microseconds, "
                        + "wrong precision = " + dorisType.getPrecision());
            }
            return LocalDateTime.ofInstant(Instant.ofEpochSecond(seconds, nanoseconds), zoneId);
        }
    }

    @Override
    public boolean canGetStringAsBytes() {
        return false;
    }

    @Override
    public boolean isNull() {
        return fieldData == null;
    }

    @Override
    public BigInteger getBigInteger() {
        throw new UnsupportedOperationException("Hoodie type does not support largeint");
    }

    @Override
    public byte[] getStringAsBytes() {
        throw new UnsupportedOperationException("Hoodie type does not support getStringAsBytes");
    }

    @Override
    public void unpackArray(List<ColumnValue> values) {
        ListObjectInspector inspector = (ListObjectInspector) fieldInspector;
        List<?> items = inspector.getList(fieldData);
        ObjectInspector itemInspector = inspector.getListElementObjectInspector();
        for (int i = 0; i < items.size(); i++) {
            Object item = items.get(i);
            HadoopHudiColumnValue childValue = new HadoopHudiColumnValue(zoneId);
            childValue.setRow(item);
            childValue.setField(dorisType.getChildTypes().get(0), itemInspector);
            values.add(childValue);
        }
    }

    @Override
    public void unpackMap(List<ColumnValue> keys, List<ColumnValue> values) {
        MapObjectInspector inspector = (MapObjectInspector) fieldInspector;
        ObjectInspector keyObjectInspector = inspector.getMapKeyObjectInspector();
        ObjectInspector valueObjectInspector = inspector.getMapValueObjectInspector();
        for (Map.Entry kv : inspector.getMap(fieldData).entrySet()) {
            HadoopHudiColumnValue key = new HadoopHudiColumnValue(zoneId);
            key.setRow(kv.getKey());
            key.setField(dorisType.getChildTypes().get(0), keyObjectInspector);
            keys.add(key);

            HadoopHudiColumnValue value = new HadoopHudiColumnValue(zoneId);
            value.setRow(kv.getValue());
            value.setField(dorisType.getChildTypes().get(1), valueObjectInspector);
            values.add(value);
        }
    }

    @Override
    public void unpackStruct(List<Integer> structFieldIndex, List<ColumnValue> values) {
        StructObjectInspector inspector = (StructObjectInspector) fieldInspector;
        List<? extends StructField> fields = inspector.getAllStructFieldRefs();
        for (int i = 0; i < structFieldIndex.size(); i++) {
            Integer idx = structFieldIndex.get(i);
            HadoopHudiColumnValue value = new HadoopHudiColumnValue(zoneId);
            Object obj = null;
            if (idx != null) {
                StructField sf = fields.get(idx);
                obj = inspector.getStructFieldData(fieldData, sf);
            }
            value.setRow(obj);
            value.setField(dorisType.getChildTypes().get(i), fields.get(i).getFieldObjectInspector());
            values.add(value);
        }
    }
}