Starting from version 2.1.3, Doris supports DDL and DML operations for Hive. Users can directly create databases and tables in Hive through Doris and write data into Hive tables. With this feature, users can perform complete data queries and write operations on Hive through Doris, further helping to simplify the data lake integrated architecture.

This article introduces Hive operations supported in Doris, including syntax and usage notes.

Hive - 图1tip

This is an experimental feature.

Hive - 图2tip

Before using, please set:

set global enable_nereids_planner = true;

set global enable_fallback_to_original_planner = false;

For clusters upgraded from old versions, these variables may change.

Metadata Creation and Deletion

Catalog

  • Create

    1. CREATE CATALOG [IF NOT EXISTS] hive PROPERTIES (
    2. "type"="hms",
    3. "hive.metastore.uris" = "thrift://172.21.16.47:7004",
    4. "hadoop.username" = "hadoop",
    5. "fs.defaultFS" = "hdfs://172.21.16.47:4007"
    6. );

    Note, if you need to create Hive tables or write data through Doris, you must explicitly include the fs.defaultFS property in the Catalog properties. If creating the Catalog is only for querying, this parameter can be omitted.

    For more parameters, please refer to Hive Catalog

  • Drop

    1. DROP CATALOG [IF EXISTS] hive;

    Deleting a Catalog does not delete any database or table information in Hive. It merely removes the mapping to this Hive cluster in Doris.

Database

  • Create

    You can switch to the corresponding Catalog and execute the CREATE DATABASE statement:

    1. SWITCH hive;
    2. CREATE DATABASE [IF NOT EXISTS] hive_db;

    You can also create using the fully qualified name or specify the location, as:

    1. CREATE DATABASE [IF NOT EXISTS] hive.hive_db;
    2. CREATE DATABASE [IF NOT EXISTS] hive.hive_db
    3. PROPERTIES ('location'='hdfs://172.21.16.47:4007/path/to/db/');

    Later, you can view the Database’s Location information using the SHOW CREATE DATABASE command:

    1. mysql> SHOW CREATE DATABASE hive_db;
    2. +----------+---------------------------------------------------------------------------------------------+
    3. | Database | Create Database |
    4. +----------+---------------------------------------------------------------------------------------------+
    5. | hive_db | CREATE DATABASE `hive_db` LOCATION 'hdfs://172.21.16.47:4007/usr/hive/warehouse/hive_db.db' |
    6. +----------+---------------------------------------------------------------------------------------------+
  • Drop

    1. DROP DATABASE [IF EXISTS] hive.hive_db;

    Note that for Hive Databases, all tables within the Database must be deleted first, otherwise an error will occur. This operation will also delete the corresponding Database in Hive.

Table

  • Create

    Doris supports creating partitioned or non-partitioned tables in Hive.

    1. -- Create unpartitioned hive table
    2. CREATE TABLE unpartitioned_table (
    3. `col1` BOOLEAN COMMENT 'col1',
    4. `col2` INT COMMENT 'col2',
    5. `col3` BIGINT COMMENT 'col3',
    6. `col4` CHAR(10) COMMENT 'col4',
    7. `col5` FLOAT COMMENT 'col5',
    8. `col6` DOUBLE COMMENT 'col6',
    9. `col7` DECIMAL(9,4) COMMENT 'col7',
    10. `col8` VARCHAR(11) COMMENT 'col8',
    11. `col9` STRING COMMENT 'col9'
    12. ) ENGINE=hive
    13. PROPERTIES (
    14. 'file_format'='parquet'
    15. );
    16. -- Create partitioned hive table
    17. -- The partition columns must be in table's column definition list
    18. CREATE TABLE partition_table (
    19. `col1` BOOLEAN COMMENT 'col1',
    20. `col2` INT COMMENT 'col2',
    21. `col3` BIGINT COMMENT 'col3',
    22. `col4` DECIMAL(2,1) COMMENT 'col4',
    23. `pt1` VARCHAR COMMENT 'pt1',
    24. `pt2` VARCHAR COMMENT 'pt2'
    25. ) ENGINE=hive
    26. PARTITION BY LIST (pt1, pt2) ()
    27. PROPERTIES (
    28. 'file_format'='orc',
    29. 'compression'='zlib'
    30. );

    After creation, you can view the Hive table creation statement using the SHOW CREATE TABLE command.

    Note, unlike Hive’s table creation statements. In Doris, when creating a Hive partitioned table, the partition columns must also be included in the Table’s Schema. At the same time, the partition columns must be at the end of all schemas and in the same order.

    :::tip

    For some Hive clusters that enable ACID transaction features by default, after using Doris to create a table, the table attribute transactional will be true. However, Doris only supports some features of Hive transaction tables, which may cause the problem that Doris itself cannot read the Hive created by Doris. Therefore, it is necessary to explicitly add: "transactional" = "false" in the table creation properties to create a non-transactional Hive table:

    1. CREATE TABLE non_acid_table(
    2. `col1` BOOLEAN COMMENT 'col1',
    3. `col2` INT COMMENT 'col2',
    4. `col3` BIGINT COMMENT 'col3'
    5. ) ENGINE=hive
    6. PROPERTIES (
    7. 'transactional'='false',
    8. );

    :::

  • Drop

    You can drop a Hive table using the DROP TABLE statement. Currently, deleting the table also removes the data, including partition data.

  • Column Types

    The column types used when creating Hive tables in Doris correspond to those in Hive as follows:

    DorisHive
    BOOLEANBOOLEAN
    TINYINTTINYINT
    SMALLINTSMALLINT
    INTINT
    BIGINTBIGINT
    DATEDATE
    DATETIMETIMESTAMP
    FLOATFLOAT
    DOUBLEDOUBLE
    CHARCHAR
    VARCHARSTRING
    STRINGSTRING
    DECIMALDECIMAL
    ARRAYARRAY
    MAPMAP
    STRUCTSTRUCT
    • Column types can only be nullable by default, NOT NULL is not supported.

    • Hive 3.0 supports setting default values. If you need to set default values, you need to explicitly add "hive.version" = "3.0.0" in the Catalog properties.

    • After inserting data, if the types are not compatible, such as 'abc' being inserted into a numeric type, it will be converted to a null value before insertion.

  • Partitions

    The partition types in Hive correspond to the List partition in Doris. Therefore, when creating a Hive partitioned table in Doris, you need to use the List partition table creation statement, but there is no need to explicitly enumerate each partition. When writing data, Doris will automatically create the corresponding Hive partition based on the values of the data.

    Supports creating single-column or multi-column partitioned tables.

  • File Formats

    • Parquet
    • ORC (default)
  • Compression Formats

    • Parquet: snappy(default), zstd, plain. (plain means no compression is used.)
    • ORC: snappy, zlib(default), zstd, plain. (plain means no compression is used.)
  • Storage Medium

    • HDFS
    • Object Storage

Data Operations

Data can be written into Hive tables through INSERT statements.

Supports writing to Hive tables created by Doris or existing Hive tables with supported format.

For partitioned tables, data will automatically be written to the corresponding partition or new partitions will be created.

Currently, writing to specific partitions is not supported.

INSERT

The INSERT operation appends data to the target table.

  1. INSERT INTO hive_tbl values (val1, val2, val3, val4);
  2. INSERT INTO hive.hive_db.hive_tbl SELECT col1, col2 FROM internal.db1.tbl1;
  3. INSERT INTO hive_tbl(col1, col2) values (val1, val2);
  4. INSERT INTO hive_tbl(col1, col2, partition_col1, partition_col2) values (1, 2, "beijing", "2023-12-12");

INSERT OVERWRITE

The INSERT OVERWRITE operation completely overwrites the existing data in the table with new data.

  1. INSERT OVERWRITE TABLE VALUES(val1, val2, val3, val4)
  2. INSERT OVERWRITE TABLE hive.hive_db.hive_tbl(col1, col2) SELECT col1, col2 FROM internal.db1.tbl1;

CTAS (CREATE TABLE AS SELECT)

A Hive table can be created and populated with data using the CTAS (CREATE TABLE AS SELECT) statement:

  1. CREATE TABLE hive_ctas ENGINE=hive AS SELECT * FROM other_table;

CTAS supports specifying file formats, partitioning methods, and other information, such as:

  1. CREATE TABLE hive_ctas ENGINE=hive
  2. PARTITION BY LIST (pt1, pt2) ()
  3. AS SELECT col1,pt1,pt2 FROM part_ctas_src WHERE col1>0;
  4. CREATE TABLE hive.hive_db.hive_ctas (col1,col2,pt1) ENGINE=hive
  5. PARTITION BY LIST (pt1) ()
  6. PROPERTIES (
  7. "file_format"="parquet",
  8. "compression"="zstd"
  9. )
  10. AS SELECT col1,pt1 as col2,pt2 as pt1 FROM test_ctas.part_ctas_src WHERE col1>0;

Exception Data and Data Transformation

TODO

Transaction Mechanism

Write operations to Hive are placed in a separate transaction. Until the transaction is committed, the data is not visible externally. Only after committing the transaction do the table’s related operations become visible to others.

Transactions ensure the atomicity of operations—all operations within a transaction either succeed completely or fail altogether.

Transactions do not fully guarantee isolation of operations; they strive to minimize the inconsistency window by separating file system operations from metadata operations on the Hive Metastore.

For example, in a transaction involving multiple partition modifications of a Hive table, if the task is divided into two batches, and the first batch is completed but the second batch has not yet started, the partitions from the first batch are already visible externally, and can be read, but the second batch partitions cannot.

If any anomalies occur during the transaction commit process, the transaction will be directly rolled back, including modifications to HDFS files and metadata in the Hive Metastore, without requiring further action from the user.

Concurrent Writing Mechanism

Currently, Doris supports concurrent writing using multiple insert statements. However, it is important to note that users need to control concurrent writing to avoid potential conflicts.

As ordinary non-transactional Hive tables lack a complete transaction mechanism. From the Doris transaction mechanism described earlier, we know that the current implementation in Doris can only make efforts to minimize the possible inconsistency time window and cannot guarantee true ACID properties. Therefore, concurrent writing to Hive in Doris may lead to data consistency issues.

  1. INSERT Concurrent Operations

    INSERT is a data append operation. When INSERT is executed concurrently, it will not cause conflicts, and the operations will produce the expected results.

  2. INSERT OVERWRITE Concurrent Operations

    If INSERT OVERWRITE is used for concurrent writing to the same table or partition, it may lead to data loss or corruption, and the result may be uncertain.

    There are generally the following solutions:

    • For partitioned tables, data can be written into different partitions, and concurrent operations on different partitions will not cause conflicts.
    • For non-partitioned tables, INSERT can be executed simultaneously without using INSERT OVERWRITE, thus avoiding conflicts.
    • For potentially conflicting operations, users need to control on the business side to ensure that only one write operation is being performed at the same time.

HDFS File Operations

Data in Hive tables on HDFS is usually written first to a temporary directory, then operations like rename are used to commit the files finally. Here, we detail the specific operations on files in HDFS during different data operations.

The format of the temporary directory is: /tmp/.doris_staging/<username>/<uuid>

The format of the written data file names is: <query-id>_<uuid>-<index>.<compress-type>.<file-type>

Below, we describe the file operations in various cases.

  1. Non-partitioned table

    • Append

      • Target table directory: hdfs://ns/usr/hive/warehouse/example.db/table1

      • Temporary file: hdfs://ns/tmp/.doris_staging/root/f02247cb662846038baae272af5eeb05/b35fdbcea3a4e39-86d1f36987ef1492_7e3985bf-9de9-4fc7-b84e-adf11aa08756-0.orc

        During the commit phase, all temporary files are moved to the target table directory.

    • Overwrite

      • Target table directory: hdfs://ns/usr/hive/warehouse/example.db/table1

      • Temporary file: hdfs://ns/tmp/.doris_staging/root/f02247cb662846038baae272af5eeb05/b35fdbcea3a4e39-86d1f36987ef1492_7e3985bf-9de9-4fc7-b84e-adf11aa08756-0.orc

        Commit phase:

      1. The target table directory is renamed to a temporary target table directory: hdfs://ns/usr/hive/warehouse/example.db/_temp_b35fdbcea3a4e39-86d1f36987ef1492_table1
      2. The temporary directory is renamed to the target table directory.
      3. The temporary target table directory is deleted.
  2. Partitioned table

    • Add (Add to a new partition)

      • Target table directory: hdfs://ns/usr/hive/warehouse/example.db/table2/part_col=2024-01-01

      • Temporary file: hdfs://ns/tmp/.doris_staging/root/a7eac7505d7a42fdb06cb9ef1ea3e912/par1=a/d678a74d232345e0-b659e2fb58e86ffd_549ad677-ee75-4fa1-b8a6-3e821e1dae61-0.orc

        During the commit phase, the temporary directory is renamed to the target table directory.

    • Append (Write data to an existing partition)

      • Target table directory: hdfs://ns/usr/hive/warehouse/example.db/table2/part_col=2024-01-01

      • Temporary file: hdfs://ns/tmp/.doris_staging/root/a7eac7505d7a42fdb06cb9ef1ea3e912/par1=a/d678a74d232345e0-b659e2fb58e86ffd_549ad677-ee75-4fa1-b8a6-3e821e1dae61-0.orc

        During the commit phase, files from the temporary directory are moved to the target table directory.

    • Overwrite (Overwrite an existing partition)

      • Target table directory: hdfs://ns/usr/hive/warehouse/example.db/table2/part_col=2024-01-01

      • Temporary file: hdfs://ns/tmp/.doris_staging/root/a7eac7505d7a42fdb06cb9ef1ea3e912/par1=a/d678a74d232345e0-b659e2fb58e86ffd_549ad677-ee75-4fa1-b8a6-3e821e1dae61-0.orc

        Commit phase:

      1. The target table partition directory is renamed to a temporary partition directory: hdfs://ns/usr/hive/warehouse/example.db/table2/_temp_d678a74d232345e0-b659e2fb58e86ffd_part_col=2024-01-01
      2. The temporary partition directory is renamed to the target table partition directory.
      3. The temporary partition directory is deleted.

Relevant Parameters

FE

TODO

BE

Parameter NameDefault ValueDescription
hive_sink_max_file_sizeMaximum file size for data files. When the volume of written data exceeds this size, the current file is closed, and a new file is opened for continued writing.1GB
table_sink_partition_write_max_partition_nums_per_writerMaximum number of partitions that can be written by each Instance on a BE node.128
table_sink_non_partition_write_scaling_data_processed_thresholdThreshold of data volume for starting scaling-write in non-partitioned tables. For every increase of table_sink_non_partition_write_scaling_data_processed_threshold in data volume, a new writer (instance) will be engaged for writing. The scaling-write mechanism aims to use a different number of writers (instances) based on the volume of data to increase the throughput of concurrent writing. When the volume of data is small, it also saves resources and reduces the number of files produced as much as possible.25MB
table_sink_partition_write_min_data_processed_rebalance_thresholdMinimum data volume threshold for triggering rebalance in partitioned tables. If current accumulated data volume - data volume accumulated since the last rebalance or from the start >= table_sink_partition_write_min_data_processed_rebalance_threshold, rebalancing is triggered. If there is a significant difference in the final file sizes, you can reduce this threshold to increase balance. However, too small a threshold may increase the cost of rebalancing and potentially affect performance.25MB
table_sink_partition_write_min_partition_data_processed_rebalance_thresholdMinimum data volume threshold per partition for rebalancing in partitioned tables. If current partition’s data volume >= threshold * number of tasks allocated to the current partition, rebalancing for that partition begins. If there is a significant difference in the final file sizes, you can reduce this threshold to increase balance. However, too small a threshold may increase the cost of rebalancing and potentially affect performance.15MB