Hive
Last updated
Last updated
Hive source connector
Read data from Hive.
Read all the data in a split in a pollNext call. What splits are read will be saved in snapshot.
table_name
string
yes
-
metastore_uri
string
yes
-
krb5_path
string
no
/etc/krb5.conf
kerberos_principal
string
no
-
kerberos_keytab_path
string
no
-
hdfs_site_path
string
no
-
hive_site_path
string
no
-
hive.hadoop.conf
Map
no
-
hive.hadoop.conf-path
string
no
-
read_partitions
list
no
-
read_columns
list
no
-
compress_codec
string
no
none
common-options
no
-
Target Hive table name eg: db1.table1
Hive metastore uri
The path of hdfs-site.xml
, used to load ha configuration of namenodes
Properties in hadoop conf('core-site.xml', 'hdfs-site.xml', 'hive-site.xml')
The specified loading path for the 'core-site.xml', 'hdfs-site.xml', 'hive-site.xml' files
The target partitions that user want to read from hive table, if user does not set this parameter, it will read all the data from hive table.
Tips: Every partition in partitions list should have the same directory depth. For example, a hive table has two partitions: par1 and par2, if user sets it like as the following: read_partitions = [par1=xxx, par1=yyy/par2=zzz], it is illegal
The path of krb5.conf
, used to authentication kerberos
The principal of kerberos authentication
The keytab file path of kerberos authentication
The read column list of the data source, user can use it to implement field projection.
The compress codec of files and the details that supported as the following shown:
txt: lzo
none
json: lzo
none
csv: lzo
none
orc/parquet: automatically recognizes the compression type, no additional settings required.
Source plugin common parameters, please refer to Source Common Options for details
Run the case.
Run the case.