HdfsFile
Hdfs file source connector
Descriptionâ
Read data from hdfs file system.
Key featuresâ
Read all the data in a split in a pollNext call. What splits are read will be saved in snapshot.
- schema projection
- parallelism
- support user-defined split
- file format
- text
- csv
- parquet
- orc
- json
Optionsâ
name | type | required | default value |
---|---|---|---|
path | string | yes | - |
type | string | yes | - |
fs.defaultFS | string | yes | - |
hdfs_site_path | string | no | - |
delimiter | string | no | \001 |
parse_partition_from_path | boolean | no | true |
date_format | string | no | yyyy-MM-dd |
datetime_format | string | no | yyyy-MM-dd HH:mm:ss |
time_format | string | no | HH:mm:ss |
schema | config | no | - |
common-options | no | - |
path [string]â
The source file path.
delimiter [string]â
Field delimiter, used to tell connector how to slice and dice fields when reading text files
default \001
, the same as hive's default delimiter
parse_partition_from_path [boolean]â
Control whether parse the partition keys and values from file path
For example if you read a file from path hdfs://hadoop-cluster/tmp/seatunnel/parquet/name=tyrantlucifer/age=26
Every record data from file will be added these two fields:
name | age |
---|---|
tyrantlucifer | 26 |
Tips: Do not define partition fields in schema option
date_format [string]â
Date type format, used to tell connector how to convert string to date, supported as the following formats:
yyyy-MM-dd
yyyy.MM.dd
yyyy/MM/dd
default yyyy-MM-dd
datetime_format [string]â
Datetime type format, used to tell connector how to convert string to datetime, supported as the following formats:
yyyy-MM-dd HH:mm:ss
yyyy.MM.dd HH:mm:ss
yyyy/MM/dd HH:mm:ss
yyyyMMddHHmmss
default yyyy-MM-dd HH:mm:ss
time_format [string]â
Time type format, used to tell connector how to convert string to time, supported as the following formats:
HH:mm:ss
HH:mm:ss.SSS
default HH:mm:ss
type [string]â
File type, supported as the following file types:
text
csv
parquet
orc
json
If you assign file type to json
, you should also assign schema option to tell connector how to parse data to the row you want.
For example:
upstream data is the following:
{"code": 200, "data": "get success", "success": true}
You can also save multiple pieces of data in one file and split them by newline:
{"code": 200, "data": "get success", "success": true}
{"code": 300, "data": "get failed", "success": false}
you should assign schema as the following:
schema {
fields {
code = int
data = string
success = boolean
}
}
connector will generate data as the following:
code | data | success |
---|---|---|
200 | get success | true |
If you assign file type to parquet
orc
, schema option not required, connector can find the schema of upstream data automatically.
If you assign file type to text
csv
, you can choose to specify the schema information or not.
For example, upstream data is the following:
tyrantlucifer#26#male
If you do not assign data schema connector will treat the upstream data as the following:
content |
---|
tyrantlucifer#26#male |
If you assign data schema, you should also assign the option delimiter
too except CSV file type
you should assign schema and delimiter as the following:
delimiter = "#"
schema {
fields {
name = string
age = int
gender = string
}
}
connector will generate data as the following:
name | age | gender |
---|---|---|
tyrantlucifer | 26 | male |
fs.defaultFS [string]â
Hdfs cluster address.
hdfs_site_path [string]â
The path of hdfs-site.xml
, used to load ha configuration of namenodes
schema [Config]â
fields [Config]â
the schema fields of upstream data
common optionsâ
Source plugin common parameters, please refer to Source Common Options for details.
Exampleâ
HdfsFile {
path = "/apps/hive/demo/student"
type = "parquet"
fs.defaultFS = "hdfs://namenode001"
}
HdfsFile {
schema {
fields {
name = string
age = int
}
}
path = "/apps/hive/demo/student"
type = "json"
fs.defaultFS = "hdfs://namenode001"
}
Changelogâ
2.2.0-beta 2022-09-26â
- Add HDFS File Source Connector