File
File source connector
Descriptionâ
Read data from local or hdfs file.
tip
Engine Supported and plugin name
- Spark: File
- Flink: File
Optionsâ
- Spark
- Flink
name | type | required | default value |
---|---|---|---|
format | string | no | json |
path | string | yes | - |
common-options | string | yes | - |
format [string]â
Format for reading files, currently supports text, parquet, json, orc, csv.
name | type | required | default value |
---|---|---|---|
format.type | string | yes | - |
path | string | yes | - |
schema | string | yes | - |
common-options | string | no | - |
parallelism | int | no | - |
format.type [string]â
The format for reading files from the file system, currently supports csv
, json
, parquet
, orc
and text
.
schema [string]â
- csv
- The
schema
ofcsv
is a string ofjsonArray
, such as"[{\"type\":\"long\"},{\"type\":\"string\"}]"
, this can only specify the type of the field , The field name cannot be specified, and the common configuration parameterfield_name
is generally required.
- The
- json
- The
schema
parameter ofjson
is to provide ajson string
of the original data, and theschema
can be automatically generated, but the original data with the most complete content needs to be provided, otherwise the fields will be lost.
- The
- parquet
- The
schema
ofparquet
is anAvro schema string
, such as{\"type\":\"record\",\"name\":\"test\",\"fields\":[{\"name\" :\"a\",\"type\":\"int\"},{\"name\":\"b\",\"type\":\"string\"}]}
.
- The
- orc
- The
schema
oforc
is the string oforc schema
, such as"struct<name:string,addresses:array<struct<street:string,zip:smallint>>>"
.
- The
- text
- The
schema
oftext
can be filled withstring
.
- The
parallelism [Int
]â
The parallelism of an individual operator, for FileSource
path [string]â
- If read data from hdfs , the file path should start with
hdfs://
- If read data from local , the file path should start with
file://
common options [string]â
Source plugin common parameters, please refer to Source Plugin for details
Examplesâ
- Spark
- Flink
file {
path = "hdfs:///var/logs"
result_table_name = "access_log"
}
file {
path = "file:///var/logs"
result_table_name = "access_log"
}
FileSource{
path = "hdfs://localhost:9000/input/"
format.type = "json"
schema = "{\"data\":[{\"a\":1,\"b\":2},{\"a\":3,\"b\":4}],\"db\":\"string\",\"q\":{\"s\":\"string\"}}"
result_table_name = "test"
}