Sls
Sls sink connector
Support Those Engines
Spark
Flink
Seatunnel Zeta
Key Features
Description
Sink connector for Aliyun Sls.
Supported DataSource Info
In order to use the Sls connector, the following dependencies are required. They can be downloaded via install-plugin.sh or from the Maven central repository.
Datasource | Supported Versions | Maven |
---|---|---|
Sls | Universal | Download |
Source Options
Name | Type | Required | Default | Description |
---|---|---|---|---|
project | String | Yes | - | Aliyun Sls Project |
logstore | String | Yes | - | Aliyun Sls Logstore |
endpoint | String | Yes | - | Aliyun Access Endpoint |
access_key_id | String | Yes | - | Aliyun AccessKey ID |
access_key_secret | String | Yes | - | Aliyun AccessKey Secret |
source | String | No | SeaTunnel-Source | Data Source marking in sls |
topic | String | No | SeaTunnel-Topic | Data topic marking in sls |
Task Example
Simple
This example write data to the sls's logstore1.And if you have not yet installed and deployed SeaTunnel, you need to follow the instructions in Install SeaTunnel to install and deploy SeaTunnel. And if you have not yet installed and deployed SeaTunnel, you need to follow the instructions in Install SeaTunnel to install and deploy SeaTunnel. And then follow the instructions in Quick Start With SeaTunnel Engine to run this job.
Create RAM user and authorization,Please ensure thr ram user have sufficient rights to perform, reference RAM Custom Authorization Example
# Defining the runtime environment
env {
parallelism = 2
job.mode = "STREAMING"
checkpoint.interval = 30000
}
source {
FakeSource {
row.num = 10
map.size = 10
array.size = 10
bytes.length = 10
string.length = 10
schema = {
fields = {
id = "int"
name = "string"
description = "string"
weight = "string"
}
}
}
}
sink {
Sls {
endpoint = "cn-hangzhou-intranet.log.aliyuncs.com"
project = "project1"
logstore = "logstore1"
access_key_id = "xxxxxxxxxxxxxxxxxxxxxxxx"
access_key_secret = "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxx"
}
}