名称:phoenixwriter
phoenix4.12.0-HBase-1.3及以上
- connection
- 描述:数据库连接参数,包含jdbcUrl、schema、table等参数
- 必选:是
- 字段类型:List
- 示例:指定jdbcUrl、schema、table
"connection": [{
"jdbcUrl": "jdbc:phoenix:node01,node02,node03:2181",
"table": ["table"],
"schema":"public"
}]
- 默认值:无
- jdbcUrl
- 描述:针对关系型数据库的jdbc连接字符串
- 必选:是
- 字段类型:String
- 默认值:无
- schema
- 描述:数据库schema名
- 必选:否
- 字段类型:String
- 默认值:无
- table
- 描述:目的表的表名称。目前只支持配置单个表,后续会支持多表
- 必选:是
- 字段类型:List
- 默认值:无
- username
- 描述:数据源的用户名
- 必选:是
- 字段类型:String
- 默认值:无
- password
- 描述:数据源指定用户名的密码
- 必选:是
- 字段类型:String
- 默认值:无
- column
- 描述:目的表需要写入数据的字段,字段之间用英文逗号分隔。例如: "column": ["id","name","age"]
- 必选:是
- 默认值:否
- 字段类型:List
- 默认值:无
- fullcolumn
- 描述:目的表中的所有字段,字段之间用英文逗号分隔。例如: "column": ["id","name","age","hobby"],如果不配置,将在系统表中获取
- 必选:否
- 字段类型:List
- 默认值:无
- preSql
- 描述:写入数据到目的表前,会先执行这里的一组标准语句
- 必选:否
- 字段类型:String
- 默认值:无
- postSql
- 描述:写入数据到目的表后,会执行这里的一组标准语句
- 必选:否
- 字段类型:String
- 默认值:无
- writeMode
- 描述:控制写入数据到目标表采用
insert into
或者 merge into
语句
- 必选:是
- 所有选项:insert/update
- 字段类型:String
- 默认值:insert
- batchSize
- 描述:一次性批量提交的记录数大小,该值可以极大减少FlinkX与数据库的网络交互次数,并提升整体吞吐量。但是该值设置过大可能会造成FlinkX运行进程OOM情况
- 必选:否
- 字段类型:int
- 默认值:1024
- updateKey
- 描述:当写入模式为update时,需要指定此参数的值为唯一索引字段
- 注意:
- 采用
merge into
语法,对目标表进行匹配查询,匹配成功时更新,不成功时插入;
- 必选:否
- 字段类型:Map<String,List>
- 示例:"updateKey": {"key": ["id"]}
- 默认值:无
{
"job": {
"content": [{
"reader": {
"parameter": {
"sliceRecordCount": ["1"],
"column": [
{
"name": "id",
"type": "int",
"value": "400"
},
{
"name": "user_id",
"type": "int"
},
{
"name": "name",
"type": "string"
}
]
},
"name": "streamreader"
},
"writer": {
"name": "phoenixwriter",
"parameter": {
"connection": [{
"jdbcUrl": "jdbc:phoenix:node01,node02,node03:2181",
"table": [
"tableTest"
]
}],
"username": "",
"password": "",
"column": [
{
"name": "id",
"type": "BIGINT"
},
{
"name": "user_id",
"type": "BIGINT"
},
{
"name": "name",
"type": "varchar"
}],
"writeMode": "insert",
"batchSize": 1024,
"preSql": [],
"postSql": [],
"updateKey": {}
}
}
}],
"setting": {
"speed": {
"channel": 1,
"bytes": 0
},
"errorLimit": {
"record": 100
},
"restore": {
"maxRowNumForCheckpoint": 0,
"isRestore": false,
"restoreColumnName": "",
"restoreColumnIndex": 0
}
}
}
}
{
"job": {
"content": [{
"reader": {
"parameter": {
"sliceRecordCount": ["1"],
"column": [
{
"name": "id",
"type": "int",
"value": "400"
},
{
"name": "user_id",
"type": "int"
},
{
"name": "name",
"type": "string"
}
]
},
"name": "streamreader"
},
"writer": {
"name": "phoenixwriter",
"parameter": {
"connection": [{
"jdbcUrl": "jdbc:phoenix:node01,node02,node03:2181",
"table": [
"tableTest"
]
}],
"username": "",
"password": "",
"column": [
{
"name": "id",
"type": "BIGINT"
},
{
"name": "user_id",
"type": "BIGINT"
},
{
"name": "name",
"type": "varchar"
}],
"writeMode": "update",
"updateKey": {"key": ["id"]},
"batchSize": 1024,
"preSql": [],
"postSql": []
}
}
}],
"setting": {
"speed": {
"channel": 1,
"bytes": 0
},
"errorLimit": {
"record": 100
},
"restore": {
"maxRowNumForCheckpoint": 0,
"isRestore": false,
"restoreColumnName": "",
"restoreColumnIndex": 0
}
}
}
}