path |
是 |
无 |
string |
hudi表路径 |
hoodie.table.name |
文档没有 |
无 |
string |
注册到Hive元数据的表名 |
table.type |
否 |
COPY_ON_WRITE |
string |
表的类型,取值范围:[COPY_ON_WRITE,MERGE_ON_READ ],不区分大小写 |
write.operation |
否 |
upsert |
string |
写入操作,暂时看只有upsert,应该还有insert |
write.precombine.field |
否 |
ts |
string |
预合并字段,当2条记录key相同时,通过Object.compareTo方法进行比较,保留结果比较大的那个 |
write.payload.class |
否 |
org.apache.hudi .common.model .Overwrite WithLatest AvroPayload |
string |
Payload class used. Override this, if you like to roll your own merge logic, when upserting/inserting. This will render any value set for the option in-effective |
write.insert.drop.duplicates |
否/文档没有 |
false |
bool |
是否删除重复记录 |
write.retry.times |
否/文档没有 |
3 |
int |
重试次数 |
write.retry.interval.ms |
否/文档没有 |
2000 |
long |
重试间隔 |
write.ignore.failed |
否 |
true |
bool |
是否忽略异常 |
hoodie.datasource .write.recordkey.field |
否 |
uuid |
string |
Record key field. Value to be used as the recordKey component of HoodieKey . "Actual value will be obtained by invoking .toString() on the field value. Nested fields can be specified using the dot notation eg: a.b.c |
hoodie.datasource.write .partitionpath.field |
否/文档没有 |
partition-path |
string |
Partition path field. Value to be used at the partitionPath component of HoodieKey . Actual value obtained by invoking .toString() |
hoodie.datasource.write .keygenerator.class |
否 |
org.apache .hudi.keygen .SimpleAvro KeyGenerator |
|
生成key的类,从输入记录中提取key |
write.tasks |
否 |
4 |
int |
写任务并行度 |
write.batch.size.MB |
否 |
128 |
double |
写缓冲大小 |