Apache Doris 中文技术论坛
Questions Tags Users Badges

【已解决】flink-connector-doris消费kafka写入doris出现问题

Asked Apr 12, 2024 Modified May 6, 2024
Viewed 66
ingestion 2.0

1712900818935.png

edited May 6, 2024
zhb123319
asked Apr 12, 2024
1 Answers

默认的心跳超时时间为5s, 心跳停止后,FE马上abort coordinate BE 的事务。实际上be没有down
然而BE事务在导入过程中并不需要fe的参与,这个5s太敏感了,建议改成超过1分钟没心跳才abort coordinate BE 的事务。

参考pr:
https://github.com/apache/doris/pull/22781

edited May 6, 2024
徐振超@SelectDB7121
answered Apr 12, 2024
Related Questions
java中通过mybatis批量写入数据,写入速度才100/秒,远远达不到官网说的1w/秒
1 answers
Doris 2.0.15 使用 insert into 插入1kw数据到一张空表上,之后想执行full compaction压缩,但是报错
Base Compaction(基线压缩)多久执行一次?
1 answers
Doris Kafka Connector 同步问题
1 answers
Doirs Kafka Connector 字段配置
1 answers
如图中所示,我1fe 2be,测试2000万行0.2G数据,导出文件不会分散在多个节点上,enable_parallel_outfile也开了,是有什么条件吗?
1 answers

Terms of service Privacy policy

Powered by Answer - the open-source software that powers Q&A communities.
Made with love © 2025 Apache Doris 中文技术论坛.