diff --git a/pom.xml b/pom.xml
index 315fcf0..de2b189 100644
--- a/pom.xml
+++ b/pom.xml
@@ -5,8 +5,8 @@
4.0.0
com.zdjizhi
- log-completion-schema
- 220318-Nacos
+ knowledge-log
+ 202207
log-completion-schema
http://www.example.com
@@ -35,10 +35,9 @@
UTF-8
1.13.1
1.0.0
- 1.2.0
1.0.8
-
+
@@ -230,13 +229,6 @@
test
-
-
- com.alibaba.nacos
- nacos-client
- ${nacos.version}
-
-
org.jasypt
@@ -249,13 +241,8 @@
ru.yandex.clickhouse
clickhouse-jdbc
- 0.2.3
+ 0.2.6
-
-
-
-
-
org.apache.flink
flink-table-planner-blink_2.12
diff --git a/properties/service_flow_config.properties b/properties/service_flow_config.properties
index 2308097..91858cd 100644
--- a/properties/service_flow_config.properties
+++ b/properties/service_flow_config.properties
@@ -1,22 +1,22 @@
-#--------------------------------地址配置------------------------------#
-#管理kafka地址
+#--------------------------------\u5730\u5740\u914D\u7F6E------------------------------#
+#\u7BA1\u7406kafka\u5730\u5740
source.kafka.servers=192.168.45.102:9092
-#管理输出kafka地址
+#\u7BA1\u7406\u8F93\u51FAkafka\u5730\u5740
sink.kafka.servers=192.168.45.102:9092
-#zookeeper 地址 用于配置log_id
+#zookeeper \u5730\u5740 \u7528\u4E8E\u914D\u7F6Elog_id
zookeeper.servers=192.168.45.102:2181
-#hbase zookeeper地址 用于连接HBase
+#hbase zookeeper\u5730\u5740 \u7528\u4E8E\u8FDE\u63A5HBase
hbase.zookeeper.servers=192.168.45.102:2181
-#--------------------------------HTTP/定位库------------------------------#
-#定位库地址
+#--------------------------------HTTP/\u5B9A\u4F4D\u5E93------------------------------#
+#\u5B9A\u4F4D\u5E93\u5730\u5740
tools.library=D:\\workerspace\\dat\\
-#--------------------------------nacos配置------------------------------#
-#nacos 地址
+#--------------------------------nacos\u914D\u7F6E------------------------------#
+#nacos \u5730\u5740
nacos.server=192.168.45.102:8848
#nacos namespace
@@ -25,39 +25,39 @@ nacos.schema.namespace=prod
#nacos data id
nacos.data.id=session_record.json
-#--------------------------------Kafka消费/生产配置------------------------------#
+#--------------------------------Kafka\u6D88\u8D39/\u751F\u4EA7\u914D\u7F6E------------------------------#
-#kafka 接收数据topic
+#kafka \u63A5\u6536\u6570\u636Etopic
source.kafka.topic=atest
-#补全数据 输出 topic
+#\u8865\u5168\u6570\u636E \u8F93\u51FA topic
sink.kafka.topic=atest2
-#读取topic,存储该spout id的消费offset信息,可通过该拓扑命名;具体存储offset的位置,确定下次读取不重复的数据;
+#\u8BFB\u53D6topic,\u5B58\u50A8\u8BE5spout id\u7684\u6D88\u8D39offset\u4FE1\u606F\uFF0C\u53EF\u901A\u8FC7\u8BE5\u62D3\u6251\u547D\u540D;\u5177\u4F53\u5B58\u50A8offset\u7684\u4F4D\u7F6E\uFF0C\u786E\u5B9A\u4E0B\u6B21\u8BFB\u53D6\u4E0D\u91CD\u590D\u7684\u6570\u636E\uFF1B
group.id=flinktest-102
-#--------------------------------topology配置------------------------------#
+#--------------------------------topology\u914D\u7F6E------------------------------#
-#consumer 并行度
+#consumer \u5E76\u884C\u5EA6
source.parallelism=1
-#转换函数并行度
+#\u8F6C\u6362\u51FD\u6570\u5E76\u884C\u5EA6
transform.parallelism=1
-#kafka producer 并行度
+#kafka producer \u5E76\u884C\u5EA6
sink.parallelism=1
-#数据中心,取值范围(0-31)
+#\u6570\u636E\u4E2D\u5FC3\uFF0C\u53D6\u503C\u8303\u56F4(0-31)
data.center.id.num=0
-#hbase 更新时间,如填写0则不更新缓存
+#hbase \u66F4\u65B0\u65F6\u95F4\uFF0C\u5982\u586B\u51990\u5219\u4E0D\u66F4\u65B0\u7F13\u5B58
hbase.tick.tuple.freq.secs=180
-#--------------------------------默认值配置------------------------------#
-#0不需要补全原样输出日志,1需要补全
-log.need.complete=1
+#--------------------------------\u9ED8\u8BA4\u503C\u914D\u7F6E------------------------------#
+#1 connection\u65E5\u5FD7 \uFF0C2 dns\u65E5\u5FD7
+log.need.complete=2
-#生产者压缩模式 none or snappy
+#\u751F\u4EA7\u8005\u538B\u7F29\u6A21\u5F0F none or snappy
producer.kafka.compression.type=none
@@ -77,29 +77,29 @@ sink.arango.table.r.mx.domain2domain=R_MX_DOMAIN2DOMAIN
sink.arango.table.r.resolve.domain2ip=R_RESOLVE_DOMAIN2IP
sink.arango.table.r.nx.domain2domain=R_NX_DOMAIN2DOMAIN
-#clickhouse 入库
+#clickhouse \u5165\u5E93
ck.hosts=192.168.45.102:8123
ck.database=tsg_galaxy_v3
ck.username=default
ck.pin=galaxy2019
-ck.connection.timeout=100000
-ck.socket.timeout=1000000
+#\u5355\u4F4D\u6BEB\u79D2
+ck.connection.timeout=10000
+ck.socket.timeout=300000
#connection_record_log
-flink.watermark.max.orderness=100000
-#统计时间间隔 单位s
+flink.watermark.max.orderness=50
+#\u7EDF\u8BA1\u65F6\u95F4\u95F4\u9694 \u5355\u4F4Ds
log.aggregate.duration=5
log.aggregate.duration.graph=5
-#arangoDB参数配置
+#arangoDB\u53C2\u6570\u914D\u7F6E
arangoDB.host=192.168.45.102
#arangoDB.host=192.168.40.224
arangoDB.port=8529
-arangoDB.user=upsert
-arangoDB.password=galaxy2018
+arangoDB.user=root
+arangoDB.password=galaxy_2019
arangoDB.DB.name=knowledge
-#arangoDB.DB.name=tsg_galaxy_v3
arangoDB.batch=100000
arangoDB.ttl=3600
@@ -109,5 +109,5 @@ update.arango.batch=10000
thread.pool.number=10
thread.await.termination.time=10
-sink.batch.time.out=1
+sink.batch.time.out=5
sink.batch=10000
\ No newline at end of file
diff --git a/src/main/java/com/zdjizhi/common/ArangoDelayProcess.java b/src/main/java/com/zdjizhi/common/ArangoDelayProcess.java
deleted file mode 100644
index d39e6c5..0000000
--- a/src/main/java/com/zdjizhi/common/ArangoDelayProcess.java
+++ /dev/null
@@ -1,65 +0,0 @@
-package com.zdjizhi.common;
-
-import com.arangodb.entity.BaseDocument;
-import org.apache.flink.api.common.state.ListState;
-import org.apache.flink.api.common.state.ListStateDescriptor;
-import org.apache.flink.api.common.state.ValueState;
-import org.apache.flink.api.common.state.ValueStateDescriptor;
-import org.apache.flink.api.common.typeinfo.Types;
-import org.apache.flink.configuration.Configuration;
-import org.apache.flink.streaming.api.functions.ProcessFunction;
-import org.apache.flink.util.Collector;
-
-import java.util.List;
-import java.util.Map;
-import java.util.Spliterator;
-import java.util.stream.Collectors;
-import java.util.stream.StreamSupport;
-
-public class ArangoDelayProcess extends ProcessFunction> {
-
- private ValueState currentTimer;
- private ListState itemState;
- private String stateName;
-
- @Override
- public void open(Configuration parameters) throws Exception {
- currentTimer = getRuntimeContext().getState(new ValueStateDescriptor<>(getStateName() + "_timer", Types.LONG));
- ListStateDescriptor itemViewStateDesc = new ListStateDescriptor(getStateName() + "_state", Map.class);
- itemState = getRuntimeContext().getListState(itemViewStateDesc);
- }
-
- @Override
- public void processElement(BaseDocument value, Context context, Collector> collector) throws Exception {
- //判断定时器是否为空,为空则创建新的定时器
- Long curTimeStamp = currentTimer.value();
- if (curTimeStamp == null || curTimeStamp == 0) {
- long onTimer = context.timestamp() + FlowWriteConfig.SINK_BATCH_TIME_OUT * 1000;
- context.timerService().registerEventTimeTimer(onTimer);
- currentTimer.update(onTimer);
- }
- itemState.add(value);
- }
-
- @Override
- public void onTimer(long timestamp, OnTimerContext ctx, Collector> out) throws Exception {
- Spliterator spliterator = itemState.get().spliterator();
- List collect = StreamSupport.stream(spliterator, false)
- .collect(Collectors.toList());
- out.collect(collect);
- currentTimer.clear();
- itemState.clear();
- }
-
- public ArangoDelayProcess(String stateName) {
- this.stateName = stateName;
- }
-
- public String getStateName() {
- return stateName;
- }
-
- public void setStateName(String stateName) {
- this.stateName = stateName;
- }
-}
diff --git a/src/main/java/com/zdjizhi/common/ArangodbDnsWindow.java b/src/main/java/com/zdjizhi/common/ArangodbDnsWindow.java
new file mode 100644
index 0000000..373eef5
--- /dev/null
+++ b/src/main/java/com/zdjizhi/common/ArangodbDnsWindow.java
@@ -0,0 +1,36 @@
+package com.zdjizhi.common;
+
+import cn.hutool.core.util.StrUtil;
+import com.arangodb.entity.BaseEdgeDocument;
+import org.apache.flink.streaming.api.functions.windowing.AllWindowFunction;
+import org.apache.flink.streaming.api.windowing.windows.TimeWindow;
+import org.apache.flink.util.Collector;
+
+import java.util.ArrayList;
+import java.util.Iterator;
+import java.util.List;
+import java.util.Map;
+
+public class ArangodbDnsWindow implements AllWindowFunction