2020-12-25 17:32:54 +08:00
|
|
|
|
#--------------------------------地址配置------------------------------#
|
|
|
|
|
|
|
2019-08-05 17:50:15 +08:00
|
|
|
|
#管理kafka地址
|
2020-12-25 17:32:54 +08:00
|
|
|
|
input.kafka.servers=192.168.44.12:9092
|
2019-08-05 17:50:15 +08:00
|
|
|
|
|
2020-12-25 17:32:54 +08:00
|
|
|
|
#管理输出kafka地址
|
|
|
|
|
|
output.kafka.servers=192.168.44.12:9092
|
2019-08-05 17:50:15 +08:00
|
|
|
|
|
2020-12-25 17:32:54 +08:00
|
|
|
|
#zookeeper 地址 用于配置log_id
|
2021-02-01 11:05:02 +08:00
|
|
|
|
zookeeper.servers=192.168.44.12:2181
|
2020-12-25 17:32:54 +08:00
|
|
|
|
|
|
|
|
|
|
#hbase zookeeper地址 用于连接HBase
|
|
|
|
|
|
hbase.zookeeper.servers=192.168.44.12:2181
|
|
|
|
|
|
|
2021-04-23 18:06:37 +08:00
|
|
|
|
#--------------------------------HTTP/定位库------------------------------#
|
2020-12-25 17:32:54 +08:00
|
|
|
|
#定位库地址
|
2021-03-25 14:27:41 +08:00
|
|
|
|
ip.library=D:\\K18-Phase2\\tsgSpace\\dat\\
|
2021-02-01 11:05:02 +08:00
|
|
|
|
#ip.library=/home/bigdata/topology/dat/
|
2020-12-25 17:32:54 +08:00
|
|
|
|
|
|
|
|
|
|
#网关的schema位置
|
2021-02-01 11:05:02 +08:00
|
|
|
|
schema.http=http://192.168.44.12:9999/metadata/schema/v1/fields/security_event_log
|
2019-08-05 17:50:15 +08:00
|
|
|
|
|
2021-04-23 18:06:37 +08:00
|
|
|
|
#网关APP_ID 获取接口
|
|
|
|
|
|
app.id.http=http://192.168.44.67:9999/open-api/appDicList
|
|
|
|
|
|
|
|
|
|
|
|
#--------------------------------Kafka消费组信息------------------------------#
|
|
|
|
|
|
|
|
|
|
|
|
#kafka 接收数据topic
|
2021-02-01 11:05:02 +08:00
|
|
|
|
kafka.topic=test
|
2019-08-05 17:50:15 +08:00
|
|
|
|
|
2021-04-23 18:06:37 +08:00
|
|
|
|
#补全数据 输出 topic
|
|
|
|
|
|
results.output.topic=test-result
|
|
|
|
|
|
|
2019-08-05 17:50:15 +08:00
|
|
|
|
#读取topic,存储该spout id的消费offset信息,可通过该拓扑命名;具体存储offset的位置,确定下次读取不重复的数据;
|
2020-12-25 17:32:54 +08:00
|
|
|
|
group.id=connection-record-log-20200818-1-test
|
|
|
|
|
|
|
|
|
|
|
|
#生产者压缩模式 none or snappy
|
2021-04-23 18:06:37 +08:00
|
|
|
|
producer.kafka.compression.type=none
|
2020-12-25 17:32:54 +08:00
|
|
|
|
|
|
|
|
|
|
#生产者ack
|
|
|
|
|
|
producer.ack=1
|
|
|
|
|
|
|
|
|
|
|
|
#latest/earliest 从当前消 or 从头消费
|
|
|
|
|
|
auto.offset.reset=latest
|
2019-08-05 17:50:15 +08:00
|
|
|
|
|
2021-04-23 18:06:37 +08:00
|
|
|
|
#接收自kafka的消费者 client-id
|
|
|
|
|
|
consumer.client.id=consumer-connection-record
|
|
|
|
|
|
|
|
|
|
|
|
#回写给kafka的生产者 client-id
|
|
|
|
|
|
producer.client.id=producer-connection-record
|
2020-12-25 17:32:54 +08:00
|
|
|
|
|
|
|
|
|
|
#--------------------------------topology配置------------------------------#
|
2019-08-05 17:50:15 +08:00
|
|
|
|
|
|
|
|
|
|
#storm topology workers
|
2020-12-25 17:32:54 +08:00
|
|
|
|
topology.workers=2
|
2019-08-05 17:50:15 +08:00
|
|
|
|
|
|
|
|
|
|
#spout并行度 建议与kafka分区数相同
|
2020-02-25 15:12:56 +08:00
|
|
|
|
spout.parallelism=3
|
2019-08-05 17:50:15 +08:00
|
|
|
|
|
|
|
|
|
|
#处理补全操作的bolt并行度-worker的倍数
|
2020-12-25 17:32:54 +08:00
|
|
|
|
completion.bolt.parallelism=6
|
2019-08-05 17:50:15 +08:00
|
|
|
|
|
2020-12-25 17:32:54 +08:00
|
|
|
|
#写入kafka的并行度10
|
|
|
|
|
|
kafka.bolt.parallelism=6
|
2019-08-05 17:50:15 +08:00
|
|
|
|
|
|
|
|
|
|
#数据中心(UID)
|
2019-11-12 16:02:50 +08:00
|
|
|
|
data.center.id.num=15
|
2019-08-05 17:50:15 +08:00
|
|
|
|
|
2020-12-25 17:32:54 +08:00
|
|
|
|
#hbase 更新时间
|
|
|
|
|
|
hbase.tick.tuple.freq.secs=60
|
|
|
|
|
|
|
2021-04-23 18:06:37 +08:00
|
|
|
|
#app_id 更新时间
|
|
|
|
|
|
app.tick.tuple.freq.secs=60
|
2020-12-25 17:32:54 +08:00
|
|
|
|
|
|
|
|
|
|
#--------------------------------默认值配置------------------------------#
|
|
|
|
|
|
|
2019-08-05 17:50:15 +08:00
|
|
|
|
#当bolt性能受限时,限制spout接收速度,理论看ack开启才有效
|
|
|
|
|
|
topology.config.max.spout.pending=150000
|
|
|
|
|
|
|
2020-12-25 17:32:54 +08:00
|
|
|
|
#hbase table name
|
|
|
|
|
|
hbase.table.name=subscriber_info
|
2019-08-05 17:50:15 +08:00
|
|
|
|
|
2021-04-23 18:06:37 +08:00
|
|
|
|
#ack设置 1启动ack 0不启动ack
|
|
|
|
|
|
topology.num.acks=0
|
|
|
|
|
|
|
|
|
|
|
|
#kafka批量条数
|
|
|
|
|
|
batch.insert.num=2000
|
|
|
|
|
|
|
|
|
|
|
|
#tick时钟频率
|
|
|
|
|
|
topology.tick.tuple.freq.secs=5
|
|
|
|
|
|
|
2019-08-05 17:50:15 +08:00
|
|
|
|
#spout接收睡眠时间
|
|
|
|
|
|
topology.spout.sleep.time=1
|
|
|
|
|
|
|
|
|
|
|
|
#允许发送kafka最大失败数
|
|
|
|
|
|
max.failure.num=20
|
|
|
|
|
|
|
2020-12-25 17:32:54 +08:00
|
|
|
|
#邮件默认编码
|
|
|
|
|
|
mail.default.charset=UTF-8
|
|
|
|
|
|
|
|
|
|
|
|
#需不要补全,不需要则原样日志输出
|
|
|
|
|
|
log.need.complete=yes
|
|
|
|
|
|
|