This repository has been archived on 2025-09-14. You can view files and clone it, but cannot push or open issues or pull requests.
Files
galaxy-tsg-olap-storm-log-s…/properties/service_flow_config.properties

104 lines
2.5 KiB
Properties
Raw Normal View History

2020-12-25 17:32:54 +08:00
#--------------------------------地址配置------------------------------#
2019-08-05 17:50:15 +08:00
#管理kafka地址
2020-12-25 17:32:54 +08:00
input.kafka.servers=192.168.44.12:9092
2019-08-05 17:50:15 +08:00
2020-12-25 17:32:54 +08:00
#管理输出kafka地址
output.kafka.servers=192.168.44.12:9092
2019-08-05 17:50:15 +08:00
2020-12-25 17:32:54 +08:00
#zookeeper 地址 用于配置log_id
2021-02-01 11:05:02 +08:00
zookeeper.servers=192.168.44.12:2181
2020-12-25 17:32:54 +08:00
#hbase zookeeper地址 用于连接HBase
hbase.zookeeper.servers=192.168.44.12:2181
#--------------------------------HTTP/定位库------------------------------#
2020-12-25 17:32:54 +08:00
#定位库地址
2021-03-25 14:27:41 +08:00
ip.library=D:\\K18-Phase2\\tsgSpace\\dat\\
2021-02-01 11:05:02 +08:00
#ip.library=/home/bigdata/topology/dat/
2020-12-25 17:32:54 +08:00
#网关的schema位置
2021-02-01 11:05:02 +08:00
schema.http=http://192.168.44.12:9999/metadata/schema/v1/fields/security_event_log
2019-08-05 17:50:15 +08:00
#网关APP_ID 获取接口
app.id.http=http://192.168.44.67:9999/open-api/appDicList
#--------------------------------Kafka消费组信息------------------------------#
#kafka 接收数据topic
2021-02-01 11:05:02 +08:00
kafka.topic=test
2019-08-05 17:50:15 +08:00
#补全数据 输出 topic
results.output.topic=test-result
2019-08-05 17:50:15 +08:00
#读取topic,存储该spout id的消费offset信息可通过该拓扑命名;具体存储offset的位置确定下次读取不重复的数据
2020-12-25 17:32:54 +08:00
group.id=connection-record-log-20200818-1-test
#生产者压缩模式 none or snappy
producer.kafka.compression.type=none
2020-12-25 17:32:54 +08:00
#生产者ack
producer.ack=1
#latest/earliest 从当前消 or 从头消费
auto.offset.reset=latest
2019-08-05 17:50:15 +08:00
#接收自kafka的消费者 client-id
consumer.client.id=consumer-connection-record
#回写给kafka的生产者 client-id
producer.client.id=producer-connection-record
2020-12-25 17:32:54 +08:00
#--------------------------------topology配置------------------------------#
2019-08-05 17:50:15 +08:00
#storm topology workers
2020-12-25 17:32:54 +08:00
topology.workers=2
2019-08-05 17:50:15 +08:00
#spout并行度 建议与kafka分区数相同
spout.parallelism=3
2019-08-05 17:50:15 +08:00
#处理补全操作的bolt并行度-worker的倍数
2020-12-25 17:32:54 +08:00
completion.bolt.parallelism=6
2019-08-05 17:50:15 +08:00
2020-12-25 17:32:54 +08:00
#写入kafka的并行度10
kafka.bolt.parallelism=6
2019-08-05 17:50:15 +08:00
#数据中心UID
2019-11-12 16:02:50 +08:00
data.center.id.num=15
2019-08-05 17:50:15 +08:00
2020-12-25 17:32:54 +08:00
#hbase 更新时间
hbase.tick.tuple.freq.secs=60
#app_id 更新时间
app.tick.tuple.freq.secs=60
2020-12-25 17:32:54 +08:00
#--------------------------------默认值配置------------------------------#
2019-08-05 17:50:15 +08:00
#当bolt性能受限时限制spout接收速度理论看ack开启才有效
topology.config.max.spout.pending=150000
2020-12-25 17:32:54 +08:00
#hbase table name
hbase.table.name=subscriber_info
2019-08-05 17:50:15 +08:00
#ack设置 1启动ack 0不启动ack
topology.num.acks=0
#kafka批量条数
batch.insert.num=2000
#tick时钟频率
topology.tick.tuple.freq.secs=5
2019-08-05 17:50:15 +08:00
#spout接收睡眠时间
topology.spout.sleep.time=1
#允许发送kafka最大失败数
max.failure.num=20
2020-12-25 17:32:54 +08:00
#邮件默认编码
mail.default.charset=UTF-8
#需不要补全,不需要则原样日志输出
log.need.complete=yes