21.08终版

This commit is contained in:
jianlong
2021-09-08 11:33:00 +08:00
parent 350bdb9f1c
commit 93522be84e
94 changed files with 1446 additions and 398 deletions

346
.idea/workspace.xml generated
View File

@@ -3,7 +3,80 @@
<component name="ChangeListManager">
<list default="true" id="55c8c5b9-9b57-431e-a0ed-0064b85979fa" name="Default Changelist" comment="">
<change beforePath="$PROJECT_DIR$/.idea/workspace.xml" beforeDir="false" afterPath="$PROJECT_DIR$/.idea/workspace.xml" afterDir="false" />
<change beforePath="$PROJECT_DIR$/configurations/components.yml" beforeDir="false" afterPath="$PROJECT_DIR$/configurations/components.yml" afterDir="false" />
<change beforePath="$PROJECT_DIR$/configurations/config.yml" beforeDir="false" afterPath="$PROJECT_DIR$/configurations/config.yml" afterDir="false" />
<change beforePath="$PROJECT_DIR$/configurations/hosts" beforeDir="false" afterPath="$PROJECT_DIR$/configurations/hosts" afterDir="false" />
<change beforePath="$PROJECT_DIR$/configurations/services.yml" beforeDir="false" afterPath="$PROJECT_DIR$/configurations/services.yml" afterDir="false" />
<change beforePath="$PROJECT_DIR$/install.sh" beforeDir="false" afterPath="$PROJECT_DIR$/install.sh" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/init-galaxy-hos-service.yml" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/init-galaxy-hos-service.yml" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/menu/appMenu" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/menu/appMenu" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/menu/commonMenu" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/menu/commonMenu" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/check/check-services/tasks/main.yml" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/check/check-services/tasks/main.yml" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/components/clickhouse/common/templates/config.xml.j2" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/components/clickhouse/common/templates/config.xml.j2" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/components/druid/files/mysql" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/components/druid/files/mysql" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/components/druid/tasks/install_druid.yml" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/components/druid/tasks/install_druid.yml" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/components/druid/templates/keepdruidall.j2" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/components/druid/templates/keepdruidall.j2" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/components/druid/templates/keepdruiddata.j2" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/components/druid/templates/keepdruiddata.j2" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/components/druid/templates/keepdruidquery.j2" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/components/druid/templates/keepdruidquery.j2" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/components/druid/templates/unload_druid.sh.j2" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/components/druid/templates/unload_druid.sh.j2" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/components/flink/tasks/install_flink.yml" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/components/flink/tasks/install_flink.yml" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/components/flink/templates/set_flink_env.sh.j2" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/components/flink/templates/set_flink_env.sh.j2" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/components/flink/templates/unload_flink.sh.j2" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/components/flink/templates/unload_flink.sh.j2" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/components/hadoop/tasks/install_hadoop.yml" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/components/hadoop/tasks/install_hadoop.yml" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/components/hadoop/templates/set_hadoop_env.sh.j2" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/components/hadoop/templates/set_hadoop_env.sh.j2" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/components/hadoop/templates/unload_hadoop.sh.j2" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/components/hadoop/templates/unload_hadoop.sh.j2" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/components/hbase/tasks/install_hbase.yml" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/components/hbase/tasks/install_hbase.yml" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/components/hbase/templates/set_hbase_env.sh.j2" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/components/hbase/templates/set_hbase_env.sh.j2" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/components/hbase/templates/unload_hbase.sh.j2" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/components/hbase/templates/unload_hbase.sh.j2" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/components/kafka/tasks/install_kafka.yml" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/components/kafka/tasks/install_kafka.yml" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/components/kafka/templates/set_kafka_env.sh.j2" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/components/kafka/templates/set_kafka_env.sh.j2" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/components/kafka/templates/unload_kafka.sh.j2" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/components/kafka/templates/unload_kafka.sh.j2" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/components/mariadb/install-mariadb-cluster/tasks/install-mariadb.yml" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/components/mariadb/install-mariadb-cluster/tasks/install-mariadb.yml" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/components/nacos/files/mysql" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/components/nacos/files/mysql" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/components/nacos/tasks/install_nacos.yml" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/components/nacos/tasks/install_nacos.yml" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/components/nacos/templates/set-nacos-env.sh.j2" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/components/nacos/templates/set-nacos-env.sh.j2" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/components/nacos/templates/unload_nacos.sh.j2" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/components/nacos/templates/unload_nacos.sh.j2" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/components/spark/templates/set_spark_env.sh.j2" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/components/spark/templates/set_spark_env.sh.j2" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/components/storm/templates/set_storm_env.sh.j2" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/components/storm/templates/set_storm_env.sh.j2" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/components/storm/templates/unload_storm.sh.j2" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/components/storm/templates/unload_storm.sh.j2" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/components/zookeeper/tasks/install_zk.yml" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/components/zookeeper/tasks/install_zk.yml" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/components/zookeeper/templates/keepzkalive.j2" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/components/zookeeper/templates/keepzkalive.j2" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/components/zookeeper/templates/set_zk_env.sh.j2" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/components/zookeeper/templates/set_zk_env.sh.j2" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/components/zookeeper/templates/unload_zk.sh.j2" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/components/zookeeper/templates/unload_zk.sh.j2" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/dependencies/environments/install_docker/docker/containerd" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/dependencies/environments/install_docker/docker/containerd" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/dependencies/environments/install_docker/docker/ctr" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/dependencies/environments/install_docker/docker/ctr" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/dependencies/environments/install_docker/docker/docker" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/dependencies/environments/install_docker/docker/docker" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/dependencies/environments/install_docker/docker/dockerd" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/dependencies/environments/install_docker/docker/dockerd" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/init/clickhouse/tasks/init_clickhouse.yml" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/init/clickhouse/tasks/init_clickhouse.yml" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/init/dos-baseline/tasks/init_dos.yml" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/init/dos-baseline/tasks/init_dos.yml" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/init/druid/files/druid_topology/rule/supervisor-manger" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/init/druid/files/druid_topology/rule/supervisor-manger" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/init/druid/tasks/main.yml" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/init/druid/tasks/main.yml" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/init/druid/templates/ingestionTask.sh.j2" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/init/druid/templates/ingestionTask.sh.j2" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/init/druid/templates/post_rule.sh.j2" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/init/druid/templates/post_rule.sh.j2" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/init/flink/tasks/main.yml" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/init/flink/tasks/main.yml" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/init/galaxy-hos-service/tasks/main.yml" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/init/galaxy-hos-service/tasks/main.yml" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/init/galaxy-job-service/files/mysql" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/init/galaxy-job-service/files/mysql" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/init/hbase/tasks/main.yml" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/init/hbase/tasks/main.yml" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/init/hbase/templates/create-hbase-table.sh.j2" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/init/hbase/templates/create-hbase-table.sh.j2" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/services/galaxy-chproxy/tasks/load_chproxy.yml" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/services/galaxy-chproxy/tasks/load_chproxy.yml" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/services/galaxy-gateway-keepalive/tasks/main.yml" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/services/galaxy-gateway-keepalive/tasks/main.yml" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/services/galaxy-gateway-keepalive/templates/installKeepAlived.sh.j2" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/services/galaxy-gateway-keepalive/templates/installKeepAlived.sh.j2" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/services/galaxy-gateway-nginx/tasks/main.yml" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/services/galaxy-gateway-nginx/tasks/main.yml" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/services/galaxy-hos-keepalive/tasks/main.yml" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/services/galaxy-hos-keepalive/tasks/main.yml" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/services/galaxy-hos-keepalive/templates/installKeepAlived.sh.j2" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/services/galaxy-hos-keepalive/templates/installKeepAlived.sh.j2" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/services/galaxy-hos-nginx/tasks/main.yml" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/services/galaxy-hos-nginx/tasks/main.yml" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/services/galaxy-hos-nginx/templates/docker-compose.yml.j2" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/services/galaxy-hos-nginx/templates/docker-compose.yml.j2" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/services/galaxy-hos-service/tasks/load_hos.yml" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/services/galaxy-hos-service/tasks/load_hos.yml" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/services/galaxy-hos-service/templates/application.yml.j2" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/services/galaxy-hos-service/templates/application.yml.j2" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/services/galaxy-job-service/files/mysql" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/services/galaxy-job-service/files/mysql" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/services/galaxy-job-service/tasks/load_admin.yml" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/services/galaxy-job-service/tasks/load_admin.yml" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/services/galaxy-job-service/tasks/load_job.yml" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/services/galaxy-job-service/tasks/load_job.yml" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/services/galaxy-job-service/templates/docker-compose.yml.j2" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/services/galaxy-job-service/templates/docker-compose.yml.j2" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/services/galaxy-qgw-service/files/dat/ip_v4.mmdb" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/services/galaxy-qgw-service/files/dat/ip_v4.mmdb" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/services/galaxy-qgw-service/tasks/load_qgw.yml" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/services/galaxy-qgw-service/tasks/load_qgw.yml" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/services/galaxy-qgw-service/templates/push_config.sh.j2" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/services/galaxy-qgw-service/templates/push_config.sh.j2" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/roles/services/galaxy-report-service/tasks/load_report.yml" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/roles/services/galaxy-report-service/tasks/load_report.yml" afterDir="false" />
<change beforePath="$PROJECT_DIR$/parcels/test1.yml" beforeDir="false" afterPath="$PROJECT_DIR$/parcels/test1.yml" afterDir="false" />
</list>
<option name="EXCLUDED_CONVERTED_TO_IGNORED" value="true" />
<option name="SHOW_DIALOG" value="false" />
@@ -16,44 +89,81 @@
<file pinned="false" current-in-tab="false">
<entry file="file://$PROJECT_DIR$/install.sh">
<provider selected="true" editor-type-id="text-editor">
<state relative-caret-position="301">
<caret line="651" column="19" selection-start-line="651" selection-start-column="19" selection-end-line="651" selection-end-column="19" />
<state relative-caret-position="420">
<caret line="28" column="15" selection-start-line="28" selection-start-column="15" selection-end-line="28" selection-end-column="15" />
</state>
</provider>
</entry>
</file>
<file pinned="false" current-in-tab="false">
<entry file="file://$PROJECT_DIR$/configurations/components.yml">
<provider selected="true" editor-type-id="text-editor">
<state relative-caret-position="412">
<caret line="146" column="26" selection-start-line="146" selection-start-column="26" selection-end-line="146" selection-end-column="26" />
</state>
</provider>
</entry>
</file>
<file pinned="false" current-in-tab="true">
<entry file="file://$PROJECT_DIR$/parcels/roles/components/clickhouse/common/templates/config.xml.j2">
<provider selected="true" editor-type-id="text-editor">
<state relative-caret-position="570">
<caret line="379" column="24" lean-forward="true" selection-start-line="379" selection-start-column="24" selection-end-line="379" selection-end-column="24" />
</state>
</provider>
<entry file="file://$PROJECT_DIR$/parcels/Clickhouse.yml">
<provider selected="true" editor-type-id="text-editor" />
</entry>
</file>
<file pinned="false" current-in-tab="false">
<entry file="file://$PROJECT_DIR$/parcels/roles/components/clickhouse/common/tasks/install_clickhouse.yml">
<provider selected="true" editor-type-id="text-editor">
<state relative-caret-position="292">
<caret line="66" column="63" selection-start-line="66" selection-start-column="44" selection-end-line="66" selection-end-column="63" />
<state relative-caret-position="15">
<caret line="1" column="6" selection-start-line="1" selection-start-column="2" selection-end-line="1" selection-end-column="6" />
</state>
</provider>
</entry>
</file>
<file pinned="false" current-in-tab="false">
<entry file="file://$PROJECT_DIR$/parcels/roles/components/clickhouse/data/tasks/install_clickhouse.yml">
<entry file="file://$PROJECT_DIR$/parcels/roles/components/mariadb/install-mariadb-cluster/tasks/install-mariadb.yml">
<provider selected="true" editor-type-id="text-editor">
<state relative-caret-position="578">
<caret line="40" column="57" selection-start-line="40" selection-start-column="57" selection-end-line="40" selection-end-column="57" />
<state relative-caret-position="1590">
<caret line="106" column="39" selection-start-line="106" selection-start-column="39" selection-end-line="106" selection-end-column="39" />
</state>
</provider>
</entry>
</file>
<file pinned="false" current-in-tab="true">
<entry file="file://$PROJECT_DIR$/parcels/roles/components/flink/tasks/install_flink.yml">
<provider selected="true" editor-type-id="text-editor">
<state relative-caret-position="104">
<caret line="112" column="13" lean-forward="true" selection-start-line="112" selection-start-column="13" selection-end-line="112" selection-end-column="13" />
</state>
</provider>
</entry>
</file>
<file pinned="false" current-in-tab="false">
<entry file="file://$PROJECT_DIR$/parcels/Arangodb.yml">
<provider selected="true" editor-type-id="text-editor">
<state relative-caret-position="75">
<caret line="5" column="16" selection-end-line="7" selection-end-column="45" />
</state>
</provider>
</entry>
</file>
<file pinned="false" current-in-tab="false">
<entry file="file://$PROJECT_DIR$/configurations/config.yml">
<provider selected="true" editor-type-id="text-editor">
<state relative-caret-position="285">
<caret line="19" column="4" selection-start-line="18" selection-end-line="20" />
</state>
</provider>
</entry>
</file>
<file pinned="false" current-in-tab="false">
<entry file="file://$PROJECT_DIR$/parcels/roles/services/galaxy-hos-keepalive/tasks/main.yml">
<provider selected="true" editor-type-id="text-editor" />
</entry>
</file>
<file pinned="false" current-in-tab="false">
<entry file="file://$PROJECT_DIR$/parcels/roles/services/galaxy-gateway-nginx/tasks/main.yml">
<provider selected="true" editor-type-id="text-editor">
<state relative-caret-position="90">
<caret line="6" column="17" selection-start-line="6" selection-start-column="11" selection-end-line="6" selection-end-column="17" />
</state>
</provider>
</entry>
</file>
<file pinned="false" current-in-tab="false">
<entry file="file://$PROJECT_DIR$/parcels/roles/services/galaxy-gateway-keepalive/tasks/main.yml">
<provider selected="true" editor-type-id="text-editor">
<state relative-caret-position="30">
<caret line="2" column="177" selection-start-line="2" selection-start-column="177" selection-end-line="2" selection-end-column="177" />
</state>
</provider>
</entry>
@@ -64,6 +174,21 @@
<findStrings>
<find>clickhouse_ssd_path</find>
<find>pol</find>
<find>home</find>
<find>abnorma</find>
<find>abnormalCombination</find>
<find>checkManager</find>
<find>checkCluster</find>
<find>installcombination</find>
<find>yn</find>
<find>common_list</find>
<find>set</find>
<find>file</find>
<find>failed</find>
<find>run_</find>
<find>or</find>
<find>defi</find>
<find>absent</find>
</findStrings>
</component>
<component name="Git.Settings">
@@ -76,21 +201,24 @@
<option value="$PROJECT_DIR$/README.md" />
<option value="$PROJECT_DIR$/configurations/config.yml" />
<option value="$PROJECT_DIR$/configurations/components.yml" />
<option value="$PROJECT_DIR$/parcels/roles/components/clickhouse/common/tasks/install_clickhouse.yml" />
<option value="$PROJECT_DIR$/parcels/roles/components/clickhouse/data/tasks/install_clickhouse.yml" />
<option value="$PROJECT_DIR$/parcels/roles/components/clickhouse/common/templates/config.xml.j2" />
<option value="$PROJECT_DIR$/parcels/roles/services/galaxy-gateway-keepalive/tasks/main.yml" />
<option value="$PROJECT_DIR$/parcels/roles/components/clickhouse/common/tasks/install_clickhouse.yml" />
</list>
</option>
</component>
<component name="ProjectFrameBounds" extendedState="6" fullScreen="true">
<option name="width" value="2560" />
<option name="height" value="1440" />
<component name="ProjectFrameBounds">
<option name="y" value="25" />
<option name="width" value="1440" />
<option name="height" value="810" />
</component>
<component name="ProjectView">
<navigator proportions="" version="1">
<foldersAlwaysOnTop value="true" />
</navigator>
<panes>
<pane id="Scope" />
<pane id="ProjectPane">
<subPane>
<expand>
@@ -103,11 +231,45 @@
<item name="Galaxy-auto-deploy-cluster" type="462c0819:PsiDirectoryNode" />
<item name="configurations" type="462c0819:PsiDirectoryNode" />
</path>
<path>
<item name="Galaxy-auto-deploy-cluster" type="b2602c69:ProjectViewProjectNode" />
<item name="Galaxy-auto-deploy-cluster" type="462c0819:PsiDirectoryNode" />
<item name="parcels" type="462c0819:PsiDirectoryNode" />
</path>
<path>
<item name="Galaxy-auto-deploy-cluster" type="b2602c69:ProjectViewProjectNode" />
<item name="Galaxy-auto-deploy-cluster" type="462c0819:PsiDirectoryNode" />
<item name="parcels" type="462c0819:PsiDirectoryNode" />
<item name="roles" type="462c0819:PsiDirectoryNode" />
</path>
<path>
<item name="Galaxy-auto-deploy-cluster" type="b2602c69:ProjectViewProjectNode" />
<item name="Galaxy-auto-deploy-cluster" type="462c0819:PsiDirectoryNode" />
<item name="parcels" type="462c0819:PsiDirectoryNode" />
<item name="roles" type="462c0819:PsiDirectoryNode" />
<item name="components" type="462c0819:PsiDirectoryNode" />
</path>
<path>
<item name="Galaxy-auto-deploy-cluster" type="b2602c69:ProjectViewProjectNode" />
<item name="Galaxy-auto-deploy-cluster" type="462c0819:PsiDirectoryNode" />
<item name="parcels" type="462c0819:PsiDirectoryNode" />
<item name="roles" type="462c0819:PsiDirectoryNode" />
<item name="components" type="462c0819:PsiDirectoryNode" />
<item name="flink" type="462c0819:PsiDirectoryNode" />
</path>
<path>
<item name="Galaxy-auto-deploy-cluster" type="b2602c69:ProjectViewProjectNode" />
<item name="Galaxy-auto-deploy-cluster" type="462c0819:PsiDirectoryNode" />
<item name="parcels" type="462c0819:PsiDirectoryNode" />
<item name="roles" type="462c0819:PsiDirectoryNode" />
<item name="components" type="462c0819:PsiDirectoryNode" />
<item name="flink" type="462c0819:PsiDirectoryNode" />
<item name="tasks" type="462c0819:PsiDirectoryNode" />
</path>
</expand>
<select />
</subPane>
</pane>
<pane id="Scope" />
</panes>
</component>
<component name="PropertiesComponent">
@@ -146,7 +308,10 @@
<workItem from="1629854869947" duration="503000" />
<workItem from="1630466216768" duration="33000" />
<workItem from="1630466362401" duration="1497000" />
<workItem from="1630476153920" duration="1245000" />
<workItem from="1630476153920" duration="3228000" />
<workItem from="1630891852958" duration="429000" />
<workItem from="1630914648031" duration="4146000" />
<workItem from="1631066183840" duration="1779000" />
</task>
<task id="LOCAL-00001" summary="1. 新增自动检测安装ansible(本地yum源的方式实现)&#10;2. 抽取后的变量放回各自的roles中(因不想对客户暴露)&#10;3. clickhous初始化脚本关于bifang-maridb的部分变量化&#10;4. 修复一些小bug">
<created>1629774644583</created>
@@ -183,17 +348,24 @@
<option name="project" value="LOCAL" />
<updated>1630481026679</updated>
</task>
<option name="localTasksCounter" value="6" />
<task id="LOCAL-00006" summary="clickhouse配置文件更改&#10;1. 去掉冷热数据配置">
<created>1630481332070</created>
<option name="number" value="00006" />
<option name="presentableId" value="LOCAL-00006" />
<option name="project" value="LOCAL" />
<updated>1630481332070</updated>
</task>
<option name="localTasksCounter" value="7" />
<servers />
</component>
<component name="TimeTrackingManager">
<option name="totallyTimeSpent" value="5797000" />
<option name="totallyTimeSpent" value="14134000" />
</component>
<component name="ToolWindowManager">
<frame x="0" y="0" width="2560" height="1440" extended-state="0" />
<frame x="0" y="25" width="1440" height="810" extended-state="0" />
<editor active="true" />
<layout>
<window_info active="true" content_ui="combo" id="Project" order="0" visible="true" weight="0.26370135" />
<window_info active="true" content_ui="combo" id="Project" order="0" visible="true" weight="0.2746781" />
<window_info id="Structure" order="1" side_tool="true" weight="0.25" />
<window_info id="Favorites" order="2" side_tool="true" />
<window_info anchor="bottom" id="Message" order="0" />
@@ -204,7 +376,7 @@
<window_info anchor="bottom" id="Inspection" order="5" weight="0.4" />
<window_info anchor="bottom" id="TODO" order="6" />
<window_info anchor="bottom" id="Docker" order="7" show_stripe_button="false" />
<window_info anchor="bottom" id="Version Control" order="8" visible="true" weight="0.30145985" />
<window_info anchor="bottom" id="Version Control" order="8" visible="true" weight="0.3005618" />
<window_info anchor="bottom" id="Database Changes" order="9" />
<window_info anchor="bottom" id="Event Log" order="10" side_tool="true" />
<window_info anchor="bottom" id="Terminal" order="11" weight="0.329927" />
@@ -227,28 +399,6 @@
<option name="LAST_COMMIT_MESSAGE" value="clickhouse配置文件更改&#10;1. 去掉冷热数据配置" />
</component>
<component name="editorHistoryManager">
<entry file="file://$PROJECT_DIR$/README.md" />
<entry file="file://$PROJECT_DIR$/configurations/config.yml">
<provider selected="true" editor-type-id="text-editor">
<state relative-caret-position="472">
<caret line="64" lean-forward="true" selection-start-line="64" selection-end-line="64" />
</state>
</provider>
</entry>
<entry file="file://$PROJECT_DIR$/install.sh">
<provider selected="true" editor-type-id="text-editor">
<state relative-caret-position="301">
<caret line="651" column="19" selection-start-line="651" selection-start-column="19" selection-end-line="651" selection-end-column="19" />
</state>
</provider>
</entry>
<entry file="file://$PROJECT_DIR$/parcels/roles/components/clickhouse/common/tasks/install_clickhouse.yml">
<provider selected="true" editor-type-id="text-editor">
<state relative-caret-position="292">
<caret line="66" column="63" selection-start-line="66" selection-start-column="44" selection-end-line="66" selection-end-column="63" />
</state>
</provider>
</entry>
<entry file="file://$PROJECT_DIR$/parcels/roles/components/clickhouse/data/tasks/install_clickhouse.yml">
<provider selected="true" editor-type-id="text-editor">
<state relative-caret-position="578">
@@ -256,13 +406,6 @@
</state>
</provider>
</entry>
<entry file="file://$PROJECT_DIR$/configurations/components.yml">
<provider selected="true" editor-type-id="text-editor">
<state relative-caret-position="412">
<caret line="146" column="26" selection-start-line="146" selection-start-column="26" selection-end-line="146" selection-end-column="26" />
</state>
</provider>
</entry>
<entry file="file://$PROJECT_DIR$/parcels/roles/components/clickhouse/common/templates/config.xml.j2">
<provider selected="true" editor-type-id="text-editor">
<state relative-caret-position="570">
@@ -270,5 +413,84 @@
</state>
</provider>
</entry>
<entry file="file://$PROJECT_DIR$/configurations/components.yml">
<provider selected="true" editor-type-id="text-editor">
<state relative-caret-position="412">
<caret line="146" column="26" selection-start-line="146" selection-start-column="26" selection-end-line="146" selection-end-column="26" />
</state>
</provider>
</entry>
<entry file="file://$PROJECT_DIR$/README.md">
<provider selected="true" editor-type-id="split-provider[text-editor;markdown-preview-editor]">
<state split_layout="SPLIT">
<first_editor relative-caret-position="105">
<caret line="7" selection-start-line="7" selection-end-line="7" />
</first_editor>
<second_editor />
</state>
</provider>
</entry>
<entry file="file://$PROJECT_DIR$/install.sh">
<provider selected="true" editor-type-id="text-editor">
<state relative-caret-position="420">
<caret line="28" column="15" selection-start-line="28" selection-start-column="15" selection-end-line="28" selection-end-column="15" />
</state>
</provider>
</entry>
<entry file="file://$PROJECT_DIR$/parcels/Clickhouse.yml">
<provider selected="true" editor-type-id="text-editor" />
</entry>
<entry file="file://$PROJECT_DIR$/parcels/roles/components/clickhouse/common/tasks/install_clickhouse.yml">
<provider selected="true" editor-type-id="text-editor">
<state relative-caret-position="15">
<caret line="1" column="6" selection-start-line="1" selection-start-column="2" selection-end-line="1" selection-end-column="6" />
</state>
</provider>
</entry>
<entry file="file://$PROJECT_DIR$/parcels/roles/components/mariadb/install-mariadb-cluster/tasks/install-mariadb.yml">
<provider selected="true" editor-type-id="text-editor">
<state relative-caret-position="1590">
<caret line="106" column="39" selection-start-line="106" selection-start-column="39" selection-end-line="106" selection-end-column="39" />
</state>
</provider>
</entry>
<entry file="file://$PROJECT_DIR$/parcels/Arangodb.yml">
<provider selected="true" editor-type-id="text-editor">
<state relative-caret-position="75">
<caret line="5" column="16" selection-end-line="7" selection-end-column="45" />
</state>
</provider>
</entry>
<entry file="file://$PROJECT_DIR$/configurations/config.yml">
<provider selected="true" editor-type-id="text-editor">
<state relative-caret-position="285">
<caret line="19" column="4" selection-start-line="18" selection-end-line="20" />
</state>
</provider>
</entry>
<entry file="file://$PROJECT_DIR$/parcels/roles/services/galaxy-hos-keepalive/tasks/main.yml">
<provider selected="true" editor-type-id="text-editor" />
</entry>
<entry file="file://$PROJECT_DIR$/parcels/roles/services/galaxy-gateway-keepalive/tasks/main.yml">
<provider selected="true" editor-type-id="text-editor">
<state relative-caret-position="30">
<caret line="2" column="177" selection-start-line="2" selection-start-column="177" selection-end-line="2" selection-end-column="177" />
</state>
</provider>
</entry>
<entry file="file://$PROJECT_DIR$/parcels/roles/services/galaxy-gateway-nginx/tasks/main.yml">
<provider selected="true" editor-type-id="text-editor">
<state relative-caret-position="90">
<caret line="6" column="17" selection-start-line="6" selection-start-column="11" selection-end-line="6" selection-end-column="17" />
</state>
</provider>
</entry>
<entry file="file://$PROJECT_DIR$/parcels/roles/components/flink/tasks/install_flink.yml">
<provider selected="true" editor-type-id="text-editor">
<state relative-caret-position="104">
<caret line="112" column="13" lean-forward="true" selection-start-line="112" selection-start-column="13" selection-end-line="112" selection-end-column="13" />
</state>
</provider>
</entry>
</component>
</project>

View File

@@ -122,6 +122,12 @@ topic_file_bytes: 1073741824
#缓冲池字节大小 单位单位MB
mariadb_innodb_buffer_pool_size: 1024
#mariadb 端口
galaxy_mariadb_port: 3306
#mariadb 用户
galaxy_mariadb_username: root
#galaxy mariadb key
galaxy_mariadb_pin: galaxy2019
@@ -134,7 +140,10 @@ spark_worker_cores: 30
#===========================Nacos===============================#
#Nacos 内存配置
nacos_java_opt: '-Xms1024m -Xmx1024m -Xmn1024m'
nacos_java_opt: '-Xms256m -Xmx256m -Xmn256m'
#Galaxy的 Nacos pin,(默认账户为:nacos)用于galaxy-hos-service, galaxy-job-service, galaxy-qgw-service, galaxy-report-service连接nacos获取配置信息
nacos_pin: nacos
#===========================Storm===============================#
#单个supervisor可使用worker数量一般为CPU的一半。
@@ -149,25 +158,18 @@ storm_worker_min_mem: 1024
#===========================Flink================================#
#网络缓存大小
taskmanager_memory_network_min: 512m
taskmanager_memory_network_max: 1280m
taskmanager_memory_network_max: 512m
#taskmanager堆外内存
taskmanager_memory_managed_size: 256m
#TaskManager进程占用的所有与Flink相关的内存
taskmanager_memory_flink_size: 5120m
taskmanager_memory_flink_size: 3072m
#JobManager进程占用的所有与Flink相关的内存
jobmanager_memory_flink_size: 1024m
jobmanager_memory_flink_size: 512m
#===========================Zookeeper===============================#
#zookeeper 进程 启动内存大小 单位MB
zookeeper_max_mem: 1024
#===========================Mariadb===============================#
#mariadb 端口
galaxy_mariadb_port: 3306
#mariadb 用户
galaxy_mariadb_username: root

View File

@@ -22,7 +22,7 @@ package_path: "{{ playbook_dir | dirname }}/software_packages"
#============================bifang-Mariadb===============================#
#Bifang MariaDB 地址用于galaxy-qgw-service、galaxy-report-service、flink-dos的业务使用。
bifang_mariadb_host: 192.168.40.153
bifang_mariadb_host: 127.0.0.1
#Bifang mariadb 端口
bifang_mariadb_port: 3306
@@ -31,14 +31,14 @@ bifang_mariadb_port: 3306
bifang_mariadb_database: tsg-bifang
#Bifang mariadb root's pin
bifang_mariadb_pin: 111111
bifang_mariadb_pin: ******
#============================虚IP===============================#
#-------------apps------------------#
#Galaxy 业务组件keepalive IP地址,填写的IP只能是和hosts-services配置的IP是同段的且不能跟同段内已有的IP相同。
gateway_keepalive_host: 192.168.45.252
gateway_keepalive_host: 127.0.0.*
#Galaxy 业务组件keepalive绑定网卡名称不支持两个服务器网卡名称不同的情况此情况需要安装完成后手动修改配置文件
gateway_keepalive_interface: eth0
@@ -53,7 +53,7 @@ hos_keepalive_need: "yes"
#galaxy-hos-service服务keepalive IP地址 若 hos_keepalive_need为no则此处填写gateway_keepalive_host对应的IP
#填写的IP只能是和hosts-hoskeepalive配置的IP是同段的且不能跟同段内已有的IP相同。
hos_keepalive_host: 192.168.45.253
hos_keepalive_host: 127.0.0.*
#galaxy-hos-service服务keepalive绑定网卡名称不支持两个服务器网卡名称不同的情况此情况需要安装完成后手>动修改配置文件
#若 hos_keepalive_need 配置no则此处 可不进行修改
@@ -65,18 +65,15 @@ hos_keepalive_router_id: 62
#============================Kafka===============================#
#任务消费kafka地址,若无特殊情况为本安装集群kafka地址格式:kafkaip1:9092,kafkaip2:9092,kafkaip3:9092.....
kafka_source_servers: 192.168.45.37:9092,192.168.45.38:9092,192.168.45.39:9092
kafka_source_servers: 127.0.0.1:9092
#任务结果写入kafka地址,若无特殊情况为本安装集群kafka地址格式:kafkaip1:9092,kafkaip2:9092,kafkaip3:9092.....
kafka_sink_servers: 192.168.45.37:9092,192.168.45.38:9092,192.168.45.39:9092
kafka_sink_servers: 127.0.0.1:9092
#============================默认配置===============================#
#自研服务组件使用nacos 命名空间名称
services_config_namespace: prod
#Galaxy的 Nacos pin
nacos_pin: nacos
#Nacos 使用的mariadb 数据库名称
mariadb_nacos_database: nacos

View File

@@ -1,97 +1,54 @@
#此标签指定的IP用于chproxy、galaxy-app-nginx、galaxy-job-service、galaxy-qgw-service、galaxy-report-service、galaxy-app-keepalive的安装
#最少需要指定两台。
[services]
192.168.45.42
192.168.45.43
#此标签指定的IP用于galaxy-hos-service 的安装
[hos]
192.168.45.40
192.168.45.41
#此标签指定的IP用于galaxy-hos-keepalive和galaxy-hos-nginx的安装
#inventories/pro/group_vars/all内的hos_keepalive_need配置是no则此处不写任何IPyes就写部署galaxy-hos-service服务keepalive的服务器地址。
#不能与services指定的IP相同,且最多为两台。
[hoskeepalive]
192.168.45.40
192.168.45.41
#此标签指定的IP用于ArangoDB的安装
[arangodb]
192.168.45.42
#此标签指定的IP用于Mariadb的安装目前版本下仅可配置两个IP且需要与services标签的前两个IP相同。
#现没有负载需求,保留后期做负载的功能需求。
[mariadb]
192.168.45.42
192.168.45.43
#此标签指定的IP用于Nacos的安装
[nacos]
192.168.45.37
192.168.45.38
192.168.45.39
#此标签指定的IP用于Apache Zookeeper的安装
[zookeeper]
192.168.45.37
192.168.45.38
192.168.45.39
#此标签指定的IP用于Apache Kafka的安装集群第一台同时部署Kafka-Manager服务。
[kafka]
192.168.45.37
192.168.45.38
192.168.45.39
#此标签指定的IP用于Apache Storm的安装根据内部约定若集群整体小于3台则第一台为numbus节点大于三台则前三台为nimbus节点。
#[storm]
#192.168.45.37
#192.168.45.38
#192.168.45.39
#此标签指定的IP用于Apache Flink的安装根据内部约定前两台台为master节点所有节点为worker节点。
[flink]
192.168.45.37
192.168.45.38
192.168.45.39
#此标签指定的IP用于Apache Hadoop的安装
[hadoop]
192.168.45.37
192.168.45.38
192.168.45.39
#此标签指定的IP用于Apache HBase的安装
[hbase]
192.168.45.37
192.168.45.38
192.168.45.39
#此标签指定的IP用于Apache Spark的安装根据内部约定若集群小于3台则worker在所有节点上大于3台则第一个节点为master节点。
[spark]
192.168.45.37
192.168.45.38
192.168.45.39
#此标签指定的IP用于Galaxy-Gohangout的安装
[gohangout]
192.168.45.40
#此标签指定的IP用于Apache Druid的安装根据内部约定若集群整体小于4台则每台为全部节点大于4台则前两台默认为query节点。
[druid]
192.168.45.37
192.168.45.38
192.168.45.39
#此标签指定的IP用于Clickhouse的安装,根据内部约定前两台默认为query节点。
[clickhouse]
192.168.45.40
192.168.45.41
192.168.45.42
192.168.45.43
#dos检测的离线generate-baseline程序
[dos_baseline]
192.168.45.43

View File

@@ -11,8 +11,8 @@ qgw_java_opts: "-Xmx512m -Xms512m"
#=======================galaxy-job-service==========================#
#galaxy-job-service 服务内存设置
job_java_opts: "-Xmx512m -Xms512m"
job_java_opts: "-Xmx128m -Xms128m"
#======================galaxy-report-service=========================#
#galaxy-report-service 服务内存设置
report_java_opts: "-Xmx512m -Xms512m"
report_java_opts: "-Xmx128m -Xms128m"

View File

@@ -1,6 +1,5 @@
#!/bin/bash
#######################基础变量######################
#脚本目录
bin_path=$(cd `dirname $0`; pwd)
@@ -195,7 +194,7 @@ clear
i=0
#ins_names=(Zookeeper Mariadb galaxy-gateway-keepalive Nacos Kafka Storm Hadoop HBase Clickhouse Druid Spark Arangodb)
#替换Storm成Flink
ins_names=(Zookeeper Mariadb galaxy-gateway-keepalive Nacos Kafka Flink Hadoop HBase Clickhouse Druid Spark Arangodb)
ins_names=(Zookeeper Mariadb galaxy-gateway-keepalive Nacos Kafka Hadoop HBase Flink Clickhouse Druid Spark Arangodb)
echo ${ins_names[@]} > $NUM_DIR/common_list
#根据组合进行顺序安装
@@ -338,9 +337,9 @@ clear
i=0
hoskeep=`cat ../configurations/config.yml | grep -vE "^#|^$" | grep "hos_keepalive_need" | grep yes | wc -l`
if [[ $hoskeep -eq "1" ]]; then
ins_names=(galaxy-gateway-nginx galaxy-qgw-service galaxy-job-service galaxy-report-service galaxy-hos-service galaxy-chproxy galaxy-hos-nginx galaxy-hos-keepalive galaxy-gohangout)
ins_names=(galaxy-gateway-nginx galaxy-chproxy galaxy-qgw-service galaxy-job-service galaxy-report-service galaxy-hos-nginx galaxy-hos-keepalive galaxy-hos-service galaxy-gohangout)
else
ins_names=(galaxy-gateway-nginx galaxy-qgw-service galaxy-job-service galaxy-report-service galaxy-hos-service galaxy-chproxy galaxy-gohangout)
ins_names=(galaxy-gateway-nginx galaxy-chproxy galaxy-qgw-service galaxy-job-service galaxy-report-service galaxy-hos-service galaxy-gohangout)
fi
echo ${ins_names[@]} > $NUM_DIR/app_list
@@ -648,9 +647,31 @@ fi
clear
cat ../parcels/menu/homePage
if [ -f $NUM_DIR/stepNum ]; then
step_num=`cat $NUM_DIR/stepNum`
next_step_num=`expr ${step_num} + 1`
#if [ ${next_step_num} -eq '5' ]; then
# echo -e "\033[33m You have successfully installed, exiting ! ! \033[0m"
# break;
#fi
#if [ ${next_step_num} -ne ${yn_main} ]; then
# #echo "Then next step should be ${next_step_num}"
# echo -e "\033[33m Then next step should be ${next_step_num} \033[0m"
# sleep 3s
#continue
#fi
echo -e "\033[33m Then next step should be ${next_step_num}, but you can chose other number of step if you want ! ! \033[0m"
else
echo -e "\033[33m Then next step should be 1, but you can chose other number of step if you want ! ! \033[0m"
#sleep 3s
#continue
fi
read -p "
Selection(1-4)? " yn
case $yn in
Selection(1-4)? " yn_main
case $yn_main in
[1])
while true; do
clear
@@ -664,39 +685,42 @@ echo " *
*********************************************************************************************************
"
read -p "Enter [Y] or [N] " yn
read -p "Enter [yY] or [nN] " yn
case $yn in
[Yy]* )
installcombination
echo ${yn_main} > $NUM_DIR/stepNum
break;;
[Nn]*)
break;;
* )
echo "Please Enter [Y] or [N].";;
echo "Please Enter [yY] or [nN].";;
esac
done;;
[2] )
while true; do
clear
#installApps
cat ../parcels/menu/appMenu
read -p "Enter [Y] or [N] " yn
read -p "Enter [yY] or [nN] " yn
case $yn in
[Yy]* )
installApps
echo ${yn_main} > $NUM_DIR/stepNum
break;;
[Nn]* )
break;;
* )
echo "Please Enter [Y] or [N].";;
echo "Please Enter [yY] or [nN].";;
esac
done;;
[3] )
installInit
echo ${yn_main} > $NUM_DIR/stepNum
;;
[4] )
installCheck
echo ${yn_main} > $NUM_DIR/stepNum
;;
* )
echo "Please Enter (1-4)."

View File

@@ -1,4 +1,5 @@
- hosts: services
- hosts:
- hos
remote_user: root
roles:
- init/galaxy-hos-service

View File

@@ -1,17 +1,16 @@
Service Components
┌──────────────────────────────────────
galaxy-qgw-service
galaxy-job-service
galaxy-report-service ┤
galaxy-hos-service
galaxy-chproxy
galaxy-gohangout
galaxy-gateway-nginx
galaxy-nginx-hos
galaxy-gateway-keepalive 
├ galaxy-hos-keepalive ┤
└──────────────────────────────────────┘  
┌───────────────────────────────────┐
[*] galaxy-gateway-nginx
[*] galaxy-chproxy
[*] galaxy-qgw-service ┤
[*] galaxy-job-service
[*] galaxy-report-service
[*] galaxy-hos-nginx
[*] galaxy-hos-keepalive
[*] galaxy-hos-service
[*] galaxy-gohangout
└───────────────────────────────────┘  
*********************************************************
* Press Ctrl+C or N to exit, Enter or Y to continue. *

View File

@@ -5,9 +5,9 @@ Open Source Software
    [*] Mariadb        
    [*] Nacos ┤
    [*] Apache Kafka ┤
├ [*] Apache Flink ┤
├ [*] Apache Hadoop ┤
├ [*] Apache HBase ┤
├ [*] Apache Flink ┤
├ [*] Clickhouse ┤
├ [*] Apache Druid ┤
├ [*] Apache Spark ┤

View File

@@ -1,51 +1,51 @@
- name: Checking Components Heartbeats
shell: "curl -s http://{{ inventory_hostname }}:8183/monitor/health | grep DOWN | grep -v grep | wc -l"
shell: "curl --retry-delay 3 --retry 3 -s http://{{ inventory_hostname }}:8183/monitor/health | grep DOWN | grep -v grep | wc -l"
register: health
- name: Checking Components Heartbeats
fail:
msg: "组件心跳检测异常,请通过接口 curl -s http://{{ inventory_hostname }}:8183/monitor/health 检查具体DOWN组件。"
msg: "组件心跳检测异常,请通过接口 curl --retry-delay 3 --retry 3 -s http://{{ inventory_hostname }}:8183/monitor/health 检查具体DOWN组件。"
when: health.stdout != '0'
- name: Checking Metadata
shell: "curl -s http://{{ inventory_hostname }}:8183/diagnosis/metadata | grep '\"status\":200' | grep -v grep | wc -l"
shell: "curl --retry-delay 3 --retry 3 -s http://{{ inventory_hostname }}:8183/diagnosis/metadata | grep '\"status\":200' | grep -v grep | wc -l"
register: metadata
- name: Checking Metadata
fail:
msg: "元数据验证异常,请通过接口 curl -s http://{{ inventory_hostname }}:8183/diagnosis/metadata 检查具体异常信息。"
msg: "元数据验证异常,请通过接口 curl --retry-delay 3 --retry 3 -s http://{{ inventory_hostname }}:8183/diagnosis/metadata 检查具体异常信息。"
when: metadata.stdout != '1'
- name: Checking SQL
shell: "curl -s http://{{ inventory_hostname }}:8183/diagnosis/runSql | grep '\"status\":200' | grep -v grep | wc -l"
shell: "curl --retry-delay 3 --retry 3 -s http://{{ inventory_hostname }}:8183/diagnosis/runSql | grep '\"status\":200' | grep -v grep | wc -l"
register: sql
- name: Checking SQL
fail:
msg: "系统内置的SQL数据集执行异常请通过接口 curl -s http://{{ inventory_hostname }}:8183/diagnosis/runSql 检查具体异常信息。"
msg: "系统内置的SQL数据集执行异常请通过接口 curl --retry-delay 3 --retry 3 -s http://{{ inventory_hostname }}:8183/diagnosis/runSql 检查具体异常信息。"
when: sql.stdout != '1'
#TODO 因服务版本问题于21.07版本开启验证
#- name: Checking {{ groups.hos[0] }} galaxy-hos-service
# shell: "curl --location -s -w %{http_code} http://{{ groups.hos[0] }}:8186/monitor | grep 200 | wc -l"
# shell: "curl --retry-delay 3 --retry 3 --location -s -w %{http_code} http://{{ groups.hos[0] }}:8186/monitor | grep 200 | wc -l"
# register: monitor_out
# run_once: true
# delegate_facts: true
# delegate_to: '{{ groups.hos[0] }}'
#
#- fail:
# msg: "HOS Monitor检测异常,请通过接口 curl --location -s -w %{http_code} http://{{ groups.hos[0] }}:8186/monitor 检查具体异常信息"
# msg: "HOS Monitor检测异常,请通过接口 curl --retry-delay 3 --retry 3 --location -s -w %{http_code} http://{{ groups.hos[0] }}:8186/monitor 检查具体异常信息"
# when: monitor_out.stdout != '1'
#
#- name: Checking {{ groups.hos[1] }} galaxy-hos-service
# shell: "curl --location -s -w %{http_code} http://{{ groups.hos[1] }}:8186/monitor | grep 200 | wc -l"
# shell: "curl --retry-delay 3 --retry 3 --location -s -w %{http_code} http://{{ groups.hos[1] }}:8186/monitor | grep 200 | wc -l"
# register: monitor_out
# run_once: true
# delegate_facts: true
# delegate_to: '{{ groups.hos[1] }}'
#
#- fail:
# msg: "HOS Monitor检测异常,请通过接口 curl --location -s -w %{http_code} http://{{ groups.hos[1] }}:8186/monitor 检查具体异常信息"
# msg: "HOS Monitor检测异常,请通过接口 curl --retry-delay 3 --retry 3 --location -s -w %{http_code} http://{{ groups.hos[1] }}:8186/monitor 检查具体异常信息"
# when: monitor_out.stdout != '1'
- name: wait gohangout start ,sleep 30s

View File

@@ -368,7 +368,7 @@
-->
<format_schema_path>{{ install_path }}/clickhouse/format_schemas/</format_schema_path>
<!--
<!--
<storage_configuration>
<disks>
<ssd>

View File

@@ -37,9 +37,11 @@
dest: '{{ install_path }}/unload_druid.sh'
force: true
mode: 0755
when: check_out.stdout >= '1'
# when: check_out.stdout >= '1'
- name: unload Druid
shell: cd {{ install_path }} && sh unload_druid.sh
- name: unload Druid
shell: "if [ `ps aux | grep druid | grep -vE 'grep|ansible|install' | awk '{print $2}' | wc -l` -ne 0 ] ; then ps aux | grep druid | grep -vE 'grep|ansible|install' | awk '{print $2}' | xargs kill -9 ; fi"
when: check_out.stdout >= '1'
- name: check if {{ mariadb_druid_database }} exist
shell: mysql -s -uroot -p{{ galaxy_mariadb_pin }} -P3306 -h{{ gateway_keepalive_host }} -e "select count(1) from information_schema.schemata where schema_name='{{ mariadb_druid_database }}'" |grep 1|wc -l
@@ -228,8 +230,9 @@
dest: '{{ install_path }}/{{ druid_version }}/bin/dae-druid-data.sh'
mode: 0755
backup: yes
when: node_nums > cluster_limit and inventory_hostname != '{{master_ip}}' and inventory_hostname != '{{slave1_ip}}'
#when: node_nums > cluster_limit and inventory_hostname != '{{master_ip}}' and inventory_hostname != '{{slave1_ip}}'
when: node_nums > cluster_limit and inventory_hostname != master_ip and inventory_hostname != slave1_ip
#拷贝druid 配置文件
- name: copy Druid data node keep shell
template:
@@ -237,7 +240,8 @@
dest: '/etc/init.d/keepdruiddata'
mode: 0755
backup: yes
when: node_nums > cluster_limit and inventory_hostname != '{{master_ip}}' and inventory_hostname != '{{slave1_ip}}'
#when: node_nums > cluster_limit and inventory_hostname != '{{master_ip}}' and inventory_hostname != '{{slave1_ip}}'
when: node_nums > cluster_limit and inventory_hostname != master_ip and inventory_hostname != slave1_ip
#拷贝set_druid_env.sh
- name: copy set_druid_env.sh
@@ -253,7 +257,7 @@
#拷贝druid 配置文件
- name: start druid query node
shell: source /etc/profile && sh /etc/init.d/keepdruidquery 'start'
shell: source /etc/profile && /bin/bash /etc/init.d/keepdruidquery 'start'
run_once: true
delegate_facts: true
delegate_to: "{{ item.ip }}"
@@ -264,12 +268,13 @@
#拷贝druid 配置文件
- name: start druid data node
shell: source /etc/profile && sh /etc/init.d/keepdruiddata 'start'
when: node_nums > cluster_limit and inventory_hostname != '{{master_ip}}' and inventory_hostname != '{{slave1_ip}}'
shell: source /etc/profile && /bin/bash /etc/init.d/keepdruiddata 'start'
#when: node_nums > cluster_limit and inventory_hostname != '{{master_ip}}' and inventory_hostname != '{{slave1_ip}}'
when: node_nums > cluster_limit and inventory_hostname != master_ip and inventory_hostname != slave1_ip
#拷贝druid 配置文件
- name: start druid all node
shell: source /etc/profile && sh /etc/init.d/keepdruidall 'start'
shell: source /etc/profile && /bin/bash /etc/init.d/keepdruidall 'start'
when: node_nums <= cluster_limit
- name: Ansible delete {{ druid_version }}.tar.gz

View File

@@ -0,0 +1,10 @@
-server
-Xms{{ broker_mem }}m
-Xmx{{ broker_mem }}m
-XX:MaxDirectMemorySize={{ broker_MaxDirectMemorySize }}m
-Duser.timezone=UTC
-Dfile.encoding=UTF-8
-Djava.io.tmpdir=var/tmp
-Dlogfile.name=broker
-Djava.util.logging.manager=org.apache.logging.log4j.jul.LogManager

View File

@@ -0,0 +1,42 @@
#
# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
#
druid.service=druid/broker
druid.plaintextPort=8082
druid.sql.planner.sqlTimeZone={{ local_time }}
# HTTP server settings
druid.server.http.numThreads=60
# HTTP client settings
druid.broker.http.numConnections=50
druid.broker.http.maxQueuedBytes=10000000
# Processing threads and buffers
druid.processing.buffer.sizeBytes={{ broker_sizeBytes }}
druid.processing.numMergeBuffers={{ broker_numMergeBuffers }}
druid.processing.numThreads={{ broker_numThreads }}
druid.processing.tmpDir=var/druid/processing
# Query cache disabled -- push down caching and merging instead
druid.broker.cache.useCache=false
druid.broker.cache.populateCache=false
druid.query.groupBy.maxMergingDictionarySize=10000000000
druid.query.groupBy.maxOnDiskStorage=10000000000

View File

@@ -0,0 +1,151 @@
#
# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
#
# Extensions specified in the load list will be loaded by Druid
# We are using local fs for deep storage - not recommended for production - use S3, HDFS, or NFS instead
# We are using local derby for the metadata store - not recommended for production - use MySQL or Postgres instead
# If you specify `druid.extensions.loadList=[]`, Druid won't load any extension from file system.
# If you don't specify `druid.extensions.loadList`, Druid will load all the extensions under root extension directory.
# More info: https://druid.apache.org/docs/latest/operations/including-extensions.html
druid.extensions.loadList=["druid-hdfs-storage", "druid-kafka-indexing-service", "druid-datasketches", "mysql-metadata-storage", "druid-group-uniq-extension"]
# If you have a different version of Hadoop, place your Hadoop client jar files in your hadoop-dependencies directory
# and uncomment the line below to point to your directory.
#druid.extensions.hadoopDependenciesDir=/my/dir/hadoop-dependencies
#
# Hostname
#
druid.host={{ inventory_hostname }}
#
# Logging
#
# Log all runtime properties on startup. Disable to avoid logging properties on startup:
druid.startup.logging.logProperties=true
#
# Zookeeper
#
druid.zk.service.host={{ zookeeper_servers }}
druid.zk.paths.base=/druid
#
# Metadata storage
#
# For Derby server on your Druid Coordinator (only viable in a cluster with a single Coordinator, no fail-over):
#druid.metadata.storage.type=derby
#druid.metadata.storage.connector.connectURI=jdbc:derby://localhost:1527/var/druid/metadata.db;create=true
#druid.metadata.storage.connector.host=localhost
#druid.metadata.storage.connector.port=1527
# For MySQL (make sure to include the MySQL JDBC driver on the classpath):
druid.metadata.storage.type=mysql
druid.metadata.storage.connector.connectURI=jdbc:mysql://{{ gateway_keepalive_host }}:3306/{{ mariadb_druid_database }}
druid.metadata.storage.connector.user=root
druid.metadata.storage.connector.password={{ galaxy_mariadb_pin }}
# For PostgreSQL:
#druid.metadata.storage.type=postgresql
#druid.metadata.storage.connector.connectURI=jdbc:postgresql://db.example.com:5432/druid
#druid.metadata.storage.connector.user=...
#druid.metadata.storage.connector.password=...
#
# Deep storage
#
# For local disk (only viable in a cluster if this is a network mount):
#druid.storage.type=local
#druid.storage.storageDirectory=var/druid/segments
# For HDFS:
druid.storage.type=hdfs
druid.storage.storageDirectory=/druid/segments
# For S3:
#druid.storage.type=s3
#druid.storage.bucket=your-bucket
#druid.storage.baseKey=druid/segments
#druid.s3.accessKey=...
#druid.s3.secretKey=...
#
# Indexing service logs
#
# For local disk (only viable in a cluster if this is a network mount):
#druid.indexer.logs.type=file
#druid.indexer.logs.directory=var/druid/indexing-logs
# For HDFS:
druid.indexer.logs.type=hdfs
druid.indexer.logs.directory=/druid/indexing-logs
druid.indexer.logs.kill.enabled=true
druid.indexer.logs.kill.durationToRetain=604800000
druid.indexer.logs.kill.delay=21600000
# For S3:
#druid.indexer.logs.type=s3
#druid.indexer.logs.s3Bucket=your-bucket
#druid.indexer.logs.s3Prefix=druid/indexing-logs
#
# Service discovery
#
druid.selectors.indexing.serviceName=druid/overlord
druid.selectors.coordinator.serviceName=druid/coordinator
#
# Monitoring
#
druid.monitoring.monitors=["org.apache.druid.java.util.metrics.SysMonitor","org.apache.druid.java.util.metrics.JvmMonitor"]
druid.emitter=http
druid.emitter.logging.logLevel=info
druid.emitter.http.recipientBaseUrl=http://{{ inventory_hostname }}:9903
# Storage type of double columns
# ommiting this will lead to index double as float at the storage layer
druid.indexing.doubleStorage=double
#
# Security
#
druid.server.hiddenProperties=["druid.s3.accessKey","druid.s3.secretKey","druid.metadata.storage.connector.password"]
#
# SQL
#
druid.sql.enable=true
#
# Lookups
#
druid.lookup.enableLookupSyncOnStartup=false

View File

@@ -0,0 +1,11 @@
-server
-Xms{{ coordinator_mem }}m
-Xmx{{ coordinator_mem }}m
-XX:+UseG1GC
-Duser.timezone=UTC
-Dfile.encoding=UTF-8
-Djava.io.tmpdir=var/tmp
-Dlogfile.name=coordinator
-Djava.util.logging.manager=org.apache.logging.log4j.jul.LogManager
-Dderby.stream.error.file=var/druid/derby.log

View File

@@ -0,0 +1,31 @@
#!/bin/bash
source /etc/profile
BASH_DIR={{ install_path }}
VERSION={{ druid_version }}
while true ; do
metrics=`ps -ef | grep druid-metrics | grep -v grep | wc -l`
druid=`ps -ef | grep druid | grep -v grep | grep -v json | grep cluster.conf| wc -l`
if [ $druid -eq "0" ];then
setsid nohup $BASH_DIR/$VERSION/bin/start-cluster-all-server > /dev/null 2>&1 &
OLD_NUM=`cat $BASH_DIR/$VERSION/protect/restartsum/all-server`
RESTART_NUM=`expr $OLD_NUM + 1`
echo $RESTART_NUM > $BASh_DIR/$VERSION/protect/restartsum/all-server
echo "`date +%Y-%m-%d` `date +%H:%M:%S` - restart start-cluster-all-server - 重启次数 -> $RESTART_NUM." >> $BASH_DIR/$VERSION/protect/protecthdfs.log
fi
if [ $metrics -eq "0" ];then
nohup java -jar -Xmx1024m -Xms1024m $BASH_DIR/$VERSION/monitor/druid-metrics.jar 9903 > $BASH_DIR/$VERSION/monitor/metrics.log 2>&1 &
OLD_NUM=`cat $BASH_DIR/$VERSION/protect/restartsum/metrics`
RESTART_NUM=`expr $OLD_NUM + 1`
echo $RESTART_NUM > $BASH_DIR/$VERSION/protect/restartsum/metrics
echo "`date +%Y-%m-%d` `date +%H:%M:%S` - restart druid-metrics.jar - 重启次数 -> $RESTART_NUM." >> $BASH_DIR/$VERSION/protect/protecthdfs.log
fi
sleep 60
done

View File

@@ -0,0 +1,31 @@
#!/bin/bash
source /etc/profile
BASH_DIR={{ install_path }}
VERSION={{ druid_version }}
while true ; do
metrics=`ps -ef | grep druid-metrics | grep -v grep | wc -l`
druid=`ps -ef | grep druid | grep -v grep | grep -v json | grep data.conf | wc -l`
if [ $druid -eq "0" ];then
setsid nohup $BASH_DIR/$VERSION/bin/start-cluster-data-server > /dev/null 2>&1 &
OLD_NUM=`cat $BASH_DIR/$VERSION/protect/restartsum/data-server`
RESTART_NUM=`expr $OLD_NUM + 1`
echo $RESTART_NUM > $BASh_DIR/$VERSION/protect/restartsum/data-server
echo "`date +%Y-%m-%d` `date +%H:%M:%S` - restart start-cluster-data-server - 重启次数 -> $RESTART_NUM." >> $BASH_DIR/$VERSION/protect/protecthdfs.log
fi
if [ $metrics -eq "0" ];then
nohup java -jar -Xmx1024m -Xms1024m $BASH_DIR/$VERSION/monitor/druid-metrics.jar 9903 > $BASH_DIR/$VERSION/monitor/metrics.log 2>&1 &
OLD_NUM=`cat $BASH_DIR/$VERSION/protect/restartsum/metrics`
RESTART_NUM=`expr $OLD_NUM + 1`
echo $RESTART_NUM > $BASH_DIR/$VERSION/protect/restartsum/metrics
echo "`date +%Y-%m-%d` `date +%H:%M:%S` - restart druid-metrics.jar - 重启次数 -> $RESTART_NUM." >> $BASH_DIR/$VERSION/protect/protecthdfs.log
fi
sleep 60
done

View File

@@ -0,0 +1,31 @@
#!/bin/bash
source /etc/profile
BASH_DIR={{ install_path }}
VERSION={{ druid_version }}
while true ; do
metrics=`ps -ef | grep druid-metrics | grep -v grep | wc -l`
druid=`ps -ef | grep druid | grep -v grep | grep -v json | grep master-with-query.conf | wc -l`
if [ $druid -eq "0" ];then
setsid nohup $BASH_DIR/$VERSION/bin/start-cluster-query-server > /dev/null 2>&1 &
OLD_NUM=`cat $BASH_DIR/$VERSION/protect/restartsum/query-server`
RESTART_NUM=`expr $OLD_NUM + 1`
echo $RESTART_NUM > $BASh_DIR/$VERSION/protect/restartsum/query-server
echo "`date +%Y-%m-%d` `date +%H:%M:%S` - restart start-cluster-query-server - 重启次数 -> $RESTART_NUM." >> $BASH_DIR/$VERSION/protect/protecthdfs.log
fi
if [ $metrics -eq "0" ];then
nohup java -jar -Xmx1024m -Xms1024m $BASH_DIR/$VERSION/monitor/druid-metrics.jar 9903 > $BASH_DIR/$VERSION/monitor/metrics.log 2>&1 &
OLD_NUM=`cat $BASH_DIR/$VERSION/protect/restartsum/metrics`
RESTART_NUM=`expr $OLD_NUM + 1`
echo $RESTART_NUM > $BASH_DIR/$VERSION/protect/restartsum/metrics
echo "`date +%Y-%m-%d` `date +%H:%M:%S` - restart druid-metrics.jar - 重启次数 -> $RESTART_NUM." >> $BASH_DIR/$VERSION/protect/protecthdfs.log
fi
sleep 60
done

View File

@@ -0,0 +1,10 @@
-server
-Xms{{ historical_mem }}m
-Xmx{{ historical_mem }}m
-XX:MaxDirectMemorySize={{ historical_MaxDirectMemorySize }}m
-Duser.timezone=UTC
-Dfile.encoding=UTF-8
-Djava.io.tmpdir=var/tmp
-Dlogfile.name=historical
-Djava.util.logging.manager=org.apache.logging.log4j.jul.LogManager

View File

@@ -0,0 +1,45 @@
#
# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
#
druid.service=druid/historical
druid.plaintextPort=8083
# HTTP server threads
druid.server.http.numThreads=60
# Processing threads and buffers
druid.processing.buffer.sizeBytes={{ historical_buffer_sizeBytes }}
druid.processing.numMergeBuffers={{ historical_numMergeBuffers }}
druid.processing.numThreads={{ historical_numThreads }}
druid.processing.tmpDir=var/druid/processing
# Segment storage
#druid.segmentCache.locations=[{"path":"var/druid/segment-cache","maxSize":3000000000000}]
#druid.server.maxSize=35000000000000
druid.segmentCache.locations=[{"path":"var/druid/segment-cache","maxSize":{{ segmentCache_max_size }}}]
druid.server.maxSize={{ server_disk_maxsize }}
# Query cache
druid.historical.cache.useCache=true
druid.historical.cache.populateCache=true
druid.cache.type=caffeine
druid.cache.sizeInBytes=256000000
druid.query.groupBy.maxMergingDictionarySize=10000000000
druid.query.groupBy.maxOnDiskStorage=10000000000

View File

@@ -0,0 +1,43 @@
#!/bin/bash
#
# netconsole This loads the netconsole module with the configured parameters.
#
# chkconfig:123456 40 60
# description: keepdruidall
source /etc/profile
PRO_NAME=keepdruidall
INS_DIR={{ install_path }}
#版本
VERSION={{ druid_version }}
case $1 in
start)
master=`ps -ef | grep dae-druid-all.sh | grep -v grep | wc -l`
if [ $master -lt 1 ];then
nohup $INS_DIR/$VERSION/bin/dae-druid-all.sh > /dev/null 2>&1 &
fi
;;
stop)
ps -ef | grep dae-druid-all.sh | grep -v grep | awk '{print $2}' | xargs kill -9 2>/dev/null
#keeppid=`ps -ef | grep dae-druid-all.sh | grep -v grep | awk '{print $2}' | xargs kill -9 2>/dev/null`
#echo "守护进程PID$keeppid"
#kill -9 $keeppid
echo "关闭所有druid进程"
#ps aux | grep druid | grep -v grep | awk '{print $2}' | xargs kill -9
livenum=`ps -ef |grep "druid" | grep -v grep | grep -v json | wc -l`
if [ $livenum -ne 0 ];then
ps aux|grep "druid" |grep -v grep | awk '{print $2}'| xargs kill -9 2>/dev/null
fi
;;
status)
ps -ef | grep druid | grep -v grep | grep -v json
;;
* )
echo "use keepdruidall [start|stop|status]"
;;
esac

View File

@@ -0,0 +1,42 @@
#!/bin/bash
#
# netconsole This loads the netconsole module with the configured parameters.
#
# chkconfig:123456 40 60
# description: keepdruiddata
source /etc/profile
PRO_NAME=keepdruiddata
INS_DIR={{ install_path }}
#版本
VERSION={{ druid_version }}
case $1 in
start)
master=`ps -ef | grep dae-druid-data.sh | grep -v grep | wc -l`
if [ $master -lt 1 ];then
nohup $INS_DIR/$VERSION/bin/dae-druid-data.sh > /dev/null 2>&1 &
fi
;;
stop)
keeppid=`ps -ef | grep dae-druid-data.sh | grep -v grep | awk '{print $2}'`
echo "守护进程PID$keeppid"
kill -9 $keeppid
echo "关闭所有druid进程"
#ps aux | grep druid | grep -v grep | awk '{print $2}' | xargs kill -9
livenum=`ps -ef |grep "druid" | grep -v grep | grep -v json | wc -l`
if [ $livenum -ne 0 ];then
ps aux|grep "druid" |grep -v grep | awk '{print $2}'| xargs kill -9
fi
;;
status)
ps -ef | grep druid | grep -v grep | grep -v json
;;
* )
echo "use keepdruiddata [start|stop|status]"
;;
esac

View File

@@ -0,0 +1,42 @@
#!/bin/bash
#
# netconsole This loads the netconsole module with the configured parameters.
#
# chkconfig:123456 40 60
# description: keepdruidquery
source /etc/profile
PRO_NAME=keepdruidquery
INS_DIR={{ install_path }}
#版本
VERSION={{ druid_version }}
case $1 in
start)
master=`ps -ef | grep dae-druid-query.sh | grep -v grep | wc -l`
if [ $master -lt 1 ];then
nohup $INS_DIR/$VERSION/bin/dae-druid-query.sh > /dev/null 2>&1 &
fi
;;
stop)
keeppid=`ps -ef | grep dae-druid-query.sh | grep -v grep | awk '{print $2}'`
echo "守护进程PID$keeppid"
kill -9 $keeppid
echo "关闭所有druid进程"
#ps aux | grep druid | grep -v grep | awk '{print $2}' | xargs kill -9
livenum=`ps -ef |grep "druid" | grep -v grep | grep -v json | wc -l`
if [ $livenum -ne 0 ];then
ps aux|grep "druid" |grep -v grep | awk '{print $2}'| xargs kill -9
fi
;;
status)
ps -ef | grep druid | grep -v grep | grep -v json
;;
* )
echo "use keepdruidquery [start|stop|status]"
;;
esac

View File

@@ -0,0 +1,76 @@
<?xml version="1.0" encoding="UTF-8" ?>
<!--
~ Licensed to the Apache Software Foundation (ASF) under one
~ or more contributor license agreements. See the NOTICE file
~ distributed with this work for additional information
~ regarding copyright ownership. The ASF licenses this file
~ to you under the Apache License, Version 2.0 (the
~ "License"); you may not use this file except in compliance
~ with the License. You may obtain a copy of the License at
~
~ http://www.apache.org/licenses/LICENSE-2.0
~
~ Unless required by applicable law or agreed to in writing,
~ software distributed under the License is distributed on an
~ "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
~ KIND, either express or implied. See the License for the
~ specific language governing permissions and limitations
~ under the License.
-->
<Configuration status="WARN">
<properties>
<property name="pattern">%d{yyyy-MM-dd HH:mm:ss.SSS} %c{1.} %t [%p] %msg%n</property>
<property name="patternNoTime">%msg%n</property>
<property name="patternMetrics">%d %-8r %m%n</property>
<property name="logPath">{{ install_path }}/{{ druid_version }}/var/sv</property>
</properties>
<appenders>
<RollingFile name="STDOUT"
fileName="${logPath}/${sys:logfile.name}.log"
filePattern="${logPath}/history/${sys:logfile.name}-%d{yyyy-MM-dd}.log.%i.gz">
<PatternLayout>
<pattern>${pattern}</pattern>
</PatternLayout>
<Policies>
<SizeBasedTriggeringPolicy size="100 MB"/> <!-- Or every 100 MB -->
</Policies>
<DefaultRolloverStrategy max="10">
<Delete basePath="${LOG_PATH}/history" maxDepth="1">
<IfFileName glob="*.gz">
<IfLastModified age="7d">
<!--<IfAny>
<IfAccumulatedFileSize exceeds="200 GB" />
</IfAny>-->
</IfLastModified>
</IfFileName>
</Delete>
</DefaultRolloverStrategy>
</RollingFile>
<!--
<RollingFile name="METRICS"
fileName="${logPath}/${sys:logfile.name}.metrics"
filePattern="${logPath}/old/${sys:logfile.name}-%d{yyyy-MM-dd}.metrics.%i.gz">
<PatternLayout>
<pattern>${patternMetrics}</pattern>
</PatternLayout>
<Policies>
<SizeBasedTriggeringPolicy size="100 MB"/>
</Policies>
<DefaultRolloverStrategy max="6"/>
</RollingFile>
-->
</appenders>
<loggers>
<Root level="error">
<appender-ref ref="STDOUT"/>
</Root>
</loggers>
</Configuration>

View File

@@ -0,0 +1,43 @@
#
# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
#
druid.service=druid/middleManager
druid.plaintextPort=8091
# Number of tasks per middleManager
druid.worker.capacity=200
# Task launch parameters
druid.indexer.runner.javaOpts=-server {{ middlemanager_runner_javaOpts }} -Duser.timezone=UTC -Dfile.encoding=UTF-8 -Djava.util.logging.manager=org.apache.logging.log4j.jul.LogManager -Dlog4j.configurationFile=conf/druid/cluster/_common/log4j2-task.xml
druid.indexer.task.baseTaskDir=var/druid/task
# HTTP server threads
druid.server.http.numThreads=60
# Processing threads and buffers on Peons
druid.indexer.fork.property.druid.processing.numMergeBuffers={{ middlemanager_numMergeBuffers }}
druid.indexer.fork.property.druid.processing.buffer.sizeBytes={{ middlemanager_buffer_sizeBytes }}
druid.indexer.fork.property.druid.processing.numThreads={{ middlemanager_numThreads }}
# Hadoop indexing
druid.indexer.task.hadoopWorkingPath=var/druid/hadoop-tmp
druid.query.groupBy.maxMergingDictionarySize=10000000000
druid.query.groupBy.maxOnDiskStorage=10000000000

View File

@@ -0,0 +1,30 @@
#!/bin/bash
#num=`cat /etc/profile | grep druid | wc -l`
#if [ $num -eq "0" ];then
# echo -e "\n#druid" >> /etc/profile
# echo -e "export DRUID_HOME={{ install_path }}/{{ druid_version }}" >> /etc/profile
# echo -e "export PATH=\$DRUID_HOME/bin:\$PATH" >> /etc/profile
# source /etc/profile
#fi
keeppath='/etc/init.d/keepdruidall'
if [ -x $keeppath ];then
chkconfig --add keepdruidall
chkconfig keepdruidall on
fi
keeppath='/etc/init.d/keepdruiddata'
if [ -x $keeppath ];then
chkconfig --add keepdruiddata
chkconfig keepdruiddata on
fi
keeppath='/etc/init.d/keepdruidquery'
if [ -x $keeppath ];then
chkconfig --add keepdruidquery
chkconfig keepdruidquery on
fi

View File

@@ -0,0 +1,43 @@
#!/bin/bash
source /etc/profile
function killService(){
keeppath='/etc/init.d/keepdruidall'
if [ -x $keeppath ];then
service keepdruidall stop
chkconfig keepdruidall off
systemctl daemon-reload
rm -rf /etc/init.d/keepdruidall
fi
keeppath='/etc/init.d/keepdruiddata'
if [ -x $keeppath ];then
service keepdruiddata stop
chkconfig keepdruiddata off
systemctl daemon-reload
rm -rf /etc/init.d/keepdruiddata
fi
keeppath='/etc/init.d/keepdruidquery'
if [ -x $keeppath ];then
service keepdruidquery stop
chkconfig keepdruidquery off
systemctl daemon-reload
rm -rf /etc/init.d/keepdruidquery
fi
}
function killPid(){
livenum=`ps -ef |grep "druid" | grep -v grep | grep -v json | wc -l`
if [ $livenum -ne 0 ];then
ps aux|grep "druid" |grep -v grep | awk '{print $2}'| xargs kill -9
fi
}
killService
sleep 15
killPid
rm -rf {{ install_path }}/{{ druid_version }}

View File

@@ -20,16 +20,18 @@ fi
;;
stop)
ps -ef | grep dae-druid-all.sh | grep -v grep | awk '{print $2}' | xargs kill -9 2>/dev/null
#keeppid=`ps -ef | grep dae-druid-all.sh | grep -v grep | awk '{print $2}' | xargs kill -9 2>/dev/null`
#echo "守护进程PID$keeppid"
#kill -9 $keeppid
keeppid=`ps -ef | grep dae-druid-all.sh | grep -v grep | awk '{print $2}'`
if [ `ps -ef | grep dae-druid-all.sh | grep -v grep | awk '{print $2}'` -ne 0 ]
then
echo "守护进程PID$keeppid"
kill -9 $keeppid
echo "关闭所有druid进程"
#ps aux | grep druid | grep -v grep | awk '{print $2}' | xargs kill -9
livenum=`ps -ef |grep "druid" | grep -v grep | grep -v json | wc -l`
if [ $livenum -ne 0 ];then
ps aux|grep "druid" |grep -v grep | awk '{print $2}'| xargs kill -9 2>/dev/null
fi
#livenum=`ps -ef |grep "druid" | grep -v grep | grep -v json | wc -l`
#if [ $livenum -ne 0 ];then
#ps aux|grep "druid" |grep -v grep | awk '{print $2}'| xargs kill -9
#fi
;;
status)

View File

@@ -21,14 +21,17 @@ fi
stop)
keeppid=`ps -ef | grep dae-druid-data.sh | grep -v grep | awk '{print $2}'`
if [ `ps -ef | grep dae-druid-all.sh | grep -v grep | awk '{print $2}'` -ne 0 ]
then
echo "守护进程PID$keeppid"
kill -9 $keeppid
echo "关闭所有druid进程"
#ps aux | grep druid | grep -v grep | awk '{print $2}' | xargs kill -9
livenum=`ps -ef |grep "druid" | grep -v grep | grep -v json | wc -l`
if [ $livenum -ne 0 ];then
ps aux|grep "druid" |grep -v grep | awk '{print $2}'| xargs kill -9
fi
#livenum=`ps -ef |grep "druid" | grep -v grep | grep -v json | wc -l`
#if [ $livenum -ne 0 ];then
#ps aux|grep "druid" |grep -v grep | awk '{print $2}'| xargs kill -9
#fi
;;
status)

View File

@@ -21,14 +21,18 @@ fi
stop)
keeppid=`ps -ef | grep dae-druid-query.sh | grep -v grep | awk '{print $2}'`
if [ `ps -ef | grep dae-druid-all.sh | grep -v grep | awk '{print $2}'` -ne 0 ]
then
echo "守护进程PID$keeppid"
kill -9 $keeppid
echo "关闭所有druid进程"
#ps aux | grep druid | grep -v grep | awk '{print $2}' | xargs kill -9
livenum=`ps -ef |grep "druid" | grep -v grep | grep -v json | wc -l`
if [ $livenum -ne 0 ];then
ps aux|grep "druid" |grep -v grep | awk '{print $2}'| xargs kill -9
fi
#livenum=`ps -ef |grep "druid" | grep -v grep | grep -v json | wc -l`
#if [ $livenum -ne 0 ];then
#ps aux|grep "druid" |grep -v grep | awk '{print $2}'| xargs kill -9
#fi
;;
status)

View File

@@ -38,6 +38,6 @@ fi
}
killService
sleep 15
killPid
#sleep 15
#killPid
rm -rf {{ install_path }}/{{ druid_version }}

View File

@@ -120,7 +120,7 @@
#启动flink守护
- name: start keepflinkalive service
shell: source /etc/profile && sh /etc/init.d/keepflinkalive start
shell: source /etc/profile && /bin/bash /etc/init.d/keepflinkalive start
when: hostvars[inventory_hostname]['ansible_env'].SSH_CONNECTION.split(' ')[2] == groups.flink[0]

View File

@@ -1,13 +1,9 @@
#!/bin/bash
num=`cat /etc/profile | grep flink | wc -l`
if [ $num -eq "0" ];then
echo -e "\n#flink" >> /etc/profile
echo -e "export FLINK_HOME={{ install_path }}/{{ flink_version }}" >> /etc/profile
echo -e "export PATH=\$FLINK_HOME/bin:\$PATH" >> /etc/profile
source /etc/profile
fi
echo -e "\n#flink\nexport FLINK_HOME={{ install_path }}/{{ flink_version }}\nexport PATH=\$FLINK_HOME/bin:\$PATH" >> /etc/profile.d/flink.sh
chmod +x /etc/profile.d/flink.sh
source /etc/profile
keeppath='/etc/init.d/keepflinkalive'
if [ -x $keeppath ];then

View File

@@ -22,4 +22,5 @@ killService
sleep 5
killPid
rm -rf {{ install_path }}/{{ flink_version }}
rm -rf /etc/profile.d/flink.sh
sleep 5

View File

@@ -202,7 +202,7 @@
shell: cd {{ install_path }}/{{ hadoop_version }}/bin/ && ./set_hadoop_env.sh
- name: start hadoop journal
shell: source /etc/profile && sh /etc/init.d/keephdfsjournal start
shell: source /etc/profile && /bin/bash /etc/init.d/keephdfsjournal start
run_once: true
delegate_facts: true
delegate_to: "{{ item.ip }}"
@@ -283,7 +283,7 @@
delegate_to: "{{ master_ip }}"
- name: start hadoop-master
shell: service keephdfsmaster start
shell: source /etc/profile && /bin/bash /etc/init.d/keephdfsmaster start
run_once: true
delegate_facts: true
delegate_to: "{{ master_ip }}"
@@ -298,13 +298,13 @@
delegate_to: "{{ slave1_ip }}"
- name: start hadoop-slave
shell: service keephdfsslave start
shell: source /etc/profile && /bin/bash /etc/init.d/keephdfsslave start
run_once: true
delegate_facts: true
delegate_to: "{{ slave1_ip }}"
- name: start hadoop-worker
shell: service keephdfsworker start
shell: source /etc/profile && /bin/bash /etc/init.d/keephdfsworker start
- name: deleted hadoop.tar.gz
shell: rm -rf {{ install_path }}/hadoop-2.7.1.tar.gz

View File

@@ -1,15 +1,12 @@
#!/bin/bash
num=`cat /etc/profile | grep hadoop | wc -l`
if [ $num -eq "0" ];then
echo -e "\n#hadoop" >> /etc/profile
echo -e "export HADOOP_HOME={{ install_path }}/{{ hadoop_version }}" >> /etc/profile
echo -e "export PATH=\$HADOOP_HOME/sbin:\$PATH" >> /etc/profile
echo -e "export PATH=\$HADOOP_HOME/bin:\$PATH" >> /etc/profile
source /etc/profile
fi
echo -e "\n#hadoop\nexport HADOOP_HOME={{ install_path }}/{{ hadoop_version }}\nexport PATH=\$HADOOP_HOME/sbin:\$PATH\nexport PATH=\$HADOOP_HOME/bin:\$PATH" >> /etc/profile.d/hadoop.sh
chmod +x /etc/profile.d/hadoop.sh
source /etc/profile
keeppath='/etc/init.d/keephdfsmaster'
if [ -x $keeppath ];then

View File

@@ -67,4 +67,4 @@ sleep 15
killPid
rm -rf {{ install_path }}/{{ hadoop_version }}
rm -rf {{ install_path }}/hadoop
rm -rf /etc/profile.d/hadoop.sh

View File

@@ -210,7 +210,7 @@
shell: cd {{ install_path }}/{{ hbase_version }}/bin/ && sh set_hbase_env.sh
- name: start hbase master
shell: source /etc/profile && sh /etc/init.d/keephbasemaster 'start'
shell: source /etc/profile && /bin/bash /etc/init.d/keephbasemaster 'start'
run_once: true
delegate_facts: true
delegate_to: "{{ item.ip }}"
@@ -220,7 +220,7 @@
- { ip: '{{ slave2_ip }}' }
- name: start hbase region
shell: source /etc/profile && sh /etc/init.d/keephbaseregion 'start'
shell: source /etc/profile && /bin/bash /etc/init.d/keephbaseregion 'start'
- name: Ansible delete {{ install_path }}/{{ hbase_version }}.tar.gz
file:

View File

@@ -1,14 +1,8 @@
#!/bin/bash
num=`cat /etc/profile | grep hbase | wc -l`
if [ $num -eq "0" ];then
echo -e "\n#hbase" >> /etc/profile
echo -e "export HBASE_HOME={{ data_path }}/{{ hbase_version }}" >> /etc/profile
echo -e "export PATH=\$HBASE_HOME/bin:\$PATH" >> /etc/profile
source /etc/profile
fi
echo -e "\n#hbase\nexport HBASE_HOME={{ data_path }}/{{ hbase_version }}\nexport PATH=\$HBASE_HOME/bin:\$PATH" >> /etc/profile.d/hbase.sh
chmod +x /etc/profile.d/hbase.sh
source /etc/profile
keeppath='/etc/init.d/keephbasemaster'
if [ -x $keeppath ];then

View File

@@ -37,4 +37,4 @@ killService
sleep 15
killPid
rm -rf {{ install_path }}/{{ hbase_version }}
rm -rf /etc/profile.d/hbase.sh

View File

@@ -209,7 +209,7 @@
#启动kafka守护
- name: start keepkafalive service
shell: source /etc/profile && sh /etc/init.d/keepkafalive 'start'
shell: source /etc/profile && /bin/bash /etc/init.d/keepkafalive 'start'
- name: Ansible delete {{ kafka_version }}.tar.gz
file:

View File

@@ -1,13 +1,8 @@
#!/bin/bash
num=`cat /etc/profile | grep kafka | wc -l`
if [ $num -eq "0" ];then
echo -e "\n#kafka" >> /etc/profile
echo -e "export KAFKA_HOME={{ install_path }}/{{ kafka_version }}" >> /etc/profile
echo -e "export PATH=\$KAFKA_HOME/bin:\$PATH" >> /etc/profile
source /etc/profile
fi
echo -e "\n#kafka\nexport KAFKA_HOME={{ install_path }}/{{ kafka_version }}\nexport PATH=\$KAFKA_HOME/bin:\$PATH" >> /etc/profile.d/kafka.sh
chmod +x /etc/profile.d/kafka.sh
source /etc/profile
keeppath='/etc/init.d/keepkafalive'
if [ -x $keeppath ];then

View File

@@ -24,4 +24,5 @@ sleep 5
killPid
rm -rf {{ install_path }}/{{ kafka_version }}
rm -rf {{ data_path }}/{{ kafka_version }}
rm -rf /etc/profile.d/kafka.sh
sleep 5

View File

@@ -45,39 +45,39 @@
when: (allowed_unload) == "no"
- name: check mariadb version; if mariadb does not exist, it will be print error;Don't Worry!!!
shell: mysql -uroot
ignore_errors: True
#shell: mysql -uroot
shell: ps aux | grep mysqld | grep -v grep | wc -l
register: result
- name: create directory path:/data/mariadb/logs
file:
state: directory
path: '/data/mariadb/logs'
when: result.stderr is defined and result.stderr != '' #有错误信息,说明mariadb服务不存在
when: result.stdout == '0' #有错误信息,说明mariadb服务不存在
- name: create directory path:/data/mariadb/sql
file:
state: directory
path: '/data/mariadb/sql'
when: result.stderr is defined and result.stderr != '' #有错误信息,说明mariadb服务不存在
when: result.stdout == '0' #有错误信息,说明mariadb服务不存在
- name: create directory path:{{ install_path }}/mariadb/sql
file:
state: directory
path: '{{ install_path }}/mariadb'
when: result.stderr is defined and result.stderr != ''
when: result.stdout == '0' #有错误信息,说明mariadb服务不存在
- name: copy
copy: src=files/slow_query.log dest=/data/mariadb/logs force=true backup=yes
when: result.stderr is defined and result.stderr != '' #有错误信息,说明mariadb服务不存在
when: result.stdout == '0' #有错误信息,说明mariadb服务不存在
- name: mariadb service not exist; copy -rpm.zip and unzip mariadb.zip to {{ install_path }}/mariadb
unarchive: src={{ package_path }}/mariadb.zip dest={{ install_path }}/mariadb copy=yes
when: result.stderr is defined and result.stderr != '' #有错误信息,说明mariadb服务不存在
when: result.stdout == '0' #有错误信息,说明mariadb服务不存在
- name: mariadb service not exist; install mariadb
shell: rpm -ivh {{ install_path }}/mariadb/*.rpm --force --nodeps && systemctl start mysql && systemctl enable mysql
when: result.stderr is defined and result.stderr != '' #有错误信息,说明mariadb服务不存在
when: result.stdout == '0' #有错误信息,说明mariadb服务不存在
register: install_mariadb_result
- name: copy my.cnf
@@ -87,7 +87,7 @@
force: yes
run_once: true
delegate_to: '{{ groups.mariadb[0] }}'
when: result.stderr is defined and result.stderr != '' #有错误信息,说明mariadb服务不存在
when: result.stdout == '0' #有错误信息,说明mariadb服务不存在
- name: copy my.cnf
template:
@@ -96,15 +96,15 @@
force: yes
run_once: true
delegate_to: '{{ groups.mariadb[1] }}'
when: result.stderr is defined and result.stderr != '' #有错误信息,说明mariadb服务不存在
when: result.stdout == '0' #有错误信息,说明mariadb服务不存在
- name: change mariadb root password
shell: mysql -uroot -e "ALTER USER 'root'@'localhost' IDENTIFIED BY '{{ galaxy_mariadb_pin }}';"
when: result.stderr is defined and result.stderr != '' #有错误信息,说明mariadb服务不存在
when: result.stdout == '0' #有错误信息,说明mariadb服务不存在
- name: change mariadb remote authority
shell: mysql -uroot -p{{ galaxy_mariadb_pin }} -e"use mysql;grant all privileges on *.* to 'root'@'%' identified by '{{ galaxy_mariadb_pin }}' with grant option;FLUSH PRIVILEGES;"
when: result.stderr is defined and result.stderr != '' #有错误信息,说明mariadb服务不存在
when: result.stdout == '0' #有错误信息,说明mariadb服务不存在
- name: restart mariadb
shell: systemctl restart mysql

View File

@@ -156,7 +156,7 @@
#启动守护
- name: start keepnacosalive service
shell: source /etc/profile && sh /etc/init.d/keepnacosalive 'start'
shell: source /etc/profile && /bin/bash /etc/init.d/keepnacosalive 'start'
- name: Ansible delete {{ nacos_version }}.tar.gz
file:

View File

@@ -1,13 +1,9 @@
#!/bin/bash
num=`cat /etc/profile | grep nacos | wc -l`
if [ $num -eq "0" ];then
echo -e "\n#nacos" >> /etc/profile
echo -e "export NACOS_HOME={{ install_path }}/{{ nacos_version }}" >> /etc/profile
echo -e "export PATH=\$NACOS_HOME/bin:\$PATH" >> /etc/profile
fi
echo -e "\n#nacos\nexport NACOS_HOME={{ install_path }}/{{ nacos_version }}\nexport PATH=\$NACOS_HOME/bin:\$PATH" >> /etc/profile.d/nacos.sh
chmod +x /etc/profile.d/nacos.sh
source /etc/profile
keeppath='/etc/init.d/keepnacosalive'
if [ -x $keeppath ];then
chkconfig --add keepnacosalive

View File

@@ -23,4 +23,6 @@ killService
sleep 10
killPid
rm -rf {{ install_path }}/{{ nacos_version }}
rm -rf /etc/profile.d/nacos.sh
sleep 5

View File

@@ -1,14 +1,8 @@
#!/bin/bash
num=`cat /etc/profile | grep spark | wc -l`
if [ $num -eq "0" ];then
echo -e "\n#spark" >> /etc/profile
echo -e "export SPARK_HOME={{ install_path }}/{{ spark_version }}" >> /etc/profile
echo -e "export PATH=\$SPARK_HOME/sbin:\$PATH" >> /etc/profile
echo -e "export PATH=\$SPARK_HOME/bin:\$PATH" >> /etc/profile
source /etc/profile
fi
echo -e "\n#spark\nexport SPARK_HOME={{ install_path }}/{{ spark_version }}\nexport PATH=\$SPARK_HOME/sbin:\$PATH\nexport PATH=\$SPARK_HOME/bin:\$PATH" >> /etc/profile.d/spark.sh
chmod +x /etc/profile.d/spark.sh
source /etc/profile
keeppath='/etc/init.d/keepsparkall'
if [ -x $keeppath ];then

View File

@@ -1,12 +1,8 @@
#!/bin/bash
num=`cat /etc/profile | grep storm | wc -l`
if [ $num -eq "0" ];then
echo -e "\n#storm" >> /etc/profile
echo -e "export STORM_HOME={{ install_path }}/{{ storm_version }}" >> /etc/profile
echo -e "export PATH=\$STORM_HOME/bin:\$PATH" >> /etc/profile
source /etc/profile
fi
echo -e "\n#storm\nexport STORM_HOME={{ install_path }}/{{ storm_version }}\nexport PATH=\$STORM_HOME/bin:\$PATH" >> /etc/profile.d/storm.sh
chmod +x /etc/profile.d/storm.sh
source /etc/profile
keeppath='/etc/init.d/keepnimalive'
if [ -x $keeppath ];then

View File

@@ -37,3 +37,4 @@ killService
sleep 5
killPid
rm -rf {{ install_path }}/{{ storm_version }}
rm -rf /etc/profile.d/storm.sh

View File

@@ -157,7 +157,7 @@
force_source: yes
- name: start keepzkalive service
shell: source /etc/profile && sh /etc/init.d/keepzkalive 'start'
shell: source /etc/profile && /bin/bash /etc/init.d/keepzkalive 'start'
- name: waiting zookeeper start sleep 10 s
shell: sleep 10
@@ -167,7 +167,7 @@
register: zkstatus_out
- name: start keepzkalive service again
shell: source /etc/profile && sh /etc/init.d/keepzkalive 'start'
shell: source /etc/profile && /bin/bash /etc/init.d/keepzkalive 'start'
when: zkstatus_out.stdout != '1'
- name: start zookeeper_exporter

View File

@@ -12,7 +12,7 @@ start)
keepnum=`ps -ef | grep dae-zookeeper.sh | grep -v grep | wc -l`
if [ $keepnum -eq "0" ];then
nohup {{ install_path }}/{{ zookeeper_version }}/bin/dae-zookeeper.sh /home/bigdata > /dev/null 2>&1 &
nohup {{ install_path }}/{{ zookeeper_version }}/bin/dae-zookeeper.sh > /dev/null 2>&1 &
fi
;;

View File

@@ -1,13 +1,7 @@
#!/bin/bash
num=`cat /etc/profile | grep zookeeper | wc -l`
if [ $num -eq "0" ];then
echo -e "\n#zookeeper" >> /etc/profile
echo -e "export ZOOKEEPER_HOME={{ install_path }}/{{ zookeeper_version }}" >> /etc/profile
echo -e "export PATH=\$ZOOKEEPER_HOME/bin:\$PATH" >> /etc/profile
source /etc/profile
fi
echo -e "\n#zookeeper\nexport ZOOKEEPER_HOME={{ install_path }}/{{ zookeeper_version }}\nexport PATH=\$ZOOKEEPER_HOME/bin:\$PATH" >> /etc/profile.d/zookeeper.sh
chmod +x /etc/profile.d/zookeeper.sh
keeppsth='/etc/init.d/keepzkalive'
if [ -x $keeppsth ];then

View File

@@ -8,7 +8,7 @@ livenum=`jps -l | grep "org.apache.zookeeper.server.quorum.QuorumPeerMain" | wc
if [ $livenum -ne 0 ];then
keeppid=`jps -l |grep zookeeper | awk '{print $1}'`
kill -9 $keeppid
rm -rf /etc/profile.d/zookeeper.sh
rm -rf {{ install_path }}/{{ zookeeper_version }}
rm -rf {{ data_path }}/{{ zookeeper_version }}
chkconfig keepzkalive off

View File

@@ -12,6 +12,12 @@
delegate_to: "{{ query_ip }}"
- name: create clickhouse tables
shell: clickhouse-client -h {{ inventory_hostname }} --port 9001 -m -u default --password {{ root_key }} --multiquery < {{ install_path }}/create_ck_table.sql 2>/dev/null
run_once: true
ignore_errors: True
delegate_to: "{{ query_ip }}"
- name: check clickhouse tables
shell: clickhouse-client -h {{ inventory_hostname }} --port 9001 -m -u default --password {{ root_key }} --multiquery < {{ install_path }}/create_ck_table.sql
run_once: true
delegate_to: "{{ query_ip }}"

View File

@@ -51,7 +51,8 @@
cron:
name: 'dos-baseline'
hour: "3"
weekday: "1"
job: '/bin/sh {{ install_path }}/dos-baseline/start.sh'
user: root
delegate_to: "{{ master_ip }}"

View File

@@ -18,31 +18,31 @@ case $1 in
resetAll)
for var in ${common_task_name[@]};
do
curl -X 'POST' http://${router_ip}:8089/druid/indexer/v1/supervisor/${var}/reset
curl --retry-delay 3 --retry 3 -X 'POST' http://${router_ip}:8089/druid/indexer/v1/supervisor/${var}/reset
sleep 2
done
for var in ${hot_task_name[@]};
do
curl -X 'POST' http://${router_ip}:8089/druid/indexer/v1/supervisor/${var}/reset
curl --retry-delay 3 --retry 3 -X 'POST' http://${router_ip}:8089/druid/indexer/v1/supervisor/${var}/reset
sleep 2
done
;;
terminateAll)
curl -X 'POST' http://${router_ip}:8089/druid/indexer/v1/supervisor/terminateAll
curl --retry-delay 3 --retry 3 -X 'POST' http://${router_ip}:8089/druid/indexer/v1/supervisor/terminateAll
;;
shutdownAllTasks)
for var in ${common_task_name[@]};
do
curl -X 'POST' http://${router_ip}:8089/druid/indexer/v1/datasources/${var}/shutdownAllTasks
curl --retry-delay 3 --retry 3 -X 'POST' http://${router_ip}:8089/druid/indexer/v1/datasources/${var}/shutdownAllTasks
sleep 2
done
for var in ${hot_task_name[@]};
do
curl -X 'POST' http://${router_ip}:8089/druid/indexer/v1/datasources/${var}/shutdownAllTasks
curl --retry-delay 3 --retry 3 -X 'POST' http://${router_ip}:8089/druid/indexer/v1/datasources/${var}/shutdownAllTasks
sleep 2
done
;;
@@ -52,7 +52,7 @@ case $1 in
echo "Usage: supervisor-manger reset <dataSource>"
exit 1
fi
curl -X 'POST' http://${router_ip}:8089/druid/indexer/v1/supervisor/$2/reset
curl --retry-delay 3 --retry 3 -X 'POST' http://${router_ip}:8089/druid/indexer/v1/supervisor/$2/reset
;;
terminate)
@@ -60,7 +60,7 @@ case $1 in
echo "Usage: supervisor-manger terminate <dataSource>"
exit 1
fi
curl -X 'POST' http://${router_ip}:8089/druid/indexer/v1/supervisor/$2/terminate
curl --retry-delay 3 --retry 3 -X 'POST' http://${router_ip}:8089/druid/indexer/v1/supervisor/$2/terminate
;;
shutdownTasks)
@@ -68,7 +68,7 @@ case $1 in
echo "Usage: supervisor-manger shutdownTasks <dataSource>"
exit 1
fi
curl -X 'POST' http://${router_ip}:8089/druid/indexer/v1/datasources/$2/shutdownAllTasks
curl --retry-delay 3 --retry 3 -X 'POST' http://${router_ip}:8089/druid/indexer/v1/datasources/$2/shutdownAllTasks
;;
*)
echo "Usage: supervisor-manger {resetAll|terminateAll|shutdownAllTasks}"

View File

@@ -38,11 +38,13 @@
run_once: true
delegate_to: '{{ hadoop_ip }}'
- name: check if {{ mariadb_druid_database }} exist
shell: mysql -s -uroot -p{{ galaxy_mariadb_pin }} -P3306 -h{{ gateway_keepalive_host }} -e "SELECT COUNT(1) FROM druid.druid_segments WHERE start = '3000-01-02T00:00:00.000Z'"
register: segments_nums
run_once: true
delegate_to: "127.0.0.1"
#- name: check if {{ mariadb_druid_database }} exist, if 'FAILED message', please ignore ! ! !
# shell: mysql -s -uroot -p{{ galaxy_mariadb_pin }} -P3306 -h{{ gateway_keepalive_host }} -e "SELECT COUNT(1) FROM druid.druid_segments WHERE start = '3000-01-02T00:00:00.000Z'"
# register: segments_nums
# run_once: true
# delegate_to: "127.0.0.1"
# ignore_errors: True
# failed_when: "'FAILED' in segments_nums.stderr or segments_nums.stdout != '20'"
- name: copy segments.sql to ~
copy:
@@ -51,13 +53,13 @@
force: true
run_once: true
delegate_to: '127.0.0.1'
when: segments_nums.stdout != '20'
#when: segments_nums is defined
- name: insert segments to mariadb
shell: 'mysql -uroot -p{{ galaxy_mariadb_pin }} -P3306 -h{{ gateway_keepalive_host }} {{ mariadb_druid_database }} < ~/druid_segments-tsg3.0.sql'
run_once: true
delegate_to: '127.0.0.1'
when: segments_nums.stdout != '20'
#when: segments_nums is defined
- name: copy druid_topology to {{ install_path }}
copy:

View File

@@ -6,11 +6,11 @@ task_name=`cat topology_list.txt`
kafka_host={{ kafka_source_servers }}
druid_host={{ groups.druid[0] }}
curl -i -XGET 'http://'$druid_host':8081/druid/indexer/v1/leader'
curl --retry-delay 3 --retry 3 -i -XGET 'http://'$druid_host':8088/druid/indexer/v1/leader'
num=$?
while [ $num -gt "0" ];do
sleep 5
curl -s -XGET 'http://'$druid_host':8081/druid/indexer/v1/leader'
curl --retry-delay 3 --retry 3 -s -XGET 'http://'$druid_host':8088/druid/indexer/v1/leader'
num=$?
done
cp ./tasks/* ./
@@ -18,7 +18,7 @@ cp ./tasks/* ./
for var in ${task_name[@]};
do
sed -i 's/kafkabootstrap/'$kafka_host'/' ${var}
curl -X 'POST' -H 'Content-Type:application/json' -d @${var} http://$druid_host:8081/druid/indexer/v1/supervisor
curl --retry-delay 3 --retry 3 -X 'POST' -H 'Content-Type:application/json' -d @${var} http://$druid_host:8088/druid/indexer/v1/supervisor
echo "'${var}' 任务启动成功"
sleep 2
done

View File

@@ -14,14 +14,14 @@ hot_rule_file="$base_dir/hot_data_rule.json"
for var in ${common_task_name[@]};
do
curl -X 'POST' -H 'Content-Type:application/json' -d @$common_rule_file http://${router_ip}:8089/druid/coordinator/v1/rules/${var}
curl --retry-delay 3 --retry 3 -X 'POST' -H 'Content-Type:application/json' -d @$common_rule_file http://${router_ip}:8089/druid/coordinator/v1/rules/${var}
echo "'${var}' 任务启动成功"
sleep 2
done
for var in ${hot_task_name[@]};
do
curl -X 'POST' -H 'Content-Type:application/json' -d @$hot_rule_file http://${router_ip}:8089/druid/coordinator/v1/rules/${var}
curl --retry-delay 3 --retry 3 -X 'POST' -H 'Content-Type:application/json' -d @$hot_rule_file http://${router_ip}:8089/druid/coordinator/v1/rules/${var}
echo "'${var}' 任务启动成功"
sleep 2
done

View File

@@ -15,7 +15,7 @@
src: '{{ package_path }}/topology'
dest: '{{ install_path }}/'
force: true
backup: yes
backup: false
#其他服务器上的包就当备份
#delegate_to: "{{ master_ip }}"

View File

@@ -7,59 +7,59 @@
port: 8186
timeout: 60
- name: init hos service
shell: "curl --location -s -w %{http_code} --request PUT 'http://{{ master_ip }}:8186/admin/initialize' --header 'token: c21f969b5f03d33d43e04f8f136e7682'"
- name: Initialize Galaxy-hos-service
shell: "curl --retry-delay 3 --retry 3 --location -s -w %{http_code} --request PUT 'http://{{ master_ip }}:8186/admin/initialize' --header 'token: c21f969b5f03d33d43e04f8f136e7682'"
register: initialize
run_once: true
delegate_facts: true
delegate_to: '{{ master_ip }}'
- name: check hos init status
- name: Checking Initialize status
fail:
msg: "HOS 内部初始化接口执行失败!请检查"
when: initialize.stdout != '200'
- name: delete firewall_hos_bucket
shell: "curl --location -s -w %{http_code} --request DELETE 'http://{{ master_ip }}:8186/hos/firewall_hos_bucket' --header 'token: c21f969b5f03d33d43e04f8f136e7682'"
register: delete
run_once: true
delegate_facts: true
delegate_to: '{{ master_ip }}'
- name: check delete firewall_hos_bucket status
fail:
msg: "HOS 删除 firewall_hos_bucket 桶异常"
when: delete.stdout != '204'
- name: delete proxy_hos_bucket
shell: "curl --location -s -w %{http_code} --request DELETE 'http://{{ master_ip }}:8186/hos/proxy_hos_bucket' --header 'token: c21f969b5f03d33d43e04f8f136e7682'"
register: delete
run_once: true
delegate_facts: true
delegate_to: '{{ master_ip }}'
- name: check delete proxy_hos_bucket status
fail:
msg: "HOS 删除 proxy_hos_bucket 桶异常"
when: delete.stdout != '204'
- name: delete session_record_hos_bucket
shell: "curl --location -s -w %{http_code} --request DELETE 'http://{{ master_ip }}:8186/hos/session_record_hos_bucket' --header 'token: c21f969b5f03d33d43e04f8f136e7682'"
register: delete
run_once: true
delegate_facts: true
delegate_to: '{{ master_ip }}'
- name: check delete session_record_hos_bucket status
fail:
msg: "HOS 删除 session_record_hos_bucket 桶异常"
when: delete.stdout != '204'
#- name: delete firewall_hos_bucket
# shell: "curl --retry-delay 3 --retry 3 --location -s -w %{http_code} --request DELETE 'http://{{ master_ip }}:8186/hos/firewall_hos_bucket' --header 'token: c21f969b5f03d33d43e04f8f136e7682'"
# register: delete
# run_once: true
# delegate_facts: true
# delegate_to: '{{ master_ip }}'
#
#- name: check delete firewall_hos_bucket status
# fail:
# msg: "HOS 删除 firewall_hos_bucket 桶异常"
# when: delete.stdout != '204'
#
#
#- name: delete proxy_hos_bucket
# shell: "curl --retry-delay 3 --retry 3 --location -s -w %{http_code} --request DELETE 'http://{{ master_ip }}:8186/hos/proxy_hos_bucket' --header 'token: c21f969b5f03d33d43e04f8f136e7682'"
# register: delete
# run_once: true
# delegate_facts: true
# delegate_to: '{{ master_ip }}'
#
#- name: check delete proxy_hos_bucket status
# fail:
# msg: "HOS 删除 proxy_hos_bucket 桶异常"
# when: delete.stdout != '204'
#
#
#- name: delete session_record_hos_bucket
# shell: "curl --retry-delay 3 --retry 3 --location -s -w %{http_code} --request DELETE 'http://{{ master_ip }}:8186/hos/session_record_hos_bucket' --header 'token: c21f969b5f03d33d43e04f8f136e7682'"
# register: delete
# run_once: true
# delegate_facts: true
# delegate_to: '{{ master_ip }}'
#
#- name: check delete session_record_hos_bucket status
# fail:
# msg: "HOS 删除 session_record_hos_bucket 桶异常"
# when: delete.stdout != '204'
- name: create firewall_hos_bucket
shell: "curl --location -s -w %{http_code} --request PUT 'http://{{ master_ip }}:8186/hos/firewall_hos_bucket' --header 'token: c21f969b5f03d33d43e04f8f136e7682'"
shell: "curl --retry-delay 3 --retry 3 --location -s -w %{http_code} --request PUT 'http://{{ master_ip }}:8186/hos/firewall_hos_bucket' --header 'token: c21f969b5f03d33d43e04f8f136e7682'"
register: table1
run_once: true
delegate_facts: true
@@ -71,7 +71,7 @@
when: table1.stdout.find('200') == '-1' and table1.stdout.find('409') == '-1'
- name: create proxy_hos_bucket
shell: "curl --location -s -w %{http_code} --request PUT 'http://{{ master_ip }}:8186/hos/proxy_hos_bucket' --header 'token: c21f969b5f03d33d43e04f8f136e7682'"
shell: "curl --retry-delay 3 --retry 3 --location -s -w %{http_code} --request PUT 'http://{{ master_ip }}:8186/hos/proxy_hos_bucket' --header 'token: c21f969b5f03d33d43e04f8f136e7682'"
register: table2
run_once: true
delegate_facts: true
@@ -83,7 +83,7 @@
when: table2.stdout.find('200') == '-1' and table2.stdout.find('409') == '-1'
- name: create session_record_hos_bucket
shell: "curl --location -s -w %{http_code} --request PUT 'http://{{ master_ip }}:8186/hos/session_record_hos_bucket' --header 'token: c21f969b5f03d33d43e04f8f136e7682'"
shell: "curl --retry-delay 3 --retry 3 --location -s -w %{http_code} --request PUT 'http://{{ master_ip }}:8186/hos/session_record_hos_bucket' --header 'token: c21f969b5f03d33d43e04f8f136e7682'"
register: table3
run_once: true
delegate_facts: true
@@ -95,7 +95,7 @@
when: table3.stdout.find('200') == '-1' and table3.stdout.find('409') == '-1'
- name: Does proxy_hos_bucket exist
shell: "curl --location -s -w %{http_code} --request GET 'http://{{ master_ip }}:8186/hos/' --header 'token: c21f969b5f03d33d43e04f8f136e7682' |grep proxy_hos_bucket | wc -l"
shell: "curl --retry-delay 3 --retry 3 --location -s -w %{http_code} --request GET 'http://{{ master_ip }}:8186/hos/' --header 'token: c21f969b5f03d33d43e04f8f136e7682' |grep proxy_hos_bucket | wc -l"
register: bucketa_out
- fail:
@@ -103,7 +103,7 @@
when: bucketa_out.stdout != '1'
- name: Does session_record_hos_bucket exist
shell: "curl --location -s -w %{http_code} --request GET 'http://{{ master_ip }}:8186/hos/' --header 'token: c21f969b5f03d33d43e04f8f136e7682' |grep session_record_hos_bucket | wc -l"
shell: "curl --retry-delay 3 --retry 3 --location -s -w %{http_code} --request GET 'http://{{ master_ip }}:8186/hos/' --header 'token: c21f969b5f03d33d43e04f8f136e7682' |grep session_record_hos_bucket | wc -l"
register: bucketb_out
- fail:
@@ -111,7 +111,7 @@
when: bucketb_out.stdout != '1'
- name: Does firewall_hos_bucket exist
shell: "curl --location -s -w %{http_code} --request GET 'http://{{ master_ip }}:8186/hos/' --header 'token: c21f969b5f03d33d43e04f8f136e7682' |grep firewall_hos_bucket | wc -l"
shell: "curl --retry-delay 3 --retry 3 --location -s -w %{http_code} --request GET 'http://{{ master_ip }}:8186/hos/' --header 'token: c21f969b5f03d33d43e04f8f136e7682' |grep firewall_hos_bucket | wc -l"
register: bucketc_out
- fail:

View File

@@ -15,8 +15,10 @@
backup: yes
- name: create hbase table
shell: cd {{ install_path }}/{{ hbase_version }}/bin/ && ./create-hbase-table.sh
shell: cd {{ install_path }}/{{ hbase_version }}/bin/ && ./create-hbase-table.sh | grep ERROR | grep -v "already exists"
register: result
run_once: true
delegate_facts: true
delegate_to: "{{ master_ip }}"
failed_when: "'ERROR' in result.stdout"

View File

@@ -5,12 +5,12 @@ source /etc/profile
exec {{ install_path }}/{{ hbase_version }}/bin/hbase shell <<EOF
create_namespace 'tsg'
create_namespace 'sub'
create_namespace 'dos'
create 'tsg:report_result', {NAME => 'response', VERSIONS => 1,COMPRESSION => 'GZ',IS_MOB => true, MOB_THRESHOLD => 0}, {NAME => 'detail',COMPRESSION => 'GZ',VERSIONS => 1}
create 'sub:subscriber_info', {NAME => 'subscriber_id', VERSIONS => 1}
create 'ddos_traffic_baselines', 'TCP SYN Flood','UDP Flood','ICMP Flood', 'DNS Amplification'
create 'dos:ddos_traffic_baselines', 'TCP SYN Flood','UDP Flood','ICMP Flood','DNS Amplification'
list
EOF

View File

@@ -1,3 +1,6 @@
- name: stop keepalived
shell: if [ `ps aux | grep kp_daemon | grep -v grep | awk '{print $2}' | wc -l` -ne '0' ] ; then ps aux | grep kp_daemon | grep -v grep | awk '{print $2}' | xargs kill -9 ; fi && service keepalived stop
- name: stop and remove {{ chproxy_image_container_name }} container
docker_container:
name: '{{ chproxy_image_container_name }}'
@@ -129,3 +132,6 @@
with_items:
- { ip: '{{ groups.services[0] }}' }
- { ip: '{{ groups.services[1] }}' }
- name: start keepalived
shell: 'nohup /bin/bash /etc/keepalived/kp_daemon.sh >/dev/null 2>&1 &'

View File

@@ -1,7 +1,32 @@
#关闭旧的自启进程
- name: kill kp_daemon.sh
shell: ps aux | grep kp_daemon | grep -v grep | awk '{print $2}' | xargs kill 2>/dev/null
shell: "if [ `ps aux | grep kp_daemon | grep -v grep | awk '{print $2}' | wc -l` -ne 0 ] ; then ps aux | grep kp_daemon | grep -v grep | awk '{print $2}' | xargs kill -9 ; fi"
- block:
- name: Check if the Keepalived service already exists
shell: source /etc/profile && rpm -qa | grep keepalived | wc -l
register: check_out
- name: copy unload_keepalived.sh
template:
src: unload_keepalived.sh
dest: /root/
- name: unload keepalived
shell: sh /root/unload_keepalived.sh | grep -v "warning"
when: check_out.stdout >= '1'
- name: Check if the keepalived service already exists
shell: source /etc/profile && rpm -qa | grep keepalived | wc -l
register: check_out
- name: delete unload_keepalived.sh
file:
path: "/root/unload_keepalived.sh"
state: absent
- name: To terminate execution
fail:
msg: "卸载失败,组件可能非本安装部署,请联系开发确认或手动卸载后继续安装"
run_once: true
delegate_to: 127.0.0.1
when: check_out.stdout >= '1'
when: (allowed_unload) == "yes"
#创建文件夹
- name: create keepalived package path:{{ keepalived_package_path }}

View File

@@ -67,13 +67,18 @@ installKP(){
chmod +x $keepalivedInstallPath/$keepalivedDaeonName
echo "将$keepalivedInstallPath/$keepalivedDaeonName守护进程添加到开机自启动中"
chmod +x /etc/rc.d/rc.local
upStartNum=`cat /etc/rc.d/rc.local | grep "$keepalivedInstallPath" | wc -l`
if [ $upStartNum -eq "0" ];then
echo -e "\n#设置$keepalivedDaeonName守护脚本开机自启动" >> /etc/rc.d/rc.local
echo "nohup $keepalivedInstallPath/$keepalivedDaeonName >/dev/null 2>&1 &" >> /etc/rc.d/rc.local
echo "将$keepalivedInstallPath/$keepalivedDaeonName守护进程添加到开机自启动成功"
fi
echo "开始启动$keepalivedInstallPath/$keepalivedDaeonName守护进程"
# nohup $keepalivedInstallPath/$keepalivedDaeonName >/dev/null 2>&1 &
sed -i -e 's#. /etc/sysconfig/keepalived#. /usr/local/etc/sysconfig/keepalived#g' /etc/init.d/keepalived
systemctl daemon-reload
service keepalived start
rm -rf $keepalivedBagPath
}

View File

@@ -0,0 +1,9 @@
#!/bin/bash
#for i in `rpm -qa | grep keepalived` ; do rpm -e --nodeps $i ; done
if [ `ps aux | grep -E "keepalived|kp_daemon" | grep -v grep | wc -l` -ne '0' ] ; then ps aux | grep -E "keepalived|kp_daemon" | grep -v grep | awk '{print $2}' | xargs kill -9 ; fi
rm -rf /etc/keepalived
rm -rf /etc/init.d/keepalived

View File

@@ -1,3 +1,6 @@
- name: stop keepalived
shell: if [ `ps aux | grep kp_daemon | grep -v grep | awk '{print $2}' | wc -l` -ne '0' ] ; then ps aux | grep kp_daemon | grep -v grep | awk '{print $2}' | xargs kill -9 ; fi && service keepalived stop
- name: stop and remove {{ gateway_image_container_name }} container
docker_container:
name: '{{ gateway_image_container_name }}'
@@ -130,3 +133,7 @@
- { ip: '{{ groups.services[0] }}' }
- { ip: '{{ groups.services[1] }}' }
#- name: start keepalived
# shell: nohup /bin/bash /etc/keepalived/kp_daemon.sh >/dev/null 2>&1 &
#shell: systemctl start keepalived
# shell: service keepalived start

View File

@@ -1,3 +1,34 @@
#关闭旧的自启进程
- name: kill kp_daemon.sh
shell: "if [ `ps aux | grep kp_daemon | grep -v grep | awk '{print $2}' | wc -l` -ne 0 ] ; then ps aux | grep kp_daemon | grep -v grep | awk '{print $2}' | xargs kill -9 ; fi"
- block:
- name: Check if the Keepalived service already exists
shell: source /etc/profile && rpm -qa | grep keepalived | wc -l
register: check_out
- name: copy unload_keepalived.sh
template:
src: unload_keepalived.sh
dest: /root/
mode: 0755
- name: unload keepalived
shell: sh /root/unload_keepalived.sh
when: check_out.stdout >= '1'
- name: Check if the keepalived service already exists
shell: source /etc/profile && rpm -qa | grep keepalived | grep -v 'keepalived-1.3.5-1.el7.x86_64' | wc -l
register: check_out
- name: delete unload_keepalived.sh
file:
path: "/root/unload_keepalived.sh"
state: absent
- name: To terminate execution
fail:
msg: "卸载失败,组件可能非本安装部署,请联系开发确认或手动卸载后继续安装"
run_once: true
delegate_to: 127.0.0.1
when: check_out.stdout >= '1'
when: (allowed_unload) == "yes"
#创建文件夹
- name: create keepalived package path:{{ keepalived_package_path }}
file:

View File

@@ -67,12 +67,17 @@ installKP(){
chmod +x $keepalivedInstallPath/$keepalivedDaeonName
echo "将$keepalivedInstallPath/$keepalivedDaeonName守护进程添加到开机自启动中"
chmod +x /etc/rc.d/rc.local
echo -e "\n#设置$keepalivedDaeonName守护脚本开机自启动" >> /etc/rc.d/rc.local
echo "nohup $keepalivedInstallPath/$keepalivedDaeonName >/dev/null 2>&1 &" >> /etc/rc.d/rc.local
echo "将$keepalivedInstallPath/$keepalivedDaeonName守护进程添加到开机自启动成功"
chmod +x /etc/rc.d/rc.local
upStartNum=`cat /etc/rc.d/rc.local | grep "$keepalivedInstallPath" | wc -l`
if [ $upStartNum -eq "0" ];then
echo -e "\n#设置$keepalivedDaeonName守护脚本开机自启动" >> /etc/rc.d/rc.local
echo "nohup $keepalivedInstallPath/$keepalivedDaeonName >/dev/null 2>&1 &" >> /etc/rc.d/rc.local
fi
echo "开始启动$keepalivedInstallPath/$keepalivedDaeonName守护进程"
sed -i -e 's#. /etc/sysconfig/keepalived#. /usr/local/etc/sysconfig/keepalived#g' /etc/init.d/keepalived
systemctl daemon-reload
nohup $keepalivedInstallPath/$keepalivedDaeonName >/dev/null 2>&1 &
rm -rf $keepalivedBagPath

View File

@@ -0,0 +1,9 @@
#!/bin/bash
#for i in `rpm -qa | grep keepalived` ; do rpm -e --nodeps $i ; done
if [ `ps aux | grep -E "keepalived|kp_daemon" | grep -vE "grep|unload_keepalived" | wc -l` -ne '0' ] ; then ps aux | grep -E "keepalived|kp_daemon" | grep -vE "grep|unload_keepalived" | awk '{print $2}' | xargs kill -9 ; fi
rm -rf /etc/keepalived
rm -rf /etc/init.d/keepalived

View File

@@ -1,3 +1,6 @@
- name: stop keepalived
shell: if [ `ps aux | grep kp_daemon | grep -v grep | awk '{print $2}' | wc -l` -ne '0' ] ; then ps aux | grep kp_daemon | grep -v grep | awk '{print $2}' | xargs kill -9 ; fi && service keepalived stop
- name: stop and remove {{ hos_nginx_image_container_name }} container
docker_container:
name: '{{ hos_nginx_image_container_name }}'
@@ -6,20 +9,20 @@
delegate_facts: true
delegate_to: "{{ item.ip }}"
with_items:
- { ip: '{{ groups.hosnginx[0] }}' }
- { ip: '{{ groups.hosnginx[1] }}' }
- { ip: '{{ groups.hoskeepalive[0] }}' }
- { ip: '{{ groups.hoskeepalive[1] }}' }
- name: remove old {{ hos_nginx_hos_service_image_name }} image
- name: remove old {{ hos_nginx_image_name }} image
docker_image:
name: '{{ hos_nginx_hos_service_image_name }}'
tag: '{{ hos_nginx_hos_service_image_tag_name }}'
name: '{{ hos_nginx_image_name }}'
tag: '{{ hos_nginx_image_tag_name }}'
state: absent
run_once: true
delegate_facts: true
delegate_to: "{{ item.ip }}"
with_items:
- { ip: '{{ groups.hosnginx[0] }}' }
- { ip: '{{ groups.hosnginx[1] }}' }
- { ip: '{{ groups.hoskeepalive[0] }}' }
- { ip: '{{ groups.hoskeepalive[1] }}' }
- name: create {{ hos_nginx_volume_path }}
file:
@@ -29,8 +32,8 @@
delegate_facts: true
delegate_to: "{{ item.ip }}"
with_items:
- { ip: '{{ groups.hosnginx[0] }}' }
- { ip: '{{ groups.hosnginx[1] }}' }
- { ip: '{{ groups.hoskeepalive[0] }}' }
- { ip: '{{ groups.hoskeepalive[1] }}' }
- name: create {{ install_path }}/{{ hos_nginx_soft_home_path }}
file:
@@ -40,12 +43,12 @@
delegate_facts: true
delegate_to: "{{ item.ip }}"
with_items:
- { ip: '{{ groups.hosnginx[0] }}' }
- { ip: '{{ groups.hosnginx[1] }}' }
- { ip: '{{ groups.hoskeepalive[0] }}' }
- { ip: '{{ groups.hoskeepalive[1] }}' }
- name: copy {{ hos_nginx_hos_service_image_tar_name }}
- name: copy {{ hos_nginx_image_tar_name }}
copy:
src: '{{ package_path }}/{{ hos_nginx_hos_service_image_tar_name }}'
src: '{{ package_path }}/{{ hos_nginx_image_tar_name }}'
dest: '{{ install_path }}/{{ hos_nginx_soft_home_path }}/'
force: true
backup: yes
@@ -53,14 +56,14 @@
delegate_facts: true
delegate_to: "{{ item.ip }}"
with_items:
- { ip: '{{ groups.hosnginx[0] }}' }
- { ip: '{{ groups.hosnginx[1] }}' }
- { ip: '{{ groups.hoskeepalive[0] }}' }
- { ip: '{{ groups.hoskeepalive[1] }}' }
- name: load new {{ hos_nginx_hos_service_image_name }} image from {{ hos_nginx_hos_service_image_tar_name }}
- name: load new {{ hos_nginx_image_name }} image from {{ hos_nginx_image_tar_name }}
docker_image:
name: '{{ hos_nginx_hos_service_image_name }}'
tag: '{{ hos_nginx_hos_service_image_tag_name }}'
load_path: '{{ install_path }}/{{ hos_nginx_soft_home_path }}/{{ hos_nginx_hos_service_image_tar_name }}'
name: '{{ hos_nginx_image_name }}'
tag: '{{ hos_nginx_image_tag_name }}'
load_path: '{{ install_path }}/{{ hos_nginx_soft_home_path }}/{{ hos_nginx_image_tar_name }}'
source: load
force_tag: yes
force_source: yes
@@ -68,19 +71,19 @@
delegate_facts: true
delegate_to: "{{ item.ip }}"
with_items:
- { ip: '{{ groups.hosnginx[0] }}' }
- { ip: '{{ groups.hosnginx[1] }}' }
- { ip: '{{ groups.hoskeepalive[0] }}' }
- { ip: '{{ groups.hoskeepalive[1] }}' }
- name: change the image tag
shell: docker tag nginx-metrics:{{ hos_nginx_hos_service_image_tag_name }} {{ hos_nginx_hos_service_image_name }}:{{ hos_nginx_hos_service_image_tag_name }}
shell: docker tag nginx-metrics:{{ hos_nginx_image_tag_name }} {{ hos_nginx_image_name }}:{{ hos_nginx_image_tag_name }}
run_once: true
delegate_facts: true
delegate_to: "{{ item.ip }}"
with_items:
- { ip: '{{ groups.hosnginx[0] }}' }
- { ip: '{{ groups.hosnginx[1] }}' }
- { ip: '{{ groups.hoskeepalive[0] }}' }
- { ip: '{{ groups.hoskeepalive[1] }}' }
- name: copy {{ hos_nginx_hos_service_image_name }} docker-compose.yml
- name: copy {{ hos_nginx_image_name }} docker-compose.yml
template:
src: docker-compose.yml.j2
dest: '{{ install_path }}/{{ hos_nginx_soft_home_path }}/docker-compose.yml'
@@ -89,10 +92,10 @@
delegate_facts: true
delegate_to: "{{ item.ip }}"
with_items:
- { ip: '{{ groups.hosnginx[0] }}' }
- { ip: '{{ groups.hosnginx[1] }}' }
- { ip: '{{ groups.hoskeepalive[0] }}' }
- { ip: '{{ groups.hoskeepalive[1] }}' }
- name: copy {{ hos_nginx_hos_service_image_name }} nginx.conf
- name: copy {{ hos_nginx_image_name }} nginx.conf
template:
src: nginx.conf.j2
dest: '{{ install_path }}/{{ hos_nginx_volume_path }}/nginx.conf'
@@ -101,8 +104,8 @@
delegate_facts: true
delegate_to: "{{ item.ip }}"
with_items:
- { ip: '{{ groups.hosnginx[0] }}' }
- { ip: '{{ groups.hosnginx[1] }}' }
- { ip: '{{ groups.hoskeepalive[0] }}' }
- { ip: '{{ groups.hoskeepalive[1] }}' }
- name: start {{ hos_nginx_image_container_name }} container
docker_compose:
@@ -111,5 +114,6 @@
delegate_facts: true
delegate_to: "{{ item.ip }}"
with_items:
- { ip: '{{ groups.hosnginx[0] }}' }
- { ip: '{{ groups.hosnginx[1] }}' }
- { ip: '{{ groups.hoskeepalive[0] }}' }
- { ip: '{{ groups.hoskeepalive[1] }}' }

View File

@@ -1,7 +1,7 @@
version: '3'
services:
nginx:
image: {{ hos_nginx_hos_service_image_name }}:{{ hos_nginx_hos_service_image_tag_name }}
image: {{ hos_nginx_image_name }}:{{ hos_nginx_image_tag_name }}
container_name: {{ hos_nginx_image_container_name }}
restart: always
ports:

View File

@@ -67,7 +67,7 @@
backup: yes
- name: push config
shell: 'curl --data-urlencode content="`cat {{ data_path }}/{{ hos_service_soft_home_path }}/galaxy-hos-service`" --request POST "http://{{ gateway_keepalive_host }}:8848/nacos/v1/cs/configs?username=nacos&password={{ nacos_pin }}&tenant={{ services_config_namespace }}&group=Galaxy&dataId=galaxy-hos-service.yml&appName=galaxy-hos-service&type=yaml"'
shell: 'curl --retry-delay 3 --retry 3 --data-urlencode content="`cat {{ data_path }}/{{ hos_service_soft_home_path }}/galaxy-hos-service`" --request POST "http://{{ gateway_keepalive_host }}:8848/nacos/v1/cs/configs?username=nacos&password={{ nacos_pin }}&tenant={{ services_config_namespace }}&group=Galaxy&dataId=galaxy-hos-service.yml&appName=galaxy-hos-service&type=yaml"'
register: change_out
- fail:
@@ -77,3 +77,20 @@
- name: start {{ hos_service_image_container_name }} container
docker_compose:
project_src: '{{ data_path }}/{{ hos_service_soft_home_path }}'
##获取 ip 列表
#- name: get storm numbus iplist
# shell: echo "{{ ansible_play_hosts }}" | grep -E -o "[0-9]+\.[0-9]+\.[0-9]+\.[0-9]+" > {{ data_path }}/{{ hos_service_soft_home_path }}/ip_list
#
#- name: copy {{ hos_service_image_name }} registerinstance.sh.j2
# template:
# src: registerinstance.sh.j2
# dest: '{{ data_path }}/{{ hos_service_soft_home_path }}/registerinstance.sh'
# backup: yes
#
#- name: change the image tag
# shell: cd {{ data_path }}/{{ hos_service_soft_home_path }} && chmod +x registerinstance.sh && ./registerinstance.sh

View File

@@ -3,7 +3,7 @@ nacos:
type: yaml
server-addr: {{ gateway_keepalive_host }}:8848
namespace: {{ services_config_namespace }}
data-id: galaxy-hos-service
data-id: galaxy-hos-service.yml
auto-refresh: true
group: Galaxy
username: nacos

View File

@@ -0,0 +1,19 @@
#!/bin/bash
basedir=`pwd`
serverAddr="http://{{ gateway_keepalive_host }}:8848"
namespace="{{ services_config_namespace }}"
group="Galaxy"
username="nacos"
password="{{ nacos_pin }}"
serviceName="hos"
hosPort="8186"
for ip in `cat $basedir/hosiplist`
do
issuccess=`curl --retry-delay 3 --retry 3 -X POST ''$serverAddr'/nacos/v1/ns/instance?serviceName='$serviceName'&ip='$ip'&port='$hosPort'&namespaceId='$namespace'&groupName='$group'&ephemeral=false&username='$username'&password='$password''`
if [ `echo $issuccess | grep ok | wc -l` -eq 0 ];then
echo "register $ip error"
fi
done

View File

@@ -34,6 +34,19 @@
tag: '{{ admin_image_tag_name }}'
state: absent
- name: stop and remove {{ job_image_container_name }} container
docker_container:
name: '{{ job_image_container_name }}'
state: absent
- name: remove old {{ job_image_name }} image
docker_image:
name: '{{ job_image_name }}'
tag: '{{ job_image_tag_name }}'
state: absent
- name: create {{ admin_volume_path }}
file:
state: directory

View File

@@ -91,7 +91,7 @@
backup: yes
- name: push config
shell: 'curl --data-urlencode content="`cat {{ install_path }}/{{ job_soft_home_path }}/galaxy-job-executor`" --request POST "http://{{ gateway_keepalive_host }}:8848/nacos/v1/cs/configs?username=nacos&password={{ nacos_pin }}&tenant={{ services_config_namespace }}&group=Galaxy&dataId=galaxy-job-executor.properties&appName=galaxy-job-executor&type=properties"'
shell: 'curl --retry-delay 3 --retry 3 --data-urlencode content="`cat {{ install_path }}/{{ job_soft_home_path }}/galaxy-job-executor`" --request POST "http://{{ gateway_keepalive_host }}:8848/nacos/v1/cs/configs?username=nacos&password={{ nacos_pin }}&tenant={{ services_config_namespace }}&group=Galaxy&dataId=galaxy-job-executor.properties&appName=galaxy-job-executor&type=properties"'
register: change_out
- fail:
@@ -99,7 +99,7 @@
when: change_out.stdout != 'true'
- name: push config
shell: 'curl --data-urlencode content="`cat {{ install_path }}/{{ job_soft_home_path }}/galaxy-job-admin`" --request POST "http://{{ gateway_keepalive_host }}:8848/nacos/v1/cs/configs?username=nacos&password={{ nacos_pin }}&tenant={{ services_config_namespace }}&group=Galaxy&dataId=galaxy-job-admin.properties&appName=galaxy-job-admin&type=properties"'
shell: 'curl --retry-delay 3 --retry 3 --data-urlencode content="`cat {{ install_path }}/{{ job_soft_home_path }}/galaxy-job-admin`" --request POST "http://{{ gateway_keepalive_host }}:8848/nacos/v1/cs/configs?username=nacos&password={{ nacos_pin }}&tenant={{ services_config_namespace }}&group=Galaxy&dataId=galaxy-job-admin.properties&appName=galaxy-job-admin&type=properties"'
register: change_out
- fail:

View File

@@ -3,7 +3,7 @@ version: '2'
services:
galaxy-job-executor:
image: {{ job_image_name }}:{{ job_image_tag_name }}
container_name: galaxy-job
container_name: {{ job_image_container_name }}
environment:
JAVA_OPTS: "{{ job_java_opts }}"
ports:
@@ -18,7 +18,7 @@ services:
galaxy-job-admin:
image: {{ admin_image_name }}:{{ admin_image_tag_name }}
container_name: xxl-job-admin
container_name: {{ admin_image_container_name }}
environment:
JAVA_OPTS: "{{ job_java_opts }}"
ports:

Binary file not shown.

Before

Width:  |  Height:  |  Size: 80 MiB

After

Width:  |  Height:  |  Size: 80 MiB

View File

@@ -1,3 +1,14 @@
- name: start keepalived service
shell: service keepalived start
- name: check keepalived
shell: ps aux | grep kp_daemon | grep -v grep | wc -l
register: daemon_stat
- name: start keepalived
shell: nohup /bin/bash /etc/keepalived/kp_daemon.sh >/dev/null 2>&1 &
when: daemon_stat.stdout == '0'
- name: get zookeeper_servers to ansible variable
set_fact: zookeeper_servers="{{groups.zookeeper[0]}}:2181,{{groups.zookeeper[1]}}:2181,{{groups.zookeeper[2]}}:2181"
when: '(groups.zookeeper|length) == 3'
@@ -88,7 +99,7 @@
backup: yes
#- name: push config
# shell: 'curl --data-urlencode content="`cat {{ install_path }}/{{ qgw_soft_home_path }}/galaxy-qgw-service`" --request POST "http://{{ gateway_keepalive_host }}:8848/nacos/v1/cs/configs?username=nacos&password={{ nacos_pin }}&tenant={{ services_config_namespace }}&group=Galaxy&dataId=galaxy-qgw-service.yml&appName=galaxy-qgw-service&type=yaml"'
# shell: 'curl --retry-delay 3 --retry 3 --data-urlencode content="`cat {{ install_path }}/{{ qgw_soft_home_path }}/galaxy-qgw-service`" --request POST "http://{{ gateway_keepalive_host }}:8848/nacos/v1/cs/configs?username=nacos&password={{ nacos_pin }}&tenant={{ services_config_namespace }}&group=Galaxy&dataId=galaxy-qgw-service.yml&appName=galaxy-qgw-service&type=yaml"'
# register: change_out
#- debug:

View File

@@ -31,7 +31,7 @@ do
#3.提交config配置
for (( i = 0; i < 10; i++ )); do
if [ "$suffix" = "yaml" ]||[ "$suffix" = "json" ]||[ "$suffix" = "text" ]||[ "$suffix" = "properties" ];then
result=$(curl -sw '%{http_code}' -o /dev/null --request POST "$nacos_push_url&dataId=$data_id&appName=$appName&type=$suffix" --data-urlencode content="`cat $config_path_file`")
result=$(curl --retry-delay 3 --retry 3 -sw '%{http_code}' -o /dev/null --request POST "$nacos_push_url&dataId=$data_id&appName=$appName&type=$suffix" --data-urlencode content="`cat $config_path_file`")
echo "push config response code "$result
if [[ $result -eq '200' ]];then
i=10

View File

@@ -63,7 +63,7 @@
backup: yes
- name: push config
shell: 'curl --data-urlencode content="`cat {{ install_path }}/{{ report_soft_home_path }}/galaxy-report-service`" --request POST "http://{{ gateway_keepalive_host }}:8848/nacos/v1/cs/configs?username=nacos&password={{ nacos_pin }}&tenant={{ services_config_namespace }}&group=Galaxy&dataId=galaxy-report-service.yml&appName=galaxy-report-service&type=yaml"'
shell: 'curl --retry-delay 3 --retry 3 --data-urlencode content="`cat {{ install_path }}/{{ report_soft_home_path }}/galaxy-report-service`" --request POST "http://{{ gateway_keepalive_host }}:8848/nacos/v1/cs/configs?username=nacos&password={{ nacos_pin }}&tenant={{ services_config_namespace }}&group=Galaxy&dataId=galaxy-report-service.yml&appName=galaxy-report-service&type=yaml"'
register: change_out
- fail:

View File

@@ -27,6 +27,18 @@
# msg: "{{groups['flink'].index}}"
when: index_no >= 0
- name: test mysql
shell: which mysqlk
register: result
failed_when: "'FAILED' in result.stderr or result.stdout != '20'"
ignore_errors: True
- name: output result
debug:
msg: "shuchu: {{ result }}"
#when: result.stderr is defined and result.stderr != '' #有错误信息,说明mariadb服务不存在
when: result is undefined or result.stdout != '20'
vars_files:
- ../configurations/config.yml
- ../configurations/components.yml