Merge branch 'feature-181031' into develop
This commit is contained in:
2
pom.xml
2
pom.xml
@@ -472,7 +472,7 @@
|
|||||||
<dependency>
|
<dependency>
|
||||||
<groupId>org.apache.hive</groupId>
|
<groupId>org.apache.hive</groupId>
|
||||||
<artifactId>hive-jdbc</artifactId>
|
<artifactId>hive-jdbc</artifactId>
|
||||||
<version>2.1.1</version>
|
<version>2.3.3</version>
|
||||||
<exclusions>
|
<exclusions>
|
||||||
<exclusion>
|
<exclusion>
|
||||||
<groupId>org.apache.logging.log4j</groupId>
|
<groupId>org.apache.logging.log4j</groupId>
|
||||||
|
|||||||
@@ -8,7 +8,6 @@ import org.springframework.beans.BeansException;
|
|||||||
import org.springframework.beans.factory.config.ConfigurableListableBeanFactory;
|
import org.springframework.beans.factory.config.ConfigurableListableBeanFactory;
|
||||||
import org.springframework.beans.factory.config.PropertyPlaceholderConfigurer;
|
import org.springframework.beans.factory.config.PropertyPlaceholderConfigurer;
|
||||||
|
|
||||||
|
|
||||||
public class PropertyPlaceholderConfigurerCrypt extends PropertyPlaceholderConfigurer {
|
public class PropertyPlaceholderConfigurerCrypt extends PropertyPlaceholderConfigurer {
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
@@ -16,89 +15,29 @@ public class PropertyPlaceholderConfigurerCrypt extends PropertyPlaceholderConfi
|
|||||||
throws BeansException {
|
throws BeansException {
|
||||||
|
|
||||||
try {
|
try {
|
||||||
String productPassword = props.getProperty("jdbc.product.password");
|
// mysql
|
||||||
String productScretKey = props.getProperty("jdbc.product.key");
|
|
||||||
if (null != productPassword) {
|
|
||||||
props.setProperty("jdbc.product.password",
|
|
||||||
new String(AESUtil.decrypt(Base64.decodeBase64(productPassword), productScretKey)));
|
|
||||||
}
|
|
||||||
|
|
||||||
String devlopPassword = props.getProperty("jdbc.devlop.password");
|
|
||||||
String devlopScretKey = props.getProperty("jdbc.devlop.key");
|
|
||||||
if (null != devlopPassword) {
|
|
||||||
props.setProperty("jdbc.devlop.password",
|
|
||||||
new String(AESUtil.decrypt(Base64.decodeBase64(devlopPassword), devlopScretKey)));
|
|
||||||
}
|
|
||||||
|
|
||||||
String logPassword = props.getProperty("jdbc.log.password");
|
String logPassword = props.getProperty("jdbc.log.password");
|
||||||
String logScretKey = props.getProperty("jdbc.log.key");
|
String logScretKey = props.getProperty("jdbc.log.key");
|
||||||
if (null != logPassword) {
|
if (null != logPassword) {
|
||||||
props.setProperty("jdbc.log.password",
|
props.setProperty("jdbc.log.password",
|
||||||
new String(AESUtil.decrypt(Base64.decodeBase64(logPassword), logScretKey)));
|
new String(AESUtil.decrypt(Base64.decodeBase64(logPassword), logScretKey)));
|
||||||
}
|
}
|
||||||
|
|
||||||
// 日志A版
|
|
||||||
String logAPassword = props.getProperty("jdbc.logA.password");
|
|
||||||
String logAScretKey = props.getProperty("jdbc.logA.key");
|
|
||||||
if (null != logAPassword) {
|
|
||||||
props.setProperty("jdbc.logA.password",
|
|
||||||
new String(AESUtil.decrypt(Base64.decodeBase64(logAPassword), logAScretKey)));
|
|
||||||
}
|
|
||||||
// 日志A版
|
|
||||||
String logCPassword = props.getProperty("jdbc.logC.password");
|
|
||||||
String logCScretKey = props.getProperty("jdbc.logC.key");
|
|
||||||
if (null != logAPassword) {
|
|
||||||
props.setProperty("jdbc.logC.password",
|
|
||||||
new String(AESUtil.decrypt(Base64.decodeBase64(logCPassword), logCScretKey)));
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
// 测试使用,后期会删除
|
// clickhouse
|
||||||
String testPassword = props.getProperty("jdbc.test.password");
|
|
||||||
String testScretKey = props.getProperty("jdbc.test.key");
|
|
||||||
if (null != testPassword) {
|
|
||||||
props.setProperty("jdbc.test.password",
|
|
||||||
new String(AESUtil.decrypt(Base64.decodeBase64(testPassword), testScretKey)));
|
|
||||||
}
|
|
||||||
|
|
||||||
String jkPzPassword = props.getProperty("jdbc.jk.password");
|
|
||||||
String jkPzScretKey = props.getProperty("jdbc.jk.key");
|
|
||||||
if (null != jkPzPassword) {
|
|
||||||
props.setProperty("jdbc.jk.password",
|
|
||||||
new String(AESUtil.decrypt(Base64.decodeBase64(jkPzPassword), jkPzScretKey)));
|
|
||||||
}
|
|
||||||
//A版hive库
|
|
||||||
String hiveAPassword = props.getProperty("jdbc.hiveA.password");
|
|
||||||
String hiveAScretKey = props.getProperty("jdbc.hiveA.key");
|
|
||||||
if (null != hiveAPassword) {
|
|
||||||
props.setProperty("jdbc.hiveA.password",
|
|
||||||
new String(AESUtil.decrypt(Base64.decodeBase64(hiveAPassword), hiveAScretKey)));
|
|
||||||
}
|
|
||||||
|
|
||||||
//B版hive库
|
|
||||||
String hiveBPassword = props.getProperty("jdbc.hiveB.password");
|
|
||||||
String hiveBScretKey = props.getProperty("jdbc.hiveB.key");
|
|
||||||
if (null != hiveBPassword) {
|
|
||||||
props.setProperty("jdbc.hiveB.password",
|
|
||||||
new String(AESUtil.decrypt(Base64.decodeBase64(hiveBPassword), hiveBScretKey)));
|
|
||||||
}
|
|
||||||
|
|
||||||
//神通数据库库
|
|
||||||
String clusterPassword = props.getProperty("jdbc.log.cluster.password");
|
|
||||||
String clusterScretKey = props.getProperty("jdbc.log.cluster.key");
|
|
||||||
if (null != clusterPassword) {
|
|
||||||
props.setProperty("jdbc.log.cluster.password",
|
|
||||||
new String(AESUtil.decrypt(Base64.decodeBase64(clusterPassword), clusterScretKey)));
|
|
||||||
}
|
|
||||||
|
|
||||||
//clickhouse
|
|
||||||
String clickHousePassword = props.getProperty("jdbc.clickhouse.password");
|
String clickHousePassword = props.getProperty("jdbc.clickhouse.password");
|
||||||
String clickHouseScretKey = props.getProperty("jdbc.clickhouse.key");
|
String clickHouseScretKey = props.getProperty("jdbc.clickhouse.key");
|
||||||
if (null != clickHousePassword) {
|
if (null != clickHousePassword) {
|
||||||
props.setProperty("jdbc.clickhouse.password",
|
props.setProperty("jdbc.clickhouse.password",
|
||||||
new String(AESUtil.decrypt(Base64.decodeBase64(clickHousePassword), clickHouseScretKey)));
|
new String(AESUtil.decrypt(Base64.decodeBase64(clickHousePassword), clickHouseScretKey)));
|
||||||
}
|
}
|
||||||
|
// hive
|
||||||
|
String hivePassword = props.getProperty("jdbc.hive.password");
|
||||||
|
String hiveScretKey = props.getProperty("jdbc.hive.key");
|
||||||
|
if (null != clickHousePassword) {
|
||||||
|
props.setProperty("jdbc.hive.password",
|
||||||
|
new String(AESUtil.decrypt(Base64.decodeBase64(hivePassword), hiveScretKey)));
|
||||||
|
}
|
||||||
|
|
||||||
} catch (Exception e) {
|
} catch (Exception e) {
|
||||||
e.printStackTrace();
|
e.printStackTrace();
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1409,7 +1409,7 @@
|
|||||||
<result column="found_Time" jdbcType="TIMESTAMP" property="foundTime" />
|
<result column="found_Time" jdbcType="TIMESTAMP" property="foundTime" />
|
||||||
<result column="recv_Time" jdbcType="TIMESTAMP" property="recvTime" />
|
<result column="recv_Time" jdbcType="TIMESTAMP" property="recvTime" />
|
||||||
<result column="cap_ip" jdbcType="VARCHAR" property="capIp" />
|
<result column="cap_ip" jdbcType="VARCHAR" property="capIp" />
|
||||||
<result column="voip_rotocol" jdbcType="VARCHAR" property="voipProtocol" />
|
<result column="voip_protocol" jdbcType="VARCHAR" property="voipProtocol" />
|
||||||
<result column="rtp_d_ip" jdbcType="VARCHAR" property="rtpDIp" />
|
<result column="rtp_d_ip" jdbcType="VARCHAR" property="rtpDIp" />
|
||||||
<result column="rtp_s_ip" jdbcType="VARCHAR" property="rtpSIp" />
|
<result column="rtp_s_ip" jdbcType="VARCHAR" property="rtpSIp" />
|
||||||
<result column="rtp_d_port" jdbcType="VARCHAR" property="rtpDPort" />
|
<result column="rtp_d_port" jdbcType="VARCHAR" property="rtpDPort" />
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
package com.nis.util;
|
package com.nis.web.dao.impl;
|
||||||
|
|
||||||
import java.beans.BeanInfo;
|
import java.beans.BeanInfo;
|
||||||
import java.beans.Introspector;
|
import java.beans.Introspector;
|
||||||
@@ -20,9 +20,11 @@ import org.apache.ibatis.mapping.ResultMapping;
|
|||||||
import org.apache.ibatis.session.SqlSessionFactory;
|
import org.apache.ibatis.session.SqlSessionFactory;
|
||||||
import org.slf4j.Logger;
|
import org.slf4j.Logger;
|
||||||
import org.slf4j.LoggerFactory;
|
import org.slf4j.LoggerFactory;
|
||||||
|
import org.springframework.stereotype.Repository;
|
||||||
|
|
||||||
import com.alibaba.druid.pool.DruidDataSource;
|
import com.alibaba.druid.pool.DruidDataSource;
|
||||||
import com.nis.domain.Page;
|
import com.nis.domain.Page;
|
||||||
|
import com.nis.util.Constants;
|
||||||
import com.nis.web.service.SpringContextHolder;
|
import com.nis.web.service.SpringContextHolder;
|
||||||
import com.zdjizhi.utils.StringUtil;
|
import com.zdjizhi.utils.StringUtil;
|
||||||
|
|
||||||
@@ -42,6 +44,7 @@ import com.zdjizhi.utils.StringUtil;
|
|||||||
* @date 2018年8月20日
|
* @date 2018年8月20日
|
||||||
*
|
*
|
||||||
*/
|
*/
|
||||||
|
@Repository
|
||||||
public class LogJDBCByDruid {
|
public class LogJDBCByDruid {
|
||||||
private final static Logger logger = LoggerFactory.getLogger(LogJDBCByDruid.class);
|
private final static Logger logger = LoggerFactory.getLogger(LogJDBCByDruid.class);
|
||||||
static DruidDataSource datasource = null;
|
static DruidDataSource datasource = null;
|
||||||
@@ -2,9 +2,6 @@ package com.nis.web.service;
|
|||||||
|
|
||||||
import java.lang.reflect.Field;
|
import java.lang.reflect.Field;
|
||||||
import java.lang.reflect.Method;
|
import java.lang.reflect.Method;
|
||||||
import java.sql.Connection;
|
|
||||||
import java.sql.ResultSet;
|
|
||||||
import java.sql.Statement;
|
|
||||||
import java.text.SimpleDateFormat;
|
import java.text.SimpleDateFormat;
|
||||||
import java.util.ArrayList;
|
import java.util.ArrayList;
|
||||||
import java.util.HashMap;
|
import java.util.HashMap;
|
||||||
@@ -14,13 +11,13 @@ import java.util.Map;
|
|||||||
import org.apache.ibatis.mapping.ResultMap;
|
import org.apache.ibatis.mapping.ResultMap;
|
||||||
import org.apache.ibatis.mapping.ResultMapping;
|
import org.apache.ibatis.mapping.ResultMapping;
|
||||||
import org.apache.ibatis.session.SqlSessionFactory;
|
import org.apache.ibatis.session.SqlSessionFactory;
|
||||||
|
import org.springframework.beans.factory.annotation.Autowired;
|
||||||
import org.springframework.stereotype.Service;
|
import org.springframework.stereotype.Service;
|
||||||
|
|
||||||
import com.alibaba.druid.pool.DruidDataSource;
|
|
||||||
import com.nis.domain.Page;
|
import com.nis.domain.Page;
|
||||||
import com.nis.util.Configurations;
|
import com.nis.util.Configurations;
|
||||||
import com.nis.util.Constants;
|
import com.nis.util.Constants;
|
||||||
import com.nis.util.LogJDBCByDruid;
|
import com.nis.web.dao.impl.LogJDBCByDruid;
|
||||||
import com.zdjizhi.utils.StringUtil;
|
import com.zdjizhi.utils.StringUtil;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -33,10 +30,9 @@ import com.zdjizhi.utils.StringUtil;
|
|||||||
public class LogDataService {
|
public class LogDataService {
|
||||||
// private final static Logger logger =
|
// private final static Logger logger =
|
||||||
// LoggerFactory.getLogger(LogDataService.class);
|
// LoggerFactory.getLogger(LogDataService.class);
|
||||||
static DruidDataSource datasource = null;
|
|
||||||
Connection conn = null;
|
@Autowired
|
||||||
ResultSet rs = null;
|
private LogJDBCByDruid logJDBCByDruid;
|
||||||
Statement st = null;
|
|
||||||
private static SimpleDateFormat sdf = new SimpleDateFormat("yyyy-MM-dd HH:mm:ss");
|
private static SimpleDateFormat sdf = new SimpleDateFormat("yyyy-MM-dd HH:mm:ss");
|
||||||
private static SimpleDateFormat sdf2 = new SimpleDateFormat("yyyyMMdd");
|
private static SimpleDateFormat sdf2 = new SimpleDateFormat("yyyyMMdd");
|
||||||
private static Map<String, Map<String, String>> col2col = new HashMap<String, Map<String, String>>();
|
private static Map<String, Map<String, String>> col2col = new HashMap<String, Map<String, String>>();
|
||||||
@@ -189,7 +185,7 @@ public class LogDataService {
|
|||||||
}
|
}
|
||||||
Integer startNum = (page.getPageNo() - 1) * page.getPageSize();
|
Integer startNum = (page.getPageNo() - 1) * page.getPageSize();
|
||||||
StringBuffer foundTimeSql = new StringBuffer();
|
StringBuffer foundTimeSql = new StringBuffer();
|
||||||
foundTimeSql.append("select distinct found_time from " + tableName + " where ");
|
foundTimeSql.append("select found_time from " + tableName + " where ");
|
||||||
if (whereSB.length() == 0) {// 没有其他查询条件只有默认的found_time条件
|
if (whereSB.length() == 0) {// 没有其他查询条件只有默认的found_time条件
|
||||||
if (whereFoundTime.length() > 0) {
|
if (whereFoundTime.length() > 0) {
|
||||||
int indexOf = whereFoundTime.indexOf("and") + "and".length();
|
int indexOf = whereFoundTime.indexOf("and") + "and".length();
|
||||||
@@ -365,10 +361,11 @@ public class LogDataService {
|
|||||||
*/
|
*/
|
||||||
private <T> void searchFromDataCenter(Page<T> page, Object bean, StringBuffer selSql, StringBuffer countSql)
|
private <T> void searchFromDataCenter(Page<T> page, Object bean, StringBuffer selSql, StringBuffer countSql)
|
||||||
throws Exception {
|
throws Exception {
|
||||||
new LogJDBCByDruid().getTableData(page, selSql.toString(), bean.getClass());
|
// new LogJDBCByDruid().getTableData(page, selSql.toString(), bean.getClass());
|
||||||
|
logJDBCByDruid.getTableData(page, selSql.toString(), bean.getClass());
|
||||||
if (Constants.ISOPENLOGCOUNTANDLAST) {
|
if (Constants.ISOPENLOGCOUNTANDLAST) {
|
||||||
if (page.getList() != null && page.getList().size() > 0) {
|
if (page.getList() != null && page.getList().size() > 0) {
|
||||||
new LogJDBCByDruid().getCount(page, countSql.toString().toLowerCase());
|
logJDBCByDruid.getCount(page, countSql.toString().toLowerCase());
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -16,7 +16,6 @@ import com.nis.domain.restful.ConfigPzIdSource;
|
|||||||
import com.zdjizhi.utils.StringUtil;
|
import com.zdjizhi.utils.StringUtil;
|
||||||
import com.nis.web.dao.ConfigPzIdDao;
|
import com.nis.web.dao.ConfigPzIdDao;
|
||||||
import com.nis.web.service.BaseLogService;
|
import com.nis.web.service.BaseLogService;
|
||||||
import com.sun.tools.internal.xjc.reader.xmlschema.bindinfo.BIConversion.Static;
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* @ClassName:ConfigPzIdService
|
* @ClassName:ConfigPzIdService
|
||||||
|
|||||||
47
src/main/resources/applicationLog-hive.properties.xinghan
Normal file
47
src/main/resources/applicationLog-hive.properties.xinghan
Normal file
@@ -0,0 +1,47 @@
|
|||||||
|
#由于数据中心hive日志表的表名会变动,所以本系统中将日志的表名提取到配置文件中方便后期修改,key是bean的名称+HiveTable,value是hive的实际表名
|
||||||
|
NtcIpLogHiveTable=singhand_NTC_IP_LOG
|
||||||
|
NtcHttpLogHiveTable=singhand_NTC_HTTP_LOG
|
||||||
|
NtcDnsLogHiveTable=singhand_NTC_DNS_LOG
|
||||||
|
NtcMailLogHiveTable=singhand_NTC_MAIL_LOG
|
||||||
|
NtcSslLogHiveTable=singhand_NTC_SSL_LOG
|
||||||
|
NtcPptpLogHiveTable=singhand_NTC_PPTP_LOG
|
||||||
|
NtcL2tpLogHiveTable=singhand_NTC_L2TP_LOG
|
||||||
|
NtcOpenvpnLogHiveTable=singhand_NTC_OPENVPN_LOG
|
||||||
|
NtcIpsecLogHiveTable=singhand_NTC_IPSEC_LOG
|
||||||
|
NtcSshLogHiveTable=singhand_NTC_SSH_LOG
|
||||||
|
NtcFtpLogHiveTable=singhand_NTC_FTP_LOG
|
||||||
|
NtcAppLogHiveTable=singhand_NTC_APP_LOG
|
||||||
|
NtcDdosLogHiveTable=singhand_NTC_DDOS_LOG
|
||||||
|
NtcP2pLogHiveTable=singhand_NTC_P2P_LOG
|
||||||
|
NtcBgpLogHiveTable=singhand_NTC_BGP_LOG
|
||||||
|
DkBehaviorLogHiveTable=singhand_DK_BEHAVIOR_LOG
|
||||||
|
MmAvIpLogHiveTable=singhand_MM_AV_IP_LOG
|
||||||
|
MmAvUrlLogHiveTable=singhand_MM_AV_URL_LOG
|
||||||
|
MmPicIpLogHiveTable=singhand_MM_PIC_IP_LOG
|
||||||
|
MmPicUrlLogHiveTable=singhand_MM_PIC_URL_LOG
|
||||||
|
MmVoipIpLogHiveTable=singhand_MM_VOIP_IP_LOG
|
||||||
|
MmVoipAccountLogHiveTable=singhand_MM_VOIP_ACCOUNT_LOG
|
||||||
|
MmSampleAudioLogHiveTable=singhand_MM_SAMPLE_AUDIO_LOG
|
||||||
|
MmSampleVideoLogHiveTable=singhand_MM_SAMPLE_VIDEO_LOG
|
||||||
|
MmPornAudioLevelLogHiveTable=singhand_MM_PORN_AUDIO_LEVEL_LOG
|
||||||
|
MmPornVideoLevelLogHiveTable=singhand_MM_PORN_VIDEO_LEVEL_LOG
|
||||||
|
MmSamplePicLogHiveTable=singhand_MM_SAMPLE_PIC_LOG
|
||||||
|
MmSampleVoipLogHiveTable=singhand_MM_SAMPLE_VOIP_LOG
|
||||||
|
PxyHttpLogHiveTable=singhand_PXY_HTTP_LOG
|
||||||
|
MmFaceRecognizationLogHiveTable=singhand_MM_FACE_RECOGNIZATION_LOG
|
||||||
|
MmLogoDetectionLogHiveTable=singhand_MM_LOGO_DETECTION_LOG
|
||||||
|
MmSpeakerRecognizationLogHiveTable=singhand_MM_SPEAKER_RECOGNIZATION_LOG
|
||||||
|
NtcVoipLogHiveTable=singhand_NTC_VOIP_LOG
|
||||||
|
NtcStreamingMediaLogHiveTable=singhand_NTC_STREAMING_MEDIA_LOG
|
||||||
|
MmFileDigestLogHiveTable=singhand_MM_FILE_DIGEST_LOG
|
||||||
|
NtcKeywordsUrlLogHiveTable=singhand_NTC_KEYWORDS_URL_LOG
|
||||||
|
NtcCollectVoipLogHiveTable=singhand_NTC_COLLECT_VOIP_LOG
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
@@ -40,7 +40,7 @@ bonecp.statementsCacheSize=100
|
|||||||
#数据中心hive接口配置
|
#数据中心hive接口配置
|
||||||
############################################################################################################################################
|
############################################################################################################################################
|
||||||
#数据中心hive日志库数据库名称,程序中每次查询时使用的数据库名称 use dbName
|
#数据中心hive日志库数据库名称,程序中每次查询时使用的数据库名称 use dbName
|
||||||
jdbc.hive.DBName=maat
|
#jdbc.hive.DBName=maat
|
||||||
#A版日志库
|
#A版日志库
|
||||||
jdbc.hive.driver=org.apache.hive.jdbc.HiveDriver
|
jdbc.hive.driver=org.apache.hive.jdbc.HiveDriver
|
||||||
#元辰鑫内网
|
#元辰鑫内网
|
||||||
@@ -48,16 +48,27 @@ jdbc.hive.driver=org.apache.hive.jdbc.HiveDriver
|
|||||||
#华严
|
#华严
|
||||||
#jdbc.hive.url=jdbc:hive2://192.168.11.243:2181,192.168.10.76:2181,192.168.10.77:2181/maat;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=hiveserver2
|
#jdbc.hive.url=jdbc:hive2://192.168.11.243:2181,192.168.10.76:2181,192.168.10.77:2181/maat;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=hiveserver2
|
||||||
#元辰鑫外网
|
#元辰鑫外网
|
||||||
jdbc.hive.url=jdbc:hive2://192.168.10.22:10000/maat
|
#jdbc.hive.url=jdbc:hive2://192.168.10.22:10000/maat
|
||||||
#亦庄演示环境高可用集群连接,注意使用此种方法需要配置hosts,主机名与ip的对应关系
|
#亦庄演示环境高可用集群连接,注意使用此种方法需要配置hosts,主机名与ip的对应关系
|
||||||
#jdbc.hive.url=jdbc:hive2://10.3.48.2:2181,10.3.48.3:2181,10.3.48.4:2181/maat;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=hiveserver2
|
#jdbc.hive.url=jdbc:hive2://10.3.48.2:2181,10.3.48.3:2181,10.3.48.4:2181/maat;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=hiveserver2
|
||||||
jdbc.hive.username=xa_z2_mesa
|
#jdbc.hive.username=xa_z2_mesa
|
||||||
jdbc.hive.key=aC/8fTC9vfPVhCk+CDzbAQ==
|
#jdbc.hive.key=aC/8fTC9vfPVhCk+CDzbAQ==
|
||||||
#加密后密码
|
#加密后密码
|
||||||
jdbc.hive.password=V3GyFlG8Mg01bTt8ykFVaA==
|
#jdbc.hive.password=V3GyFlG8Mg01bTt8ykFVaA==
|
||||||
#实际密码
|
#实际密码
|
||||||
#jdbc.hiveA.password=123!@#qwe
|
#jdbc.hiveA.password=123!@#qwe
|
||||||
|
|
||||||
|
|
||||||
|
#星汉hive数据源
|
||||||
|
jdbc.hive.DBName=singhand_ntcstore
|
||||||
|
jdbc.hive.url=jdbc:hive2://218.76.55.165:30000/singhand_ntcstore
|
||||||
|
jdbc.hive.username=singhand
|
||||||
|
jdbc.hive.key=uKYTzWsRN1+X5Tt3W+CA4A==
|
||||||
|
jdbc.hive.password=f+IYYX73TPEqUp3fxLzxow==
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
bonecp.hive.idleMaxAgeInMinutes=60
|
bonecp.hive.idleMaxAgeInMinutes=60
|
||||||
bonecp.hive.idleConnectionTestPeriodInMinutes=240
|
bonecp.hive.idleConnectionTestPeriodInMinutes=240
|
||||||
bonecp.hive.maxConnectionsPerPartition=20
|
bonecp.hive.maxConnectionsPerPartition=20
|
||||||
|
|||||||
Reference in New Issue
Block a user