为日志查询添加默认查询时间,防止调用接口时不传时间条件导致全表扫描
This commit is contained in:
@@ -181,21 +181,23 @@ public class HiveJDBC {
|
||||
Integer startNum = (page.getPageNo() - 1) * page.getPageSize();
|
||||
Integer endNum = startNum - 1 + page.getPageSize() + 1;
|
||||
if (listString.size() >= startNum) {
|
||||
if (listString.size() >= endNum) {
|
||||
mapList.put("str", listString.subList(startNum, endNum));
|
||||
} else {
|
||||
mapList.put("str", listString.subList(startNum, listString.size()));
|
||||
}
|
||||
// if (listString.size() >= endNum) {
|
||||
// mapList.put("str", listString.subList(startNum, endNum));
|
||||
// } else {
|
||||
// mapList.put("str", listString.subList(startNum, listString.size()));
|
||||
// }
|
||||
|
||||
mapList.put("str", listString);
|
||||
} else {
|
||||
mapList.put("str", new ArrayList());
|
||||
}
|
||||
if (listObject.size() >= startNum) {
|
||||
if (listObject.size() >= endNum) {
|
||||
mapList.put("obj", listObject.subList(startNum, endNum));
|
||||
} else {
|
||||
mapList.put("obj", listObject.subList(startNum, listObject.size()));
|
||||
}
|
||||
|
||||
// if (listObject.size() >= endNum) {
|
||||
// mapList.put("obj", listObject.subList(startNum, endNum));
|
||||
// } else {
|
||||
// mapList.put("obj", listObject.subList(startNum, listObject.size()));
|
||||
// }
|
||||
mapList.put("str", listObject);
|
||||
} else {
|
||||
mapList.put("obj", new ArrayList());
|
||||
}
|
||||
|
||||
@@ -71,8 +71,11 @@ public class LogController extends BaseRestController {
|
||||
request, null);
|
||||
Page<NtcIpLog> ntcIpLogPage = null;
|
||||
try {
|
||||
resetTime(ntcIpLog);
|
||||
testService.queryConditionCheck(auditLogThread, start, ntcIpLog, NtcIpLog.class, page);
|
||||
ntcIpLogPage = new Page<NtcIpLog>();
|
||||
ntcIpLogPage.setPageNo(page.getPageNo());
|
||||
ntcIpLogPage.setPageSize(page.getPageSize());
|
||||
String orderBy = "";
|
||||
if (null != page.getOrderBy() && !page.getOrderBy().equals("")) {
|
||||
orderBy = Page.getOrderBySql(NtcIpLog.class.getSimpleName(), page.getOrderBy());
|
||||
@@ -99,9 +102,15 @@ public class LogController extends BaseRestController {
|
||||
NtcIpLog.class);
|
||||
ntcIpLogPage.setList(List);
|
||||
ntcIpLogPage.setCount(List.size());
|
||||
|
||||
// ntcIpLogPage
|
||||
// .setCount(
|
||||
// HiveSqlService.getHivePageCount(ntcIpLog, null,
|
||||
// Configurations.getStringProperty(
|
||||
// NtcIpLog.class.getSimpleName() + "HiveTable", "ntc_ip_log"),
|
||||
// getCol2Col(), null));
|
||||
} else {
|
||||
ntcIpLogPage.setList(new ArrayList());
|
||||
ntcIpLogPage.setCount(0l);
|
||||
}
|
||||
}
|
||||
} catch (Exception e) {
|
||||
@@ -127,8 +136,11 @@ public class LogController extends BaseRestController {
|
||||
|
||||
Page<NtcHttpLog> ntcHttpLogPage = null;
|
||||
try {
|
||||
resetTime(ntcHttpLog);
|
||||
testService.queryConditionCheck(auditLogThread, start, ntcHttpLog, NtcHttpLog.class, page);
|
||||
ntcHttpLogPage = new Page<NtcHttpLog>();
|
||||
ntcHttpLogPage.setPageNo(page.getPageNo());
|
||||
ntcHttpLogPage.setPageSize(page.getPageSize());
|
||||
String orderBy = "";
|
||||
if (null != page.getOrderBy() && !page.getOrderBy().equals("")) {
|
||||
orderBy = Page.getOrderBySql(NtcHttpLog.class.getSimpleName(), page.getOrderBy());
|
||||
@@ -184,8 +196,11 @@ public class LogController extends BaseRestController {
|
||||
|
||||
Page<NtcDnsLog> ntcDnsLogPage = null;
|
||||
try {
|
||||
resetTime(ntcDnsLog);
|
||||
testService.queryConditionCheck(auditLogThread, start, ntcDnsLog, NtcDnsLog.class, page);
|
||||
ntcDnsLogPage = new Page<NtcDnsLog>();
|
||||
ntcDnsLogPage.setPageNo(page.getPageNo());
|
||||
ntcDnsLogPage.setPageSize(page.getPageSize());
|
||||
String orderBy = "";
|
||||
if (null != page.getOrderBy() && !page.getOrderBy().equals("")) {
|
||||
orderBy = Page.getOrderBySql(NtcDnsLog.class.getSimpleName(), page.getOrderBy());
|
||||
@@ -241,8 +256,11 @@ public class LogController extends BaseRestController {
|
||||
|
||||
Page<NtcMailLog> ntcMailLogPage = null;
|
||||
try {
|
||||
resetTime(ntcMailLog);
|
||||
testService.queryConditionCheck(auditLogThread, start, ntcMailLog, NtcMailLog.class, page);
|
||||
ntcMailLogPage = new Page<NtcMailLog>();
|
||||
ntcMailLogPage.setPageNo(page.getPageNo());
|
||||
ntcMailLogPage.setPageSize(page.getPageSize());
|
||||
String orderBy = "";
|
||||
if (null != page.getOrderBy() && !page.getOrderBy().equals("")) {
|
||||
orderBy = Page.getOrderBySql(NtcMailLog.class.getSimpleName(), page.getOrderBy());
|
||||
@@ -298,8 +316,11 @@ public class LogController extends BaseRestController {
|
||||
|
||||
Page<NtcSslLog> ntcSslLogPage = null;
|
||||
try {
|
||||
resetTime(ntcSslLog);
|
||||
testService.queryConditionCheck(auditLogThread, start, ntcSslLog, NtcSslLog.class, page);
|
||||
ntcSslLogPage = new Page<NtcSslLog>();
|
||||
ntcSslLogPage.setPageNo(page.getPageNo());
|
||||
ntcSslLogPage.setPageSize(page.getPageSize());
|
||||
String orderBy = "";
|
||||
if (null != page.getOrderBy() && !page.getOrderBy().equals("")) {
|
||||
orderBy = Page.getOrderBySql(NtcSslLog.class.getSimpleName(), page.getOrderBy());
|
||||
@@ -355,8 +376,11 @@ public class LogController extends BaseRestController {
|
||||
|
||||
Page<NtcPptpLog> ntcPptpLogPage = null;
|
||||
try {
|
||||
resetTime(ntcPptpLog);
|
||||
testService.queryConditionCheck(auditLogThread, start, ntcPptpLog, NtcPptpLog.class, page);
|
||||
ntcPptpLogPage = new Page<NtcPptpLog>();
|
||||
ntcPptpLogPage.setPageNo(page.getPageNo());
|
||||
ntcPptpLogPage.setPageSize(page.getPageSize());
|
||||
String orderBy = "";
|
||||
if (null != page.getOrderBy() && !page.getOrderBy().equals("")) {
|
||||
orderBy = Page.getOrderBySql(NtcPptpLog.class.getSimpleName(), page.getOrderBy());
|
||||
@@ -411,8 +435,11 @@ public class LogController extends BaseRestController {
|
||||
|
||||
Page<NtcL2tpLog> ntcL2tpLogPage = null;
|
||||
try {
|
||||
resetTime(ntcL2tpLog);
|
||||
testService.queryConditionCheck(auditLogThread, start, ntcL2tpLog, NtcL2tpLog.class, page);
|
||||
ntcL2tpLogPage = new Page<NtcL2tpLog>();
|
||||
ntcL2tpLogPage.setPageNo(page.getPageNo());
|
||||
ntcL2tpLogPage.setPageSize(page.getPageSize());
|
||||
String orderBy = "";
|
||||
if (null != page.getOrderBy() && !page.getOrderBy().equals("")) {
|
||||
orderBy = Page.getOrderBySql(NtcL2tpLog.class.getSimpleName(), page.getOrderBy());
|
||||
@@ -468,8 +495,11 @@ public class LogController extends BaseRestController {
|
||||
|
||||
Page<NtcOpenvpnLog> ntcOpenvpnLogPage = null;
|
||||
try {
|
||||
resetTime(ntcOpenvpnLog);
|
||||
testService.queryConditionCheck(auditLogThread, start, ntcOpenvpnLog, NtcOpenvpnLog.class, page);
|
||||
ntcOpenvpnLogPage = new Page<NtcOpenvpnLog>();
|
||||
ntcOpenvpnLogPage.setPageNo(page.getPageNo());
|
||||
ntcOpenvpnLogPage.setPageSize(page.getPageSize());
|
||||
String orderBy = "";
|
||||
if (null != page.getOrderBy() && !page.getOrderBy().equals("")) {
|
||||
orderBy = Page.getOrderBySql(NtcOpenvpnLog.class.getSimpleName(), page.getOrderBy());
|
||||
@@ -525,8 +555,11 @@ public class LogController extends BaseRestController {
|
||||
|
||||
Page<NtcIpsecLog> ntcIpsecLogPage = null;
|
||||
try {
|
||||
resetTime(ntcIpsecLog);
|
||||
testService.queryConditionCheck(auditLogThread, start, ntcIpsecLog, NtcIpsecLog.class, page);
|
||||
ntcIpsecLogPage = new Page<NtcIpsecLog>();
|
||||
ntcIpsecLogPage.setPageNo(page.getPageNo());
|
||||
ntcIpsecLogPage.setPageSize(page.getPageSize());
|
||||
String orderBy = "";
|
||||
if (null != page.getOrderBy() && !page.getOrderBy().equals("")) {
|
||||
orderBy = Page.getOrderBySql(NtcIpsecLog.class.getSimpleName(), page.getOrderBy());
|
||||
@@ -582,8 +615,11 @@ public class LogController extends BaseRestController {
|
||||
|
||||
Page<NtcSshLog> ntcSshLogPage = null;
|
||||
try {
|
||||
resetTime(ntcSshLog);
|
||||
testService.queryConditionCheck(auditLogThread, start, ntcSshLog, NtcSshLog.class, page);
|
||||
ntcSshLogPage = new Page<NtcSshLog>();
|
||||
ntcSshLogPage.setPageNo(page.getPageNo());
|
||||
ntcSshLogPage.setPageSize(page.getPageSize());
|
||||
String orderBy = "";
|
||||
if (null != page.getOrderBy() && !page.getOrderBy().equals("")) {
|
||||
orderBy = Page.getOrderBySql(NtcSshLog.class.getSimpleName(), page.getOrderBy());
|
||||
@@ -639,8 +675,11 @@ public class LogController extends BaseRestController {
|
||||
|
||||
Page<NtcFtpLog> ntcFtpLogPage = null;
|
||||
try {
|
||||
resetTime(ntcFtpLog);
|
||||
testService.queryConditionCheck(auditLogThread, start, ntcFtpLog, NtcFtpLog.class, page);
|
||||
ntcFtpLogPage = new Page<NtcFtpLog>();
|
||||
ntcFtpLogPage.setPageNo(page.getPageNo());
|
||||
ntcFtpLogPage.setPageSize(page.getPageSize());
|
||||
String orderBy = "";
|
||||
if (null != page.getOrderBy() && !page.getOrderBy().equals("")) {
|
||||
orderBy = Page.getOrderBySql(NtcFtpLog.class.getSimpleName(), page.getOrderBy());
|
||||
|
||||
@@ -21,7 +21,6 @@ import com.nis.domain.Page;
|
||||
import com.nis.util.Configurations;
|
||||
import com.nis.util.Constants;
|
||||
import com.nis.util.HiveDataSource;
|
||||
import com.nis.util.HiveJDBC;
|
||||
import com.nis.util.StringUtil;
|
||||
import com.nis.util.redis.SaveRedisThread;
|
||||
|
||||
@@ -71,10 +70,10 @@ public class HiveSqlService {
|
||||
if (col2col.containsKey(key)) {
|
||||
value = sdf.parse(value.toString().trim()).getTime() / 1000;
|
||||
if (key.toLowerCase().equals("searchfoundstarttime")) {
|
||||
foundTimePartStart = Long.parseLong(value.toString()) / 3600;
|
||||
foundTimePartStart = Long.parseLong(value.toString()) / 3600L/ 24L;
|
||||
}
|
||||
if (key.toLowerCase().equals("searchfoundendtime")) {
|
||||
foundTimePartEnd = Long.parseLong(value.toString()) / 3600;
|
||||
foundTimePartEnd = Long.parseLong(value.toString()) / 3600L/ 24L;
|
||||
}
|
||||
if (col2col.get(key).get("start") != null) {
|
||||
// sql.append(" and " +
|
||||
@@ -115,7 +114,6 @@ public class HiveSqlService {
|
||||
|
||||
}
|
||||
}
|
||||
if (null != searchActiveSys) {// 添加分区字段
|
||||
if (null != foundTimePartStart) {
|
||||
// sql.append(" and found_time_partition>=" + foundTimePartStart + "L");
|
||||
sql.append(" and found_time_partition>=" + foundTimePartStart);
|
||||
@@ -124,7 +122,6 @@ public class HiveSqlService {
|
||||
// sql.append(" and found_time_partition<" + foundTimePartEnd + "L");
|
||||
sql.append(" and found_time_partition<" + foundTimePartEnd);
|
||||
}
|
||||
}
|
||||
|
||||
}
|
||||
}
|
||||
@@ -143,10 +140,9 @@ public class HiveSqlService {
|
||||
public static Long getHivePageCount(Object bean, String countKey, String tableName,
|
||||
Map<String, Map<String, String>> col2col, String searchActiveSys) throws Exception {
|
||||
tableName = tableName.toLowerCase();
|
||||
tableName = Configurations.getStringProperty(tableName, "t_" + tableName).trim();
|
||||
StringBuffer countSql = new StringBuffer();
|
||||
StringBuffer sql = new StringBuffer();
|
||||
Map<String, String> filedAndColumnMap = getFiledAndColumnMap(bean.getClass());
|
||||
countSql.append("select count(1) from " + tableName + " where 1=1 ");
|
||||
sql.append("select count(1) from " + tableName + " where 1=1 ");
|
||||
if (bean != null) {
|
||||
Class<?> clazz = bean.getClass();
|
||||
for (; clazz != Object.class; clazz = clazz.getSuperclass()) {
|
||||
@@ -156,8 +152,10 @@ public class HiveSqlService {
|
||||
Long foundTimePartStart = null;
|
||||
Long foundTimePartEnd = null;
|
||||
for (int i = 0; i < fields.length; i++) {
|
||||
// 现在gwall日志表结构中只有数值和字符串两种类型,数值都是int类型没有bigint所以不需要加L,
|
||||
Field f = fields[i];
|
||||
String key = f.getName();// 获取字段名
|
||||
String typeName = f.getType().getName();
|
||||
if (f.getType().getName().equals("java.lang.String") && key.startsWith("search")) {
|
||||
Object value = getFieldValue(bean, key);
|
||||
if (value != null) {
|
||||
@@ -166,17 +164,27 @@ public class HiveSqlService {
|
||||
if (col2col.containsKey(key)) {
|
||||
value = sdf.parse(value.toString().trim()).getTime() / 1000;
|
||||
if (key.toLowerCase().equals("searchfoundstarttime")) {
|
||||
foundTimePartStart = Long.parseLong(value.toString()) / 3600;
|
||||
foundTimePartStart = Long.parseLong(value.toString()) / 3600L/ 24L;
|
||||
}
|
||||
if (key.toLowerCase().equals("searchfoundendtime")) {
|
||||
foundTimePartEnd = Long.parseLong(value.toString()) / 3600;
|
||||
foundTimePartEnd = Long.parseLong(value.toString()) / 3600L/ 24L;
|
||||
}
|
||||
if (col2col.get(key).get("start") != null) {
|
||||
countSql.append(" and " + filedAndColumnMap.get(col2col.get(key).get("start"))
|
||||
+ ">=" + value + "L");
|
||||
// sql.append(" and " +
|
||||
// filedAndColumnMap.get(col2col.get(key).get("start"))
|
||||
// + ">=to_date('" +
|
||||
// value.toString().trim()
|
||||
// + "','yyyy-mm-dd HH24:mi:ss')");
|
||||
sql.append(" and " + filedAndColumnMap.get(col2col.get(key).get("start")) + ">="
|
||||
+ value);
|
||||
} else {
|
||||
countSql.append(" and " + filedAndColumnMap.get(col2col.get(key).get("end"))
|
||||
+ "<" + value + "L");
|
||||
// sql.append(" and " +
|
||||
// filedAndColumnMap.get(col2col.get(key).get("end"))
|
||||
// + "<=to_date('" +
|
||||
// value.toString().trim()
|
||||
// + "','yyyy-mm-dd HH24:mi:ss')");
|
||||
sql.append(" and " + filedAndColumnMap.get(col2col.get(key).get("end")) + "<"
|
||||
+ value);
|
||||
}
|
||||
}
|
||||
} else {
|
||||
@@ -185,48 +193,35 @@ public class HiveSqlService {
|
||||
key = key.substring(0, 1).toLowerCase() + key.substring(1);
|
||||
}
|
||||
|
||||
if (!value.toString().trim().equals("") && filedAndColumnMap.containsKey(key)
|
||||
&& (key.toLowerCase().equals("cfgid")
|
||||
|| key.toLowerCase().equals("entranceid"))) {
|
||||
countSql.append(
|
||||
if (typeName.equals("java.lang.String")) {
|
||||
sql.append(" and " + filedAndColumnMap.get(key) + "='" + value.toString().trim()
|
||||
+ "'");
|
||||
} else if (typeName.equals("java.lang.Integer") || typeName.equals("int")) {
|
||||
sql.append(" and " + filedAndColumnMap.get(key) + "=" + value.toString().trim());
|
||||
|
||||
} else if (typeName.equals("java.lang.Long") || typeName.equals("long")) {
|
||||
sql.append(
|
||||
" and " + filedAndColumnMap.get(key) + "=" + value.toString().trim() + "L");
|
||||
} else if (!value.toString().trim().equals("") && filedAndColumnMap.containsKey(key)
|
||||
&& (key.toLowerCase().equals("protocol") || key.toLowerCase().equals("serverip")
|
||||
|| key.toLowerCase().equals("clientip")
|
||||
|| key.toLowerCase().equals("url")
|
||||
|| key.toLowerCase().equals("mailfrom")
|
||||
|| key.toLowerCase().equals("mailto")
|
||||
|| key.toLowerCase().equals("encryptmode")
|
||||
|| key.toLowerCase().equals("exprotocol")
|
||||
|| key.toLowerCase().equals("cljip"))) {
|
||||
countSql.append(" and " + filedAndColumnMap.get(key) + "='"
|
||||
+ value.toString().trim() + "'");
|
||||
} else if (!value.toString().trim().equals("") && filedAndColumnMap.containsKey(key)
|
||||
&& key.toLowerCase().equals("servicetype")) {
|
||||
countSql.append(
|
||||
" and " + filedAndColumnMap.get(key) + "=" + value.toString().trim());
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
}
|
||||
}
|
||||
// if (null != searchActiveSys &&
|
||||
// !searchActiveSys.equals(Constants.ACTIVESYS_A)) {//
|
||||
// B版数据库才有found_time_partition字段,A版毛衣found_time_partition分区字段
|
||||
if (null != searchActiveSys) {
|
||||
if (null != foundTimePartStart) {
|
||||
countSql.append(" and found_time_partition>=" + foundTimePartStart + "L");
|
||||
// sql.append(" and found_time_partition>=" + foundTimePartStart + "L");
|
||||
sql.append(" and found_time_partition>=" + foundTimePartStart);
|
||||
}
|
||||
if (null != foundTimePartEnd) {
|
||||
countSql.append(" and found_time_partition<" + foundTimePartEnd + "L");
|
||||
// sql.append(" and found_time_partition<" + foundTimePartEnd + "L");
|
||||
sql.append(" and found_time_partition<" + foundTimePartEnd);
|
||||
}
|
||||
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
logger.info("获取数据中心日志总条数sql==================" + countSql.toString());
|
||||
logger.info("获取数据中心日志总条数sql==================" + sql.toString());
|
||||
// ResultSet countRs = HiveJDBC.query(countSql.toString());
|
||||
ResultSet countRs = HiveDataSource.query(countSql.toString());
|
||||
ResultSet countRs = HiveDataSource.query(sql.toString());
|
||||
String countStr = null;
|
||||
while (countRs.next()) {
|
||||
countStr = countRs.getObject(1).toString();
|
||||
@@ -248,7 +243,6 @@ public class HiveSqlService {
|
||||
public static ResultSet getResultSet2(Page page, Object bean, String tableName,
|
||||
Map<String, Map<String, String>> col2col, String orderBy, String searchActiveSys) throws Exception {
|
||||
tableName = tableName.toLowerCase();
|
||||
tableName = Configurations.getStringProperty(tableName, "t_" + tableName).trim();
|
||||
String showColmun = getFiledsSql(bean.getClass().getSimpleName(), page.getFields());
|
||||
StringBuffer sql = new StringBuffer();
|
||||
Map<String, String> filedAndColumnMap = getFiledAndColumnMap(bean.getClass());
|
||||
@@ -277,8 +271,10 @@ public class HiveSqlService {
|
||||
Long foundTimePartStart = null;
|
||||
Long foundTimePartEnd = null;
|
||||
for (int i = 0; i < fields.length; i++) {
|
||||
// 现在gwall日志表结构中只有数值和字符串两种类型,数值都是int类型没有bigint所以不需要加L,
|
||||
Field f = fields[i];
|
||||
String key = f.getName();// 获取字段名
|
||||
String typeName = f.getType().getName();
|
||||
if (f.getType().getName().equals("java.lang.String") && key.startsWith("search")) {
|
||||
Object value = getFieldValue(bean, key);
|
||||
if (value != null) {
|
||||
@@ -287,10 +283,10 @@ public class HiveSqlService {
|
||||
if (col2col.containsKey(key)) {
|
||||
value = sdf.parse(value.toString().trim()).getTime() / 1000;
|
||||
if (key.toLowerCase().equals("searchfoundstarttime")) {
|
||||
foundTimePartStart = Long.parseLong(value.toString()) / 3600;
|
||||
foundTimePartStart = Long.parseLong(value.toString()) / 3600L/ 24L;
|
||||
}
|
||||
if (key.toLowerCase().equals("searchfoundendtime")) {
|
||||
foundTimePartEnd = Long.parseLong(value.toString()) / 3600;
|
||||
foundTimePartEnd = Long.parseLong(value.toString()) / 3600L/ 24L;
|
||||
}
|
||||
if (col2col.get(key).get("start") != null) {
|
||||
// sql.append(" and " +
|
||||
@@ -299,7 +295,7 @@ public class HiveSqlService {
|
||||
// value.toString().trim()
|
||||
// + "','yyyy-mm-dd HH24:mi:ss')");
|
||||
sql.append(" and " + filedAndColumnMap.get(col2col.get(key).get("start")) + ">="
|
||||
+ value + "L");
|
||||
+ value);
|
||||
} else {
|
||||
// sql.append(" and " +
|
||||
// filedAndColumnMap.get(col2col.get(key).get("end"))
|
||||
@@ -307,7 +303,7 @@ public class HiveSqlService {
|
||||
// value.toString().trim()
|
||||
// + "','yyyy-mm-dd HH24:mi:ss')");
|
||||
sql.append(" and " + filedAndColumnMap.get(col2col.get(key).get("end")) + "<"
|
||||
+ value + "L");
|
||||
+ value);
|
||||
}
|
||||
}
|
||||
} else {
|
||||
@@ -316,43 +312,38 @@ public class HiveSqlService {
|
||||
key = key.substring(0, 1).toLowerCase() + key.substring(1);
|
||||
}
|
||||
|
||||
if (!value.toString().trim().equals("") && filedAndColumnMap.containsKey(key)
|
||||
&& (key.toLowerCase().equals("cfgid")
|
||||
|| key.toLowerCase().equals("entranceid"))) {
|
||||
sql.append(
|
||||
" and " + filedAndColumnMap.get(key) + "=" + value.toString().trim() + "L");
|
||||
} else if (!value.toString().trim().equals("") && filedAndColumnMap.containsKey(key)
|
||||
&& (key.toLowerCase().equals("protocol") || key.toLowerCase().equals("serverip")
|
||||
|| key.toLowerCase().equals("clientip")
|
||||
|| key.toLowerCase().equals("cljip"))) {
|
||||
if (typeName.equals("java.lang.String")) {
|
||||
sql.append(" and " + filedAndColumnMap.get(key) + "='" + value.toString().trim()
|
||||
+ "'");
|
||||
} else if (!value.toString().trim().equals("") && filedAndColumnMap.containsKey(key)
|
||||
&& key.toLowerCase().equals("servicetype")) {
|
||||
} else if (typeName.equals("java.lang.Integer") || typeName.equals("int")) {
|
||||
sql.append(" and " + filedAndColumnMap.get(key) + "=" + value.toString().trim());
|
||||
|
||||
} else if (typeName.equals("java.lang.Long") || typeName.equals("long")) {
|
||||
sql.append(
|
||||
" and " + filedAndColumnMap.get(key) + "=" + value.toString().trim() + "L");
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
}
|
||||
}
|
||||
if (null != searchActiveSys && !searchActiveSys.equals(Constants.ACTIVESYS_A)) {// B版数据库才有found_time_partition字段,A版毛衣found_time_partition分区字段
|
||||
if (null != foundTimePartStart) {
|
||||
sql.append(" and found_time_partition>=" + foundTimePartStart + "L");
|
||||
// sql.append(" and found_time_partition>=" + foundTimePartStart + "L");
|
||||
sql.append(" and found_time_partition>=" + foundTimePartStart);
|
||||
}
|
||||
if (null != foundTimePartEnd) {
|
||||
sql.append(" and found_time_partition<" + foundTimePartEnd + "L");
|
||||
}
|
||||
// sql.append(" and found_time_partition<" + foundTimePartEnd + "L");
|
||||
sql.append(" and found_time_partition<" + foundTimePartEnd);
|
||||
}
|
||||
|
||||
}
|
||||
}
|
||||
Integer startNum = (page.getPageNo() - 1) * page.getPageSize() + 1;
|
||||
Integer endNum = startNum - 1 + page.getPageSize();
|
||||
sql.append(
|
||||
" order by " + orderBy + " limit 10000) t1) t2 where row_Num between " + startNum + " and " + endNum);
|
||||
//sql.append(" order by " + orderBy + " limit 10000) t1) t2 where row_Num between " + startNum + " and " + endNum);
|
||||
sql.append(" order by " + orderBy + " ) t1) t2 where row_Num between " + startNum + " and " + endNum);
|
||||
logger.info("获取数据中心日志sql===================" + sql);
|
||||
ResultSet query = HiveJDBC.query(sql.toString(), searchActiveSys);
|
||||
ResultSet query = HiveDataSource.query(sql.toString());
|
||||
logger.info("获取数据中心日志成功");
|
||||
return query;
|
||||
}
|
||||
@@ -434,7 +425,7 @@ public class HiveSqlService {
|
||||
}
|
||||
}
|
||||
logger.info("获取数据中心日志总条数sql==================" + countSql.toString());
|
||||
ResultSet countRs = HiveJDBC.query(countSql.toString(), searchActiveSys);
|
||||
ResultSet countRs = HiveDataSource.query(countSql.toString());
|
||||
String countStr = null;
|
||||
while (countRs.next()) {
|
||||
countStr = countRs.getObject(1).toString();
|
||||
@@ -446,7 +437,7 @@ public class HiveSqlService {
|
||||
}
|
||||
Long count = Long.valueOf(countStr);
|
||||
logger.info("获取数据中心日志总条数成功总共===================" + count + "条配置");
|
||||
HiveJDBC.closeConn();
|
||||
HiveDataSource.closeConn();
|
||||
return count;
|
||||
}
|
||||
|
||||
|
||||
@@ -183,7 +183,7 @@ selFromHive=false
|
||||
isGetHiveCount=false
|
||||
|
||||
#每次获取数据中心多少条数据,咱们在对获取的数据进行分页处理
|
||||
everyGetHiveDataNum=100
|
||||
everyGetHiveDataNum=100000
|
||||
|
||||
#oracle数据库有问题不从oracle查询数据,所有日志数据均从数据中心查询
|
||||
onlySelFromHive=false
|
||||
|
||||
Reference in New Issue
Block a user