bug修复

This commit is contained in:
yinjiangyi
2021-08-02 16:54:24 +08:00
parent 857474babc
commit d8e4775d73
11 changed files with 1709 additions and 74 deletions

File diff suppressed because one or more lines are too long

View File

@@ -48,6 +48,6 @@ public class ApplicationConfig {
public static final Double BASELINE_KALMAN_R = ConfigUtils.getDoubleProperty("baseline.kalman.r"); public static final Double BASELINE_KALMAN_R = ConfigUtils.getDoubleProperty("baseline.kalman.r");
public static final Integer LOG_WRITE_COUNT = ConfigUtils.getIntProperty("log.write.count"); public static final Integer LOG_WRITE_COUNT = ConfigUtils.getIntProperty("log.write.count");
public static final Integer GENERATE_BATCH_NUM= ConfigUtils.getIntProperty("generate.batch.number"); public static final Integer GENERATE_BATCH_SIZE = ConfigUtils.getIntProperty("generate.batch.size");
} }

View File

@@ -52,14 +52,13 @@ public class DruidData {
return druidData; return druidData;
} }
public ArrayList<String> getServerIpList(String attackType) { public ArrayList<String> getServerIpList() {
Long startQueryIPLIstTime = System.currentTimeMillis(); Long startQueryIPLIstTime = System.currentTimeMillis();
ArrayList<String> serverIPs = new ArrayList<String>(); ArrayList<String> serverIPs = new ArrayList<String>();
String sql = "SELECT distinct " + ApplicationConfig.DRUID_SERVERIP_COLUMN_NAME String sql = "SELECT distinct " + ApplicationConfig.DRUID_SERVERIP_COLUMN_NAME
+ " FROM " + ApplicationConfig.DRUID_TABLE + " FROM " + ApplicationConfig.DRUID_TABLE
+ " WHERE " + ApplicationConfig.DRUID_ATTACKTYPE_COLUMN_NAME + " = '" + attackType + "'" + " WHERE " + timeFilter
+ " AND " + timeFilter + " LIMIT 10000";// FOR TEST
+ " LIMIT 10"; // FOR TEST
try{ try{
ResultSet resultSet = DruidUtils.executeQuery(connection,sql); ResultSet resultSet = DruidUtils.executeQuery(connection,sql);
while(resultSet.next()){ while(resultSet.next()){
@@ -75,39 +74,12 @@ public class DruidData {
return serverIPs; return serverIPs;
} }
public ArrayList<String> getServerIpList(String attackType, String test) {
ArrayList<String> serverIPs = new ArrayList<String>();
serverIPs.add("153.99.250.54");
return serverIPs;
}
public List<Map<String, Object>> getTimeSeriesData(String ip, String attackType){
List<Map<String, Object>> rsList = null;
String sql = "SELECT "+ ApplicationConfig.DRUID_SERVERIP_COLUMN_NAME
+ ", "+ ApplicationConfig.BASELINE_METRIC_TYPE
+ ", " + ApplicationConfig.DRUID_RECVTIME_COLUMN_NAME
+ " FROM " + ApplicationConfig.DRUID_TABLE
+ " WHERE " + ApplicationConfig.DRUID_SERVERIP_COLUMN_NAME
+ " = '" + ip + "'"
+ " AND " + ApplicationConfig.DRUID_ATTACKTYPE_COLUMN_NAME
+ " = '" + attackType + "'"
+ " AND " + timeFilter;
try{
ResultSet resultSet = DruidUtils.executeQuery(connection,sql);
ResultSetToListService service = new ResultSetToListServiceImp();
rsList = service.selectAll(resultSet);
} catch (Exception e){
e.printStackTrace();
}
return rsList;
}
public List<Map<String, Object>> getTimeSeriesData(List<Map<String, Object>> allData, String ip, String attackType){ public List<Map<String, Object>> getTimeSeriesData(List<Map<String, Object>> allData, String ip, String attackType){
List<Map<String, Object>> rsList = new ArrayList<>(); List<Map<String, Object>> rsList = new ArrayList<>();
try{ try{
rsList = allData.stream(). rsList = allData.stream().
filter(i->(i.get(ApplicationConfig.DRUID_SERVERIP_COLUMN_NAME).equals(ip))) filter(i->((i.get(ApplicationConfig.DRUID_SERVERIP_COLUMN_NAME).equals(ip))
)&&(i.get(ApplicationConfig.DRUID_ATTACKTYPE_COLUMN_NAME).equals(attackType)))
.collect(Collectors.toList()); .collect(Collectors.toList());
} catch (NullPointerException e){ } catch (NullPointerException e){
} }
@@ -119,6 +91,7 @@ public class DruidData {
ipList = ipList.stream().map( ip -> "\'"+ip+"\'").collect(Collectors.toList()); ipList = ipList.stream().map( ip -> "\'"+ip+"\'").collect(Collectors.toList());
String ipString = "(" + StringUtils.join(ipList, ",").toString() + ")"; String ipString = "(" + StringUtils.join(ipList, ",").toString() + ")";
String sql = "SELECT "+ ApplicationConfig.DRUID_SERVERIP_COLUMN_NAME String sql = "SELECT "+ ApplicationConfig.DRUID_SERVERIP_COLUMN_NAME
+ ", "+ ApplicationConfig.DRUID_ATTACKTYPE_COLUMN_NAME
+ ", "+ ApplicationConfig.BASELINE_METRIC_TYPE + ", "+ ApplicationConfig.BASELINE_METRIC_TYPE
+ ", " + ApplicationConfig.DRUID_RECVTIME_COLUMN_NAME + ", " + ApplicationConfig.DRUID_RECVTIME_COLUMN_NAME
+ " FROM " + ApplicationConfig.DRUID_TABLE + " FROM " + ApplicationConfig.DRUID_TABLE

View File

@@ -5,6 +5,7 @@ import cn.mesalab.dao.DruidData;
import cn.mesalab.service.BaselineService.KalmanFilter; import cn.mesalab.service.BaselineService.KalmanFilter;
import cn.mesalab.utils.HbaseUtils; import cn.mesalab.utils.HbaseUtils;
import cn.mesalab.utils.SeriesUtils; import cn.mesalab.utils.SeriesUtils;
import com.google.common.collect.Lists;
import com.google.common.util.concurrent.ThreadFactoryBuilder; import com.google.common.util.concurrent.ThreadFactoryBuilder;
import org.apache.commons.math3.stat.StatUtils; import org.apache.commons.math3.stat.StatUtils;
import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Put;
@@ -30,6 +31,12 @@ public class BaselineGeneration {
private static Table hbaseTable; private static Table hbaseTable;
private static List<Map<String, Object>> batchDruidData = new ArrayList<>(); private static List<Map<String, Object>> batchDruidData = new ArrayList<>();
private static List<String> attackTypeList = Arrays.asList(
ApplicationConfig.DRUID_ATTACKTYPE_TCP_SYN_FLOOD,
ApplicationConfig.DRUID_ATTACKTYPE_ICMP_FLOOD,
ApplicationConfig.DRUID_ATTACKTYPE_UDP_FLOOD,
ApplicationConfig.DRUID_ATTACKTYPE_DNS_AMPL
);
private static final Integer BASELINE_POINT_NUM = ApplicationConfig.BASELINE_RANGE_DAYS * 24 * (60/ApplicationConfig.HISTORICAL_GRAD); private static final Integer BASELINE_POINT_NUM = ApplicationConfig.BASELINE_RANGE_DAYS * 24 * (60/ApplicationConfig.HISTORICAL_GRAD);
public static void perform() { public static void perform() {
@@ -42,10 +49,7 @@ public class BaselineGeneration {
LOG.info("Druid 成功建立连接"); LOG.info("Druid 成功建立连接");
try{ try{
generateBaselinesThread(ApplicationConfig.DRUID_ATTACKTYPE_TCP_SYN_FLOOD); generateBaselinesThread();
generateBaselines(ApplicationConfig.DRUID_ATTACKTYPE_UDP_FLOOD);
// generateBaselines(ApplicationConfig.DRUID_ATTACKTYPE_ICMP_FLOOD);
// generateBaselines(ApplicationConfig.DRUID_ATTACKTYPE_DNS_AMPL);
long last = System.currentTimeMillis(); long last = System.currentTimeMillis();
LOG.warn("运行时间:" + (last - start)); LOG.warn("运行时间:" + (last - start));
@@ -60,11 +64,11 @@ public class BaselineGeneration {
System.exit(0); System.exit(0);
} }
private static void generateBaselinesThread(String attackType) throws InterruptedException { private static void generateBaselinesThread() throws InterruptedException {
int threadNum = Runtime.getRuntime().availableProcessors(); int threadNum = Runtime.getRuntime().availableProcessors();
ThreadFactory namedThreadFactory = new ThreadFactoryBuilder() ThreadFactory namedThreadFactory = new ThreadFactoryBuilder()
.setNameFormat(attackType+"-baseline-demo-%d").build(); .setNameFormat("baseline-demo-%d").build();
// 创建线程池 // 创建线程池
ThreadPoolExecutor executor = new ThreadPoolExecutor( ThreadPoolExecutor executor = new ThreadPoolExecutor(
@@ -77,57 +81,49 @@ public class BaselineGeneration {
new ThreadPoolExecutor.AbortPolicy()); new ThreadPoolExecutor.AbortPolicy());
// baseline 生成及写入 // baseline 生成及写入
ArrayList<String> destinationIps = druidData.getServerIpList(attackType); ArrayList<String> destinationIps = druidData.getServerIpList();
LOG.info("查询到服务端ip " +destinationIps.size() + ""); LOG.info("查询到服务端ip " +destinationIps.size() + "");
LOG.info("Baseline batch 大小: " + ApplicationConfig.GENERATE_BATCH_SIZE);
int batchCount = destinationIps.size() / ApplicationConfig.GENERATE_BATCH_NUM; List<List<String>> batchIpLists = Lists.partition(destinationIps, ApplicationConfig.GENERATE_BATCH_SIZE);
for (List<String> batchIps: batchIpLists){
for (int batchCurrent = 0; batchCurrent <batchCount; batchCurrent++){
List<String> batchIps = destinationIps.subList(batchCurrent*ApplicationConfig.GENERATE_BATCH_NUM,
(batchCurrent+1)*ApplicationConfig.GENERATE_BATCH_NUM);
if(batchIps.size()>0){ if(batchIps.size()>0){
executor.execute(() -> generateBaselines(batchIps, attackType)); executor.execute(() -> generateBaselines(batchIps));
} }
} }
executor.shutdown(); executor.shutdown();
executor.awaitTermination(10L, TimeUnit.SECONDS); executor.awaitTermination(10L, TimeUnit.HOURS);
}
LOG.info("BaselineGeneration 完成:" + attackType); static void generateBaselines(){
ArrayList<String> destinationIps = druidData.getServerIpList();
generateBaselines(destinationIps);
LOG.info("BaselineGeneration 共写入数据条数:" + destinationIps.size()); LOG.info("BaselineGeneration 共写入数据条数:" + destinationIps.size());
} }
static void generateBaselines(String attackType){ public static void generateBaselines(List<String> ipList){
ArrayList<String> destinationIps = druidData.getServerIpList(attackType);
generateBaselines(destinationIps, attackType);
LOG.info("BaselineGeneration 完成:" + attackType);
LOG.info("BaselineGeneration 共写入数据条数:" + destinationIps.size());
}
public static void generateBaselines(List<String> ipList, String attackType){
Long startGenerateTime = System.currentTimeMillis(); Long startGenerateTime = System.currentTimeMillis();
druidData = DruidData.getInstance(); druidData = DruidData.getInstance();
batchDruidData = druidData.readFromDruid(ipList); batchDruidData = druidData.readFromDruid(ipList);
List<Put> putList = new ArrayList<>(); List<Put> putList = new ArrayList<>();
for(String ip: ipList){ for(String attackType: attackTypeList){
int[] ipBaseline = generateSingleIpBaseline(ip, attackType); for(String ip: ipList){
if (ipBaseline==null){ int[] ipBaseline = generateSingleIpBaseline(ip, attackType);
break; if (!(ipBaseline ==null)){
putList = hbaseUtils.cachedInPut(putList, ip, ipBaseline, attackType, ApplicationConfig.BASELINE_METRIC_TYPE);
}
} }
putList = hbaseUtils.cachedInPut(putList, ip, ipBaseline, attackType, ApplicationConfig.BASELINE_METRIC_TYPE);
} }
Long endGenerateTime = System.currentTimeMillis(); Long endGenerateTime = System.currentTimeMillis();
// LOG.info("性能测试Baseline生成耗时——"+(endGenerateTime-startGenerateTime));
try { try {
hbaseTable.put(putList); hbaseTable.put(putList);
LOG.info("HBase 处理数据条数 " + ApplicationConfig.GENERATE_BATCH_NUM); LOG.info("Baseline 线程 " + Thread.currentThread().getId() + " 成功写入Baseline条数共计 " + putList.size());
} catch (IOException e) { } catch (IOException e) {
e.printStackTrace(); e.printStackTrace();
} }
@@ -178,6 +174,8 @@ public class BaselineGeneration {
Long endGenerateSingleIPTime = System.currentTimeMillis(); Long endGenerateSingleIPTime = System.currentTimeMillis();
//LOG.info("性能测试单个baseline生成耗时——"+(endGenerateSingleIPTime-endQuerySingleIPTime)); //LOG.info("性能测试单个baseline生成耗时——"+(endGenerateSingleIPTime-endQuerySingleIPTime));
//System.out.println(ip);
//System.out.println(attackType + Arrays.toString(baselineArr));
return baselineArr; return baselineArr;
} }

View File

@@ -1 +1 @@
package cn.mesalab.utils; package cn.mesalab.utils;

View File

@@ -7,8 +7,8 @@ druid.table=top_server_ip_test_log
#字段映射 #字段映射
druid.attacktype.tcpsynflood=sessions druid.attacktype.tcpsynflood=sessions
druid.attacktype.udpflood=bytes druid.attacktype.udpflood=bytes
#druid.attacktype.icmpflood=IPv6_UDP druid.attacktype.icmpflood=packets
#druid.attacktype.dnsamplification=IPv4_UDP druid.attacktype.dnsamplification=packets
druid.serverip.columnname=destination druid.serverip.columnname=destination
druid.attacktype.columnname=order_by druid.attacktype.columnname=order_by
druid.recvtime.columnname=__time druid.recvtime.columnname=__time
@@ -54,5 +54,5 @@ baseline.kalman.r=0.002
# 每更新1000个记录打印log # 每更新1000个记录打印log
log.write.count=10000 log.write.count=10000
# FOR TEST # FOR TEST
generate.batch.number=5 generate.batch.size=1000

View File

@@ -7,8 +7,8 @@ druid.table=top_server_ip_test_log
#字段映射 #字段映射
druid.attacktype.tcpsynflood=sessions druid.attacktype.tcpsynflood=sessions
druid.attacktype.udpflood=bytes druid.attacktype.udpflood=bytes
#druid.attacktype.icmpflood=IPv6_UDP druid.attacktype.icmpflood=packets
#druid.attacktype.dnsamplification=IPv4_UDP druid.attacktype.dnsamplification=packets
druid.serverip.columnname=destination druid.serverip.columnname=destination
druid.attacktype.columnname=order_by druid.attacktype.columnname=order_by
druid.recvtime.columnname=__time druid.recvtime.columnname=__time
@@ -54,5 +54,5 @@ baseline.kalman.r=0.002
# 每更新1000个记录打印log # 每更新1000个记录打印log
log.write.count=10000 log.write.count=10000
# FOR TEST # FOR TEST
generate.batch.number=5 generate.batch.size=1000