修改
This commit is contained in:
@@ -49,7 +49,7 @@ public class ApplicationConfig {
|
|||||||
|
|
||||||
public static final Integer BASELINE_GENERATE_BATCH_SIZE = ConfigUtils.getIntProperty("baseline.generate.batch.size");
|
public static final Integer BASELINE_GENERATE_BATCH_SIZE = ConfigUtils.getIntProperty("baseline.generate.batch.size");
|
||||||
public static final Long DRUID_READ_BATCH_TIME_GRAD_HOUR = ConfigUtils.getLongProperty("druid.read.batch.time.grad.hour");
|
public static final Long DRUID_READ_BATCH_TIME_GRAD_HOUR = ConfigUtils.getLongProperty("druid.read.batch.time.grad.hour");
|
||||||
public static final Integer THREAD_MAX_NUM = ConfigUtils.getIntProperty("thread.max.num");
|
public static final Integer THREAD_POOL_NUM = ConfigUtils.getIntProperty("thread.pool.num");
|
||||||
|
|
||||||
// http config
|
// http config
|
||||||
public static final Integer HTTP_REQUEST_TIMEOUT = ConfigUtils.getIntProperty("http.request.timeout");
|
public static final Integer HTTP_REQUEST_TIMEOUT = ConfigUtils.getIntProperty("http.request.timeout");
|
||||||
|
|||||||
@@ -35,7 +35,6 @@ public class BaselineGeneration {
|
|||||||
|
|
||||||
private static final Tuple2<Long, Long> START_END_TIMES = DruidData.getTimeLimit();
|
private static final Tuple2<Long, Long> START_END_TIMES = DruidData.getTimeLimit();
|
||||||
private static final Map<String, List<Map<String, Object>>> allFromDruid = new HashMap<>();
|
private static final Map<String, List<Map<String, Object>>> allFromDruid = new HashMap<>();
|
||||||
private static int threadNum = ApplicationConfig.THREAD_MAX_NUM;
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* 程序执行
|
* 程序执行
|
||||||
@@ -65,19 +64,19 @@ public class BaselineGeneration {
|
|||||||
long start = System.currentTimeMillis();
|
long start = System.currentTimeMillis();
|
||||||
|
|
||||||
long timeGrad = 3600000 * ApplicationConfig.DRUID_READ_BATCH_TIME_GRAD_HOUR;
|
long timeGrad = 3600000 * ApplicationConfig.DRUID_READ_BATCH_TIME_GRAD_HOUR;
|
||||||
int threadPoolNum = (int) ((START_END_TIMES._2-START_END_TIMES._1)/timeGrad);
|
int loadDataThreadPoolNum = (int) ((START_END_TIMES._2-START_END_TIMES._1)/timeGrad);
|
||||||
ArrayList<Future<Map<String, List<Map<String, Object>>>>> resultList = new ArrayList<>();
|
ArrayList<Future<Map<String, List<Map<String, Object>>>>> resultList = new ArrayList<>();
|
||||||
CountDownLatch loadDataCountDownLatch = new CountDownLatch(threadPoolNum);
|
CountDownLatch loadDataCountDownLatch = new CountDownLatch(loadDataThreadPoolNum);
|
||||||
|
|
||||||
ThreadFactory loadDataThreadFactory = new ThreadFactoryBuilder()
|
ThreadFactory loadDataThreadFactory = new ThreadFactoryBuilder()
|
||||||
.setNameFormat("baseline-load-data-%d").build();
|
.setNameFormat("baseline-load-data-%d").build();
|
||||||
ThreadPoolExecutor loadDataExecutor = new ThreadPoolExecutor(
|
ThreadPoolExecutor loadDataExecutor = new ThreadPoolExecutor(
|
||||||
threadNum, threadNum, 0L,
|
loadDataThreadPoolNum, loadDataThreadPoolNum, 0L,
|
||||||
TimeUnit.MILLISECONDS, new LinkedBlockingQueue<>(1024), loadDataThreadFactory,
|
TimeUnit.MILLISECONDS, new LinkedBlockingQueue<>(1024), loadDataThreadFactory,
|
||||||
new ThreadPoolExecutor.AbortPolicy());
|
new ThreadPoolExecutor.AbortPolicy());
|
||||||
|
|
||||||
// 按ip数分区
|
// 按ip数分区
|
||||||
for (int i = 0; i < threadPoolNum; i++) {
|
for (int i = 0; i < loadDataThreadPoolNum; i++) {
|
||||||
String sql = DruidData.getDruidQuerySql(ATTACK_TYPE_LIST, START_END_TIMES._1, i, timeGrad);
|
String sql = DruidData.getDruidQuerySql(ATTACK_TYPE_LIST, START_END_TIMES._1, i, timeGrad);
|
||||||
ReadHistoricalDruidData readHistoricalDruidData = new ReadHistoricalDruidData(
|
ReadHistoricalDruidData readHistoricalDruidData = new ReadHistoricalDruidData(
|
||||||
sql,
|
sql,
|
||||||
@@ -115,17 +114,19 @@ public class BaselineGeneration {
|
|||||||
* @throws InterruptedException
|
* @throws InterruptedException
|
||||||
*/
|
*/
|
||||||
private static void baselineGenration() throws InterruptedException {
|
private static void baselineGenration() throws InterruptedException {
|
||||||
|
List<Map<String, List<Map<String, Object>>>> batchDruidDataLists = DruidData.splitMap(allFromDruid, ApplicationConfig.BASELINE_GENERATE_BATCH_SIZE);
|
||||||
|
int generationThreadPoolNum = batchDruidDataLists.size();
|
||||||
|
CountDownLatch generateCountDownLatch = new CountDownLatch(generationThreadPoolNum);
|
||||||
|
|
||||||
ThreadFactory generationThreadFactory = new ThreadFactoryBuilder()
|
ThreadFactory generationThreadFactory = new ThreadFactoryBuilder()
|
||||||
.setNameFormat("baseline-generate-%d").build();
|
.setNameFormat("baseline-generate-%d").build();
|
||||||
ThreadPoolExecutor generationExecutor = new ThreadPoolExecutor(
|
ThreadPoolExecutor generationExecutor = new ThreadPoolExecutor(
|
||||||
threadNum, threadNum, 0L,
|
generationThreadPoolNum, generationThreadPoolNum, 0L,
|
||||||
TimeUnit.MILLISECONDS, new LinkedBlockingQueue<>(1024), generationThreadFactory,
|
TimeUnit.MILLISECONDS, new LinkedBlockingQueue<>(1024), generationThreadFactory,
|
||||||
new ThreadPoolExecutor.AbortPolicy());
|
new ThreadPoolExecutor.AbortPolicy());
|
||||||
|
|
||||||
LOG.info("Baseline batch 大小: " + ApplicationConfig.BASELINE_GENERATE_BATCH_SIZE);
|
LOG.info("Baseline batch 大小: " + ApplicationConfig.BASELINE_GENERATE_BATCH_SIZE);
|
||||||
|
|
||||||
List<Map<String, List<Map<String, Object>>>> batchDruidDataLists = DruidData.splitMap(allFromDruid, ApplicationConfig.BASELINE_GENERATE_BATCH_SIZE);
|
|
||||||
CountDownLatch generateCountDownLatch = new CountDownLatch(batchDruidDataLists.size());
|
|
||||||
for (Map<String, List<Map<String, Object>>>batchDruidData: batchDruidDataLists){
|
for (Map<String, List<Map<String, Object>>>batchDruidData: batchDruidDataLists){
|
||||||
if(batchDruidData.size()>0){
|
if(batchDruidData.size()>0){
|
||||||
BaselineSingleThread baselineSingleThread = new BaselineSingleThread(
|
BaselineSingleThread baselineSingleThread = new BaselineSingleThread(
|
||||||
|
|||||||
@@ -1,67 +0,0 @@
|
|||||||
package cn.mesalab.utils;
|
|
||||||
|
|
||||||
import com.google.common.util.concurrent.ThreadFactoryBuilder;
|
|
||||||
import org.omg.PortableInterceptor.INACTIVE;
|
|
||||||
import org.slf4j.Logger;
|
|
||||||
import org.slf4j.LoggerFactory;
|
|
||||||
|
|
||||||
import java.util.concurrent.*;
|
|
||||||
|
|
||||||
public class ExecutorThreadPool {
|
|
||||||
private static final Logger LOG = LoggerFactory.getLogger(ExecutorThreadPool.class);
|
|
||||||
private static ExecutorService pool = null ;
|
|
||||||
private static ExecutorThreadPool poolExecutor = null;
|
|
||||||
private int threadPoolNum;
|
|
||||||
|
|
||||||
static {
|
|
||||||
getThreadPool();
|
|
||||||
}
|
|
||||||
|
|
||||||
public ExecutorThreadPool(Integer threadPoolNum){
|
|
||||||
this.threadPoolNum.
|
|
||||||
}
|
|
||||||
|
|
||||||
private static void getThreadPool(){
|
|
||||||
ThreadFactory namedThreadFactory = new ThreadFactoryBuilder()
|
|
||||||
.setNameFormat("iplearning-application-pool-%d").build();
|
|
||||||
pool = new ThreadPoolExecutor(ApplicationConfig.THREAD_POOL_NUMBER, ApplicationConfig.THREAD_POOL_NUMBER,
|
|
||||||
0L, TimeUnit.SECONDS,
|
|
||||||
new LinkedBlockingQueue<>(1024), namedThreadFactory, new ThreadPoolExecutor.AbortPolicy());
|
|
||||||
}
|
|
||||||
|
|
||||||
public static ExecutorThreadPool getInstance(){
|
|
||||||
if (null == poolExecutor){
|
|
||||||
poolExecutor = new ExecutorThreadPool();
|
|
||||||
}
|
|
||||||
return poolExecutor;
|
|
||||||
}
|
|
||||||
|
|
||||||
public void executor(Runnable command){
|
|
||||||
pool.execute(command);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Deprecated
|
|
||||||
public void awaitThreadTask(){
|
|
||||||
try {
|
|
||||||
while (!pool.awaitTermination(ApplicationConfig.THREAD_AWAIT_TERMINATION_TIME, TimeUnit.SECONDS)) {
|
|
||||||
LOG.warn("线程池没有关闭");
|
|
||||||
}
|
|
||||||
} catch (InterruptedException e) {
|
|
||||||
e.printStackTrace();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
public void shutdown(){
|
|
||||||
pool.shutdown();
|
|
||||||
}
|
|
||||||
|
|
||||||
@Deprecated
|
|
||||||
public static Long getThreadNumber(){
|
|
||||||
String name = Thread.currentThread().getName();
|
|
||||||
String[] split = name.split("-");
|
|
||||||
return Long.parseLong(split[3]);
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
}
|
|
||||||
@@ -53,7 +53,7 @@ baseline.kalman.r=4
|
|||||||
# FOR TEST
|
# FOR TEST
|
||||||
baseline.generate.batch.size=100
|
baseline.generate.batch.size=100
|
||||||
druid.read.batch.time.grad.hour=4
|
druid.read.batch.time.grad.hour=4
|
||||||
thread.max.num=10
|
thread.pool.num=10
|
||||||
|
|
||||||
|
|
||||||
# http client配置
|
# http client配置
|
||||||
|
|||||||
@@ -1,71 +0,0 @@
|
|||||||
package cn.mesalab.utils;
|
|
||||||
|
|
||||||
import cn.mesalab.config.ApplicationConfig;
|
|
||||||
import cn.mesalab.dao.DruidData;
|
|
||||||
import com.google.common.collect.Maps;
|
|
||||||
import com.zdjizhi.utils.JsonMapper;
|
|
||||||
import sun.net.util.URLUtil;
|
|
||||||
|
|
||||||
import java.net.URL;
|
|
||||||
import java.util.Map;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* @author yjy
|
|
||||||
* @version 1.0
|
|
||||||
* @date 2021/8/3 4:43 下午
|
|
||||||
*/
|
|
||||||
public class HttpClientUtilsTest {
|
|
||||||
private static HttpClientUtils httpClientUtils = new HttpClientUtils();
|
|
||||||
|
|
||||||
public static void main(String[] args) {
|
|
||||||
executeHttpPost("select * from top_server_ip_test_log limit 10");
|
|
||||||
}
|
|
||||||
|
|
||||||
private static Map<String, String> executeHttpPost(String sql){
|
|
||||||
String queryUrl = "http://192.168.44.12:8082/druid/v2/sql";
|
|
||||||
DruidQueryParam druidQueryParam = getDruidQueryParam(sql);
|
|
||||||
int socketTimeout = ApplicationConfig.HTTP_RESPONSE_TIMEOUT;
|
|
||||||
Map<String, String> stringStringMap = httpClientUtils.httpPost(queryUrl, JsonMapper.toJsonString(druidQueryParam), socketTimeout);
|
|
||||||
System.out.println(stringStringMap.toString());
|
|
||||||
return stringStringMap;
|
|
||||||
}
|
|
||||||
|
|
||||||
public static DruidQueryParam getDruidQueryParam(String sql) {
|
|
||||||
DruidQueryParam druidQueryParam = new DruidQueryParam();
|
|
||||||
druidQueryParam.setQuery(sql);
|
|
||||||
druidQueryParam.getContext().put("skipEmptyBuckets", "true");
|
|
||||||
druidQueryParam.setResultFormat("object");
|
|
||||||
return druidQueryParam;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
class DruidQueryParam {
|
|
||||||
private String query;
|
|
||||||
private Map<String, String> context = Maps.newHashMap();
|
|
||||||
private String resultFormat;
|
|
||||||
|
|
||||||
public String getQuery() {
|
|
||||||
return query;
|
|
||||||
}
|
|
||||||
|
|
||||||
public void setQuery(String query) {
|
|
||||||
this.query = query;
|
|
||||||
}
|
|
||||||
|
|
||||||
public Map<String, String> getContext() {
|
|
||||||
return context;
|
|
||||||
}
|
|
||||||
|
|
||||||
public void setContext(Map<String, String> context) {
|
|
||||||
this.context = context;
|
|
||||||
}
|
|
||||||
|
|
||||||
public String getResultFormat() {
|
|
||||||
return resultFormat;
|
|
||||||
}
|
|
||||||
|
|
||||||
public void setResultFormat(String resultFormat) {
|
|
||||||
this.resultFormat = resultFormat;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
Reference in New Issue
Block a user