feature: consume all packets and free all sessions before exit

This commit is contained in:
luwenpeng
2024-08-27 16:19:20 +08:00
parent 11bf852c15
commit 79e70f7145
10 changed files with 103 additions and 96 deletions

View File

@@ -61,6 +61,7 @@ tcp_reassembly_max_segments = 256 # range: [2, 4096]
# Note: free_expired_session_interval determines the precision of session_manager timeout # Note: free_expired_session_interval determines the precision of session_manager timeout
free_expired_session_interval = 50 # range: [1, 60000] (ms) free_expired_session_interval = 50 # range: [1, 60000] (ms)
free_expired_session_batch = 1000 # range: [1, 60000] free_expired_session_batch = 1000 # range: [1, 60000]
froce_session_expire_before_exit = 0 # 1: force session to expire before exit, 0: wait for session to naturally expire before exit.
# Note: free_expired_ip_frag_interval determines the precision of ip_reassembly timeout # Note: free_expired_ip_frag_interval determines the precision of ip_reassembly timeout
free_expired_ip_frag_interval = 50 # range: [1, 60000] (ms) free_expired_ip_frag_interval = 50 # range: [1, 60000] (ms)

View File

@@ -473,6 +473,19 @@ static int parse_schedule_options(toml_table_t *root, struct schedule_options *o
return -1; return -1;
} }
ptr = toml_raw_in(table, "froce_session_expire_before_exit");
if (ptr == NULL)
{
CONFIG_LOG_ERROR("config file missing schedule->froce_session_expire_before_exit");
return -1;
}
opts->froce_session_expire_before_exit = atoll(ptr);
if (opts->froce_session_expire_before_exit != 0 && opts->froce_session_expire_before_exit != 1)
{
CONFIG_LOG_ERROR("config file invalid schedule->froce_session_expire_before_exit %ld, range [0, 1]", opts->froce_session_expire_before_exit);
return -1;
}
ptr = toml_raw_in(table, "free_expired_ip_frag_interval"); ptr = toml_raw_in(table, "free_expired_ip_frag_interval");
if (ptr == NULL) if (ptr == NULL)
{ {

View File

@@ -14,6 +14,7 @@ struct schedule_options
// Note: free_expired_session_interval determines the precision of session_manager timeout // Note: free_expired_session_interval determines the precision of session_manager timeout
uint64_t free_expired_session_interval; // range: [1, 60000] (ms) uint64_t free_expired_session_interval; // range: [1, 60000] (ms)
uint64_t free_expired_session_batch; // range: [1, 60000] uint64_t free_expired_session_batch; // range: [1, 60000]
uint64_t froce_session_expire_before_exit; // range: [0, 1]
// Note: free_expired_ip_frag_interval determines the precision of ip_reassembly timeout // Note: free_expired_ip_frag_interval determines the precision of ip_reassembly timeout
uint64_t free_expired_ip_frag_interval; // range: [1, 60000] (ms) uint64_t free_expired_ip_frag_interval; // range: [1, 60000] (ms)

View File

@@ -37,28 +37,15 @@ static const char logo_str[] =
" \\__ \\ | |_ | __/ | | | | | (_| | | |\n" " \\__ \\ | |_ | __/ | | | | | (_| | | |\n"
" |___/ \\__| \\___| |_| |_| \\__,_| |_|\n"; " |___/ \\__| \\___| |_| |_| \\__,_| |_|\n";
struct schedule_data
{
uint64_t last_free_expired_session_timestamp;
uint64_t last_free_expired_ip_frag_timestamp;
uint64_t last_merge_thread_stat_timestamp;
uint64_t free_expired_session_interval;
uint64_t free_expired_session_batch;
uint64_t free_expired_ip_frag_interval;
uint64_t free_expired_ip_frag_batch;
uint64_t merge_stat_interval;
uint64_t packet_io_yield_interval;
};
struct stellar_thread struct stellar_thread
{ {
pthread_t tid; pthread_t tid;
uint16_t idx; uint16_t idx;
uint64_t is_runing; uint64_t is_runing;
struct schedule_data sched_data; uint64_t need_exit;
uint64_t last_free_expired_session_timestamp;
uint64_t last_free_expired_ip_frag_timestamp;
uint64_t last_merge_thread_stat_timestamp;
struct snowflake *snowflake; struct snowflake *snowflake;
struct ip_reassembly *ip_mgr; struct ip_reassembly *ip_mgr;
struct session_manager *sess_mgr; struct session_manager *sess_mgr;
@@ -202,11 +189,19 @@ static void *work_thread(void *arg)
struct stellar_thread *thread = (struct stellar_thread *)arg; struct stellar_thread *thread = (struct stellar_thread *)arg;
struct ip_reassembly *ip_reass = thread->ip_mgr; struct ip_reassembly *ip_reass = thread->ip_mgr;
struct session_manager *sess_mgr = thread->sess_mgr; struct session_manager *sess_mgr = thread->sess_mgr;
struct session_manager_stat *sess_stat = session_manager_stat(sess_mgr);
struct stellar *st = thread->st; struct stellar *st = thread->st;
struct stellar_config *config = &st->config;
struct stellar_runtime *runtime = &st->runtime; struct stellar_runtime *runtime = &st->runtime;
struct schedule_data *sched_data = &thread->sched_data;
struct packet_io *packet_io = runtime->packet_io; struct packet_io *packet_io = runtime->packet_io;
struct plugin_manager_schema *plug_mgr = runtime->plug_mgr; struct plugin_manager_schema *plug_mgr = runtime->plug_mgr;
uint64_t free_expired_session_interval = config->sched_opts.free_expired_session_interval;
uint64_t free_expired_session_batch = config->sched_opts.free_expired_session_batch;
uint64_t froce_session_expire_before_exit = config->sched_opts.froce_session_expire_before_exit;
uint64_t free_expired_ip_frag_interval = config->sched_opts.free_expired_ip_frag_interval;
uint64_t free_expired_ip_frag_batch = config->sched_opts.free_expired_ip_frag_batch;
uint64_t merge_stat_interval = config->sched_opts.merge_stat_interval;
uint64_t packet_io_yield_interval = config->sched_opts.packet_io_yield_interval;
uint16_t thr_idx = thread->idx; uint16_t thr_idx = thread->idx;
__current_thread_idx = thr_idx; __current_thread_idx = thr_idx;
@@ -232,7 +227,7 @@ static void *work_thread(void *arg)
ATOMIC_SET(&thread->is_runing, 1); ATOMIC_SET(&thread->is_runing, 1);
CORE_LOG_FATAL("worker thread %d runing", thr_idx); CORE_LOG_FATAL("worker thread %d runing", thr_idx);
while (ATOMIC_READ(&runtime->need_exit) == 0) while (ATOMIC_READ(&thread->need_exit) == 0)
{ {
/* /*
* We use the system's real time instead of monotonic time for the following reasons: * We use the system's real time instead of monotonic time for the following reasons:
@@ -345,29 +340,43 @@ static void *work_thread(void *arg)
plugin_manager_on_polling(plug_mgr); plugin_manager_on_polling(plug_mgr);
// per free_expired_session_interval MAX free_expired_session_batch sessions are released // per free_expired_session_interval MAX free_expired_session_batch sessions are released
if (now_ms - sched_data->last_free_expired_session_timestamp >= sched_data->free_expired_session_interval) if (now_ms - thread->last_free_expired_session_timestamp >= free_expired_session_interval)
{ {
free_expired_sessions(sess_mgr, sched_data->free_expired_session_batch, now_ms); free_expired_sessions(sess_mgr, free_expired_session_batch, now_ms);
sched_data->last_free_expired_session_timestamp = now_ms; thread->last_free_expired_session_timestamp = now_ms;
} }
// per merge_stat_interval merge thread stat // per merge_stat_interval merge thread stat
if (now_ms - sched_data->last_merge_thread_stat_timestamp >= sched_data->merge_stat_interval) if (now_ms - thread->last_merge_thread_stat_timestamp >= merge_stat_interval)
{ {
merge_thread_stat(thread); merge_thread_stat(thread);
sched_data->last_merge_thread_stat_timestamp = now_ms; thread->last_merge_thread_stat_timestamp = now_ms;
} }
// per free_expired_ip_frag_interval MAX free_expired_ip_frag_batch ip fragments are released // per free_expired_ip_frag_interval MAX free_expired_ip_frag_batch ip fragments are released
if (now_ms - sched_data->last_free_expired_ip_frag_timestamp >= sched_data->free_expired_ip_frag_interval) if (now_ms - thread->last_free_expired_ip_frag_timestamp >= free_expired_ip_frag_interval)
{ {
ip_reassembly_expire(ip_reass, sched_data->free_expired_ip_frag_batch, now_ms); ip_reassembly_expire(ip_reass, free_expired_ip_frag_batch, now_ms);
sched_data->last_free_expired_ip_frag_timestamp = now_ms; thread->last_free_expired_ip_frag_timestamp = now_ms;
} }
if (nr_recv == 0) if (nr_recv == 0)
{ {
packet_io_yield(packet_io, thr_idx, sched_data->packet_io_yield_interval); packet_io_yield(packet_io, thr_idx, packet_io_yield_interval);
}
}
if (froce_session_expire_before_exit)
{
free_expired_sessions(sess_mgr, UINT64_MAX, UINT64_MAX);
}
else
{
while (sess_stat->tcp_sess_used > 0 || sess_stat->udp_sess_used > 0)
{
now_ms = clock_get_real_time_ms();
free_expired_sessions(sess_mgr, free_expired_session_batch, now_ms);
usleep(1000); // 1ms
} }
} }
@@ -384,27 +393,6 @@ static void *work_thread(void *arg)
* Stellar Main Function * Stellar Main Function
******************************************************************************/ ******************************************************************************/
static int all_session_have_freed(struct stellar_runtime *runtime, struct stellar_config *config)
{
for (int i = 0; i < config->pkt_io_opts.nr_threads; i++)
{
struct session_manager *sess_mgr = runtime->threads[i].sess_mgr;
struct session_manager_stat *sess_stat = session_manager_stat(sess_mgr);
if (ATOMIC_READ(&sess_stat->tcp_sess_used) != 0)
{
return 0;
}
if (ATOMIC_READ(&sess_stat->udp_sess_used) != 0)
{
return 0;
}
}
return 1;
}
static int stellar_thread_init(struct stellar *st) static int stellar_thread_init(struct stellar *st)
{ {
struct stellar_runtime *runtime = &st->runtime; struct stellar_runtime *runtime = &st->runtime;
@@ -417,18 +405,9 @@ static int stellar_thread_init(struct stellar *st)
thread->idx = i; thread->idx = i;
thread->is_runing = 0; thread->is_runing = 0;
thread->sched_data.last_free_expired_session_timestamp = now_ms; thread->last_free_expired_session_timestamp = now_ms;
thread->sched_data.last_free_expired_ip_frag_timestamp = now_ms; thread->last_free_expired_ip_frag_timestamp = now_ms;
thread->sched_data.last_merge_thread_stat_timestamp = now_ms; thread->last_merge_thread_stat_timestamp = now_ms;
thread->sched_data.free_expired_session_interval = config->sched_opts.free_expired_session_interval;
thread->sched_data.free_expired_session_batch = config->sched_opts.free_expired_session_batch;
thread->sched_data.free_expired_ip_frag_interval = config->sched_opts.free_expired_ip_frag_interval;
thread->sched_data.free_expired_ip_frag_batch = config->sched_opts.free_expired_ip_frag_batch;
thread->sched_data.merge_stat_interval = config->sched_opts.merge_stat_interval;
thread->sched_data.packet_io_yield_interval = config->sched_opts.packet_io_yield_interval;
thread->snowflake = snowflake_new(i, config->snowflake_opts.snowflake_base, config->snowflake_opts.snowflake_offset); thread->snowflake = snowflake_new(i, config->snowflake_opts.snowflake_base, config->snowflake_opts.snowflake_offset);
if (thread->snowflake == NULL) if (thread->snowflake == NULL)
@@ -623,16 +602,21 @@ void stellar_run(struct stellar *st)
} }
usleep(1000); // 1ms usleep(1000); // 1ms
// only available in dump file mode, automatically exits when all sessions have been released // only available in dumpfile mode
if (packet_io_isbreak(runtime->packet_io) && all_session_have_freed(runtime, config)) if (packet_io_isbreak(runtime->packet_io))
{ {
CORE_LOG_FATAL("all sessions have been released, notify threads to exit"); for (uint16_t i = 0; i < config->pkt_io_opts.nr_threads; i++)
{
struct stellar_thread *thread = &runtime->threads[i];
ATOMIC_SET(&thread->need_exit, 1);
}
CORE_LOG_FATAL("notify worker thread to exit");
stellar_stat_output(runtime->stat); // flush stat stellar_stat_output(runtime->stat); // flush stat
ATOMIC_SET(&runtime->need_exit, 1); break;
} }
} }
// berfore exit, output last stat stellar_thread_join(st);
stellar_stat_output(runtime->stat); stellar_stat_output(runtime->stat);
} }
@@ -642,7 +626,6 @@ void stellar_free(struct stellar *st)
{ {
struct stellar_runtime *runtime = &st->runtime; struct stellar_runtime *runtime = &st->runtime;
stellar_thread_join(st);
stellar_thread_clean(st); stellar_thread_clean(st);
packet_io_free(runtime->packet_io); packet_io_free(runtime->packet_io);
plugin_manager_exit(runtime->plug_mgr); plugin_manager_exit(runtime->plug_mgr);

View File

@@ -13,7 +13,7 @@ struct ip_reassembly_options
uint32_t bucket_num; // range: [1, 4294967295] uint32_t bucket_num; // range: [1, 4294967295]
}; };
struct ip_reassembly_stat struct __attribute__((aligned(64))) ip_reassembly_stat
{ {
// IPv4 frag stat // IPv4 frag stat
uint64_t ip4_defrags_expected; uint64_t ip4_defrags_expected;

View File

@@ -35,6 +35,9 @@ struct dumpfile_io
uint64_t io_thread_need_exit; uint64_t io_thread_need_exit;
uint64_t io_thread_is_runing; uint64_t io_thread_is_runing;
uint64_t io_thread_wait_exit; uint64_t io_thread_wait_exit;
uint64_t read_pcap_files;
uint64_t read_pcap_pkts;
}; };
struct pcap_pkt struct pcap_pkt
@@ -121,12 +124,6 @@ static void packet_queue_pop(struct packet_queue *queue, void **data)
queue->head = (queue->head + 1) % queue->size; queue->head = (queue->head + 1) % queue->size;
} }
static int packet_queue_isempty(struct packet_queue *queue)
{
uint64_t read = ATOMIC_READ(&queue->queue[queue->head]);
return read == 0;
}
/****************************************************************************** /******************************************************************************
* Private API -- utils * Private API -- utils
******************************************************************************/ ******************************************************************************/
@@ -146,6 +143,7 @@ static void pcap_pkt_handler(u_char *user, const struct pcap_pkthdr *h, const u_
pcap_pkt->len = h->caplen; pcap_pkt->len = h->caplen;
pcap_pkt->ts = h->ts; pcap_pkt->ts = h->ts;
memcpy((char *)pcap_pkt->data, bytes, h->caplen); memcpy((char *)pcap_pkt->data, bytes, h->caplen);
ATOMIC_INC(&handle->read_pcap_pkts);
// calculate packet hash // calculate packet hash
struct packet pkt; struct packet pkt;
@@ -188,6 +186,7 @@ static int dumpfile_handler(struct dumpfile_io *handle, const char *pcap_file)
PACKET_IO_LOG_ERROR("unable to open pcap file: %s, %s", resolved_path, pcap_errbuf); PACKET_IO_LOG_ERROR("unable to open pcap file: %s, %s", resolved_path, pcap_errbuf);
return -1; return -1;
} }
handle->read_pcap_files++;
pcap_loop(handle->pcap, -1, pcap_pkt_handler, (u_char *)handle); pcap_loop(handle->pcap, -1, pcap_pkt_handler, (u_char *)handle);
pcap_close(handle->pcap); pcap_close(handle->pcap);
@@ -196,17 +195,23 @@ static int dumpfile_handler(struct dumpfile_io *handle, const char *pcap_file)
return 0; return 0;
} }
static int all_packet_processed(struct dumpfile_io *handle) static int all_packet_consumed(struct dumpfile_io *handle)
{ {
uint64_t consumed_pkts = 0;
uint64_t read_pcap_pkts = ATOMIC_READ(&handle->read_pcap_pkts);
for (uint16_t i = 0; i < handle->nr_threads; i++) for (uint16_t i = 0; i < handle->nr_threads; i++)
{ {
if (!packet_queue_isempty(handle->queue[i])) consumed_pkts += ATOMIC_READ(&handle->stat[i].pkts_rx);
}
if (consumed_pkts == read_pcap_pkts)
{
return 1;
}
else
{ {
return 0; return 0;
} }
} }
return 1;
}
static void *dumpfile_thread(void *arg) static void *dumpfile_thread(void *arg)
{ {
@@ -264,7 +269,7 @@ static void *dumpfile_thread(void *arg)
erro_out: erro_out:
while (ATOMIC_READ(&handle->io_thread_need_exit) == 0) while (ATOMIC_READ(&handle->io_thread_need_exit) == 0)
{ {
if (all_packet_processed(handle)) if (all_packet_consumed(handle))
{ {
ATOMIC_SET(&handle->io_thread_wait_exit, 1); ATOMIC_SET(&handle->io_thread_wait_exit, 1);
} }
@@ -330,6 +335,8 @@ void dumpfile_io_free(struct dumpfile_io *handle)
usleep(1000); usleep(1000);
} }
PACKET_IO_LOG_FATAL("dumpfile io thread read pcap files %lu, read pcap pkts %lu", handle->read_pcap_files, ATOMIC_READ(&handle->read_pcap_pkts));
struct pcap_pkt *pcap_pkt = NULL; struct pcap_pkt *pcap_pkt = NULL;
for (uint16_t i = 0; i < handle->nr_threads; i++) for (uint16_t i = 0; i < handle->nr_threads; i++)
{ {
@@ -380,7 +387,7 @@ uint16_t dumpfile_io_ingress(struct dumpfile_io *handle, uint16_t thr_idx, struc
} }
else else
{ {
stat->pkts_rx++; ATOMIC_INC(&stat->pkts_rx);
stat->bytes_rx += pcap_pkt->len; stat->bytes_rx += pcap_pkt->len;
stat->raw_pkts_rx++; stat->raw_pkts_rx++;

View File

@@ -48,7 +48,7 @@ struct session_manager_options
uint32_t tcp_reassembly_max_segments; // range: [2, 512] uint32_t tcp_reassembly_max_segments; // range: [2, 512]
}; };
struct session_manager_stat struct __attribute__((aligned(64))) session_manager_stat
{ {
// TCP session // TCP session
uint64_t history_tcp_sessions; uint64_t history_tcp_sessions;

View File

@@ -140,10 +140,10 @@ int main(int argc, char *argv[])
::testing::InitGoogleTest(&argc, argv); ::testing::InitGoogleTest(&argc, argv);
struct stellar *st = stellar_new("./conf/stellar.toml", "./plugin/spec.toml", "./conf/log.toml"); struct stellar *st = stellar_new("./conf/stellar.toml", "./plugin/spec.toml", "./conf/log.toml");
stellar_run(st); stellar_run(st);
stellar_free(st);
if (result_json_path != NULL) if (result_json_path != NULL)
{ {
ret = RUN_ALL_TESTS(); ret = RUN_ALL_TESTS();
} }
stellar_free(st);
return ret; return ret;
} }

View File

@@ -61,6 +61,7 @@ tcp_reassembly_max_segments = 256 # range: [2, 4096]
# Note: free_expired_session_interval determines the precision of session_manager timeout # Note: free_expired_session_interval determines the precision of session_manager timeout
free_expired_session_interval = 50 # range: [1, 60000] (ms) free_expired_session_interval = 50 # range: [1, 60000] (ms)
free_expired_session_batch = 100 # range: [1, 60000] free_expired_session_batch = 100 # range: [1, 60000]
froce_session_expire_before_exit = 0 # 1: force session to expire before exit, 0: wait for session to naturally expire before exit.
# Note: free_expired_ip_frag_interval determines the precision of ip_reassembly timeout # Note: free_expired_ip_frag_interval determines the precision of ip_reassembly timeout
free_expired_ip_frag_interval = 50 # range: [1, 60000] (ms) free_expired_ip_frag_interval = 50 # range: [1, 60000] (ms)

View File

@@ -61,6 +61,7 @@ tcp_reassembly_max_segments = 128 # range: [2, 4096]
# Note: free_expired_session_interval determines the precision of session_manager timeout # Note: free_expired_session_interval determines the precision of session_manager timeout
free_expired_session_interval = 50 # range: [1, 60000] (ms) free_expired_session_interval = 50 # range: [1, 60000] (ms)
free_expired_session_batch = 1000 # range: [1, 60000] free_expired_session_batch = 1000 # range: [1, 60000]
froce_session_expire_before_exit = 0 # 1: force session to expire before exit, 0: wait for session to naturally expire before exit.
# Note: free_expired_ip_frag_interval determines the precision of ip_reassembly timeout # Note: free_expired_ip_frag_interval determines the precision of ip_reassembly timeout
free_expired_ip_frag_interval = 50 # range: [1, 60000] (ms) free_expired_ip_frag_interval = 50 # range: [1, 60000] (ms)