TSG-14938 TFE支持新控制报文格式; 调整代码结构
This commit is contained in:
@@ -6,22 +6,89 @@
|
||||
#include <linux/netfilter.h> // for NF_ACCEPT
|
||||
#include <libnetfilter_queue/libnetfilter_queue.h>
|
||||
#include <linux/if_tun.h>
|
||||
#include <MESA/MESA_prof_load.h>
|
||||
|
||||
#include <bpf_obj.h>
|
||||
#include <tfe_utils.h>
|
||||
#include <tfe_cmsg.h>
|
||||
#include <proxy.h>
|
||||
#include <tfe_acceptor_kni.h>
|
||||
#include "io_uring.h"
|
||||
#include "tfe_tap_rss.h"
|
||||
#include "tfe_metrics.h"
|
||||
#include "tfe_tcp_restore.h"
|
||||
#include "acceptor_kni_v4.h"
|
||||
#include "tap.h"
|
||||
#include "tfe_packet_io.h"
|
||||
#include "tfe_session_table.h"
|
||||
|
||||
|
||||
static int tfe_tap_read_per_thread(int tap_fd, char *buff, int buff_size, void *logger)
|
||||
{
|
||||
int ret = read(tap_fd, buff, buff_size);
|
||||
if (ret < 0)
|
||||
{
|
||||
if (errno != EWOULDBLOCK && errno != EAGAIN)
|
||||
{
|
||||
TFE_LOG_ERROR(g_default_logger, "%s: unable to read data from tapfd %d, aborting: %s", LOG_TAG_PKTIO, tap_fd, strerror(errno));
|
||||
}
|
||||
}
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
void acceptor_ctx_destory(struct acceptor_kni_v4 * ctx)
|
||||
{
|
||||
if (ctx)
|
||||
{
|
||||
packet_io_destory(ctx->io);
|
||||
global_metrics_destory(ctx->metrics);
|
||||
|
||||
free(ctx);
|
||||
ctx = NULL;
|
||||
}
|
||||
return;
|
||||
}
|
||||
|
||||
struct acceptor_kni_v4 *acceptor_ctx_create(const char *profile)
|
||||
{
|
||||
struct acceptor_kni_v4 *ctx = ALLOC(struct acceptor_kni_v4, 1);
|
||||
|
||||
MESA_load_profile_int_def(profile, "PACKET_IO", "firewall_sids", (int *)&(ctx->firewall_sids), 1000);
|
||||
MESA_load_profile_int_def(profile, "PACKET_IO", "proxy_sids", (int *)&(ctx->proxy_sids), 1001);
|
||||
MESA_load_profile_int_def(profile, "PACKET_IO", "service_chaining_sids", (int *)&(ctx->sce_sids), 1002);
|
||||
MESA_load_profile_int_def(profile, "PACKET_IO", "packet_io_threads", (int *)&(ctx->nr_worker_threads), 8);
|
||||
MESA_load_profile_uint_range(profile, "system", "cpu_affinity_mask", TFE_THREAD_MAX, (unsigned int *)ctx->cpu_affinity_mask);
|
||||
ctx->nr_worker_threads = MIN(ctx->nr_worker_threads, TFE_THREAD_MAX);
|
||||
|
||||
CPU_ZERO(&ctx->coremask);
|
||||
for (int i = 0; i < ctx->nr_worker_threads; i++)
|
||||
{
|
||||
int cpu_id = ctx->cpu_affinity_mask[i];
|
||||
CPU_SET(cpu_id, &ctx->coremask);
|
||||
}
|
||||
|
||||
ctx->io = packet_io_create(profile, ctx->nr_worker_threads, &ctx->coremask);
|
||||
if (ctx->io == NULL)
|
||||
{
|
||||
goto error_out;
|
||||
}
|
||||
|
||||
ctx->metrics = global_metrics_create();
|
||||
if (ctx->metrics == NULL)
|
||||
{
|
||||
goto error_out;
|
||||
}
|
||||
|
||||
return ctx;
|
||||
|
||||
error_out:
|
||||
acceptor_ctx_destory(ctx);
|
||||
return NULL;
|
||||
}
|
||||
|
||||
static void *worker_thread_cycle(void *arg)
|
||||
{
|
||||
struct acceptor_thread_ctx *thread_ctx = (struct acceptor_thread_ctx *)arg;
|
||||
struct packet_io *handle = thread_ctx->ref_io;
|
||||
struct acceptor_ctx *acceptor_ctx = thread_ctx->ref_acceptor_ctx;
|
||||
|
||||
int pkg_len = 0;
|
||||
char thread_name[16];
|
||||
@@ -38,7 +105,7 @@ static void *worker_thread_cycle(void *arg)
|
||||
goto error_out;
|
||||
}
|
||||
|
||||
if (acceptor_ctx->config->enable_iouring) {
|
||||
if (is_enable_iouring(handle)) {
|
||||
io_uring_register_read_callback(thread_ctx->tap_ctx->io_uring_fd, handle_raw_packet_from_tap, thread_ctx);
|
||||
io_uring_register_read_callback(thread_ctx->tap_ctx->io_uring_c, handle_decryption_packet_from_tap, thread_ctx);
|
||||
io_uring_register_read_callback(thread_ctx->tap_ctx->io_uring_s, handle_decryption_packet_from_tap, thread_ctx);
|
||||
@@ -52,10 +119,10 @@ static void *worker_thread_cycle(void *arg)
|
||||
|
||||
while(1) {
|
||||
n_pkt_recv_from_nf = packet_io_polling_nf_interface(handle, thread_ctx->thread_index, thread_ctx);
|
||||
if (acceptor_ctx->config->enable_iouring) {
|
||||
if (is_enable_iouring(handle)) {
|
||||
n_pkt_recv_from_tap = io_uring_peek_ready_entrys(thread_ctx->tap_ctx->io_uring_fd);
|
||||
n_pkt_recv_from_tap_c = io_uring_peek_ready_entrys(thread_ctx->tap_ctx->io_uring_c);
|
||||
n_pkt_recv_from_tap_c = io_uring_peek_ready_entrys(thread_ctx->tap_ctx->io_uring_s);
|
||||
n_pkt_recv_from_tap_s = io_uring_peek_ready_entrys(thread_ctx->tap_ctx->io_uring_s);
|
||||
}
|
||||
else {
|
||||
if ((pkg_len = tfe_tap_read_per_thread(thread_ctx->tap_ctx->tap_fd, thread_ctx->tap_ctx->buff, thread_ctx->tap_ctx->buff_size, g_default_logger)) > 0)
|
||||
@@ -63,22 +130,22 @@ static void *worker_thread_cycle(void *arg)
|
||||
handle_raw_packet_from_tap(thread_ctx->tap_ctx->buff, pkg_len, thread_ctx);
|
||||
}
|
||||
|
||||
// if ((pkg_len = tfe_tap_read_per_thread(thread_ctx->tap_ctx->tap_c, thread_ctx->tap_ctx->buff, thread_ctx->tap_ctx->buff_size, g_default_logger)) > 0)
|
||||
// {
|
||||
// handle_decryption_packet_from_tap(thread_ctx->tap_ctx->buff, pkg_len, thread_ctx);
|
||||
// }
|
||||
if ((pkg_len = tfe_tap_read_per_thread(thread_ctx->tap_ctx->tap_c, thread_ctx->tap_ctx->buff, thread_ctx->tap_ctx->buff_size, g_default_logger)) > 0)
|
||||
{
|
||||
handle_decryption_packet_from_tap(thread_ctx->tap_ctx->buff, pkg_len, thread_ctx);
|
||||
}
|
||||
|
||||
// if ((pkg_len = tfe_tap_read_per_thread(thread_ctx->tap_ctx->tap_s, thread_ctx->tap_ctx->buff, thread_ctx->tap_ctx->buff_size, g_default_logger)) > 0)
|
||||
// {
|
||||
// handle_decryption_packet_from_tap(thread_ctx->tap_ctx->buff, pkg_len, thread_ctx);
|
||||
// }
|
||||
if ((pkg_len = tfe_tap_read_per_thread(thread_ctx->tap_ctx->tap_s, thread_ctx->tap_ctx->buff, thread_ctx->tap_ctx->buff_size, g_default_logger)) > 0)
|
||||
{
|
||||
handle_decryption_packet_from_tap(thread_ctx->tap_ctx->buff, pkg_len, thread_ctx);
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
// if (n_pkt_recv_from_nf == 0)
|
||||
// {
|
||||
// packet_io_thread_wait(handle, thread_ctx, 0);
|
||||
// }
|
||||
if (n_pkt_recv_from_nf == 0 && n_pkt_recv_from_tap == 0 && n_pkt_recv_from_tap_c == 0 && n_pkt_recv_from_tap_s == 0)
|
||||
{
|
||||
packet_io_thread_wait(handle, thread_ctx, -1);
|
||||
}
|
||||
|
||||
if (__atomic_fetch_add(&thread_ctx->session_table_need_reset, 0, __ATOMIC_RELAXED) > 0)
|
||||
{
|
||||
@@ -92,17 +159,21 @@ error_out:
|
||||
return (void *)NULL;
|
||||
}
|
||||
|
||||
void acceptor_kni_v4_destroy()
|
||||
void acceptor_kni_v4_destroy(struct acceptor_kni_v4 *ctx)
|
||||
{
|
||||
if (ctx)
|
||||
{
|
||||
packet_io_destory(ctx->io);
|
||||
|
||||
free(ctx);
|
||||
ctx = NULL;
|
||||
}
|
||||
return;
|
||||
}
|
||||
|
||||
struct acceptor_kni_v4 *acceptor_kni_v4_create(struct tfe_proxy *proxy, const char *profile, void *logger)
|
||||
{
|
||||
int ret = 0;
|
||||
struct acceptor_kni_v4 *__ctx = (struct acceptor_kni_v4 *)calloc(1, sizeof(struct acceptor_kni_v4));
|
||||
|
||||
struct acceptor_ctx *acceptor_ctx = acceptor_ctx_create(profile);
|
||||
struct acceptor_kni_v4 *acceptor_ctx = acceptor_ctx_create(profile);
|
||||
if (acceptor_ctx == NULL)
|
||||
goto error_out;
|
||||
|
||||
@@ -113,27 +184,15 @@ struct acceptor_kni_v4 *acceptor_kni_v4_create(struct tfe_proxy *proxy, const ch
|
||||
acceptor_ctx->work_threads[i].ref_acceptor_ctx = acceptor_ctx;
|
||||
|
||||
acceptor_ctx->work_threads[i].tap_ctx = tfe_tap_ctx_create(&acceptor_ctx->work_threads[i]);
|
||||
if (acceptor_ctx->config->enable_iouring) {
|
||||
int eventfd = 0;
|
||||
struct tap_ctx *tap_ctx = acceptor_ctx->work_threads[i].tap_ctx;
|
||||
tap_ctx->io_uring_fd = io_uring_instance_create(tap_ctx->tap_fd, eventfd, acceptor_ctx->config->ring_size, acceptor_ctx->config->buff_size, acceptor_ctx->config->flags, acceptor_ctx->config->sq_thread_idle, acceptor_ctx->config->enable_debuglog);
|
||||
tap_ctx->io_uring_c = io_uring_instance_create(tap_ctx->tap_c, eventfd, acceptor_ctx->config->ring_size, acceptor_ctx->config->buff_size, acceptor_ctx->config->flags, acceptor_ctx->config->sq_thread_idle, acceptor_ctx->config->enable_debuglog);
|
||||
tap_ctx->io_uring_s = io_uring_instance_create(tap_ctx->tap_s, eventfd, acceptor_ctx->config->ring_size, acceptor_ctx->config->buff_size, acceptor_ctx->config->flags, acceptor_ctx->config->sq_thread_idle, acceptor_ctx->config->enable_debuglog);
|
||||
}
|
||||
if (acceptor_ctx->work_threads[i].tap_ctx == NULL)
|
||||
goto error_out;
|
||||
|
||||
acceptor_ctx->work_threads[i].session_table = session_table_create();
|
||||
acceptor_ctx->work_threads[i].ref_io = acceptor_ctx->io;
|
||||
acceptor_ctx->work_threads[i].ref_proxy = proxy;
|
||||
acceptor_ctx->work_threads[i].ref_tap_config = acceptor_ctx->config;
|
||||
acceptor_ctx->work_threads[i].ref_metrics = acceptor_ctx->metrics;
|
||||
acceptor_ctx->work_threads[i].ref_acceptor_ctx = acceptor_ctx;
|
||||
acceptor_ctx->work_threads[i].session_table_need_reset = 0;
|
||||
|
||||
if (acceptor_ctx->config->tap_rps_enable)
|
||||
{
|
||||
ret = tfe_tap_set_rps(g_default_logger, acceptor_ctx->config->tap_device, i, acceptor_ctx->config->tap_rps_mask);
|
||||
if (ret != 0)
|
||||
goto error_out;
|
||||
}
|
||||
}
|
||||
|
||||
for (int i = 0; i < acceptor_ctx->nr_worker_threads; i++) {
|
||||
@@ -144,9 +203,14 @@ struct acceptor_kni_v4 *acceptor_kni_v4_create(struct tfe_proxy *proxy, const ch
|
||||
}
|
||||
}
|
||||
|
||||
return __ctx;
|
||||
return acceptor_ctx;
|
||||
|
||||
error_out:
|
||||
acceptor_kni_v4_destroy();
|
||||
for (int i = 0; i < acceptor_ctx->nr_worker_threads; i++) {
|
||||
tfe_tap_ctx_destory(acceptor_ctx->work_threads[i].tap_ctx);
|
||||
session_table_destory(acceptor_ctx->work_threads[i].session_table);
|
||||
}
|
||||
|
||||
acceptor_kni_v4_destroy(acceptor_ctx);
|
||||
return NULL;
|
||||
}
|
||||
Reference in New Issue
Block a user