| // SPDX-License-Identifier: GPL-2.0 |
| /* |
| * tcpdevmem netcat. Works similarly to netcat but does device memory TCP |
| * instead of regular TCP. Uses udmabuf to mock a dmabuf provider. |
| * |
| * Usage: |
| * |
| * On server: |
| * ncdevmem -s <server IP> [-c <client IP>] -f eth1 -l -p 5201 |
| * |
| * On client: |
| * echo -n "hello\nworld" | \ |
| * ncdevmem -s <server IP> [-c <client IP>] -p 5201 -f eth1 |
| * |
| * Note this is compatible with regular netcat. i.e. the sender or receiver can |
| * be replaced with regular netcat to test the RX or TX path in isolation. |
| * |
| * Test data validation (devmem TCP on RX only): |
| * |
| * On server: |
| * ncdevmem -s <server IP> [-c <client IP>] -f eth1 -l -p 5201 -v 7 |
| * |
| * On client: |
| * yes $(echo -e \\x01\\x02\\x03\\x04\\x05\\x06) | \ |
| * head -c 1G | \ |
| * nc <server IP> 5201 -p 5201 |
| * |
| * Test data validation (devmem TCP on RX and TX, validation happens on RX): |
| * |
| * On server: |
| * ncdevmem -s <server IP> [-c <client IP>] -l -p 5201 -v 8 -f eth1 |
| * |
| * On client: |
| * yes $(echo -e \\x01\\x02\\x03\\x04\\x05\\x06\\x07) | \ |
| * head -c 1M | \ |
| * ncdevmem -s <server IP> [-c <client IP>] -p 5201 -f eth1 |
| */ |
| #define _GNU_SOURCE |
| #define __EXPORTED_HEADERS__ |
| |
| #include <linux/uio.h> |
| #include <stdarg.h> |
| #include <stdio.h> |
| #include <stdlib.h> |
| #include <unistd.h> |
| #include <stdbool.h> |
| #include <string.h> |
| #include <errno.h> |
| #define __iovec_defined |
| #include <fcntl.h> |
| #include <malloc.h> |
| #include <error.h> |
| #include <poll.h> |
| |
| #include <arpa/inet.h> |
| #include <sys/socket.h> |
| #include <sys/mman.h> |
| #include <sys/ioctl.h> |
| #include <sys/syscall.h> |
| #include <sys/time.h> |
| |
| #include <linux/memfd.h> |
| #include <linux/dma-buf.h> |
| #include <linux/errqueue.h> |
| #include <linux/udmabuf.h> |
| #include <linux/types.h> |
| #include <linux/netlink.h> |
| #include <linux/genetlink.h> |
| #include <linux/netdev.h> |
| #include <linux/ethtool_netlink.h> |
| #include <time.h> |
| #include <net/if.h> |
| |
| #include "netdev-user.h" |
| #include "ethtool-user.h" |
| #include <ynl.h> |
| |
| #define PAGE_SHIFT 12 |
| #define TEST_PREFIX "ncdevmem" |
| #define NUM_PAGES 16000 |
| |
| #ifndef MSG_SOCK_DEVMEM |
| #define MSG_SOCK_DEVMEM 0x2000000 |
| #endif |
| |
| #define MAX_IOV 1024 |
| |
| static size_t max_chunk; |
| static char *server_ip; |
| static char *client_ip; |
| static char *port; |
| static size_t do_validation; |
| static int start_queue = -1; |
| static int num_queues = -1; |
| static char *ifname; |
| static unsigned int ifindex; |
| static unsigned int dmabuf_id; |
| static uint32_t tx_dmabuf_id; |
| static int waittime_ms = 500; |
| |
| /* System state loaded by current_config_load() */ |
| #define MAX_FLOWS 8 |
| static int ntuple_ids[MAX_FLOWS] = { -1, -1, -1, -1, -1, -1, -1, -1, }; |
| |
| struct memory_buffer { |
| int fd; |
| size_t size; |
| |
| int devfd; |
| int memfd; |
| char *buf_mem; |
| }; |
| |
| struct memory_provider { |
| struct memory_buffer *(*alloc)(size_t size); |
| void (*free)(struct memory_buffer *ctx); |
| void (*memcpy_to_device)(struct memory_buffer *dst, size_t off, |
| void *src, int n); |
| void (*memcpy_from_device)(void *dst, struct memory_buffer *src, |
| size_t off, int n); |
| }; |
| |
| static void pr_err(const char *fmt, ...) |
| { |
| va_list args; |
| |
| fprintf(stderr, "%s: ", TEST_PREFIX); |
| |
| va_start(args, fmt); |
| vfprintf(stderr, fmt, args); |
| va_end(args); |
| |
| if (errno != 0) |
| fprintf(stderr, ": %s", strerror(errno)); |
| fprintf(stderr, "\n"); |
| } |
| |
| static struct memory_buffer *udmabuf_alloc(size_t size) |
| { |
| struct udmabuf_create create; |
| struct memory_buffer *ctx; |
| int ret; |
| |
| ctx = malloc(sizeof(*ctx)); |
| if (!ctx) |
| return NULL; |
| |
| ctx->size = size; |
| |
| ctx->devfd = open("/dev/udmabuf", O_RDWR); |
| if (ctx->devfd < 0) { |
| pr_err("[skip,no-udmabuf: Unable to access DMA buffer device file]"); |
| goto err_free_ctx; |
| } |
| |
| ctx->memfd = memfd_create("udmabuf-test", MFD_ALLOW_SEALING); |
| if (ctx->memfd < 0) { |
| pr_err("[skip,no-memfd]"); |
| goto err_close_dev; |
| } |
| |
| ret = fcntl(ctx->memfd, F_ADD_SEALS, F_SEAL_SHRINK); |
| if (ret < 0) { |
| pr_err("[skip,fcntl-add-seals]"); |
| goto err_close_memfd; |
| } |
| |
| ret = ftruncate(ctx->memfd, size); |
| if (ret == -1) { |
| pr_err("[FAIL,memfd-truncate]"); |
| goto err_close_memfd; |
| } |
| |
| memset(&create, 0, sizeof(create)); |
| |
| create.memfd = ctx->memfd; |
| create.offset = 0; |
| create.size = size; |
| ctx->fd = ioctl(ctx->devfd, UDMABUF_CREATE, &create); |
| if (ctx->fd < 0) { |
| pr_err("[FAIL, create udmabuf]"); |
| goto err_close_fd; |
| } |
| |
| ctx->buf_mem = mmap(NULL, size, PROT_READ | PROT_WRITE, MAP_SHARED, |
| ctx->fd, 0); |
| if (ctx->buf_mem == MAP_FAILED) { |
| pr_err("[FAIL, map udmabuf]"); |
| goto err_close_fd; |
| } |
| |
| return ctx; |
| |
| err_close_fd: |
| close(ctx->fd); |
| err_close_memfd: |
| close(ctx->memfd); |
| err_close_dev: |
| close(ctx->devfd); |
| err_free_ctx: |
| free(ctx); |
| return NULL; |
| } |
| |
| static void udmabuf_free(struct memory_buffer *ctx) |
| { |
| munmap(ctx->buf_mem, ctx->size); |
| close(ctx->fd); |
| close(ctx->memfd); |
| close(ctx->devfd); |
| free(ctx); |
| } |
| |
| static void udmabuf_memcpy_to_device(struct memory_buffer *dst, size_t off, |
| void *src, int n) |
| { |
| struct dma_buf_sync sync = {}; |
| |
| sync.flags = DMA_BUF_SYNC_START | DMA_BUF_SYNC_WRITE; |
| ioctl(dst->fd, DMA_BUF_IOCTL_SYNC, &sync); |
| |
| memcpy(dst->buf_mem + off, src, n); |
| |
| sync.flags = DMA_BUF_SYNC_END | DMA_BUF_SYNC_WRITE; |
| ioctl(dst->fd, DMA_BUF_IOCTL_SYNC, &sync); |
| } |
| |
| static void udmabuf_memcpy_from_device(void *dst, struct memory_buffer *src, |
| size_t off, int n) |
| { |
| struct dma_buf_sync sync = {}; |
| |
| sync.flags = DMA_BUF_SYNC_START; |
| ioctl(src->fd, DMA_BUF_IOCTL_SYNC, &sync); |
| |
| memcpy(dst, src->buf_mem + off, n); |
| |
| sync.flags = DMA_BUF_SYNC_END; |
| ioctl(src->fd, DMA_BUF_IOCTL_SYNC, &sync); |
| } |
| |
| static struct memory_provider udmabuf_memory_provider = { |
| .alloc = udmabuf_alloc, |
| .free = udmabuf_free, |
| .memcpy_to_device = udmabuf_memcpy_to_device, |
| .memcpy_from_device = udmabuf_memcpy_from_device, |
| }; |
| |
| static struct memory_provider *provider = &udmabuf_memory_provider; |
| |
| static void print_nonzero_bytes(void *ptr, size_t size) |
| { |
| unsigned char *p = ptr; |
| unsigned int i; |
| |
| for (i = 0; i < size; i++) |
| putchar(p[i]); |
| } |
| |
| int validate_buffer(void *line, size_t size) |
| { |
| static unsigned char seed = 1; |
| unsigned char *ptr = line; |
| unsigned char expected; |
| static int errors; |
| size_t i; |
| |
| for (i = 0; i < size; i++) { |
| expected = seed ? seed : '\n'; |
| if (ptr[i] != expected) { |
| fprintf(stderr, |
| "Failed validation: expected=%u, actual=%u, index=%lu\n", |
| expected, ptr[i], i); |
| errors++; |
| if (errors > 20) { |
| pr_err("validation failed"); |
| return -1; |
| } |
| } |
| seed++; |
| if (seed == do_validation) |
| seed = 0; |
| } |
| |
| fprintf(stdout, "Validated buffer\n"); |
| return 0; |
| } |
| |
| static int |
| __run_command(char *out, size_t outlen, const char *cmd, va_list args) |
| { |
| char command[256]; |
| FILE *fp; |
| |
| vsnprintf(command, sizeof(command), cmd, args); |
| |
| fprintf(stderr, "Running: %s\n", command); |
| fp = popen(command, "r"); |
| if (!fp) |
| return -1; |
| if (out) { |
| size_t len; |
| |
| if (!fgets(out, outlen, fp)) |
| return -1; |
| |
| /* Remove trailing newline if present */ |
| len = strlen(out); |
| if (len && out[len - 1] == '\n') |
| out[len - 1] = '\0'; |
| } |
| return pclose(fp); |
| } |
| |
| static int run_command(const char *cmd, ...) |
| { |
| va_list args; |
| int ret; |
| |
| va_start(args, cmd); |
| ret = __run_command(NULL, 0, cmd, args); |
| va_end(args); |
| |
| return ret; |
| } |
| |
| static int ethtool_add_flow(const char *format, ...) |
| { |
| char local_output[256], cmd[256]; |
| const char *id_start; |
| int flow_idx, ret; |
| char *endptr; |
| long flow_id; |
| va_list args; |
| |
| for (flow_idx = 0; flow_idx < MAX_FLOWS; flow_idx++) |
| if (ntuple_ids[flow_idx] == -1) |
| break; |
| if (flow_idx == MAX_FLOWS) { |
| fprintf(stderr, "Error: too many flows\n"); |
| return -1; |
| } |
| |
| snprintf(cmd, sizeof(cmd), "ethtool -N %s %s", ifname, format); |
| |
| va_start(args, format); |
| ret = __run_command(local_output, sizeof(local_output), cmd, args); |
| va_end(args); |
| |
| if (ret != 0) |
| return ret; |
| |
| /* Extract the ID from the output */ |
| id_start = strstr(local_output, "Added rule with ID "); |
| if (!id_start) |
| return -1; |
| id_start += strlen("Added rule with ID "); |
| |
| flow_id = strtol(id_start, &endptr, 10); |
| if (endptr == id_start || flow_id < 0 || flow_id > INT_MAX) |
| return -1; |
| |
| fprintf(stderr, "Added flow rule with ID %ld\n", flow_id); |
| ntuple_ids[flow_idx] = flow_id; |
| return flow_id; |
| } |
| |
| static int rxq_num(int ifindex) |
| { |
| struct ethtool_channels_get_req *req; |
| struct ethtool_channels_get_rsp *rsp; |
| struct ynl_error yerr; |
| struct ynl_sock *ys; |
| int num = -1; |
| |
| ys = ynl_sock_create(&ynl_ethtool_family, &yerr); |
| if (!ys) { |
| fprintf(stderr, "YNL: %s\n", yerr.msg); |
| return -1; |
| } |
| |
| req = ethtool_channels_get_req_alloc(); |
| ethtool_channels_get_req_set_header_dev_index(req, ifindex); |
| rsp = ethtool_channels_get(ys, req); |
| if (rsp) |
| num = rsp->rx_count + rsp->combined_count; |
| ethtool_channels_get_req_free(req); |
| ethtool_channels_get_rsp_free(rsp); |
| |
| ynl_sock_destroy(ys); |
| |
| return num; |
| } |
| |
| static void reset_flow_steering(void) |
| { |
| int i; |
| |
| for (i = 0; i < MAX_FLOWS; i++) { |
| if (ntuple_ids[i] == -1) |
| continue; |
| run_command("ethtool -N %s delete %d", |
| ifname, ntuple_ids[i]); |
| ntuple_ids[i] = -1; |
| } |
| } |
| |
| static const char *tcp_data_split_str(int val) |
| { |
| switch (val) { |
| case 0: |
| return "off"; |
| case 1: |
| return "auto"; |
| case 2: |
| return "on"; |
| default: |
| return "?"; |
| } |
| } |
| |
| static struct ethtool_rings_get_rsp *get_ring_config(void) |
| { |
| struct ethtool_rings_get_req *get_req; |
| struct ethtool_rings_get_rsp *get_rsp; |
| struct ynl_error yerr; |
| struct ynl_sock *ys; |
| |
| ys = ynl_sock_create(&ynl_ethtool_family, &yerr); |
| if (!ys) { |
| fprintf(stderr, "YNL: %s\n", yerr.msg); |
| return NULL; |
| } |
| |
| get_req = ethtool_rings_get_req_alloc(); |
| ethtool_rings_get_req_set_header_dev_index(get_req, ifindex); |
| get_rsp = ethtool_rings_get(ys, get_req); |
| ethtool_rings_get_req_free(get_req); |
| |
| ynl_sock_destroy(ys); |
| |
| return get_rsp; |
| } |
| |
| static void restore_ring_config(const struct ethtool_rings_get_rsp *config) |
| { |
| struct ethtool_rings_get_req *get_req; |
| struct ethtool_rings_get_rsp *get_rsp; |
| struct ethtool_rings_set_req *req; |
| struct ynl_error yerr; |
| struct ynl_sock *ys; |
| int ret; |
| |
| if (!config) |
| return; |
| |
| ys = ynl_sock_create(&ynl_ethtool_family, &yerr); |
| if (!ys) { |
| fprintf(stderr, "YNL: %s\n", yerr.msg); |
| return; |
| } |
| |
| req = ethtool_rings_set_req_alloc(); |
| ethtool_rings_set_req_set_header_dev_index(req, ifindex); |
| ethtool_rings_set_req_set_tcp_data_split(req, |
| ETHTOOL_TCP_DATA_SPLIT_UNKNOWN); |
| if (config->_present.hds_thresh) |
| ethtool_rings_set_req_set_hds_thresh(req, config->hds_thresh); |
| |
| ret = ethtool_rings_set(ys, req); |
| if (ret < 0) |
| fprintf(stderr, "YNL restoring HDS cfg: %s\n", ys->err.msg); |
| |
| get_req = ethtool_rings_get_req_alloc(); |
| ethtool_rings_get_req_set_header_dev_index(get_req, ifindex); |
| get_rsp = ethtool_rings_get(ys, get_req); |
| ethtool_rings_get_req_free(get_req); |
| |
| /* use explicit value if UKNOWN didn't give us the previous */ |
| if (get_rsp->tcp_data_split != config->tcp_data_split) { |
| ethtool_rings_set_req_set_tcp_data_split(req, |
| config->tcp_data_split); |
| ret = ethtool_rings_set(ys, req); |
| if (ret < 0) |
| fprintf(stderr, "YNL restoring expl HDS cfg: %s\n", |
| ys->err.msg); |
| } |
| |
| ethtool_rings_get_rsp_free(get_rsp); |
| ethtool_rings_set_req_free(req); |
| |
| ynl_sock_destroy(ys); |
| } |
| |
| static int |
| configure_headersplit(const struct ethtool_rings_get_rsp *old, bool on) |
| { |
| struct ethtool_rings_get_req *get_req; |
| struct ethtool_rings_get_rsp *get_rsp; |
| struct ethtool_rings_set_req *req; |
| struct ynl_error yerr; |
| struct ynl_sock *ys; |
| int ret; |
| |
| ys = ynl_sock_create(&ynl_ethtool_family, &yerr); |
| if (!ys) { |
| fprintf(stderr, "YNL: %s\n", yerr.msg); |
| return -1; |
| } |
| |
| req = ethtool_rings_set_req_alloc(); |
| ethtool_rings_set_req_set_header_dev_index(req, ifindex); |
| if (on) { |
| ethtool_rings_set_req_set_tcp_data_split(req, |
| ETHTOOL_TCP_DATA_SPLIT_ENABLED); |
| if (old->_present.hds_thresh) |
| ethtool_rings_set_req_set_hds_thresh(req, 0); |
| } else { |
| ethtool_rings_set_req_set_tcp_data_split(req, |
| ETHTOOL_TCP_DATA_SPLIT_UNKNOWN); |
| } |
| ret = ethtool_rings_set(ys, req); |
| if (ret < 0) |
| fprintf(stderr, "YNL failed: %s\n", ys->err.msg); |
| ethtool_rings_set_req_free(req); |
| |
| if (ret == 0) { |
| get_req = ethtool_rings_get_req_alloc(); |
| ethtool_rings_get_req_set_header_dev_index(get_req, ifindex); |
| get_rsp = ethtool_rings_get(ys, get_req); |
| ethtool_rings_get_req_free(get_req); |
| if (get_rsp) |
| fprintf(stderr, "TCP header split: %s\n", |
| tcp_data_split_str(get_rsp->tcp_data_split)); |
| ethtool_rings_get_rsp_free(get_rsp); |
| } |
| |
| ynl_sock_destroy(ys); |
| |
| return ret; |
| } |
| |
| static int configure_rss(void) |
| { |
| return run_command("ethtool -X %s equal %d >&2", ifname, start_queue); |
| } |
| |
| static void reset_rss(void) |
| { |
| run_command("ethtool -X %s default >&2", ifname, start_queue); |
| } |
| |
| static int check_changing_channels(unsigned int rx, unsigned int tx) |
| { |
| struct ethtool_channels_get_req *gchan; |
| struct ethtool_channels_set_req *schan; |
| struct ethtool_channels_get_rsp *chan; |
| struct ynl_error yerr; |
| struct ynl_sock *ys; |
| int ret; |
| |
| fprintf(stderr, "setting channel count rx:%u tx:%u\n", rx, tx); |
| |
| ys = ynl_sock_create(&ynl_ethtool_family, &yerr); |
| if (!ys) { |
| fprintf(stderr, "YNL: %s\n", yerr.msg); |
| return -1; |
| } |
| |
| gchan = ethtool_channels_get_req_alloc(); |
| if (!gchan) { |
| ret = -1; |
| goto exit_close_sock; |
| } |
| |
| ethtool_channels_get_req_set_header_dev_index(gchan, ifindex); |
| chan = ethtool_channels_get(ys, gchan); |
| ethtool_channels_get_req_free(gchan); |
| if (!chan) { |
| fprintf(stderr, "YNL get channels: %s\n", ys->err.msg); |
| ret = -1; |
| goto exit_close_sock; |
| } |
| |
| schan = ethtool_channels_set_req_alloc(); |
| if (!schan) { |
| ret = -1; |
| goto exit_free_chan; |
| } |
| |
| ethtool_channels_set_req_set_header_dev_index(schan, ifindex); |
| |
| if (chan->_present.combined_count) { |
| if (chan->_present.rx_count || chan->_present.tx_count) { |
| ethtool_channels_set_req_set_rx_count(schan, 0); |
| ethtool_channels_set_req_set_tx_count(schan, 0); |
| } |
| |
| if (rx == tx) { |
| ethtool_channels_set_req_set_combined_count(schan, rx); |
| } else if (rx > tx) { |
| ethtool_channels_set_req_set_combined_count(schan, tx); |
| ethtool_channels_set_req_set_rx_count(schan, rx - tx); |
| } else { |
| ethtool_channels_set_req_set_combined_count(schan, rx); |
| ethtool_channels_set_req_set_tx_count(schan, tx - rx); |
| } |
| |
| } else if (chan->_present.rx_count) { |
| ethtool_channels_set_req_set_rx_count(schan, rx); |
| ethtool_channels_set_req_set_tx_count(schan, tx); |
| } else { |
| fprintf(stderr, "Error: device has neither combined nor rx channels\n"); |
| ret = -1; |
| goto exit_free_schan; |
| } |
| |
| ret = ethtool_channels_set(ys, schan); |
| if (ret) { |
| fprintf(stderr, "YNL set channels: %s\n", ys->err.msg); |
| } else { |
| /* We were expecting a failure, go back to previous settings */ |
| ethtool_channels_set_req_set_combined_count(schan, |
| chan->combined_count); |
| ethtool_channels_set_req_set_rx_count(schan, chan->rx_count); |
| ethtool_channels_set_req_set_tx_count(schan, chan->tx_count); |
| |
| ret = ethtool_channels_set(ys, schan); |
| if (ret) |
| fprintf(stderr, "YNL un-setting channels: %s\n", |
| ys->err.msg); |
| } |
| |
| exit_free_schan: |
| ethtool_channels_set_req_free(schan); |
| exit_free_chan: |
| ethtool_channels_get_rsp_free(chan); |
| exit_close_sock: |
| ynl_sock_destroy(ys); |
| |
| return ret; |
| } |
| |
| static int configure_flow_steering(struct sockaddr_in6 *server_sin) |
| { |
| const char *type = "tcp6"; |
| const char *server_addr; |
| char buf[40]; |
| int flow_id; |
| |
| inet_ntop(AF_INET6, &server_sin->sin6_addr, buf, sizeof(buf)); |
| server_addr = buf; |
| |
| if (IN6_IS_ADDR_V4MAPPED(&server_sin->sin6_addr)) { |
| type = "tcp4"; |
| server_addr = strrchr(server_addr, ':') + 1; |
| } |
| |
| /* Try configure 5-tuple */ |
| flow_id = ethtool_add_flow("flow-type %s %s %s dst-ip %s %s %s dst-port %s queue %d", |
| type, |
| client_ip ? "src-ip" : "", |
| client_ip ?: "", |
| server_addr, |
| client_ip ? "src-port" : "", |
| client_ip ? port : "", |
| port, start_queue); |
| if (flow_id < 0) { |
| /* If that fails, try configure 3-tuple */ |
| flow_id = ethtool_add_flow("flow-type %s dst-ip %s dst-port %s queue %d", |
| type, server_addr, port, start_queue); |
| if (flow_id < 0) |
| /* If that fails, return error */ |
| return -1; |
| } |
| |
| return 0; |
| } |
| |
| static int bind_rx_queue(unsigned int ifindex, unsigned int dmabuf_fd, |
| struct netdev_queue_id *queues, |
| unsigned int n_queue_index, struct ynl_sock **ys) |
| { |
| struct netdev_bind_rx_req *req = NULL; |
| struct netdev_bind_rx_rsp *rsp = NULL; |
| struct ynl_error yerr; |
| |
| *ys = ynl_sock_create(&ynl_netdev_family, &yerr); |
| if (!*ys) { |
| netdev_queue_id_free(queues); |
| fprintf(stderr, "YNL: %s\n", yerr.msg); |
| return -1; |
| } |
| |
| req = netdev_bind_rx_req_alloc(); |
| netdev_bind_rx_req_set_ifindex(req, ifindex); |
| netdev_bind_rx_req_set_fd(req, dmabuf_fd); |
| __netdev_bind_rx_req_set_queues(req, queues, n_queue_index); |
| |
| rsp = netdev_bind_rx(*ys, req); |
| if (!rsp) { |
| perror("netdev_bind_rx"); |
| goto err_close; |
| } |
| |
| if (!rsp->_present.id) { |
| perror("id not present"); |
| goto err_close; |
| } |
| |
| fprintf(stderr, "got dmabuf id=%d\n", rsp->id); |
| dmabuf_id = rsp->id; |
| |
| netdev_bind_rx_req_free(req); |
| netdev_bind_rx_rsp_free(rsp); |
| |
| return 0; |
| |
| err_close: |
| fprintf(stderr, "YNL failed: %s\n", (*ys)->err.msg); |
| netdev_bind_rx_req_free(req); |
| ynl_sock_destroy(*ys); |
| return -1; |
| } |
| |
| static int bind_tx_queue(unsigned int ifindex, unsigned int dmabuf_fd, |
| struct ynl_sock **ys) |
| { |
| struct netdev_bind_tx_req *req = NULL; |
| struct netdev_bind_tx_rsp *rsp = NULL; |
| struct ynl_error yerr; |
| |
| *ys = ynl_sock_create(&ynl_netdev_family, &yerr); |
| if (!*ys) { |
| fprintf(stderr, "YNL: %s\n", yerr.msg); |
| return -1; |
| } |
| |
| req = netdev_bind_tx_req_alloc(); |
| netdev_bind_tx_req_set_ifindex(req, ifindex); |
| netdev_bind_tx_req_set_fd(req, dmabuf_fd); |
| |
| rsp = netdev_bind_tx(*ys, req); |
| if (!rsp) { |
| perror("netdev_bind_tx"); |
| goto err_close; |
| } |
| |
| if (!rsp->_present.id) { |
| perror("id not present"); |
| goto err_close; |
| } |
| |
| fprintf(stderr, "got tx dmabuf id=%d\n", rsp->id); |
| tx_dmabuf_id = rsp->id; |
| |
| netdev_bind_tx_req_free(req); |
| netdev_bind_tx_rsp_free(rsp); |
| |
| return 0; |
| |
| err_close: |
| fprintf(stderr, "YNL failed: %s\n", (*ys)->err.msg); |
| netdev_bind_tx_req_free(req); |
| ynl_sock_destroy(*ys); |
| return -1; |
| } |
| |
| static int enable_reuseaddr(int fd) |
| { |
| int opt = 1; |
| int ret; |
| |
| ret = setsockopt(fd, SOL_SOCKET, SO_REUSEPORT, &opt, sizeof(opt)); |
| if (ret) { |
| pr_err("SO_REUSEPORT failed"); |
| return -1; |
| } |
| |
| ret = setsockopt(fd, SOL_SOCKET, SO_REUSEADDR, &opt, sizeof(opt)); |
| if (ret) { |
| pr_err("SO_REUSEADDR failed"); |
| return -1; |
| } |
| |
| return 0; |
| } |
| |
| static int parse_address(const char *str, int port, struct sockaddr_in6 *sin6) |
| { |
| int ret; |
| |
| sin6->sin6_family = AF_INET6; |
| sin6->sin6_port = htons(port); |
| |
| ret = inet_pton(sin6->sin6_family, str, &sin6->sin6_addr); |
| if (ret != 1) { |
| /* fallback to plain IPv4 */ |
| ret = inet_pton(AF_INET, str, &sin6->sin6_addr.s6_addr32[3]); |
| if (ret != 1) |
| return -1; |
| |
| /* add ::ffff prefix */ |
| sin6->sin6_addr.s6_addr32[0] = 0; |
| sin6->sin6_addr.s6_addr32[1] = 0; |
| sin6->sin6_addr.s6_addr16[4] = 0; |
| sin6->sin6_addr.s6_addr16[5] = 0xffff; |
| } |
| |
| return 0; |
| } |
| |
| static struct netdev_queue_id *create_queues(void) |
| { |
| struct netdev_queue_id *queues; |
| size_t i = 0; |
| |
| queues = netdev_queue_id_alloc(num_queues); |
| for (i = 0; i < num_queues; i++) { |
| netdev_queue_id_set_type(&queues[i], NETDEV_QUEUE_TYPE_RX); |
| netdev_queue_id_set_id(&queues[i], start_queue + i); |
| } |
| |
| return queues; |
| } |
| |
| static int do_server(struct memory_buffer *mem) |
| { |
| struct ethtool_rings_get_rsp *ring_config; |
| char ctrl_data[sizeof(int) * 20000]; |
| size_t non_page_aligned_frags = 0; |
| struct sockaddr_in6 client_addr; |
| struct sockaddr_in6 server_sin; |
| size_t page_aligned_frags = 0; |
| size_t total_received = 0; |
| socklen_t client_addr_len; |
| bool is_devmem = false; |
| char *tmp_mem = NULL; |
| struct ynl_sock *ys; |
| char iobuf[819200]; |
| int ret, err = -1; |
| char buffer[256]; |
| int socket_fd; |
| int client_fd; |
| |
| ret = parse_address(server_ip, atoi(port), &server_sin); |
| if (ret < 0) { |
| pr_err("parse server address"); |
| return -1; |
| } |
| |
| ring_config = get_ring_config(); |
| if (!ring_config) { |
| pr_err("Failed to get current ring configuration"); |
| return -1; |
| } |
| |
| if (configure_headersplit(ring_config, 1)) { |
| pr_err("Failed to enable TCP header split"); |
| goto err_free_ring_config; |
| } |
| |
| /* Configure RSS to divert all traffic from our devmem queues */ |
| if (configure_rss()) { |
| pr_err("Failed to configure rss"); |
| goto err_reset_headersplit; |
| } |
| |
| /* Flow steer our devmem flows to start_queue */ |
| if (configure_flow_steering(&server_sin)) { |
| pr_err("Failed to configure flow steering"); |
| goto err_reset_rss; |
| } |
| |
| if (bind_rx_queue(ifindex, mem->fd, create_queues(), num_queues, &ys)) { |
| pr_err("Failed to bind"); |
| goto err_reset_flow_steering; |
| } |
| |
| tmp_mem = malloc(mem->size); |
| if (!tmp_mem) |
| goto err_unbind; |
| |
| socket_fd = socket(AF_INET6, SOCK_STREAM, 0); |
| if (socket_fd < 0) { |
| pr_err("Failed to create socket"); |
| goto err_free_tmp; |
| } |
| |
| if (enable_reuseaddr(socket_fd)) |
| goto err_close_socket; |
| |
| fprintf(stderr, "binding to address %s:%d\n", server_ip, |
| ntohs(server_sin.sin6_port)); |
| |
| ret = bind(socket_fd, &server_sin, sizeof(server_sin)); |
| if (ret) { |
| pr_err("Failed to bind"); |
| goto err_close_socket; |
| } |
| |
| ret = listen(socket_fd, 1); |
| if (ret) { |
| pr_err("Failed to listen"); |
| goto err_close_socket; |
| } |
| |
| client_addr_len = sizeof(client_addr); |
| |
| inet_ntop(AF_INET6, &server_sin.sin6_addr, buffer, |
| sizeof(buffer)); |
| fprintf(stderr, "Waiting or connection on %s:%d\n", buffer, |
| ntohs(server_sin.sin6_port)); |
| client_fd = accept(socket_fd, &client_addr, &client_addr_len); |
| if (client_fd < 0) { |
| pr_err("Failed to accept"); |
| goto err_close_socket; |
| } |
| |
| inet_ntop(AF_INET6, &client_addr.sin6_addr, buffer, |
| sizeof(buffer)); |
| fprintf(stderr, "Got connection from %s:%d\n", buffer, |
| ntohs(client_addr.sin6_port)); |
| |
| while (1) { |
| struct iovec iov = { .iov_base = iobuf, |
| .iov_len = sizeof(iobuf) }; |
| struct dmabuf_cmsg *dmabuf_cmsg = NULL; |
| struct cmsghdr *cm = NULL; |
| struct msghdr msg = { 0 }; |
| struct dmabuf_token token; |
| ssize_t ret; |
| |
| is_devmem = false; |
| |
| msg.msg_iov = &iov; |
| msg.msg_iovlen = 1; |
| msg.msg_control = ctrl_data; |
| msg.msg_controllen = sizeof(ctrl_data); |
| ret = recvmsg(client_fd, &msg, MSG_SOCK_DEVMEM); |
| fprintf(stderr, "recvmsg ret=%ld\n", ret); |
| if (ret < 0 && (errno == EAGAIN || errno == EWOULDBLOCK)) |
| continue; |
| if (ret < 0) { |
| perror("recvmsg"); |
| if (errno == EFAULT) { |
| pr_err("received EFAULT, won't recover"); |
| goto err_close_client; |
| } |
| continue; |
| } |
| if (ret == 0) { |
| errno = 0; |
| pr_err("client exited"); |
| goto cleanup; |
| } |
| |
| for (cm = CMSG_FIRSTHDR(&msg); cm; cm = CMSG_NXTHDR(&msg, cm)) { |
| if (cm->cmsg_level != SOL_SOCKET || |
| (cm->cmsg_type != SCM_DEVMEM_DMABUF && |
| cm->cmsg_type != SCM_DEVMEM_LINEAR)) { |
| fprintf(stderr, "skipping non-devmem cmsg\n"); |
| continue; |
| } |
| |
| dmabuf_cmsg = (struct dmabuf_cmsg *)CMSG_DATA(cm); |
| is_devmem = true; |
| |
| if (cm->cmsg_type == SCM_DEVMEM_LINEAR) { |
| /* TODO: process data copied from skb's linear |
| * buffer. |
| */ |
| fprintf(stderr, |
| "SCM_DEVMEM_LINEAR. dmabuf_cmsg->frag_size=%u\n", |
| dmabuf_cmsg->frag_size); |
| |
| continue; |
| } |
| |
| token.token_start = dmabuf_cmsg->frag_token; |
| token.token_count = 1; |
| |
| total_received += dmabuf_cmsg->frag_size; |
| fprintf(stderr, |
| "received frag_page=%llu, in_page_offset=%llu, frag_offset=%llu, frag_size=%u, token=%u, total_received=%lu, dmabuf_id=%u\n", |
| dmabuf_cmsg->frag_offset >> PAGE_SHIFT, |
| dmabuf_cmsg->frag_offset % getpagesize(), |
| dmabuf_cmsg->frag_offset, |
| dmabuf_cmsg->frag_size, dmabuf_cmsg->frag_token, |
| total_received, dmabuf_cmsg->dmabuf_id); |
| |
| if (dmabuf_cmsg->dmabuf_id != dmabuf_id) { |
| pr_err("received on wrong dmabuf_id: flow steering error"); |
| goto err_close_client; |
| } |
| |
| if (dmabuf_cmsg->frag_size % getpagesize()) |
| non_page_aligned_frags++; |
| else |
| page_aligned_frags++; |
| |
| provider->memcpy_from_device(tmp_mem, mem, |
| dmabuf_cmsg->frag_offset, |
| dmabuf_cmsg->frag_size); |
| |
| if (do_validation) { |
| if (validate_buffer(tmp_mem, |
| dmabuf_cmsg->frag_size)) |
| goto err_close_client; |
| } else { |
| print_nonzero_bytes(tmp_mem, |
| dmabuf_cmsg->frag_size); |
| } |
| |
| ret = setsockopt(client_fd, SOL_SOCKET, |
| SO_DEVMEM_DONTNEED, &token, |
| sizeof(token)); |
| if (ret != 1) { |
| pr_err("SO_DEVMEM_DONTNEED not enough tokens"); |
| goto err_close_client; |
| } |
| } |
| if (!is_devmem) { |
| pr_err("flow steering error"); |
| goto err_close_client; |
| } |
| |
| fprintf(stderr, "total_received=%lu\n", total_received); |
| } |
| |
| fprintf(stderr, "%s: ok\n", TEST_PREFIX); |
| |
| fprintf(stderr, "page_aligned_frags=%lu, non_page_aligned_frags=%lu\n", |
| page_aligned_frags, non_page_aligned_frags); |
| |
| cleanup: |
| err = 0; |
| |
| err_close_client: |
| close(client_fd); |
| err_close_socket: |
| close(socket_fd); |
| err_free_tmp: |
| free(tmp_mem); |
| err_unbind: |
| ynl_sock_destroy(ys); |
| err_reset_flow_steering: |
| reset_flow_steering(); |
| err_reset_rss: |
| reset_rss(); |
| err_reset_headersplit: |
| restore_ring_config(ring_config); |
| err_free_ring_config: |
| ethtool_rings_get_rsp_free(ring_config); |
| return err; |
| } |
| |
| int run_devmem_tests(void) |
| { |
| struct ethtool_rings_get_rsp *ring_config; |
| struct netdev_queue_id *queues; |
| struct memory_buffer *mem; |
| struct ynl_sock *ys; |
| int err = -1; |
| |
| mem = provider->alloc(getpagesize() * NUM_PAGES); |
| if (!mem) { |
| pr_err("Failed to allocate memory buffer"); |
| return -1; |
| } |
| |
| ring_config = get_ring_config(); |
| if (!ring_config) { |
| pr_err("Failed to get current ring configuration"); |
| goto err_free_mem; |
| } |
| |
| /* Configure RSS to divert all traffic from our devmem queues */ |
| if (configure_rss()) { |
| pr_err("rss error"); |
| goto err_free_ring_config; |
| } |
| |
| if (configure_headersplit(ring_config, 1)) { |
| pr_err("Failed to configure header split"); |
| goto err_reset_rss; |
| } |
| |
| queues = netdev_queue_id_alloc(num_queues); |
| if (!queues) { |
| pr_err("Failed to allocate empty queues array"); |
| goto err_reset_headersplit; |
| } |
| |
| if (!bind_rx_queue(ifindex, mem->fd, queues, num_queues, &ys)) { |
| pr_err("Binding empty queues array should have failed"); |
| goto err_unbind; |
| } |
| |
| if (configure_headersplit(ring_config, 0)) { |
| pr_err("Failed to configure header split"); |
| goto err_reset_headersplit; |
| } |
| |
| queues = create_queues(); |
| if (!queues) { |
| pr_err("Failed to create queues"); |
| goto err_reset_headersplit; |
| } |
| |
| if (!bind_rx_queue(ifindex, mem->fd, queues, num_queues, &ys)) { |
| pr_err("Configure dmabuf with header split off should have failed"); |
| goto err_unbind; |
| } |
| |
| if (configure_headersplit(ring_config, 1)) { |
| pr_err("Failed to configure header split"); |
| goto err_reset_headersplit; |
| } |
| |
| queues = create_queues(); |
| if (!queues) { |
| pr_err("Failed to create queues"); |
| goto err_reset_headersplit; |
| } |
| |
| if (bind_rx_queue(ifindex, mem->fd, queues, num_queues, &ys)) { |
| pr_err("Failed to bind"); |
| goto err_reset_headersplit; |
| } |
| |
| /* Deactivating a bound queue should not be legal */ |
| if (!check_changing_channels(num_queues, num_queues)) { |
| pr_err("Deactivating a bound queue should be illegal"); |
| goto err_unbind; |
| } |
| |
| err = 0; |
| goto err_unbind; |
| |
| err_unbind: |
| ynl_sock_destroy(ys); |
| err_reset_headersplit: |
| restore_ring_config(ring_config); |
| err_reset_rss: |
| reset_rss(); |
| err_free_ring_config: |
| ethtool_rings_get_rsp_free(ring_config); |
| err_free_mem: |
| provider->free(mem); |
| return err; |
| } |
| |
| static uint64_t gettimeofday_ms(void) |
| { |
| struct timeval tv; |
| |
| gettimeofday(&tv, NULL); |
| return (tv.tv_sec * 1000ULL) + (tv.tv_usec / 1000ULL); |
| } |
| |
| static int do_poll(int fd) |
| { |
| struct pollfd pfd; |
| int ret; |
| |
| pfd.revents = 0; |
| pfd.fd = fd; |
| |
| ret = poll(&pfd, 1, waittime_ms); |
| if (ret == -1) { |
| pr_err("poll"); |
| return -1; |
| } |
| |
| return ret && (pfd.revents & POLLERR); |
| } |
| |
| static int wait_compl(int fd) |
| { |
| int64_t tstop = gettimeofday_ms() + waittime_ms; |
| char control[CMSG_SPACE(100)] = {}; |
| struct sock_extended_err *serr; |
| struct msghdr msg = {}; |
| struct cmsghdr *cm; |
| __u32 hi, lo; |
| int ret; |
| |
| msg.msg_control = control; |
| msg.msg_controllen = sizeof(control); |
| |
| while (gettimeofday_ms() < tstop) { |
| ret = do_poll(fd); |
| if (ret < 0) |
| return ret; |
| if (!ret) |
| continue; |
| |
| ret = recvmsg(fd, &msg, MSG_ERRQUEUE); |
| if (ret < 0) { |
| if (errno == EAGAIN) |
| continue; |
| pr_err("recvmsg(MSG_ERRQUEUE)"); |
| return -1; |
| } |
| if (msg.msg_flags & MSG_CTRUNC) { |
| pr_err("MSG_CTRUNC"); |
| return -1; |
| } |
| |
| for (cm = CMSG_FIRSTHDR(&msg); cm; cm = CMSG_NXTHDR(&msg, cm)) { |
| if (cm->cmsg_level != SOL_IP && |
| cm->cmsg_level != SOL_IPV6) |
| continue; |
| if (cm->cmsg_level == SOL_IP && |
| cm->cmsg_type != IP_RECVERR) |
| continue; |
| if (cm->cmsg_level == SOL_IPV6 && |
| cm->cmsg_type != IPV6_RECVERR) |
| continue; |
| |
| serr = (void *)CMSG_DATA(cm); |
| if (serr->ee_origin != SO_EE_ORIGIN_ZEROCOPY) { |
| pr_err("wrong origin %u", serr->ee_origin); |
| return -1; |
| } |
| if (serr->ee_errno != 0) { |
| pr_err("wrong errno %d", serr->ee_errno); |
| return -1; |
| } |
| |
| hi = serr->ee_data; |
| lo = serr->ee_info; |
| |
| fprintf(stderr, "tx complete [%d,%d]\n", lo, hi); |
| return 0; |
| } |
| } |
| |
| pr_err("did not receive tx completion"); |
| return -1; |
| } |
| |
| static int do_client(struct memory_buffer *mem) |
| { |
| char ctrl_data[CMSG_SPACE(sizeof(__u32))]; |
| struct sockaddr_in6 server_sin; |
| struct sockaddr_in6 client_sin; |
| struct ynl_sock *ys = NULL; |
| struct iovec iov[MAX_IOV]; |
| struct msghdr msg = {}; |
| ssize_t line_size = 0; |
| struct cmsghdr *cmsg; |
| char *line = NULL; |
| int ret, err = -1; |
| size_t len = 0; |
| int socket_fd; |
| __u32 ddmabuf; |
| int opt = 1; |
| |
| ret = parse_address(server_ip, atoi(port), &server_sin); |
| if (ret < 0) { |
| pr_err("parse server address"); |
| return -1; |
| } |
| |
| if (client_ip) { |
| ret = parse_address(client_ip, atoi(port), &client_sin); |
| if (ret < 0) { |
| pr_err("parse client address"); |
| return ret; |
| } |
| } |
| |
| socket_fd = socket(AF_INET6, SOCK_STREAM, 0); |
| if (socket_fd < 0) { |
| pr_err("create socket"); |
| return -1; |
| } |
| |
| if (enable_reuseaddr(socket_fd)) |
| goto err_close_socket; |
| |
| ret = setsockopt(socket_fd, SOL_SOCKET, SO_BINDTODEVICE, ifname, |
| strlen(ifname) + 1); |
| if (ret) { |
| pr_err("bindtodevice"); |
| goto err_close_socket; |
| } |
| |
| if (bind_tx_queue(ifindex, mem->fd, &ys)) { |
| pr_err("Failed to bind"); |
| goto err_close_socket; |
| } |
| |
| if (client_ip) { |
| ret = bind(socket_fd, &client_sin, sizeof(client_sin)); |
| if (ret) { |
| pr_err("bind"); |
| goto err_unbind; |
| } |
| } |
| |
| ret = setsockopt(socket_fd, SOL_SOCKET, SO_ZEROCOPY, &opt, sizeof(opt)); |
| if (ret) { |
| pr_err("set sock opt"); |
| goto err_unbind; |
| } |
| |
| fprintf(stderr, "Connect to %s %d (via %s)\n", server_ip, |
| ntohs(server_sin.sin6_port), ifname); |
| |
| ret = connect(socket_fd, &server_sin, sizeof(server_sin)); |
| if (ret) { |
| pr_err("connect"); |
| goto err_unbind; |
| } |
| |
| while (1) { |
| free(line); |
| line = NULL; |
| line_size = getline(&line, &len, stdin); |
| |
| if (line_size < 0) |
| break; |
| |
| if (max_chunk) { |
| msg.msg_iovlen = |
| (line_size + max_chunk - 1) / max_chunk; |
| if (msg.msg_iovlen > MAX_IOV) { |
| pr_err("can't partition %zd bytes into maximum of %d chunks", |
| line_size, MAX_IOV); |
| goto err_free_line; |
| } |
| |
| for (int i = 0; i < msg.msg_iovlen; i++) { |
| iov[i].iov_base = (void *)(i * max_chunk); |
| iov[i].iov_len = max_chunk; |
| } |
| |
| iov[msg.msg_iovlen - 1].iov_len = |
| line_size - (msg.msg_iovlen - 1) * max_chunk; |
| } else { |
| iov[0].iov_base = 0; |
| iov[0].iov_len = line_size; |
| msg.msg_iovlen = 1; |
| } |
| |
| msg.msg_iov = iov; |
| provider->memcpy_to_device(mem, 0, line, line_size); |
| |
| msg.msg_control = ctrl_data; |
| msg.msg_controllen = sizeof(ctrl_data); |
| |
| cmsg = CMSG_FIRSTHDR(&msg); |
| cmsg->cmsg_level = SOL_SOCKET; |
| cmsg->cmsg_type = SCM_DEVMEM_DMABUF; |
| cmsg->cmsg_len = CMSG_LEN(sizeof(__u32)); |
| |
| ddmabuf = tx_dmabuf_id; |
| |
| *((__u32 *)CMSG_DATA(cmsg)) = ddmabuf; |
| |
| ret = sendmsg(socket_fd, &msg, MSG_ZEROCOPY); |
| if (ret < 0) { |
| pr_err("Failed sendmsg"); |
| goto err_free_line; |
| } |
| |
| fprintf(stderr, "sendmsg_ret=%d\n", ret); |
| |
| if (ret != line_size) { |
| pr_err("Did not send all bytes %d vs %zd", ret, line_size); |
| goto err_free_line; |
| } |
| |
| if (wait_compl(socket_fd)) |
| goto err_free_line; |
| } |
| |
| fprintf(stderr, "%s: tx ok\n", TEST_PREFIX); |
| |
| err = 0; |
| |
| err_free_line: |
| free(line); |
| err_unbind: |
| ynl_sock_destroy(ys); |
| err_close_socket: |
| close(socket_fd); |
| return err; |
| } |
| |
| int main(int argc, char *argv[]) |
| { |
| struct memory_buffer *mem; |
| int is_server = 0, opt; |
| int ret, err = 1; |
| |
| while ((opt = getopt(argc, argv, "ls:c:p:v:q:t:f:z:")) != -1) { |
| switch (opt) { |
| case 'l': |
| is_server = 1; |
| break; |
| case 's': |
| server_ip = optarg; |
| break; |
| case 'c': |
| client_ip = optarg; |
| break; |
| case 'p': |
| port = optarg; |
| break; |
| case 'v': |
| do_validation = atoll(optarg); |
| break; |
| case 'q': |
| num_queues = atoi(optarg); |
| break; |
| case 't': |
| start_queue = atoi(optarg); |
| break; |
| case 'f': |
| ifname = optarg; |
| break; |
| case 'z': |
| max_chunk = atoi(optarg); |
| break; |
| case '?': |
| fprintf(stderr, "unknown option: %c\n", optopt); |
| break; |
| } |
| } |
| |
| if (!ifname) { |
| pr_err("Missing -f argument"); |
| return 1; |
| } |
| |
| ifindex = if_nametoindex(ifname); |
| |
| fprintf(stderr, "using ifindex=%u\n", ifindex); |
| |
| if (!server_ip && !client_ip) { |
| if (start_queue < 0 && num_queues < 0) { |
| num_queues = rxq_num(ifindex); |
| if (num_queues < 0) { |
| pr_err("couldn't detect number of queues"); |
| return 1; |
| } |
| if (num_queues < 2) { |
| pr_err("number of device queues is too low"); |
| return 1; |
| } |
| /* make sure can bind to multiple queues */ |
| start_queue = num_queues / 2; |
| num_queues /= 2; |
| } |
| |
| if (start_queue < 0 || num_queues < 0) { |
| pr_err("Both -t and -q are required"); |
| return 1; |
| } |
| |
| return run_devmem_tests(); |
| } |
| |
| if (start_queue < 0 && num_queues < 0) { |
| num_queues = rxq_num(ifindex); |
| if (num_queues < 2) { |
| pr_err("number of device queues is too low"); |
| return 1; |
| } |
| |
| num_queues = 1; |
| start_queue = rxq_num(ifindex) - num_queues; |
| |
| if (start_queue < 0) { |
| pr_err("couldn't detect number of queues"); |
| return 1; |
| } |
| |
| fprintf(stderr, "using queues %d..%d\n", start_queue, start_queue + num_queues); |
| } |
| |
| for (; optind < argc; optind++) |
| fprintf(stderr, "extra arguments: %s\n", argv[optind]); |
| |
| if (start_queue < 0) { |
| pr_err("Missing -t argument"); |
| return 1; |
| } |
| |
| if (num_queues < 0) { |
| pr_err("Missing -q argument"); |
| return 1; |
| } |
| |
| if (!server_ip) { |
| pr_err("Missing -s argument"); |
| return 1; |
| } |
| |
| if (!port) { |
| pr_err("Missing -p argument"); |
| return 1; |
| } |
| |
| mem = provider->alloc(getpagesize() * NUM_PAGES); |
| if (!mem) { |
| pr_err("Failed to allocate memory buffer"); |
| return 1; |
| } |
| |
| ret = is_server ? do_server(mem) : do_client(mem); |
| if (ret) |
| goto err_free_mem; |
| |
| err = 0; |
| |
| err_free_mem: |
| provider->free(mem); |
| return err; |
| } |