X-Git-Url: https://gerrit.o-ran-sc.org/r/gitweb?a=blobdiff_plain;f=fhi_lib%2Flib%2Fethernet%2Fethernet.c;h=0930665005ce6de6f5234740f253667459b64b68;hb=2de97529a4c5a1922214ba0e6f0fb84cacbd0bc7;hp=ebf997ca83627982d262e096b780aa7772a7bf75;hpb=4745e5c88ba931c6d71cb6d8c681f76cf364eac5;p=o-du%2Fphy.git diff --git a/fhi_lib/lib/ethernet/ethernet.c b/fhi_lib/lib/ethernet/ethernet.c index ebf997c..0930665 100644 --- a/fhi_lib/lib/ethernet/ethernet.c +++ b/fhi_lib/lib/ethernet/ethernet.c @@ -1,6 +1,6 @@ /****************************************************************************** * -* Copyright (c) 2019 Intel. +* Copyright (c) 2020 Intel. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. @@ -16,7 +16,6 @@ * *******************************************************************************/ - /** * @brief This file has all definitions for the Ethernet Data Interface Layer * @file ethernet.c @@ -38,12 +37,11 @@ #include #include #include - +#include #include #include #include #include -#include #include #include #include @@ -67,110 +65,103 @@ #include "ethernet.h" #include "ethdi.h" -/* Our mbuf pools. */ -struct rte_mempool *_eth_mbuf_pool; -struct rte_mempool *_eth_mbuf_pool_rx; -struct rte_mempool *_eth_mbuf_pool_small; -struct rte_mempool *_eth_mbuf_pool_big; - -/* - * Make sure the ring indexes are big enough to cover buf space x2 - * This ring-buffer maintains the property head - tail <= RINGSIZE. - * head == tail: ring buffer empty - * head - tail == RINGSIZE: ring buffer full - */ -typedef uint16_t ring_idx; -static struct { - ring_idx head; - ring_idx read_head; - ring_idx tail; - char buf[1024]; /* needs power of 2! */ -} io_ring = { {0}, 0, 0}; - -#define RINGSIZE sizeof(io_ring.buf) -#define RINGMASK (RINGSIZE - 1) - - -/* - * Display part of the message stored in the ring buffer. - * Might require multiple calls to print the full message. - * Will return 0 when nothing left to print. - */ -int xran_show_delayed_message(void) -{ - ring_idx tail = io_ring.tail; - ring_idx wlen = io_ring.read_head - tail; /* always within [0, RINGSIZE] */ - - if (wlen <= 0) - return 0; +/* mbuf pools */ +struct rte_mempool *_eth_mbuf_pool = NULL; +struct rte_mempool *_eth_mbuf_pool_indirect = NULL; +struct rte_mempool *_eth_mbuf_pool_rx = NULL; +struct rte_mempool *_eth_mbuf_pkt_gen = NULL; - tail &= RINGMASK; /* modulo the range down now that we have wlen */ +struct rte_mempool *socket_direct_pool = NULL; +struct rte_mempool *socket_indirect_pool = NULL; - /* Make sure we're not going over buffer end. Next call will wrap. */ - if (tail + wlen > RINGSIZE) - wlen = RINGSIZE - tail; +struct rte_mempool *_eth_mbuf_pool_vf_rx[16][RTE_MAX_QUEUES_PER_PORT] = {NULL}; +struct rte_mempool *_eth_mbuf_pool_vf_small[16] = {NULL}; - RTE_ASSERT(tail + wlen <= RINGSIZE); - - /* We use write() here to avoid recaculating string length in fwrite(). */ - const ssize_t written = write(STDOUT_FILENO, io_ring.buf + tail, wlen); - if (written <= 0) - return 0; /* To avoid moving tail the wrong way on error. */ - - /* Move tail up. Only we touch it. And we only print from one core. */ - io_ring.tail += written; +void +xran_init_mbuf_pool(uint32_t mtu) +{ + uint16_t data_room_size = MBUF_POOL_ELEMENT; + printf("%s: socket %d\n",__FUNCTION__, rte_socket_id()); - return written; /* next invocation will print the rest if any */ + if (mtu <= 1500) { + data_room_size = MBUF_POOL_ELM_SMALL; } - -void xran_init_mbuf_pool(void) -{ /* Init the buffer pool */ if (rte_eal_process_type() == RTE_PROC_PRIMARY) { _eth_mbuf_pool = rte_pktmbuf_pool_create("mempool", NUM_MBUFS, - MBUF_CACHE, 0, MBUF_POOL_ELEMENT, rte_socket_id()); - _eth_mbuf_pool_rx = rte_pktmbuf_pool_create("mempool_rx", NUM_MBUFS, - MBUF_CACHE, 0, MBUF_POOL_ELEMENT, rte_socket_id()); - _eth_mbuf_pool_small = rte_pktmbuf_pool_create("mempool_small", - NUM_MBUFS, MBUF_CACHE, 0, MBUF_POOL_ELM_SMALL, rte_socket_id()); - _eth_mbuf_pool_big = rte_pktmbuf_pool_create("mempool_big", - NUM_MBUFS_BIG, 0, 0, MBUF_POOL_ELM_BIG, rte_socket_id()); + MBUF_CACHE, 0, data_room_size, rte_socket_id()); + _eth_mbuf_pool_indirect = rte_pktmbuf_pool_create("mempool_indirect", NUM_MBUFS_VF, + MBUF_CACHE, 0, 0, rte_socket_id()); + _eth_mbuf_pkt_gen = rte_pktmbuf_pool_create("mempool_pkt_gen", + NUM_MBUFS, MBUF_CACHE, 0, MBUF_POOL_PKT_GEN_ELM, rte_socket_id()); } else { _eth_mbuf_pool = rte_mempool_lookup("mempool"); - _eth_mbuf_pool_rx = rte_mempool_lookup("mempool_rx"); - _eth_mbuf_pool_small = rte_mempool_lookup("mempool_small"); - _eth_mbuf_pool_big = rte_mempool_lookup("mempool_big"); + _eth_mbuf_pool_indirect = rte_mempool_lookup("mempool_indirect"); + _eth_mbuf_pkt_gen = rte_mempool_lookup("mempool_pkt_gen"); } + if (_eth_mbuf_pool == NULL) rte_panic("Cannot create mbuf pool: %s\n", rte_strerror(rte_errno)); - if (_eth_mbuf_pool_rx == NULL) + if (_eth_mbuf_pool_indirect == NULL) rte_panic("Cannot create mbuf pool: %s\n", rte_strerror(rte_errno)); - if (_eth_mbuf_pool_small == NULL) - rte_panic("Cannot create small mbuf pool: %s\n", rte_strerror(rte_errno)); - if (_eth_mbuf_pool_big == NULL) - rte_panic("Cannot create big mbuf pool: %s\n", rte_strerror(rte_errno)); + if (_eth_mbuf_pkt_gen == NULL) + rte_panic("Cannot create packet gen pool: %s\n", rte_strerror(rte_errno)); + + if (socket_direct_pool == NULL) + socket_direct_pool = _eth_mbuf_pool; + + if (socket_indirect_pool == NULL) + socket_indirect_pool = _eth_mbuf_pool_indirect; +} + +/* Configure the Rx with optional split. */ +int +rx_queue_setup(uint16_t port_id, uint16_t rx_queue_id, + uint16_t nb_rx_desc, unsigned int socket_id, + struct rte_eth_rxconf *rx_conf, struct rte_mempool *mp) +{ + unsigned int i, mp_n; + int ret; +#ifndef RTE_ETH_RX_OFFLOAD_BUFFER_SPLIT +#define RTE_ETH_RX_OFFLOAD_BUFFER_SPLIT 0x00100000 +#endif + if ((rx_conf->offloads & RTE_ETH_RX_OFFLOAD_BUFFER_SPLIT) == 0) { +#if (RTE_VER_YEAR >= 21) + rx_conf->rx_seg = NULL; + rx_conf->rx_nseg = 0; +#endif + ret = rte_eth_rx_queue_setup(port_id, rx_queue_id, + nb_rx_desc, socket_id, + rx_conf, mp); + return ret; + + } else { + printf("rx_queue_setup error\n"); + ret = -EINVAL; + return ret; + } } /* Init NIC port, then start the port */ -void xran_init_port(int p_id, struct ether_addr *p_lls_cu_addr) +void xran_init_port(int p_id, uint16_t num_rxq, uint32_t mtu) { - char buf[ETHER_ADDR_FMT_SIZE]; - struct ether_addr eth_addr; - struct rte_eth_rxmode rxmode = - { .split_hdr_size = 0, + static uint16_t nb_rxd = BURST_SIZE; + static uint16_t nb_txd = BURST_SIZE; + struct rte_ether_addr addr; + struct rte_eth_rxmode rxmode = { + .split_hdr_size = 0, .max_rx_pkt_len = MAX_RX_LEN, - .offloads=(DEV_RX_OFFLOAD_JUMBO_FRAME|DEV_RX_OFFLOAD_CRC_STRIP) + .offloads = DEV_RX_OFFLOAD_JUMBO_FRAME }; struct rte_eth_txmode txmode = { - .mq_mode = ETH_MQ_TX_NONE + .mq_mode = ETH_MQ_TX_NONE, + .offloads = DEV_TX_OFFLOAD_MULTI_SEGS }; struct rte_eth_conf port_conf = { .rxmode = rxmode, .txmode = txmode }; - struct ether_addr pDstEthAddr; - struct rte_eth_rxconf rxq_conf; struct rte_eth_txconf txq_conf; @@ -178,94 +169,135 @@ void xran_init_port(int p_id, struct ether_addr *p_lls_cu_addr) struct rte_eth_dev_info dev_info; const char *drv_name = ""; int sock_id = rte_eth_dev_socket_id(p_id); - - // ether_format_addr(buf, sizeof(buf), p_lls_cu_addr); - // printf("port %d set mac address %s\n", p_id, buf); - // rte_eth_dev_default_mac_addr_set(p_id, p_lls_cu_addr); + char rx_pool_name[32] = ""; + uint16_t data_room_size = MBUF_POOL_ELEMENT; + uint16_t qi = 0; + uint32_t num_mbufs = 0; + + if (mtu <= 1500) { + rxmode.offloads &= ~DEV_RX_OFFLOAD_JUMBO_FRAME; + rxmode.max_rx_pkt_len = RTE_ETHER_MAX_LEN; + data_room_size = MBUF_POOL_ELM_SMALL; + } rte_eth_dev_info_get(p_id, &dev_info); if (dev_info.driver_name) drv_name = dev_info.driver_name; printf("initializing port %d for TX, drv=%s\n", p_id, drv_name); - /* In order to receive packets from any server need to add broad case address - * for the port*/ - pDstEthAddr.addr_bytes[0] = 0xFF; - pDstEthAddr.addr_bytes[1] = 0xFF; - pDstEthAddr.addr_bytes[2] = 0xFF; - - pDstEthAddr.addr_bytes[3] = 0xFF; - pDstEthAddr.addr_bytes[4] = 0xFF; - pDstEthAddr.addr_bytes[5] = 0xFF; - - rte_eth_macaddr_get(p_id, ð_addr); - ether_format_addr(buf, sizeof(buf), ð_addr); - printf("port %d mac address %s\n", p_id, buf); + if (dev_info.tx_offload_capa & DEV_TX_OFFLOAD_MBUF_FAST_FREE){ + printf("set DEV_TX_OFFLOAD_MBUF_FAST_FREE\n"); + port_conf.txmode.offloads |= + DEV_TX_OFFLOAD_MBUF_FAST_FREE; + } - struct ether_addr addr; rte_eth_macaddr_get(p_id, &addr); -// rte_eth_dev_mac_addr_add(p_id, &pDstEthAddr,1); - // rte_eth_dev_mac_addr_add(p_id, &addr, 1); - printf("Port %u MAC: %02"PRIx8" %02"PRIx8" %02"PRIx8 " %02"PRIx8" %02"PRIx8" %02"PRIx8"\n", (unsigned)p_id, addr.addr_bytes[0], addr.addr_bytes[1], addr.addr_bytes[2], addr.addr_bytes[3], addr.addr_bytes[4], addr.addr_bytes[5]); + if(num_rxq > 1) { + nb_rxd = 2048; + num_mbufs = 2*nb_rxd-1; + } else { + nb_rxd = BURST_SIZE; + num_mbufs = NUM_MBUFS; + } + /* Init port */ - ret = rte_eth_dev_configure(p_id, 1, 1, &port_conf); + ret = rte_eth_dev_configure(p_id, num_rxq, 1, &port_conf); if (ret < 0) rte_panic("Cannot configure port %u (%d)\n", p_id, ret); + ret = rte_eth_dev_adjust_nb_rx_tx_desc(p_id, &nb_rxd,&nb_txd); + + if (ret < 0) { + printf("\n"); + rte_exit(EXIT_FAILURE, "Cannot adjust number of " + "descriptors: err=%d, port=%d\n", ret, p_id); + } + printf("Port %u: nb_rxd %d nb_txd %d\n", p_id, nb_rxd, nb_txd); + + for (qi = 0; qi < num_rxq; qi++) { + snprintf(rx_pool_name, RTE_DIM(rx_pool_name), "%s_p_%d_q_%d", "mp_rx_", p_id, qi); + printf("[%d] %s num blocks %d\n", p_id, rx_pool_name, num_mbufs); + _eth_mbuf_pool_vf_rx[p_id][qi] = rte_pktmbuf_pool_create(rx_pool_name, num_mbufs, + MBUF_CACHE, 0, data_room_size, rte_socket_id()); + + if (_eth_mbuf_pool_vf_rx[p_id][qi] == NULL) + rte_panic("Cannot create mbuf pool: %s\n", rte_strerror(rte_errno)); + } + + snprintf(rx_pool_name, RTE_DIM(rx_pool_name), "%s_%d", "mempool_small_", p_id); + printf("[%d] %s\n", p_id, rx_pool_name); + _eth_mbuf_pool_vf_small[p_id] = rte_pktmbuf_pool_create(rx_pool_name, NUM_MBUFS_VF, + MBUF_CACHE, 0, MBUF_POOL_ELM_SMALL_INDIRECT, rte_socket_id()); + + if (_eth_mbuf_pool_vf_small[p_id] == NULL) + rte_panic("Cannot create mbuf pool: %s\n", rte_strerror(rte_errno)); + /* Init RX queues */ + fflush(stdout); rxq_conf = dev_info.default_rxconf; - ret = rte_eth_rx_queue_setup(p_id, 0, BURST_SIZE, - sock_id, &rxq_conf, _eth_mbuf_pool_rx); + + for (qi = 0; qi < num_rxq; qi++) { + ret = rx_queue_setup(p_id, qi, nb_rxd, + sock_id, &rxq_conf, _eth_mbuf_pool_vf_rx[p_id][qi]); + } + if (ret < 0) rte_panic("Cannot init RX for port %u (%d)\n", p_id, ret); /* Init TX queues */ + fflush(stdout); txq_conf = dev_info.default_txconf; - ret = rte_eth_tx_queue_setup(p_id, 0, BURST_SIZE, sock_id, &txq_conf); + + ret = rte_eth_tx_queue_setup(p_id, 0, nb_txd, sock_id, &txq_conf); if (ret < 0) rte_panic("Cannot init TX for port %u (%d)\n", p_id, ret); + ret = rte_eth_dev_set_ptypes(p_id, RTE_PTYPE_UNKNOWN, NULL, 0); + if (ret < 0) + rte_panic("Port %d: Failed to disable Ptype parsing\n", p_id); + /* Start port */ ret = rte_eth_dev_start(p_id); if (ret < 0) rte_panic("Cannot start port %u (%d)\n", p_id, ret); - - rte_eth_promiscuous_enable(p_id); } -void xran_memdump(void *addr, int len) +void xran_init_port_mempool(int p_id, uint32_t mtu) { - int i; - char tmp_buf[len * 2 + len / 16 + 1]; - char *p = tmp_buf; - - return; -#if 0 - for (i = 0; i < len; ++i) { - sprintf(p, "%.2X ", ((uint8_t *)addr)[i]); - if (i % 16 == 15) - *p++ = '\n'; - } - *p = 0; - nlog("%s", tmp_buf); -#endif + int ret; + int sock_id = rte_eth_dev_socket_id(p_id); + char rx_pool_name[32] = ""; + uint16_t data_room_size = MBUF_POOL_ELEMENT; + + if (mtu <= 1500) { + data_room_size = MBUF_POOL_ELM_SMALL; } + snprintf(rx_pool_name, RTE_DIM(rx_pool_name), "%s_%d", "mempool_small_", p_id); + printf("[%d] %s\n", p_id, rx_pool_name); + _eth_mbuf_pool_vf_small[p_id] = rte_pktmbuf_pool_create(rx_pool_name, NUM_MBUFS_VF, + MBUF_CACHE, 0, MBUF_POOL_ELM_SMALL, rte_socket_id()); + + if (_eth_mbuf_pool_vf_small[p_id] == NULL) + rte_panic("Cannot create mbuf pool: %s\n", rte_strerror(rte_errno)); + + +} /* Prepend ethernet header, possibly vlan tag. */ -void xran_add_eth_hdr_vlan(struct ether_addr *dst, uint16_t ethertype, struct rte_mbuf *mb, uint16_t vlan_tci) +void xran_add_eth_hdr_vlan(struct rte_ether_addr *dst, uint16_t ethertype, struct rte_mbuf *mb) { /* add in the ethernet header */ - struct ether_hdr *const h = (void *)rte_pktmbuf_prepend(mb, sizeof(*h)); + struct rte_ether_hdr *h = (struct rte_ether_hdr *)rte_pktmbuf_mtod(mb, struct rte_ether_hdr*); PANIC_ON(h == NULL, "mbuf prepend of ether_hdr failed"); @@ -276,82 +308,15 @@ void xran_add_eth_hdr_vlan(struct ether_addr *dst, uint16_t ethertype, struct rt #if defined(DPDKIO_DEBUG) && DPDKIO_DEBUG > 1 { - char dst[ETHER_ADDR_FMT_SIZE] = "(empty)"; - char src[ETHER_ADDR_FMT_SIZE] = "(empty)"; + char dst[RTE_ETHER_ADDR_FMT_SIZE] = "(empty)"; + char src[RTE_ETHER_ADDR_FMT_SIZE] = "(empty)"; - nlog("*** packet for TX below (len %d) ***", rte_pktmbuf_pkt_len(mb)); - ether_format_addr(src, sizeof(src), &h->s_addr); - ether_format_addr(dst, sizeof(dst), &h->d_addr); - nlog("src: %s dst: %s ethertype: %.4X", src, dst, ethertype); + printf("*** packet for TX below (len %d) ***", rte_pktmbuf_pkt_len(mb)); + rte_ether_format_addr(src, sizeof(src), &h->s_addr); + rte_ether_format_addr(dst, sizeof(dst), &h->d_addr); + printf("src: %s dst: %s ethertype: %.4X", src, dst, ethertype); } #endif -#ifdef VLAN_SUPPORT - mb->vlan_tci = vlan_tci; - dlog("Inserting vlan tag of %d", vlan_tci); - rte_vlan_insert(&mb); -#endif } -int xran_send_message_burst(int dst_id, int pkt_type, void *body, int len) -{ - struct rte_mbuf *mbufs[BURST_SIZE]; - int i; - uint8_t *src = body; - const struct xran_ethdi_ctx *const ctx = xran_ethdi_get_ctx(); - - /* We're limited by maximum mbuf size on the receive size. - * We can change this but this would be a bigger rework. */ - RTE_ASSERT(len < MBUF_POOL_ELM_BIG); - - /* Allocate the required number of mbufs. */ - const uint8_t count = ceilf((float)len / MAX_DATA_SIZE); - if (rte_pktmbuf_alloc_bulk(_eth_mbuf_pool, mbufs, count) != 0) - rte_panic("Failed to allocate %d mbufs\n", count); - - nlog("burst transfer with data size %lu", MAX_DATA_SIZE); - for (i = 0; len > 0; ++i) { - char *p; - struct burst_hdr *bhdr; - struct ethdi_hdr *edi_hdr; - - /* Setup the ethdi_hdr. */ - edi_hdr = (void *)rte_pktmbuf_append(mbufs[i], sizeof(*edi_hdr)); - if (edi_hdr == NULL) - rte_panic("append of ethdi_hdr failed\n"); - edi_hdr->pkt_type = PKT_BURST; - /* edi_hdr->source_id setup in tx_from_ring */ - edi_hdr->dest_id = dst_id; - - /* Setup the burst header */ - bhdr = (void *)rte_pktmbuf_append(mbufs[i], sizeof(*bhdr)); - if (bhdr == NULL) /* append failed. */ - rte_panic("mbuf prepend of burst_hdr failed\n"); - bhdr->original_type = pkt_type; - bhdr->pkt_idx = i; /* save the index of the burst chunk. */ - bhdr->total_pkts = count; - - /* now copy in the actual data */ - const int curr_data_len = RTE_MIN(len, MAX_TX_LEN - - rte_pktmbuf_pkt_len(mbufs[i]) - sizeof(struct ether_hdr)); - p = (void *)rte_pktmbuf_append(mbufs[i], curr_data_len); - if (p == NULL) - rte_panic("mbuf append of %d data bytes failed\n", curr_data_len); - /* This copy is unavoidable, as we're splitting one big buffer - * into multiple mbufs. */ - rte_memcpy(p, src, curr_data_len); - - dlog("curr_data_len[%d] = %d", i, curr_data_len); - dlog("packet %d size %d", i, rte_pktmbuf_pkt_len(mbufs[i])); - - /* Update our source data pointer and remaining length. */ - len -= curr_data_len; - src += curr_data_len; - } - - /* Now enqueue the full prepared burst. */ - i = rte_ring_enqueue_bulk(ctx->tx_ring[0], (void **)mbufs, count, NULL); - PANIC_ON(i != count, "failed to enqueue all mbufs: %d/%d", i, count); - dlog("%d packets enqueued on port %d.", count, ctx->io_cfg.port); - return 1; -}