X-Git-Url: https://gerrit.o-ran-sc.org/r/gitweb?a=blobdiff_plain;f=fhi_lib%2Flib%2Fsrc%2Fxran_common.c;h=dc40ad9de2ac4f5a80e5aae86d87b1eb35ce7582;hb=HEAD;hp=d72f27aa37d6e9bb2b35afd3e5bfeea3e7c60489;hpb=cef07f74965b1749dd909fc1322e211489fea2ea;p=o-du%2Fphy.git diff --git a/fhi_lib/lib/src/xran_common.c b/fhi_lib/lib/src/xran_common.c index d72f27a..dc40ad9 100644 --- a/fhi_lib/lib/src/xran_common.c +++ b/fhi_lib/lib/src/xran_common.c @@ -1,6 +1,6 @@ /****************************************************************************** * -* Copyright (c) 2019 Intel. +* Copyright (c) 2020 Intel. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. @@ -17,35 +17,45 @@ *******************************************************************************/ /** - * @brief XRAN layer common functionality for both lls-CU and RU as well as C-plane and + * @brief XRAN layer common functionality for both O-DU and O-RU as well as C-plane and * U-plane * @file xran_common.c * @ingroup group_source_xran * @author Intel Corporation **/ +#define _GNU_SOURCE #include #include #include #include #include +#include +#include +#include +#include +#include #include "xran_common.h" #include "ethdi.h" #include "xran_pkt.h" #include "xran_pkt_up.h" +#include "xran_cp_api.h" #include "xran_up_api.h" +#include "xran_cp_proc.h" +#include "xran_dev.h" #include "xran_lib_mlog_tasks_id.h" -#include "../src/xran_printf.h" -#include +#include "xran_printf.h" #include "xran_mlog_lnx.h" -#define MBUFS_CNT 16 +static struct timespec sleeptime = {.tv_nsec = 1E3 }; /* 1 us */ -extern long interval_us; +extern int32_t first_call; + +#define MBUFS_CNT 16 -extern int xran_process_rx_sym(void *arg, +extern int32_t xran_process_rx_sym(void *arg, struct rte_mbuf *mbuf, void *iq_data_start, uint16_t size, @@ -60,7 +70,10 @@ extern int xran_process_rx_sym(void *arg, uint16_t sym_inc, uint16_t rb, uint16_t sect_id, - uint32_t *mb_free); + uint32_t *mb_free, + int8_t expect_comp, + uint8_t compMeth, + uint8_t iqWidth); extern int xran_process_prach_sym(void *arg, @@ -77,7 +90,28 @@ extern int xran_process_prach_sym(void *arg, uint16_t start_prbu, uint16_t sym_inc, uint16_t rb, - uint16_t sect_id); + uint16_t sect_id, + uint32_t *mb_free); + +extern int32_t xran_process_srs_sym(void *arg, + struct rte_mbuf *mbuf, + void *iq_data_start, + uint16_t size, + uint8_t CC_ID, + uint8_t Ant_ID, + uint8_t frame_id, + uint8_t subframe_id, + uint8_t slot_id, + uint8_t symb_id, + uint16_t num_prbu, + uint16_t start_prbu, + uint16_t sym_inc, + uint16_t rb, + uint16_t sect_id, + uint32_t *mb_free, + int8_t expect_comp, + uint8_t compMeth, + uint8_t iqWidth); extern int32_t xran_pkt_validate(void *arg, struct rte_mbuf *mbuf, @@ -89,26 +123,583 @@ extern int32_t xran_pkt_validate(void *arg, uint8_t subframe_id, uint8_t slot_id, uint8_t symb_id, - struct ecpri_seq_id *seq_id, + union ecpri_seq_id *seq_id, uint16_t num_prbu, uint16_t start_prbu, uint16_t sym_inc, uint16_t rb, uint16_t sect_id); -long rx_counter = 0; -long tx_counter = 0; +int process_mbuf_batch(struct rte_mbuf* pkt_q[], void* handle, int16_t num, struct xran_eaxc_info *p_cid, uint32_t* ret_data) +{ + struct rte_mbuf* pkt; + struct xran_device_ctx* p_dev_ctx = (struct xran_device_ctx*)handle; + void* iq_samp_buf[MBUFS_CNT]; + union ecpri_seq_id seq[MBUFS_CNT]; + static int symbol_total_bytes[XRAN_PORTS_NUM][XRAN_MAX_SECTOR_NR][XRAN_MAX_ANTENNA_NR] = { 0 }; + int num_bytes[MBUFS_CNT] = { 0 }, num_bytes_pusch[MBUFS_CNT] = { 0 }; + int16_t i, j; + + struct xran_common_counters* pCnt = &p_dev_ctx->fh_counters; + + uint8_t CC_ID[MBUFS_CNT] = { 0 }; + uint8_t Ant_ID[MBUFS_CNT] = { 0 }; + uint8_t frame_id[MBUFS_CNT] = { 0 }; + uint8_t subframe_id[MBUFS_CNT] = { 0 }; + uint8_t slot_id[MBUFS_CNT] = { 0 }; + uint8_t symb_id[MBUFS_CNT] = { 0 }; + + uint16_t num_prbu[MBUFS_CNT]; + uint16_t start_prbu[MBUFS_CNT]; + uint16_t sym_inc[MBUFS_CNT]; + uint16_t rb[MBUFS_CNT]; + uint16_t sect_id[MBUFS_CNT]; + uint16_t prb_elem_id[MBUFS_CNT] = {0}; + + uint8_t compMeth[MBUFS_CNT] = { 0 }; + uint8_t iqWidth[MBUFS_CNT] = { 0 }; + uint8_t compMeth_ini = 0; + uint8_t iqWidth_ini = 0; + + uint32_t pkt_size[MBUFS_CNT]; + + int expect_comp = (p_dev_ctx->fh_cfg.ru_conf.compMeth != XRAN_COMPMETHOD_NONE); + enum xran_comp_hdr_type staticComp = p_dev_ctx->fh_cfg.ru_conf.xranCompHdrType; + + int16_t num_pusch = 0, num_prach = 0, num_srs = 0; + int16_t pusch_idx[MBUFS_CNT] = { 0 }, prach_idx[MBUFS_CNT] = { 0 }, srs_idx[MBUFS_CNT] = { 0 }; + int8_t xran_port = xran_dev_ctx_get_port_id(p_dev_ctx); + int16_t max_ant_num = 0; + uint8_t *ptr_seq_id_num_port; + struct xran_eaxcid_config* conf; + uint8_t seq_id[MBUFS_CNT]; + uint16_t cid[MBUFS_CNT]; + + struct xran_ecpri_hdr* ecpri_hdr[MBUFS_CNT]; + struct radio_app_common_hdr* radio_hdr[MBUFS_CNT]; + struct data_section_hdr* data_hdr[MBUFS_CNT]; + struct data_section_compression_hdr* data_compr_hdr[MBUFS_CNT]; + + const int16_t ecpri_size = sizeof(struct xran_ecpri_hdr); + const int16_t rad_size = sizeof(struct radio_app_common_hdr); + const int16_t data_size = sizeof(struct data_section_hdr); + const int16_t compr_size = sizeof(struct data_section_compression_hdr); + + char* buf_start[MBUFS_CNT]; + uint16_t start_off[MBUFS_CNT]; + uint16_t iq_offset[MBUFS_CNT]; + uint16_t last[MBUFS_CNT]; + + uint32_t tti = 0; + struct rte_mbuf* mb = NULL; + struct xran_prb_map* pRbMap = NULL; + struct xran_prb_elm* prbMapElm = NULL; + //uint16_t iq_sample_size_bits; + uint16_t idxElm = 0, total_sections = 0; + +#if XRAN_MLOG_VAR + uint32_t mlogVar[10]; + uint32_t mlogVarCnt = 0; +#endif + + if (xran_port < 0) { + print_err("Invalid pHandle"); + return MBUF_FREE; + } + + if (xran_port > XRAN_PORTS_NUM) { + print_err("Invalid port - %d", xran_port); + return MBUF_FREE; + } + + if(first_call == 0) { + for(i = 0; i < num; i++ ) + ret_data[i] = MBUF_FREE; + return MBUF_FREE; + } + + conf = &(p_dev_ctx->eAxc_id_cfg); + if (conf == NULL) { + rte_panic("conf == NULL"); + } + + if (p_dev_ctx->fh_init.io_cfg.id == O_DU) + { + max_ant_num = XRAN_MAX_ANTENNA_NR * 2 + XRAN_MAX_ANT_ARRAY_ELM_NR; + ptr_seq_id_num_port = &xran_upul_seq_id_num[xran_port][0][0]; + } + else if (p_dev_ctx->fh_init.io_cfg.id == O_RU) + { + max_ant_num = XRAN_MAX_ANTENNA_NR; + ptr_seq_id_num_port = &xran_updl_seq_id_num[xran_port][0][0]; + } + else +{ + rte_panic("incorrect fh_init.io_cfg.id"); + } + + if (staticComp == XRAN_COMP_HDR_TYPE_STATIC) + { + compMeth_ini = p_dev_ctx->fh_cfg.ru_conf.compMeth; + iqWidth_ini = p_dev_ctx->fh_cfg.ru_conf.iqWidth; +} + + for (i = 0; i < MBUFS_CNT; i++) +{ + pkt_size[i] = pkt_q[i]->pkt_len; + buf_start[i] = (char*)pkt_q[i]->buf_addr; + start_off[i] = pkt_q[i]->data_off; +} + + if (expect_comp && (staticComp != XRAN_COMP_HDR_TYPE_STATIC)) + { +#pragma vector always + for (i = 0; i < MBUFS_CNT; i++) + { +#if XRAN_MLOG_VAR + mlogVarCnt = 0; +#endif + ecpri_hdr[i] = (void*)(buf_start[i] + start_off[i]); + radio_hdr[i] = (void*)(buf_start[i] + start_off[i] + ecpri_size); + data_hdr[i] = (void*)(buf_start[i] + start_off[i] + ecpri_size + rad_size); + data_compr_hdr[i] = (void*)(buf_start[i] + start_off[i] + ecpri_size + rad_size + data_size); + seq[i] = ecpri_hdr[i]->ecpri_seq_id; + seq_id[i] = seq[i].bits.seq_id; + last[i] = seq[i].bits.e_bit; + + iq_offset[i] = ecpri_size + rad_size + data_size + compr_size; + + iq_samp_buf[i] = (void*)(buf_start[i] + start_off[i] + iq_offset[i]); + num_bytes[i] = pkt_size[i] - iq_offset[i]; + + if (ecpri_hdr[i] == NULL || + radio_hdr[i] == NULL || + data_hdr[i] == NULL || + data_compr_hdr[i] == NULL || + iq_samp_buf[i] == NULL) + { + num_bytes[i] = 0; /* packet too short */ + } + +#if XRAN_MLOG_VAR + if(radio_hdr[i] != NULL && data_hdr[i] != NULL) + { + mlogVar[mlogVarCnt++] = 0xBBBBBBBB; + mlogVar[mlogVarCnt++] = xran_lib_ota_tti; + mlogVar[mlogVarCnt++] = radio_hdr[i]->frame_id; + mlogVar[mlogVarCnt++] = radio_hdr[i]->sf_slot_sym.subframe_id; + mlogVar[mlogVarCnt++] = radio_hdr[i]->sf_slot_sym.slot_id; + mlogVar[mlogVarCnt++] = radio_hdr[i]->sf_slot_sym.symb_id; + mlogVar[mlogVarCnt++] = data_hdr[i]->fields.sect_id; + mlogVar[mlogVarCnt++] = data_hdr[i]->fields.start_prbu; + mlogVar[mlogVarCnt++] = data_hdr[i]->fields.num_prbu; + mlogVar[mlogVarCnt++] = rte_pktmbuf_pkt_len(pkt_q[i]); + MLogAddVariables(mlogVarCnt, mlogVar, MLogTick()); + } +#endif + } + } + else + { +#pragma vector always + for (i = 0; i < MBUFS_CNT; i++) + { +#if XRAN_MLOG_VAR + mlogVarCnt = 0; +#endif + ecpri_hdr[i] = (void*)(buf_start[i] + start_off[i]); + radio_hdr[i] = (void*)(buf_start[i] + start_off[i] + ecpri_size); + data_hdr[i] = (void*)(buf_start[i] + start_off[i] + ecpri_size + rad_size); + seq[i] = ecpri_hdr[i]->ecpri_seq_id; + seq_id[i] = seq[i].bits.seq_id; + last[i] = seq[i].bits.e_bit; + + iq_offset[i] = ecpri_size + rad_size + data_size; + iq_samp_buf[i] = (void*)(buf_start[i] + start_off[i] + iq_offset[i]); + num_bytes[i] = pkt_size[i] - iq_offset[i]; + + if (ecpri_hdr[i] == NULL || + radio_hdr[i] == NULL || + data_hdr[i] == NULL || + iq_samp_buf[i] == NULL) + { + num_bytes[i] = 0; /* packet too short */ + } + +#if XRAN_MLOG_VAR + if (radio_hdr[i] != NULL && data_hdr[i] != NULL) + { + mlogVar[mlogVarCnt++] = 0xBBBBBBBB; + mlogVar[mlogVarCnt++] = xran_lib_ota_tti; + mlogVar[mlogVarCnt++] = radio_hdr[i]->frame_id; + mlogVar[mlogVarCnt++] = radio_hdr[i]->sf_slot_sym.subframe_id; + mlogVar[mlogVarCnt++] = radio_hdr[i]->sf_slot_sym.slot_id; + mlogVar[mlogVarCnt++] = radio_hdr[i]->sf_slot_sym.symb_id; + mlogVar[mlogVarCnt++] = data_hdr[i]->fields.sect_id; + mlogVar[mlogVarCnt++] = data_hdr[i]->fields.start_prbu; + mlogVar[mlogVarCnt++] = data_hdr[i]->fields.num_prbu; + mlogVar[mlogVarCnt++] = rte_pktmbuf_pkt_len(pkt_q[i]); + MLogAddVariables(mlogVarCnt, mlogVar, MLogTick()); + } +#endif + } + } + + for (i = 0; i < MBUFS_CNT; i++) { + if(p_cid->ccId == 0xFF && p_cid->ruPortId == 0xFF) { + cid[i] = rte_be_to_cpu_16((uint16_t)ecpri_hdr[i]->ecpri_xtc_id); + if (num_bytes[i] > 0) { + CC_ID[i] = (cid[i] & conf->mask_ccId) >> conf->bit_ccId; + Ant_ID[i] = (cid[i] & conf->mask_ruPortId) >> conf->bit_ruPortId; + } + } else { + if (num_bytes[i] > 0) { + CC_ID[i] = p_cid->ccId; + Ant_ID[i] = p_cid->ruPortId; + } + } + } + + for (i = 0; i < MBUFS_CNT; i++) + { + radio_hdr[i]->sf_slot_sym.value = rte_be_to_cpu_16(radio_hdr[i]->sf_slot_sym.value); + data_hdr[i]->fields.all_bits = rte_be_to_cpu_32(data_hdr[i]->fields.all_bits); + } + + for (i = 0; i < MBUFS_CNT; i++) + { + if (num_bytes[i] > 0) + { + compMeth[i] = compMeth_ini; + iqWidth[i] = iqWidth_ini; + + frame_id[i] = radio_hdr[i]->frame_id; + subframe_id[i] = radio_hdr[i]->sf_slot_sym.subframe_id; + slot_id[i] = radio_hdr[i]->sf_slot_sym.slot_id; + symb_id[i] = radio_hdr[i]->sf_slot_sym.symb_id; + + num_prbu[i] = data_hdr[i]->fields.num_prbu; + start_prbu[i] = data_hdr[i]->fields.start_prbu; + sym_inc[i] = data_hdr[i]->fields.sym_inc; + rb[i] = data_hdr[i]->fields.rb; + sect_id[i] = data_hdr[i]->fields.sect_id; + + if (num_prbu[i] == 0) + num_prbu[i] = p_dev_ctx->fh_cfg.nULRBs; + + if (expect_comp && (staticComp != XRAN_COMP_HDR_TYPE_STATIC)) + { + compMeth[i] = data_compr_hdr[i]->ud_comp_hdr.ud_comp_meth; + iqWidth[i] = data_compr_hdr[i]->ud_comp_hdr.ud_iq_width; + } + + if (CC_ID[i] >= XRAN_MAX_CELLS_PER_PORT || Ant_ID[i] >= max_ant_num || symb_id[i] >= XRAN_NUM_OF_SYMBOL_PER_SLOT) + { + ptr_seq_id_num_port[CC_ID[i] * max_ant_num + Ant_ID[i]] = seq_id[i]; // for next + pCnt->Rx_pkt_dupl++; +// print_err("Invalid CC ID - %d or antenna ID or Symbol ID- %d", CC_ID[i], Ant_ID[i], symb_id[i]); + } + else + { + ptr_seq_id_num_port[CC_ID[i] * max_ant_num + Ant_ID[i]]++; + } + + pCnt->rx_counter++; + pCnt->Rx_on_time++; + pCnt->Total_msgs_rcvd++; + struct xran_prach_cp_config *PrachCfg = NULL; + if(p_dev_ctx->dssEnable){ + tti = frame_id[i] * SLOTS_PER_SYSTEMFRAME(p_dev_ctx->interval_us_local) + + subframe_id[i] * SLOTNUM_PER_SUBFRAME(p_dev_ctx->interval_us_local) + slot_id[i]; + int techSlot = (tti % p_dev_ctx->dssPeriod); + if(p_dev_ctx->technology[techSlot] == 1) + PrachCfg = &(p_dev_ctx->PrachCPConfig); + else + PrachCfg = &(p_dev_ctx->PrachCPConfigLTE); + } + else{ + PrachCfg = &(p_dev_ctx->PrachCPConfig); + } + + if (Ant_ID[i] >= p_dev_ctx->srs_cfg.eAxC_offset && p_dev_ctx->fh_cfg.srsEnable) + { + Ant_ID[i] -= p_dev_ctx->srs_cfg.eAxC_offset; + if (last[i] == 1) + { + srs_idx[num_srs] = i; + num_srs += 1; + pCnt->rx_srs_packets++; + } + } + else if (Ant_ID[i] >= PrachCfg->eAxC_offset && p_dev_ctx->fh_cfg.prachEnable) + { + Ant_ID[i] -= PrachCfg->eAxC_offset; + if (last[i] == 1) + { + prach_idx[num_prach] = i; + num_prach += 1; + pCnt->rx_prach_packets[Ant_ID[i]]++; + } + } + else + { + if (last[i] == 1) + { + pusch_idx[num_pusch] = i; + num_pusch += 1; + pCnt->rx_pusch_packets[Ant_ID[i]]++; + } + } + symbol_total_bytes[xran_port][CC_ID[i]][Ant_ID[i]] += num_bytes[i]; + num_bytes_pusch[i] = symbol_total_bytes[xran_port][CC_ID[i]][Ant_ID[i]]; + if (last[i] == 1) + symbol_total_bytes[xran_port][CC_ID[i]][Ant_ID[i]] = 0; + } + } + + for (j = 0; j < num_prach; j++) + { + i = prach_idx[j]; + pkt = pkt_q[i]; + + print_dbg("Completed receiving PRACH symbol %d, size=%d bytes\n", symb_id[i], num_bytes[i]); + + xran_process_prach_sym(p_dev_ctx, + pkt, + iq_samp_buf[i], + num_bytes[i], + CC_ID[i], + Ant_ID[i], + frame_id[i], + subframe_id[i], + slot_id[i], + symb_id[i], + num_prbu[i], + start_prbu[i], + sym_inc[i], + rb[i], + sect_id[i], + &ret_data[i]); + } + + for (j = 0; j < num_srs; j++) + { + i = srs_idx[j]; + pkt = pkt_q[i]; + + print_dbg("SRS receiving symbol %d, size=%d bytes\n", + symb_id[i], symbol_total_bytes[p_dev_ctx->xran_port_id][CC_ID[i]][Ant_ID[i]]); + + uint64_t t1 = MLogXRANTick(); + xran_process_srs_sym(p_dev_ctx, + pkt, + iq_samp_buf[i], + num_bytes[i], + CC_ID[i], + Ant_ID[i], + frame_id[i], + subframe_id[i], + slot_id[i], + symb_id[i], + num_prbu[i], + start_prbu[i], + sym_inc[i], + rb[i], + sect_id[i], + &ret_data[i], + expect_comp, + compMeth[i], + iqWidth[i]); + MLogXRANTask(PID_PROCESS_UP_PKT_SRS, t1, MLogXRANTick()); + } + + if (num_pusch == MBUFS_CNT) + { + for (i = 0; i < MBUFS_CNT; i++) + { + //iq_sample_size_bits = 16; + //if (expect_comp) + // iq_sample_size_bits = iqWidth[i]; + + tti = frame_id[i] * SLOTS_PER_SYSTEMFRAME(p_dev_ctx->interval_us_local) + + subframe_id[i] * SLOTNUM_PER_SUBFRAME(p_dev_ctx->interval_us_local) + slot_id[i]; + + pRbMap = (struct xran_prb_map*)p_dev_ctx->sFrontHaulRxPrbMapBbuIoBufCtrl[tti % XRAN_N_FE_BUF_LEN][CC_ID[i]][Ant_ID[i]].sBufferList.pBuffers->pData; + + if (pRbMap) + { + /** Get the prb_elem_id */ + total_sections=0; + if(pRbMap->prbMap[0].bf_weight.extType == 1) + { + for(idxElm=0 ; idxElm < pRbMap->nPrbElm ; idxElm++) + { + total_sections += pRbMap->prbMap[idxElm].bf_weight.numSetBFWs; + if(total_sections >= (sect_id[i] + 1)) + { + prb_elem_id[i] = idxElm; + break; + } + } + } + else + { + prb_elem_id[i] = sect_id[i]; + } + + if (prb_elem_id[i] >= pRbMap->nPrbElm) + { + print_err("sect_id %d, prb_elem_id %d !=pRbMap->nPrbElm %d\n", sect_id[i], prb_elem_id[i], pRbMap->nPrbElm); + ret_data[i] = MBUF_FREE; + continue; + } + } + else + { +// print_err("pRbMap==NULL\n"); + ret_data[i] = MBUF_FREE; + continue; + } + + if (pRbMap->nPrbElm == 1) + { + p_dev_ctx->sFrontHaulRxBbuIoBufCtrl[tti % XRAN_N_FE_BUF_LEN][CC_ID[i]][Ant_ID[i]].sBufferList.pBuffers[symb_id[i]].pData = iq_samp_buf[i]; + p_dev_ctx->sFrontHaulRxBbuIoBufCtrl[tti % XRAN_N_FE_BUF_LEN][CC_ID[i]][Ant_ID[i]].sBufferList.pBuffers[symb_id[i]].pCtrl = pkt_q[i]; + ret_data[i] = MBUF_KEEP; + } + else + { + struct xran_section_desc* p_sec_desc = NULL; + prbMapElm = &pRbMap->prbMap[prb_elem_id[i]]; + int16_t nSecDesc = prbMapElm->nSecDesc[symb_id[i]]; + p_sec_desc = &prbMapElm->sec_desc[symb_id[i]][nSecDesc]; + + if (p_sec_desc) + { + mb = p_sec_desc->pCtrl; + if (mb) { + rte_pktmbuf_free(mb); + } + p_sec_desc->pCtrl = pkt_q[i]; + p_sec_desc->pData = iq_samp_buf[i]; + p_sec_desc->start_prbu = start_prbu[i]; + p_sec_desc->num_prbu = num_prbu[i]; + p_sec_desc->iq_buffer_len = num_bytes_pusch[i]; + p_sec_desc->iq_buffer_offset = iq_offset[i]; + ret_data[i] = MBUF_KEEP; + prbMapElm->nSecDesc[symb_id[i]] += 1; + } + else +{ +// print_err("p_sec_desc==NULL tti %u ant %d symb_id %d\n", tti, Ant_ID[i], symb_id[i]); + ret_data[i] = MBUF_FREE; + } + } + } + } + else + { + for (j = 0; j < num_pusch; j++) + { + i = pusch_idx[j]; + + //iq_sample_size_bits = 16; + //if (expect_comp) + // iq_sample_size_bits = iqWidth[i]; + + tti = frame_id[i] * SLOTS_PER_SYSTEMFRAME(p_dev_ctx->interval_us_local) + + subframe_id[i] * SLOTNUM_PER_SUBFRAME(p_dev_ctx->interval_us_local) + slot_id[i]; + + pRbMap = (struct xran_prb_map*)p_dev_ctx->sFrontHaulRxPrbMapBbuIoBufCtrl[tti % XRAN_N_FE_BUF_LEN][CC_ID[i]][Ant_ID[i]].sBufferList.pBuffers->pData; + + if (pRbMap) + { + /** Get the prb_elem_id */ + total_sections=0; + if(pRbMap->prbMap[0].bf_weight.extType == 1) + { + for(idxElm=0 ; idxElm < pRbMap->nPrbElm ; idxElm++) + { + total_sections += pRbMap->prbMap[idxElm].bf_weight.numSetBFWs; + if(total_sections >= (sect_id[i] + 1)) + { + prb_elem_id[i] = idxElm; + break; + } + } + } + else + { + prb_elem_id[i] = sect_id[i]; + } + + if (prb_elem_id[i] >= pRbMap->nPrbElm) + { + print_err("sect_id %d, prb_elem_id %d !=pRbMap->nPrbElm %d\n", sect_id[i], prb_elem_id[i], pRbMap->nPrbElm); + ret_data[i] = MBUF_FREE; + continue; + } + } + else + { +// print_err("pRbMap==NULL\n"); + ret_data[i] = MBUF_FREE; + continue; + } + + if (pRbMap->nPrbElm == 1) + { + p_dev_ctx->sFrontHaulRxBbuIoBufCtrl[tti % XRAN_N_FE_BUF_LEN][CC_ID[i]][Ant_ID[i]].sBufferList.pBuffers[symb_id[i]].pData = iq_samp_buf[i]; + p_dev_ctx->sFrontHaulRxBbuIoBufCtrl[tti % XRAN_N_FE_BUF_LEN][CC_ID[i]][Ant_ID[i]].sBufferList.pBuffers[symb_id[i]].pCtrl = pkt_q[i]; + ret_data[i] = MBUF_KEEP; + } + else + { + struct xran_section_desc* p_sec_desc = NULL; + prbMapElm = &pRbMap->prbMap[prb_elem_id[i]]; + int16_t nSecDesc = prbMapElm->nSecDesc[symb_id[i]]; + p_sec_desc = &prbMapElm->sec_desc[symb_id[i]][nSecDesc]; + + if (p_sec_desc) + { + mb = p_sec_desc->pCtrl; + if (mb) { + rte_pktmbuf_free(mb); + } + p_sec_desc->pCtrl = pkt_q[i]; + p_sec_desc->pData = iq_samp_buf[i]; + p_sec_desc->start_prbu = start_prbu[i]; + p_sec_desc->num_prbu = num_prbu[i]; + p_sec_desc->iq_buffer_len = num_bytes_pusch[i]; + p_sec_desc->iq_buffer_offset = iq_offset[i]; + ret_data[i] = MBUF_KEEP; + prbMapElm->nSecDesc[symb_id[i]] += 1; + } + else + { +// print_err("p_sec_desc==NULL tti %u ant %d symb_id %d\n", tti, Ant_ID[i], symb_id[i]); + ret_data[i] = MBUF_FREE; + } + } + } + } + return MBUF_FREE; +} -int process_mbuf(struct rte_mbuf *pkt) +int +process_mbuf(struct rte_mbuf *pkt, void* handle, struct xran_eaxc_info *p_cid) { + uint64_t tt1 = MLogXRANTick(); + struct xran_device_ctx *p_dev_ctx = (struct xran_device_ctx *)handle; void *iq_samp_buf; - struct ecpri_seq_id seq; - static int symbol_total_bytes = 0; + union ecpri_seq_id seq; + static int symbol_total_bytes[XRAN_PORTS_NUM][XRAN_MAX_SECTOR_NR][XRAN_MAX_ANTENNA_NR] = {0}; int num_bytes = 0; - struct xran_device_ctx * p_x_ctx = xran_dev_get_ctx(); - uint8_t CC_ID = 0; - uint8_t Ant_ID = 0; + struct xran_common_counters *pCnt = &p_dev_ctx->fh_counters; + + uint8_t CC_ID = p_cid->ccId; + uint8_t Ant_ID = p_cid->ruPortId; uint8_t frame_id = 0; uint8_t subframe_id = 0; uint8_t slot_id = 0; @@ -119,124 +710,168 @@ int process_mbuf(struct rte_mbuf *pkt) uint16_t sym_inc; uint16_t rb; uint16_t sect_id; - void *pHandle = NULL; - uint8_t num_eAxc = xran_get_num_eAxc(pHandle); + + uint8_t compMeth = 0; + uint8_t iqWidth = 0; + int ret = MBUF_FREE; uint32_t mb_free = 0; int32_t valid_res = 0; + int expect_comp = (p_dev_ctx->fh_cfg.ru_conf.compMeth != XRAN_COMPMETHOD_NONE); + enum xran_comp_hdr_type staticComp = p_dev_ctx->fh_cfg.ru_conf.xranCompHdrType; + if(first_call == 0) + return ret; - if(p_x_ctx->xran2phy_mem_ready == 0) + if (staticComp == XRAN_COMP_HDR_TYPE_STATIC) + { + compMeth = p_dev_ctx->fh_cfg.ru_conf.compMeth; + iqWidth = p_dev_ctx->fh_cfg.ru_conf.iqWidth; + } + + if(p_dev_ctx->xran2phy_mem_ready == 0 || first_call == 0) return MBUF_FREE; - num_bytes = xran_extract_iq_samples(pkt, - &iq_samp_buf, - &CC_ID, - &Ant_ID, - &frame_id, - &subframe_id, - &slot_id, - &symb_id, - &seq, - &num_prbu, - &start_prbu, - &sym_inc, - &rb, - §_id); - - if (num_bytes <= 0){ + num_bytes = xran_extract_iq_samples(pkt, &iq_samp_buf, + &CC_ID, &Ant_ID, &frame_id, &subframe_id, &slot_id, &symb_id, &seq, + &num_prbu, &start_prbu, &sym_inc, &rb, §_id, + expect_comp, staticComp, &compMeth, &iqWidth); + if (num_bytes <= 0) + { print_err("num_bytes is wrong [%d]\n", num_bytes); return MBUF_FREE; } + if (num_prbu == 0) + num_prbu = p_dev_ctx->fh_cfg.nULRBs; - valid_res = xran_pkt_validate(NULL, - pkt, - iq_samp_buf, - num_bytes, - CC_ID, - Ant_ID, - frame_id, - subframe_id, - slot_id, - symb_id, - &seq, - num_prbu, - start_prbu, - sym_inc, - rb, - sect_id); - - if(valid_res != 0) { - print_err("valid_res is wrong [%d] ant %u (%u : %u : %u : %u) seq %u num_bytes %d\n", valid_res, Ant_ID, frame_id, subframe_id, slot_id, symb_id, seq.seq_id, num_bytes); - return MBUF_FREE; - } + MLogXRANTask(PID_PROCESS_UP_PKT_PARSE, tt1, MLogXRANTick()); + /* do not validate for NDM SRS */ + if (Ant_ID >= p_dev_ctx->srs_cfg.eAxC_offset && p_dev_ctx->fh_cfg.srsEnable) + { + /* SRS packet has ruportid = 2*num_eAxc + ant_id */ + Ant_ID -= p_dev_ctx->srs_cfg.eAxC_offset; + symbol_total_bytes[p_dev_ctx->xran_port_id][CC_ID][Ant_ID] += num_bytes; + + if (seq.bits.e_bit == 1) + { + print_dbg("SRS receiving symbol %d, size=%d bytes\n", + symb_id, symbol_total_bytes[p_dev_ctx->xran_port_id][CC_ID][Ant_ID]); + + if (symbol_total_bytes[p_dev_ctx->xran_port_id][CC_ID][Ant_ID]) + { + uint64_t t1 = MLogXRANTick(); + int16_t res = xran_process_srs_sym(p_dev_ctx, + pkt, iq_samp_buf, num_bytes, + CC_ID, Ant_ID, frame_id, subframe_id, slot_id, symb_id, + num_prbu, start_prbu, sym_inc, rb, sect_id, + &mb_free, expect_comp, compMeth, iqWidth); + if(res == symbol_total_bytes[p_dev_ctx->xran_port_id][CC_ID][Ant_ID]) + ret = mb_free; + else + print_err("res != symbol_total_bytes[p_dev_ctx->xran_port_id][CC_ID][Ant_ID]\n"); + + pCnt->rx_srs_packets++; + MLogXRANTask(PID_PROCESS_UP_PKT_SRS, t1, MLogXRANTick()); + } + symbol_total_bytes[p_dev_ctx->xran_port_id][CC_ID][Ant_ID] = 0; + } + else + print_dbg("Transport layer fragmentation (eCPRI) is not supported\n"); + } /* if (Ant_ID >= p_dev_ctx->srs_cfg.eAxC_offset && p_dev_ctx->fh_cfg.srsEnable) */ - if (Ant_ID >= num_eAxc && p_x_ctx->fh_init.prachEnable) // PRACH packet has ruportid = num_eAxc + ant_id + else { - Ant_ID -= num_eAxc; - if (seq.e_bit == 1) { + valid_res = xran_pkt_validate(p_dev_ctx, + pkt, iq_samp_buf, num_bytes, + CC_ID, Ant_ID, frame_id, subframe_id, slot_id, symb_id, + &seq, num_prbu, start_prbu, sym_inc, rb, sect_id); +#ifndef FCN_ADAPT + if(valid_res != 0) + { + print_dbg("valid_res is wrong [%d] ant %u (%u : %u : %u : %u) seq %u num_bytes %d\n", valid_res, Ant_ID, frame_id, subframe_id, slot_id, symb_id, seq.bits.seq_id, num_bytes); + return MBUF_FREE; + } +#endif + int tti = 0; + struct xran_prach_cp_config *PrachCfg = NULL; + if(p_dev_ctx->dssEnable){ + tti = frame_id * SLOTS_PER_SYSTEMFRAME(p_dev_ctx->interval_us_local) + + subframe_id * SLOTNUM_PER_SUBFRAME(p_dev_ctx->interval_us_local) + slot_id; + int techSlot = (tti % p_dev_ctx->dssPeriod); + if(p_dev_ctx->technology[techSlot] == 1) + PrachCfg = &(p_dev_ctx->PrachCPConfig); + else + PrachCfg = &(p_dev_ctx->PrachCPConfigLTE); + } + else{ + PrachCfg = &(p_dev_ctx->PrachCPConfig); + } + if (Ant_ID >= PrachCfg->eAxC_offset && p_dev_ctx->fh_cfg.prachEnable) + { + /* PRACH packet has ruportid = num_eAxc + ant_id */ + Ant_ID -= PrachCfg->eAxC_offset; + symbol_total_bytes[p_dev_ctx->xran_port_id][CC_ID][Ant_ID] += num_bytes; + if (seq.bits.e_bit == 1) + { print_dbg("Completed receiving PRACH symbol %d, size=%d bytes\n", symb_id, num_bytes); - xran_process_prach_sym(NULL, - pkt, - iq_samp_buf, - num_bytes, - CC_ID, - Ant_ID, - frame_id, - subframe_id, - slot_id, - symb_id, - num_prbu, - start_prbu, - sym_inc, - rb, - sect_id); - } - else { + if (symbol_total_bytes[p_dev_ctx->xran_port_id][CC_ID][Ant_ID]) + { + int16_t res = xran_process_prach_sym(p_dev_ctx, + pkt, iq_samp_buf, num_bytes, + CC_ID, Ant_ID, frame_id, subframe_id, slot_id, symb_id, + num_prbu, start_prbu, sym_inc, rb, sect_id, &mb_free); + if(res == symbol_total_bytes[p_dev_ctx->xran_port_id][CC_ID][Ant_ID]) + ret = mb_free; + else + print_err("res != symbol_total_bytes[p_dev_ctx->xran_port_id][CC_ID][Ant_ID]\n"); + + pCnt->rx_prach_packets[Ant_ID]++; + } + symbol_total_bytes[p_dev_ctx->xran_port_id][CC_ID][Ant_ID] = 0; + } + else print_dbg("Transport layer fragmentation (eCPRI) is not supported\n"); } - } else { - symbol_total_bytes += num_bytes; + else + { + /* PUSCH */ + symbol_total_bytes[p_dev_ctx->xran_port_id][CC_ID][Ant_ID] += num_bytes; - if (seq.e_bit == 1) { + if (seq.bits.e_bit == 1) + { print_dbg("Completed receiving symbol %d, size=%d bytes\n", - symb_id, symbol_total_bytes); - - if (symbol_total_bytes){ - int res = xran_process_rx_sym(NULL, - pkt, - iq_samp_buf, - symbol_total_bytes, - CC_ID, - Ant_ID, - frame_id, - subframe_id, - slot_id, - symb_id, - num_prbu, - start_prbu, - sym_inc, - rb, - sect_id, - &mb_free); - if(res == symbol_total_bytes) - ret = mb_free; - else - print_err("res != symbol_total_bytes\n"); + symb_id, symbol_total_bytes[p_dev_ctx->xran_port_id][CC_ID][Ant_ID]); + + if (symbol_total_bytes[p_dev_ctx->xran_port_id][CC_ID][Ant_ID]) + { + uint64_t t1 = MLogXRANTick(); + int res = xran_process_rx_sym(p_dev_ctx, + pkt, iq_samp_buf, symbol_total_bytes[p_dev_ctx->xran_port_id][CC_ID][Ant_ID], + CC_ID, Ant_ID, frame_id, subframe_id, slot_id, symb_id, + num_prbu, start_prbu, sym_inc, rb, sect_id, + &mb_free, expect_comp, compMeth, iqWidth); + if(res == symbol_total_bytes[p_dev_ctx->xran_port_id][CC_ID][Ant_ID]) + ret = mb_free; + else + print_err("res != symbol_total_bytes[p_dev_ctx->xran_port_id][CC_ID][Ant_ID]\n"); + + pCnt->rx_pusch_packets[Ant_ID]++; + MLogXRANTask(PID_PROCESS_UP_PKT, t1, MLogXRANTick()); } - symbol_total_bytes = 0; - } - else { - print_dbg("Transport layer fragmentation (eCPRI) is not supported\n"); + symbol_total_bytes[p_dev_ctx->xran_port_id][CC_ID][Ant_ID] = 0; } + else + print_dbg("Transport layer fragmentation (eCPRI) is not supported\n"); } + } /* else */ return ret; } +#if 0 static int set_iq_bit_width(uint8_t iq_bit_width, struct data_section_compression_hdr *compr_hdr) { if (iq_bit_width == MAX_IQ_BIT_WIDTH) @@ -247,12 +882,15 @@ static int set_iq_bit_width(uint8_t iq_bit_width, struct data_section_compressio return 0; } +#endif /* Send a single 5G symbol over multiple packets */ -int32_t prepare_symbol_ex(enum xran_pkt_dir direction, - uint16_t section_id, +inline int32_t prepare_symbol_ex(enum xran_pkt_dir direction, + uint16_t section_id_start, struct rte_mbuf *mb, - struct rb_map *data, + uint8_t *data, + uint8_t compMeth, + uint8_t iqWidth, const enum xran_input_byte_order iq_buf_byte_order, uint8_t frame_id, uint8_t subframe_id, @@ -263,54 +901,124 @@ int32_t prepare_symbol_ex(enum xran_pkt_dir direction, uint8_t CC_ID, uint8_t RU_Port_ID, uint8_t seq_id, - uint32_t do_copy) + uint32_t do_copy, + enum xran_comp_hdr_type staticEn, + uint16_t num_sections, + uint16_t iq_offset) { - int32_t n_bytes = ((prb_num == 0) ? MAX_N_FULLBAND_SC : prb_num) * N_SC_PER_PRB * sizeof(struct rb_map); - + int32_t n_bytes , iq_len_aggr = 0; int32_t prep_bytes; + int16_t nPktSize,idx, nprb_per_section; + uint32_t curr_sect_id; + int parm_size; + struct xran_up_pkt_gen_params xp[XRAN_MAX_SECTIONS_PER_SLOT] = { 0 }; + bool prbElemBegin , prbElemEnd; + + iqWidth = (iqWidth==0) ? 16 : iqWidth; + switch(compMeth) { + case XRAN_COMPMETHOD_BLKFLOAT: parm_size = 1; break; + case XRAN_COMPMETHOD_MODULATION: parm_size = 0; break; + default: + parm_size = 0; + } - int16_t nPktSize = sizeof(struct ether_hdr) + sizeof(struct xran_ecpri_hdr) + - sizeof(struct radio_app_common_hdr)+ sizeof(struct data_section_hdr) + n_bytes; - uint32_t off; - struct xran_up_pkt_gen_no_compression_params xp = { 0 }; + nprb_per_section = prb_num/num_sections; + if(prb_num%num_sections) + nprb_per_section++; - n_bytes = RTE_MIN(n_bytes, XRAN_MAX_MBUF_LEN); + n_bytes = (3 * iqWidth + parm_size)*nprb_per_section; + // n_bytes = RTE_MIN(n_bytes, XRAN_MAX_MBUF_LEN); - /* radio app header */ - xp.app_params.data_direction = direction; - xp.app_params.payl_ver = 1; - xp.app_params.filter_id = 0; - xp.app_params.frame_id = frame_id; - xp.app_params.sf_slot_sym.subframe_id = subframe_id; - xp.app_params.sf_slot_sym.slot_id = slot_id; - xp.app_params.sf_slot_sym.symb_id = symbol_no; + for(idx=0 ; idx < num_sections ; idx++) + { + prbElemBegin = (idx == 0) ? 1 : 0; + prbElemEnd = (idx + 1 == num_sections) ? 1 : 0; + curr_sect_id = section_id_start + idx ; + + iq_len_aggr += n_bytes; + + if(prbElemBegin) + { + nPktSize = sizeof(struct rte_ether_hdr) + + sizeof(struct xran_ecpri_hdr) + + sizeof(struct radio_app_common_hdr) ; + } + + if(prbElemEnd){ + if(((idx+1)*nprb_per_section) > prb_num){ + nprb_per_section = (prb_num - idx*nprb_per_section); + // n_bytes = (3 * iqWidth + parm_size)*(nprb_per_section); + } + } + + nPktSize += sizeof(struct data_section_hdr); + + if ((compMeth != XRAN_COMPMETHOD_NONE)&&(staticEn == XRAN_COMP_HDR_TYPE_DYNAMIC)) + nPktSize += sizeof(struct data_section_compression_hdr); + + nPktSize += n_bytes; + + /** radio app header + * Setting app_params is redundant , its needed only once to create common Radio app header. + */ + xp[idx].app_params.data_feature.value = 0x10; + xp[idx].app_params.data_feature.data_direction = direction; + // xp[idx].app_params.payl_ver = 1; + // xp[idx].app_params.filter_id = 0; + xp[idx].app_params.frame_id = frame_id; + xp[idx].app_params.sf_slot_sym.subframe_id = subframe_id; + xp[idx].app_params.sf_slot_sym.slot_id = xran_slotid_convert(slot_id, 0); + xp[idx].app_params.sf_slot_sym.symb_id = symbol_no; /* convert to network byte order */ - xp.app_params.sf_slot_sym.value = rte_cpu_to_be_16(xp.app_params.sf_slot_sym.value); + xp[idx].app_params.sf_slot_sym.value = rte_cpu_to_be_16(xp[idx].app_params.sf_slot_sym.value); + + // printf("start_prbu = %d, prb_num = %d,num_sections = %d, nprb_per_section = %d,curr_sect_id = %d\n",(prb_start + idx*nprb_per_section),prb_num,num_sections,nprb_per_section,curr_sect_id); + xp[idx].sec_hdr.fields.all_bits = 0; + xp[idx].sec_hdr.fields.sect_id = curr_sect_id; + xp[idx].sec_hdr.fields.num_prbu = XRAN_CONVERT_NUMPRBC(nprb_per_section); //(uint8_t)prb_num; + xp[idx].sec_hdr.fields.start_prbu = prb_start; + xp[idx].sec_hdr.fields.sym_inc = 0; + xp[idx].sec_hdr.fields.rb = 0; + + /* compression */ + xp[idx].compr_hdr_param.ud_comp_hdr.ud_comp_meth = compMeth; + xp[idx].compr_hdr_param.ud_comp_hdr.ud_iq_width = XRAN_CONVERT_IQWIDTH(iqWidth); + xp[idx].compr_hdr_param.rsrvd = 0; + prb_start += nprb_per_section; - xp.sec_hdr.fields.sect_id = section_id; - xp.sec_hdr.fields.num_prbu = (uint8_t)prb_num; - xp.sec_hdr.fields.start_prbu = (uint8_t)prb_start; - xp.sec_hdr.fields.sym_inc = 0; - xp.sec_hdr.fields.rb = 0; +#if 0 + printf("\nidx %hu num_prbu %u sect_id %u start_prbu %u sym_inc %u curr_sec_id %u",idx,(uint32_t)xp[idx].sec_hdr.fields.num_prbu, + (uint32_t)xp[idx].sec_hdr.fields.sect_id, + (uint32_t)xp[idx].sec_hdr.fields.start_prbu, + (uint32_t)xp[idx].sec_hdr.fields.sym_inc,curr_sect_id); + +#endif /* network byte order */ - xp.sec_hdr.fields.all_bits = rte_cpu_to_be_32(xp.sec_hdr.fields.all_bits); + xp[idx].sec_hdr.fields.all_bits = rte_cpu_to_be_32(xp[idx].sec_hdr.fields.all_bits); if (mb == NULL){ MLogPrint(NULL); errx(1, "out of mbufs after %d packets", 1); } + } /* for(idx=0 ; idx < num_sections ; idx++) */ - prep_bytes = xran_prepare_iq_symbol_portion_no_comp(mb, + //printf("\niq_len_aggr %u",iq_len_aggr); + + prep_bytes = xran_prepare_iq_symbol_portion(mb, data, iq_buf_byte_order, - n_bytes, - &xp, + iq_len_aggr, + xp, CC_ID, RU_Port_ID, seq_id, - do_copy); + staticEn, + do_copy, + num_sections, + section_id_start, + iq_offset); if (prep_bytes <= 0) errx(1, "failed preparing symbol"); @@ -324,24 +1032,185 @@ int32_t prepare_symbol_ex(enum xran_pkt_dir direction, return prep_bytes; } -/* Send a single 5G symbol over multiple packets */ -int send_symbol_ex(enum xran_pkt_dir direction, - uint16_t section_id, - struct rte_mbuf *mb, - struct rb_map *data, - const enum xran_input_byte_order iq_buf_byte_order, +int32_t prepare_sf_slot_sym (enum xran_pkt_dir direction, uint8_t frame_id, uint8_t subframe_id, uint8_t slot_id, uint8_t symbol_no, - int prb_start, - int prb_num, - uint8_t CC_ID, - uint8_t RU_Port_ID, - uint8_t seq_id) + struct xran_up_pkt_gen_params *xp) +{ + /* radio app header */ + xp->app_params.data_feature.value = 0x10; + xp->app_params.data_feature.data_direction = direction; + //xp->app_params.payl_ver = 1; + //xp->app_params.filter_id = 0; + xp->app_params.frame_id = frame_id; + xp->app_params.sf_slot_sym.subframe_id = subframe_id; + xp->app_params.sf_slot_sym.slot_id = xran_slotid_convert(slot_id, 0); + xp->app_params.sf_slot_sym.symb_id = symbol_no; + + /* convert to network byte order */ + xp->app_params.sf_slot_sym.value = rte_cpu_to_be_16(xp->app_params.sf_slot_sym.value); + + return 0; +} + +int send_symbol_mult_section_ex(void *handle, + enum xran_pkt_dir direction, + uint16_t section_id, + struct rte_mbuf *mb, uint8_t *data, + uint8_t compMeth, uint8_t iqWidth, + const enum xran_input_byte_order iq_buf_byte_order, + uint8_t frame_id, uint8_t subframe_id, + uint8_t slot_id, uint8_t symbol_no, + int prb_start, int prb_num, + uint8_t CC_ID, uint8_t RU_Port_ID, uint8_t seq_id) { uint32_t do_copy = 0; - int32_t n_bytes = ((prb_num == 0) ? MAX_N_FULLBAND_SC : prb_num) * N_SC_PER_PRB * sizeof(struct rb_map); + int32_t n_bytes; + int hdr_len, parm_size; + int32_t sent=0; + uint32_t loop = 0; + struct xran_device_ctx *p_dev_ctx = (struct xran_device_ctx *)handle; + struct xran_common_counters *pCnt = &p_dev_ctx->fh_counters; + enum xran_comp_hdr_type staticEn= XRAN_COMP_HDR_TYPE_DYNAMIC; + + + if (p_dev_ctx != NULL) + { + staticEn = p_dev_ctx->fh_cfg.ru_conf.xranCompHdrType; + + hdr_len = sizeof(struct xran_ecpri_hdr) + + sizeof(struct radio_app_common_hdr) + + sizeof(struct data_section_hdr); + if ((compMeth != XRAN_COMPMETHOD_NONE)&&(staticEn == XRAN_COMP_HDR_TYPE_DYNAMIC)) + hdr_len += sizeof(struct data_section_compression_hdr); + + switch(compMeth) { + case XRAN_COMPMETHOD_BLKFLOAT: parm_size = 1; break; + case XRAN_COMPMETHOD_MODULATION: parm_size = 0; break; + default: + parm_size = 0; + } + int prb_num_pre_sec = (prb_num+2)/3; + int prb_offset = 0; + int data_offset = 0; + int prb_num_sec; + rte_iova_t ext_buff_iova = 0; + + struct rte_mbuf *send_mb; + char *p_sec_iq = NULL; + char *ext_buff = NULL; + uint16_t ext_buff_len = 0; + struct rte_mbuf_ext_shared_info * p_share_data = NULL; + struct rte_mbuf *eth_oran_hdr = NULL; + struct rte_mbuf *tmp = NULL; + for (loop = 0; loop < 3;loop++) + { + seq_id = xran_get_upul_seqid(handle, CC_ID, RU_Port_ID); + + prb_num_sec = ((loop+1)*prb_num_pre_sec > prb_num) ? (prb_num - loop*prb_num_pre_sec) : prb_num_pre_sec; + n_bytes = (3 * iqWidth + parm_size) * prb_num_sec; + char * pChar = NULL; + + send_mb = xran_ethdi_mbuf_alloc(); /* will be freede by ETH */ + if(send_mb == NULL) { + MLogPrint(NULL); + errx(1, "out of mbufs after %d packets", 1); + } + + pChar = rte_pktmbuf_append(send_mb, hdr_len + n_bytes); + if(pChar == NULL) { + MLogPrint(NULL); + errx(1, "incorrect mbuf size %d packets", 1); + } + pChar = rte_pktmbuf_prepend(send_mb, sizeof(struct rte_ether_hdr)); + if(pChar == NULL) { + MLogPrint(NULL); + errx(1, "incorrect mbuf size %d packets", 1); + } + do_copy = 1; /* new mbuf hence copy of IQs */ + pChar = rte_pktmbuf_mtod(send_mb, char*); + char *pdata_start = (pChar + sizeof(struct rte_ether_hdr) + hdr_len); + memcpy(pdata_start,data + data_offset,n_bytes); + + + sent = prepare_symbol_ex(direction, + section_id, + send_mb, + data + data_offset, + compMeth, + iqWidth, + iq_buf_byte_order, + frame_id, + subframe_id, + slot_id, + symbol_no, + prb_start+prb_offset, + prb_num_sec, + CC_ID, + RU_Port_ID, + seq_id, + do_copy, + staticEn, + 1, + 0); /*Send a single section */ + prb_offset += prb_num_sec; + data_offset += n_bytes; + if(sent) { + pCnt->tx_counter++; + pCnt->tx_bytes_counter += rte_pktmbuf_pkt_len(send_mb); + p_dev_ctx->send_upmbuf2ring(send_mb, ETHER_TYPE_ECPRI, xran_map_ecpriPcid_to_vf(p_dev_ctx, direction, CC_ID, RU_Port_ID)); + } + + } + +#ifdef DEBUG + printf("Symbol %2d sent (%d packets, %d bytes)\n", symbol_no, i, n_bytes); +#endif + } + return sent; +} + + +/* Send a single 5G symbol over multiple packets */ +int send_symbol_ex(void *handle, + enum xran_pkt_dir direction, + uint16_t section_id, + struct rte_mbuf *mb, uint8_t *data, + uint8_t compMeth, uint8_t iqWidth, + const enum xran_input_byte_order iq_buf_byte_order, + uint8_t frame_id, uint8_t subframe_id, + uint8_t slot_id, uint8_t symbol_no, + int prb_start, int prb_num, + uint8_t CC_ID, uint8_t RU_Port_ID, uint8_t seq_id) +{ + uint32_t do_copy = 0; + int32_t n_bytes; + int hdr_len, parm_size; + int32_t sent=0; + struct xran_device_ctx *p_dev_ctx = (struct xran_device_ctx *)handle; + struct xran_common_counters *pCnt = &p_dev_ctx->fh_counters; + enum xran_comp_hdr_type staticEn= XRAN_COMP_HDR_TYPE_DYNAMIC; + + + if (p_dev_ctx != NULL) + { + staticEn = p_dev_ctx->fh_cfg.ru_conf.xranCompHdrType; + + hdr_len = sizeof(struct xran_ecpri_hdr) + + sizeof(struct radio_app_common_hdr) + + sizeof(struct data_section_hdr); + if ((compMeth != XRAN_COMPMETHOD_NONE)&&(staticEn == XRAN_COMP_HDR_TYPE_DYNAMIC)) + hdr_len += sizeof(struct data_section_compression_hdr); + + switch(compMeth) { + case XRAN_COMPMETHOD_BLKFLOAT: parm_size = 1; break; + case XRAN_COMPMETHOD_MODULATION: parm_size = 0; break; + default: + parm_size = 0; + } + n_bytes = (3 * iqWidth + parm_size) * prb_num; if (mb == NULL){ char * pChar = NULL; @@ -350,25 +1219,36 @@ int send_symbol_ex(enum xran_pkt_dir direction, MLogPrint(NULL); errx(1, "out of mbufs after %d packets", 1); } - pChar = rte_pktmbuf_append(mb, sizeof(struct xran_ecpri_hdr)+ sizeof(struct radio_app_common_hdr)+ sizeof(struct data_section_hdr) + n_bytes); + pChar = rte_pktmbuf_append(mb, hdr_len + n_bytes); if(pChar == NULL){ MLogPrint(NULL); errx(1, "incorrect mbuf size %d packets", 1); } - pChar = rte_pktmbuf_prepend(mb, sizeof(struct ether_hdr)); + pChar = rte_pktmbuf_prepend(mb, sizeof(struct rte_ether_hdr)); if(pChar == NULL){ MLogPrint(NULL); errx(1, "incorrect mbuf size %d packets", 1); } do_copy = 1; /* new mbuf hence copy of IQs */ - }else { + + /**copy prach data start**/ + pChar = rte_pktmbuf_mtod(mb, char*); + char *pdata_start = (pChar + sizeof(struct rte_ether_hdr) + hdr_len); + memcpy(pdata_start,data,n_bytes); + /**copy prach data end**/ + + + } + else { rte_pktmbuf_refcnt_update(mb, 1); /* make sure eth won't free our mbuf */ } - int32_t sent = prepare_symbol_ex(direction, + sent = prepare_symbol_ex(direction, section_id, mb, data, + compMeth, + iqWidth, iq_buf_byte_order, frame_id, subframe_id, @@ -379,21 +1259,21 @@ int send_symbol_ex(enum xran_pkt_dir direction, CC_ID, RU_Port_ID, seq_id, - do_copy); + do_copy, + staticEn, + 1, + 0); /*Send a single section */ if(sent){ - tx_counter++; - xran_ethdi_mbuf_send(mb, ETHER_TYPE_ECPRI); - } else { - + pCnt->tx_counter++; + pCnt->tx_bytes_counter += rte_pktmbuf_pkt_len(mb); + p_dev_ctx->send_upmbuf2ring(mb, ETHER_TYPE_ECPRI, xran_map_ecpriPcid_to_vf(p_dev_ctx, direction, CC_ID, RU_Port_ID)); } - - #ifdef DEBUG printf("Symbol %2d sent (%d packets, %d bytes)\n", symbol_no, i, n_bytes); #endif - + } return sent; } @@ -401,31 +1281,34 @@ int send_cpmsg(void *pHandle, struct rte_mbuf *mbuf,struct xran_cp_gen_params *p struct xran_section_gen_info *sect_geninfo, uint8_t cc_id, uint8_t ru_port_id, uint8_t seq_id) { int ret = 0, nsection, i; - uint8_t frame_id = params->hdr.frameId; uint8_t subframe_id = params->hdr.subframeId; uint8_t slot_id = params->hdr.slotId; uint8_t dir = params->dir; + struct xran_device_ctx *p_dev_ctx =(struct xran_device_ctx *) pHandle; + struct xran_common_counters *pCnt = &p_dev_ctx->fh_counters; nsection = params->numSections; /* add in the ethernet header */ - struct ether_hdr *const h = (void *)rte_pktmbuf_prepend(mbuf, sizeof(*h)); - xran_ethdi_mbuf_send_cp(mbuf, ETHER_TYPE_ECPRI); - tx_counter++; + struct rte_ether_hdr *const h = (void *)rte_pktmbuf_prepend(mbuf, sizeof(*h)); + + pCnt->tx_counter++; + pCnt->tx_bytes_counter += rte_pktmbuf_pkt_len(mbuf); + p_dev_ctx->send_cpmbuf2ring(mbuf, ETHER_TYPE_ECPRI, xran_map_ecpriRtcid_to_vf(p_dev_ctx, dir, cc_id, ru_port_id)); for(i=0; iinterval_us_local))%XRAN_MAX_SECTIONDB_CTX, + sect_geninfo[i].info); return (ret); } int generate_cpmsg_dlul(void *pHandle, struct xran_cp_gen_params *params, struct xran_section_gen_info *sect_geninfo, struct rte_mbuf *mbuf, enum xran_pkt_dir dir, uint8_t frame_id, uint8_t subframe_id, uint8_t slot_id, - uint8_t startsym, uint8_t numsym, uint16_t prb_start, uint16_t prb_num, - uint16_t beam_id, uint8_t cc_id, uint8_t ru_port_id, uint8_t comp_method, uint8_t seq_id, uint8_t symInc) + uint8_t startsym, uint8_t numsym, uint16_t prb_start, uint16_t prb_num,int16_t iq_buffer_offset, int16_t iq_buffer_len, + uint16_t beam_id, uint8_t cc_id, uint8_t ru_port_id, uint8_t comp_method, uint8_t iqWidth, uint8_t seq_id, uint8_t symInc) { - int ret = 0, nsection, i; + int ret = 0, nsection, loc_sym; params->dir = dir; @@ -435,24 +1318,29 @@ int generate_cpmsg_dlul(void *pHandle, struct xran_cp_gen_params *params, struct params->hdr.subframeId = subframe_id; params->hdr.slotId = slot_id; params->hdr.startSymId = startsym; // start Symbol ID - params->hdr.iqWidth = xran_get_conf_iqwidth(pHandle); + params->hdr.iqWidth = iqWidth; params->hdr.compMeth = comp_method; nsection = 0; - sect_geninfo[nsection].info.type = params->sectionType; // for database - sect_geninfo[nsection].info.startSymId = params->hdr.startSymId; // for database - sect_geninfo[nsection].info.iqWidth = params->hdr.iqWidth; // for database - sect_geninfo[nsection].info.compMeth = params->hdr.compMeth; // for database - sect_geninfo[nsection].info.id = xran_alloc_sectionid(pHandle, dir, cc_id, ru_port_id, slot_id); - sect_geninfo[nsection].info.rb = XRAN_RBIND_EVERY; - sect_geninfo[nsection].info.symInc = symInc; - sect_geninfo[nsection].info.startPrbc = prb_start; - sect_geninfo[nsection].info.numPrbc = prb_num; - sect_geninfo[nsection].info.numSymbol = numsym; - sect_geninfo[nsection].info.reMask = 0xfff; - sect_geninfo[nsection].info.beamId = beam_id; - - sect_geninfo[nsection].info.ef = 0; + sect_geninfo[nsection].info->type = params->sectionType; // for database + sect_geninfo[nsection].info->startSymId = params->hdr.startSymId; // for database + sect_geninfo[nsection].info->iqWidth = params->hdr.iqWidth; // for database + sect_geninfo[nsection].info->compMeth = params->hdr.compMeth; // for database + sect_geninfo[nsection].info->id = xran_alloc_sectionid(pHandle, dir, cc_id, ru_port_id, subframe_id, slot_id); + sect_geninfo[nsection].info->rb = XRAN_RBIND_EVERY; + sect_geninfo[nsection].info->symInc = symInc; + sect_geninfo[nsection].info->startPrbc = prb_start; + sect_geninfo[nsection].info->numPrbc = prb_num; + sect_geninfo[nsection].info->numSymbol = numsym; + sect_geninfo[nsection].info->reMask = 0xfff; + sect_geninfo[nsection].info->beamId = beam_id; + + for (loc_sym = 0; loc_sym < XRAN_NUM_OF_SYMBOL_PER_SLOT; loc_sym++) { + sect_geninfo[0].info->sec_desc[loc_sym].iq_buffer_offset = iq_buffer_offset; + sect_geninfo[0].info->sec_desc[loc_sym].iq_buffer_len = iq_buffer_len; + } + + sect_geninfo[nsection].info->ef = 0; sect_geninfo[nsection].exDataSize = 0; // sect_geninfo[nsection].exData = NULL; nsection++; @@ -465,7 +1353,7 @@ int generate_cpmsg_dlul(void *pHandle, struct xran_cp_gen_params *params, struct return (-1); } - ret = xran_prepare_ctrl_pkt(mbuf, params, cc_id, ru_port_id, seq_id); + ret = xran_prepare_ctrl_pkt(mbuf, params, cc_id, ru_port_id, seq_id,0); if(ret < 0){ print_err("Fail to build control plane packet - [%d:%d:%d] dir=%d\n", frame_id, subframe_id, slot_id, dir); @@ -476,13 +1364,28 @@ int generate_cpmsg_dlul(void *pHandle, struct xran_cp_gen_params *params, struct } int generate_cpmsg_prach(void *pHandle, struct xran_cp_gen_params *params, struct xran_section_gen_info *sect_geninfo, struct rte_mbuf *mbuf, struct xran_device_ctx *pxran_lib_ctx, - uint8_t frame_id, uint8_t subframe_id, uint8_t slot_id, - uint16_t beam_id, uint8_t cc_id, uint8_t prach_port_id, uint8_t seq_id) + uint8_t frame_id, uint8_t subframe_id, uint8_t slot_id, int tti, + uint16_t beam_id, uint8_t cc_id, uint8_t prach_port_id, uint16_t occasionid, uint8_t seq_id) { - int i, nsection, ret; - xRANPrachCPConfigStruct *pPrachCPConfig = &(pxran_lib_ctx->PrachCPConfig); + int nsection, ret; + struct xran_prach_cp_config *pPrachCPConfig = NULL;; + int i=0; + if(pxran_lib_ctx->dssEnable){ + i = tti % pxran_lib_ctx->dssPeriod; + if(pxran_lib_ctx->technology[i]==1) { + pPrachCPConfig = &(pxran_lib_ctx->PrachCPConfig); + } + else + { + pPrachCPConfig = &(pxran_lib_ctx->PrachCPConfigLTE); + } + } + else + pPrachCPConfig = &(pxran_lib_ctx->PrachCPConfig); + uint16_t timeOffset; uint16_t nNumerology = pxran_lib_ctx->fh_cfg.frame_conf.nNumerology; + uint8_t startSymId; if(unlikely(mbuf == NULL)) { print_err("Alloc fail!\n"); @@ -500,46 +1403,81 @@ int generate_cpmsg_prach(void *pHandle, struct xran_cp_gen_params *params, struc pPrachCPConfig->freqOffset); #endif timeOffset = pPrachCPConfig->timeOffset; //this is the CP value per 38.211 tab 6.3.3.1-1&2 + startSymId = pPrachCPConfig->startSymId + occasionid * pPrachCPConfig->numSymbol; + if (startSymId > 0) + { + timeOffset += startSymId * (2048 + 144); + } + + if(XRAN_FILTERINDEX_PRACH_ABC == pPrachCPConfig->filterIdx) + { timeOffset = timeOffset >> nNumerology; //original number is Tc, convert to Ts based on mu - if (pPrachCPConfig->startSymId > 0) + if ((slot_id == 0) || (slot_id == (SLOTNUM_PER_SUBFRAME(pxran_lib_ctx->interval_us_local) >> 1))) + timeOffset += 16; + } + else { - timeOffset += (pPrachCPConfig->startSymId * 2048) >> nNumerology; - if ((slot_id == 0) || (slot_id == (SLOTNUM_PER_SUBFRAME >> 1))) - timeOffset += 16; + //when prach scs lower than 15khz, timeOffset base 15khz not need to adjust. } + params->dir = XRAN_DIR_UL; params->sectionType = XRAN_CP_SECTIONTYPE_3; params->hdr.filterIdx = pPrachCPConfig->filterIdx; params->hdr.frameId = frame_id; params->hdr.subframeId = subframe_id; params->hdr.slotId = slot_id; - params->hdr.startSymId = pPrachCPConfig->startSymId; - params->hdr.iqWidth = xran_get_conf_iqwidth(pHandle); - params->hdr.compMeth = xran_get_conf_compmethod(pHandle); + params->hdr.startSymId = startSymId; + params->hdr.iqWidth = xran_get_conf_iqwidth_prach(pHandle); + params->hdr.compMeth = xran_get_conf_compmethod_prach(pHandle); /* use timeOffset field for the CP length value for prach sequence */ params->hdr.timeOffset = timeOffset; params->hdr.fftSize = xran_get_conf_fftsize(pHandle); + /*convert to o-ran ecpri specs scs index*/ + switch(pPrachCPConfig->filterIdx) + { + case XRAN_FILTERINDEX_PRACH_012: + params->hdr.scs = 12; + break; + case XRAN_FILTERINDEX_NPRACH: + params->hdr.scs = 13; + break; + case XRAN_FILTERINDEX_PRACH_3: + params->hdr.scs = 14; + break; + case XRAN_FILTERINDEX_LTE4: + params->hdr.scs = 15; + break; + case XRAN_FILTERINDEX_PRACH_ABC: params->hdr.scs = xran_get_conf_prach_scs(pHandle); + break; + default: + print_err("prach filterIdx error - [%d:%d:%d]--%d\n", frame_id, subframe_id, slot_id,pPrachCPConfig->filterIdx); + params->hdr.scs = 0; + break; + } params->hdr.cpLength = 0; nsection = 0; - sect_geninfo[nsection].info.type = params->sectionType; // for database - sect_geninfo[nsection].info.startSymId = params->hdr.startSymId; // for database - sect_geninfo[nsection].info.iqWidth = params->hdr.iqWidth; // for database - sect_geninfo[nsection].info.compMeth = params->hdr.compMeth; // for database - sect_geninfo[nsection].info.id = xran_alloc_sectionid(pHandle, XRAN_DIR_UL, cc_id, prach_port_id, slot_id); - sect_geninfo[nsection].info.rb = XRAN_RBIND_EVERY; - sect_geninfo[nsection].info.symInc = XRAN_SYMBOLNUMBER_NOTINC; - sect_geninfo[nsection].info.startPrbc = pPrachCPConfig->startPrbc; - sect_geninfo[nsection].info.numPrbc = pPrachCPConfig->numPrbc, - sect_geninfo[nsection].info.numSymbol = pPrachCPConfig->numSymbol*pPrachCPConfig->occassionsInPrachSlot; - sect_geninfo[nsection].info.reMask = 0xfff; - sect_geninfo[nsection].info.beamId = beam_id; - sect_geninfo[nsection].info.freqOffset = pPrachCPConfig->freqOffset; - - pxran_lib_ctx->prach_last_symbol[cc_id] = sect_geninfo[nsection].info.startSymId + sect_geninfo[nsection].info.numSymbol - 1; - - sect_geninfo[nsection].info.ef = 0; + sect_geninfo[nsection].info->type = params->sectionType; // for database + sect_geninfo[nsection].info->startSymId = params->hdr.startSymId; // for database + sect_geninfo[nsection].info->iqWidth = params->hdr.iqWidth; // for database + sect_geninfo[nsection].info->compMeth = params->hdr.compMeth; // for database + sect_geninfo[nsection].info->id = xran_alloc_sectionid(pHandle, XRAN_DIR_UL, cc_id, prach_port_id, subframe_id, slot_id); + sect_geninfo[nsection].info->rb = XRAN_RBIND_EVERY; + sect_geninfo[nsection].info->symInc = XRAN_SYMBOLNUMBER_NOTINC; + sect_geninfo[nsection].info->startPrbc = pPrachCPConfig->startPrbc; + sect_geninfo[nsection].info->numPrbc = pPrachCPConfig->numPrbc, + sect_geninfo[nsection].info->numSymbol = pPrachCPConfig->numSymbol; + sect_geninfo[nsection].info->reMask = 0xfff; + sect_geninfo[nsection].info->beamId = beam_id; + sect_geninfo[nsection].info->freqOffset = pPrachCPConfig->freqOffset; + sect_geninfo[nsection].info->prbElemBegin = 1; + sect_geninfo[nsection].info->prbElemEnd = 1; + + + pxran_lib_ctx->prach_last_symbol[cc_id] = pPrachCPConfig->startSymId + pPrachCPConfig->numSymbol*pPrachCPConfig->occassionsInPrachSlot - 1; + + sect_geninfo[nsection].info->ef = 0; sect_geninfo[nsection].exDataSize = 0; // sect_geninfo[nsection].exData = NULL; nsection++; @@ -547,7 +1485,7 @@ int generate_cpmsg_prach(void *pHandle, struct xran_cp_gen_params *params, struc params->numSections = nsection; params->sections = sect_geninfo; - ret = xran_prepare_ctrl_pkt(mbuf, params, cc_id, prach_port_id, seq_id); + ret = xran_prepare_ctrl_pkt(mbuf, params, cc_id, prach_port_id, seq_id,0); if(ret < 0){ print_err("Fail to build prach control packet - [%d:%d:%d]\n", frame_id, subframe_id, slot_id); rte_pktmbuf_free(mbuf); @@ -556,72 +1494,131 @@ int generate_cpmsg_prach(void *pHandle, struct xran_cp_gen_params *params, struc } -int process_ring(struct rte_ring *r) +int process_ring(struct rte_ring *r, uint16_t ring_id, uint16_t q_id) { assert(r); struct rte_mbuf *mbufs[MBUFS_CNT]; - int i; uint32_t remaining; - uint64_t t1; + //uint64_t t1; const uint16_t dequeued = rte_ring_dequeue_burst(r, (void **)mbufs, RTE_DIM(mbufs), &remaining); if (!dequeued) return 0; - t1 = MLogTick(); - for (i = 0; i < dequeued; ++i) { - if (xran_ethdi_filter_packet(mbufs[i], 0) == MBUF_FREE) - rte_pktmbuf_free(mbufs[i]); - } - MLogTask(PID_PROCESS_UP_PKT, t1, MLogTick()); + //t1 = MLogTick(); + + xran_ethdi_filter_packet(mbufs, ring_id, q_id, dequeued); + //MLogTask(PID_PROCESS_UP_PKT, t1, MLogTick()); return remaining; } - -int32_t ring_processing_func(void) +/** FH RX AND BBDEV */ +int32_t ring_processing_func(void* args) { struct xran_ethdi_ctx *const ctx = xran_ethdi_get_ctx(); - struct xran_device_ctx *const pxran_lib_ctx = xran_dev_get_ctx(); + int16_t retPoll = 0; + int32_t i; + queueid_t qi; + uint64_t t1, t2; rte_timer_manage(); - /* UP first */ - if (process_ring(ctx->rx_ring[ETHDI_UP_VF])) - return 0; - /* CP next */ - if (process_ring(ctx->rx_ring[ETHDI_CP_VF])) - return 0; + if (ctx->bbdev_dec) { + t1 = MLogXRANTick(); + retPoll = ctx->bbdev_dec(); + if (retPoll == 1) + { + t2 = MLogXRANTick(); + MLogXRANTask(PID_XRAN_BBDEV_UL_POLL + retPoll, t1, t2); + } + } - if (pxran_lib_ctx->bbdev_dec) - pxran_lib_ctx->bbdev_dec(); + if (ctx->bbdev_enc) { + t1 = MLogXRANTick(); + retPoll = ctx->bbdev_enc(); + if (retPoll == 1) + { + t2 = MLogXRANTick(); + MLogXRANTask(PID_XRAN_BBDEV_DL_POLL + retPoll, t1, t2); + } + } - if (pxran_lib_ctx->bbdev_enc) - pxran_lib_ctx->bbdev_enc(); + for (i = 0; i < ctx->io_cfg.num_vfs && i < XRAN_VF_MAX; i++){ + for(qi = 0; qi < ctx->rxq_per_port[i]; qi++) { + if (process_ring(ctx->rx_ring[i][qi], i, qi)) + return 0; + } + } if (XRAN_STOPPED == xran_if_current_state) return -1; + return 0; + } + +/** Generic thread to perform task on specific core */ +int32_t +xran_generic_worker_thread(void *args) +{ + int32_t res = 0; + struct xran_worker_th_ctx* pThCtx = (struct xran_worker_th_ctx*)args; + struct sched_param sched_param; + struct xran_io_cfg * const p_io_cfg = &(xran_ethdi_get_ctx()->io_cfg); + + memset(&sched_param, 0, sizeof(struct sched_param)); + + printf("%s [CPU %2d] [PID: %6d]\n", __FUNCTION__, rte_lcore_id(), getpid()); + sched_param.sched_priority = XRAN_THREAD_DEFAULT_PRIO; + if ((res = pthread_setschedparam(pthread_self(), SCHED_FIFO, &sched_param))){ + printf("priority is not changed: coreId = %d, result1 = %d\n",rte_lcore_id(), res); + } + pThCtx->worker_policy = SCHED_FIFO; + if ((res = pthread_setname_np(pthread_self(), pThCtx->worker_name))) { + printf("[core %d] pthread_setname_np = %d\n",rte_lcore_id(), res); + } + + for (;;) { + if(pThCtx->task_func) { + if(pThCtx->task_func(pThCtx->task_arg) != 0) + break; + } + + if (XRAN_STOPPED == xran_if_current_state) + return -1; + + if(p_io_cfg->io_sleep) + nanosleep(&sleeptime,NULL); + } + + printf("%s worker thread finished on core %d [worker id %d]\n",pThCtx->worker_name, rte_lcore_id(), pThCtx->worker_id); return 0; } int ring_processing_thread(void *args) { struct sched_param sched_param; + struct xran_io_cfg * const p_io_cfg = &(xran_ethdi_get_ctx()->io_cfg); int res = 0; + memset(&sched_param, 0, sizeof(struct sched_param)); + printf("%s [CPU %2d] [PID: %6d]\n", __FUNCTION__, rte_lcore_id(), getpid()); sched_param.sched_priority = XRAN_THREAD_DEFAULT_PRIO; if ((res = pthread_setschedparam(pthread_self(), SCHED_FIFO, &sched_param))){ printf("priority is not changed: coreId = %d, result1 = %d\n",rte_lcore_id(), res); } - for (;;) - if(ring_processing_func() != 0) + for (;;){ + if(ring_processing_func(args) != 0) break; + /* work around for some kernel */ + if(p_io_cfg->io_sleep) + nanosleep(&sleeptime,NULL); + } + puts("Pkt processing thread finished."); return 0; } -