X-Git-Url: https://gerrit.o-ran-sc.org/r/gitweb?p=o-du%2Fphy.git;a=blobdiff_plain;f=fhi_lib%2Flib%2Fsrc%2Fxran_bfp_cplane8_snc.cpp;fp=fhi_lib%2Flib%2Fsrc%2Fxran_bfp_cplane8_snc.cpp;h=ebc2ab1127012d46bca3d4084fab797255b7e96e;hp=0000000000000000000000000000000000000000;hb=2de97529a4c5a1922214ba0e6f0fb84cacbd0bc7;hpb=81a09690b36b3a4e89b4dae34f30933de13f7f90 diff --git a/fhi_lib/lib/src/xran_bfp_cplane8_snc.cpp b/fhi_lib/lib/src/xran_bfp_cplane8_snc.cpp new file mode 100644 index 0000000..ebc2ab1 --- /dev/null +++ b/fhi_lib/lib/src/xran_bfp_cplane8_snc.cpp @@ -0,0 +1,452 @@ +/****************************************************************************** +* +* Copyright (c) 2020 Intel. +* +* Licensed under the Apache License, Version 2.0 (the "License"); +* you may not use this file except in compliance with the License. +* You may obtain a copy of the License at +* +* http://www.apache.org/licenses/LICENSE-2.0 +* +* Unless required by applicable law or agreed to in writing, software +* distributed under the License is distributed on an "AS IS" BASIS, +* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +* See the License for the specific language governing permissions and +* limitations under the License. +* +*******************************************************************************/ + +/** + * @brief xRAN BFP compression/decompression for C-plane with 8T8R + * + * @file xran_bfp_cplane8.cpp + * @ingroup group_source_xran + * @author Intel Corporation + **/ + +#include "xran_compression.hpp" +#include "xran_bfp_utils.hpp" +#include "xran_bfp_byte_packing_utils.hpp" +#include +#include +#include + + +namespace BFP_CPlane_8_SNC +{ + /// Namespace constants + const int k_numDataElements = 16; /// 16 IQ pairs + + inline __m512i + maxAbsOneReg(const __m512i maxAbs, const __m512i* inData, const int pairNum) + { + /// Compute abs of input data + const auto thisRegAbs = _mm512_abs_epi16(*inData); + /// Swap each IQ pair in each lane (via 32b rotation) and compute max of + /// each pair. + const auto maxRot16 = _mm512_rol_epi32(thisRegAbs, BlockFloatCompander::k_numBitsIQ); + const auto maxAbsIQ = _mm512_max_epi16(thisRegAbs, maxRot16); + /// Convert to 32b values + const auto maxAbsIQ32 = BlockFloatCompander::maskUpperWord(maxAbsIQ); + /// Swap 32b in each 64b chunk via rotation and compute 32b max + /// Results in blocks of 64b with 4 repeated 16b max values + const auto maxRot32 = _mm512_rol_epi64(maxAbsIQ32, BlockFloatCompander::k_numBitsIQPair); + const auto maxAbs32 = _mm512_max_epi32(maxAbsIQ32, maxRot32); + /// First 64b permute and max + /// Results in blocks of 128b with 8 repeated 16b max values + constexpr uint8_t k_perm64A = 0xB1; + const auto maxPerm64A = _mm512_permutex_epi64(maxAbs32, k_perm64A); + const auto maxAbs64 = _mm512_max_epi64(maxAbs32, maxPerm64A); + /// Second 64b permute and max + /// Results in blocks of 256b with 16 repeated 16b max values + constexpr uint8_t k_perm64B = 0x4E; + const auto maxPerm64B = _mm512_permutex_epi64(maxAbs64, k_perm64B); + const auto maxAbs128 = _mm512_max_epi64(maxAbs64, maxPerm64B); + /// Now register contains repeated max values for two compression blocks + /// Permute the desired results into maxAbs + const auto k_selectVals = _mm512_set_epi32(24, 16, 24, 16, 24, 16, 24, 16, + 24, 16, 24, 16, 24, 16, 24, 16); + constexpr uint16_t k_2ValsMsk[8] = { 0x0003, 0x000C, 0x0030, 0x00C0, 0x0300, 0x0C00, 0x3000, 0xC000 }; + return _mm512_mask_permutex2var_epi32(maxAbs, k_2ValsMsk[pairNum], k_selectVals, maxAbs128); + } + + /// Compute exponent value for a set of 16 RB from the maximum absolute value. + inline __m512i + computeExponent_16RB(const BlockFloatCompander::ExpandedData& dataIn, const __m512i totShiftBits) + { + __m512i maxAbs = __m512i(); + const __m512i* dataInAddr = reinterpret_cast(dataIn.dataExpanded); +#pragma unroll(8) + for (int n = 0; n < 8; ++n) + { + maxAbs = maxAbsOneReg(maxAbs, dataInAddr + n, n); + } + /// Calculate exponent + return BlockFloatCompander::expLzCnt(maxAbs, totShiftBits); + } + + /// Compute exponent value for a set of 4 RB from the maximum absolute value. + inline __m512i + computeExponent_4RB(const BlockFloatCompander::ExpandedData& dataIn, const __m512i totShiftBits) + { + __m512i maxAbs = __m512i(); + const __m512i* dataInAddr = reinterpret_cast(dataIn.dataExpanded); +#pragma unroll(2) + for (int n = 0; n < 2; ++n) + { + maxAbs = maxAbsOneReg(maxAbs, dataInAddr + n, n); + } + /// Calculate exponent + return BlockFloatCompander::expLzCnt(maxAbs, totShiftBits); + } + + /// Compute exponent value for 1 RB from the maximum absolute value. + inline uint8_t + computeExponent_1RB(const BlockFloatCompander::ExpandedData& dataIn, const __m512i totShiftBits) + { + __m512i maxAbs = __m512i(); + const __m512i* dataInAddr = reinterpret_cast(dataIn.dataExpanded); + maxAbs = maxAbsOneReg(maxAbs, dataInAddr, 0); + /// Calculate exponent + const auto exps = BlockFloatCompander::expLzCnt(maxAbs, totShiftBits); + return ((uint8_t*)&exps)[0]; + } + + + + /// Apply compression to one compression block + template + inline void + applyCompressionN_1RB(const __m512i* dataIn, uint8_t* outBlockAddr, + const int iqWidth, const uint8_t thisExp, const uint32_t rbWriteMask) + { + /// Store exponents first + *outBlockAddr = thisExp; + /// Apply the exponent shift + /// First Store the two exponent values in one register + const auto compData = _mm512_srai_epi16(*dataIn, thisExp); + /// Pack compressed data network byte order + const auto compDataBytePacked = networkBytePack(compData); + /// Store compressed data + _mm256_mask_storeu_epi8(outBlockAddr + 1, rbWriteMask, _mm512_extracti64x4_epi64(compDataBytePacked, 0)); + } + + /// Apply compression to two compression blocks + template + inline void + applyCompressionN_2RB(const __m512i* dataIn, uint8_t* outBlockAddr, + const int totNumBytesPerBlock, const int iqWidth, const uint8_t* theseExps, const uint32_t rbWriteMask) + { + /// Store exponents first + *outBlockAddr = theseExps[0]; + *(outBlockAddr + totNumBytesPerBlock) = theseExps[4]; + /// Apply the exponent shift + /// First Store the two exponent values in one register + __m512i thisExp = __m512i(); + constexpr uint32_t k_firstExpMask = 0x0000FFFF; + thisExp = _mm512_mask_set1_epi16(thisExp, k_firstExpMask, theseExps[0]); + constexpr uint32_t k_secondExpMask = 0xFFFF0000; + thisExp = _mm512_mask_set1_epi16(thisExp, k_secondExpMask, theseExps[4]); + const auto compData = _mm512_srav_epi16(*dataIn, thisExp); + /// Pack compressed data network byte order + const auto compDataBytePacked = networkBytePack(compData); + /// Output of network byte packing has each compression block packed in each half register + /// Store compressed data + _mm256_mask_storeu_epi8(outBlockAddr + 1, rbWriteMask, _mm512_extracti64x4_epi64(compDataBytePacked, 0)); + _mm256_mask_storeu_epi8(outBlockAddr + totNumBytesPerBlock + 1, rbWriteMask, _mm512_extracti64x4_epi64(compDataBytePacked, 1)); + } + + /// Derive and apply 9, 10, or 12bit compression to 16 compression blocks + template + inline void + compressN_16RB(const BlockFloatCompander::ExpandedData& dataIn, BlockFloatCompander::CompressedData* dataOut, + const __m512i totShiftBits, const int totNumBytesPerBlock, const uint32_t rbWriteMask) + { + const auto exponents = computeExponent_16RB(dataIn, totShiftBits); + const __m512i* dataInAddr = reinterpret_cast(dataIn.dataExpanded); +#pragma unroll(8) + for (int n = 0; n < 8; ++n) + { + applyCompressionN_2RB(dataInAddr + n, dataOut->dataCompressed + n * 2 * totNumBytesPerBlock, totNumBytesPerBlock, dataIn.iqWidth, ((uint8_t*)&exponents) + n * 8, rbWriteMask); + } + } + + /// Derive and apply 9, 10, or 12bit compression to 4 compression blocks + template + inline void + compressN_4RB(const BlockFloatCompander::ExpandedData& dataIn, BlockFloatCompander::CompressedData* dataOut, + const __m512i totShiftBits, const int totNumBytesPerBlock, const uint32_t rbWriteMask) + { + const auto exponents = computeExponent_4RB(dataIn, totShiftBits); + const __m512i* dataInAddr = reinterpret_cast(dataIn.dataExpanded); +#pragma unroll(2) + for (int n = 0; n < 2; ++n) + { + applyCompressionN_2RB(dataInAddr + n, dataOut->dataCompressed + n * 2 * totNumBytesPerBlock, totNumBytesPerBlock, dataIn.iqWidth, ((uint8_t*)&exponents) + n * 8, rbWriteMask);; + } + } + + /// Derive and apply 9, 10, or 12bit compression to 1 RB + template + inline void + compressN_1RB(const BlockFloatCompander::ExpandedData& dataIn, BlockFloatCompander::CompressedData* dataOut, + const __m512i totShiftBits, const int totNumBytesPerBlock, const uint32_t rbWriteMask) + { + const auto thisExponent = computeExponent_1RB(dataIn, totShiftBits); + const __m512i* dataInAddr = reinterpret_cast(dataIn.dataExpanded); + applyCompressionN_1RB(dataInAddr, dataOut->dataCompressed, dataIn.iqWidth, thisExponent, rbWriteMask); + } + + /// Calls compression function specific to the number of blocks to be executed. For 9, 10, or 12bit iqWidth. + template + inline void + compressByAllocN(const BlockFloatCompander::ExpandedData& dataIn, BlockFloatCompander::CompressedData* dataOut, + const __m512i totShiftBits, const int totNumBytesPerBlock, const uint32_t rbWriteMask) + { + switch (dataIn.numBlocks) + { + case 16: + compressN_16RB(dataIn, dataOut, totShiftBits, totNumBytesPerBlock, rbWriteMask); + break; + + case 4: + compressN_4RB(dataIn, dataOut, totShiftBits, totNumBytesPerBlock, rbWriteMask); + break; + + case 1: + compressN_1RB(dataIn, dataOut, totShiftBits, totNumBytesPerBlock, rbWriteMask); + break; + } + } + + + + /// Apply 8b compression to 1 compression block. + inline void + applyCompression8_1RB(const __m256i* dataIn, uint8_t* outBlockAddr, const uint8_t thisExp) + { + /// Store exponent first + *outBlockAddr = thisExp; + /// Apply the exponent shift + const auto compData = _mm256_srai_epi16(*dataIn, thisExp); + /// Truncate to 8bit and store + constexpr uint16_t k_writeMask = 0xFFFF; + _mm_mask_storeu_epi8(outBlockAddr + 1, k_writeMask, _mm256_cvtepi16_epi8(compData)); + } + + /// Derive and apply 8b compression to 16 compression blocks + inline void + compress8_16RB(const BlockFloatCompander::ExpandedData& dataIn, BlockFloatCompander::CompressedData* dataOut, const __m512i totShiftBits) + { + const auto exponents = computeExponent_16RB(dataIn, totShiftBits); + const __m256i* dataInAddr = reinterpret_cast(dataIn.dataExpanded); +#pragma unroll(16) + for (int n = 0; n < 16; ++n) + { + applyCompression8_1RB(dataInAddr + n, dataOut->dataCompressed + n * (k_numDataElements + 1), ((uint8_t*)&exponents)[n * 4]); + } + } + + /// Derive and apply 8b compression to 4 compression blocks + inline void + compress8_4RB(const BlockFloatCompander::ExpandedData& dataIn, BlockFloatCompander::CompressedData* dataOut, const __m512i totShiftBits) + { + const auto exponents = computeExponent_4RB(dataIn, totShiftBits); + const __m256i* dataInAddr = reinterpret_cast(dataIn.dataExpanded); +#pragma unroll(4) + for (int n = 0; n < 4; ++n) + { + applyCompression8_1RB(dataInAddr + n, dataOut->dataCompressed + n * (k_numDataElements + 1), ((uint8_t*)&exponents)[n * 4]); + } + } + + /// Derive and apply 8b compression to 1 compression block + inline void + compress8_1RB(const BlockFloatCompander::ExpandedData& dataIn, BlockFloatCompander::CompressedData* dataOut, const __m512i totShiftBits) + { + const auto thisExponent = computeExponent_1RB(dataIn, totShiftBits); + const __m256i* dataInAddr = reinterpret_cast(dataIn.dataExpanded); + applyCompression8_1RB(dataInAddr, dataOut->dataCompressed, thisExponent); + } + + /// Calls compression function specific to the number of RB to be executed. For 8 bit iqWidth. + inline void + compressByAlloc8(const BlockFloatCompander::ExpandedData& dataIn, BlockFloatCompander::CompressedData* dataOut, const __m512i totShiftBits) + { + switch (dataIn.numBlocks) + { + case 16: + compress8_16RB(dataIn, dataOut, totShiftBits); + break; + + case 4: + compress8_4RB(dataIn, dataOut, totShiftBits); + break; + + case 1: + compress8_1RB(dataIn, dataOut, totShiftBits); + break; + } + } + + + /// Expand 1 compression block + template + inline void + applyExpansionN_1RB(const uint8_t* expAddr, __m256i* dataOutAddr, const int maxExpShift) + { + const auto thisExpShift = maxExpShift - *expAddr; + /// Unpack network order packed data + const auto inDataUnpacked = networkByteUnpack(expAddr + 1); + /// Apply exponent scaling (by appropriate arithmetic shift right) + const auto expandedData = _mm256_srai_epi16(inDataUnpacked, thisExpShift); + /// Write expanded data to output + static constexpr uint8_t k_WriteMask = 0x0F; + _mm256_mask_storeu_epi64(dataOutAddr, k_WriteMask, expandedData); + } + + /// Calls expansion function specific to the number of blocks to be executed. For 9, 10, or 12bit iqWidth. + template + void expandByAllocN(const BlockFloatCompander::CompressedData& dataIn, BlockFloatCompander::ExpandedData* dataOut, + const int totNumBytesPerBlock, const int maxExpShift) + { + __m256i* dataOutAddr = reinterpret_cast<__m256i*>(dataOut->dataExpanded); + switch (dataIn.numBlocks) + { + case 16: +#pragma unroll(16) + for (int n = 0; n < 16; ++n) + { + applyExpansionN_1RB(dataIn.dataCompressed + n * totNumBytesPerBlock, dataOutAddr + n, maxExpShift); + } + break; + + case 4: +#pragma unroll(4) + for (int n = 0; n < 4; ++n) + { + applyExpansionN_1RB(dataIn.dataCompressed + n * totNumBytesPerBlock, dataOutAddr + n, maxExpShift); + } + break; + + case 1: + applyExpansionN_1RB(dataIn.dataCompressed, dataOutAddr, maxExpShift); + break; + } + } + + + /// Apply expansion to 2 compression block + inline void + applyExpansion8_1RB(const uint8_t* expAddr, __m256i* dataOutAddr) + { + const __m128i* rawDataIn = reinterpret_cast(expAddr + 1); + const auto compData16 = _mm256_cvtepi8_epi16(*rawDataIn); + const auto expData = _mm256_slli_epi16(compData16, *expAddr); + static constexpr uint8_t k_WriteMask = 0x0F; + _mm256_mask_storeu_epi64(dataOutAddr, k_WriteMask, expData); + } + + /// Calls expansion function specific to the number of RB to be executed. For 8 bit iqWidth. + void + expandByAlloc8(const BlockFloatCompander::CompressedData& dataIn, BlockFloatCompander::ExpandedData* dataOut) + { + __m256i* dataOutAddr = reinterpret_cast<__m256i*>(dataOut->dataExpanded); + switch (dataIn.numBlocks) + { + case 16: +#pragma unroll(16) + for (int n = 0; n < 16; ++n) + { + applyExpansion8_1RB(dataIn.dataCompressed + n * (k_numDataElements + 1), dataOutAddr + n); + } + break; + + case 4: +#pragma unroll(4) + for (int n = 0; n < 4; ++n) + { + applyExpansion8_1RB(dataIn.dataCompressed + n * (k_numDataElements + 1), dataOutAddr + n); + } + break; + + case 1: + applyExpansion8_1RB(dataIn.dataCompressed, dataOutAddr); + break; + } + } +} + + +/// Main kernel function for 8 antenna C-plane compression. +/// Starts by determining iqWidth specific parameters and functions. +void +BlockFloatCompander::BFPCompressCtrlPlane8AvxSnc(const ExpandedData& dataIn, CompressedData* dataOut) +{ + /// Compensation for extra zeros in 32b leading zero count when computing exponent + const auto totShiftBits8 = _mm512_set1_epi32(25); + const auto totShiftBits9 = _mm512_set1_epi32(24); + const auto totShiftBits10 = _mm512_set1_epi32(23); + const auto totShiftBits12 = _mm512_set1_epi32(21); + + /// Total number of data bytes per compression block is (iqWidth * numElements / 8) + 1 + const auto totNumBytesPerBlock = ((BFP_CPlane_8_SNC::k_numDataElements * dataIn.iqWidth) >> 3) + 1; + + /// Compressed data write mask for each iqWidth option + /// Compressed data write mask for each iqWidth option + constexpr uint32_t rbWriteMask9 = 0x0003FFFF; + constexpr uint32_t rbWriteMask10 = 0x000FFFFF; + constexpr uint32_t rbWriteMask12 = 0x00FFFFFF; + + switch (dataIn.iqWidth) + { + case 8: + BFP_CPlane_8_SNC::compressByAlloc8(dataIn, dataOut, totShiftBits8); + break; + + case 9: + BFP_CPlane_8_SNC::compressByAllocN(dataIn, dataOut, totShiftBits9, totNumBytesPerBlock, rbWriteMask9); + break; + + case 10: + BFP_CPlane_8_SNC::compressByAllocN(dataIn, dataOut, totShiftBits10, totNumBytesPerBlock, rbWriteMask10); + break; + + case 12: + BFP_CPlane_8_SNC::compressByAllocN(dataIn, dataOut, totShiftBits12, totNumBytesPerBlock, rbWriteMask12); + break; + } +} + + +/// Main kernel function for 8 antenna C-plane expansion. +/// Starts by determining iqWidth specific parameters and functions. +void +BlockFloatCompander::BFPExpandCtrlPlane8AvxSnc(const CompressedData& dataIn, ExpandedData* dataOut) +{ + constexpr int k_maxExpShift9 = 7; + constexpr int k_maxExpShift10 = 6; + constexpr int k_maxExpShift12 = 4; + + /// Total number of data bytes per compression block is (iqWidth * numElements / 8) + 1 + const auto totNumBytesPerBlock = ((BFP_CPlane_8_SNC::k_numDataElements * dataIn.iqWidth) >> 3) + 1; + + switch (dataIn.iqWidth) + { + case 8: + BFP_CPlane_8_SNC::expandByAlloc8(dataIn, dataOut); + break; + + case 9: + BFP_CPlane_8_SNC::expandByAllocN(dataIn, dataOut, totNumBytesPerBlock, k_maxExpShift9); + break; + + case 10: + BFP_CPlane_8_SNC::expandByAllocN(dataIn, dataOut, totNumBytesPerBlock, k_maxExpShift10); + break; + + case 12: + BFP_CPlane_8_SNC::expandByAllocN(dataIn, dataOut, totNumBytesPerBlock, k_maxExpShift12); + break; + } +}