X-Git-Url: https://gerrit.o-ran-sc.org/r/gitweb?a=blobdiff_plain;f=RIC-E2-TERMINATION%2FsctpThread.cpp;h=987bbafb0c1fbcaeee8211d7882de58369ca733e;hb=refs%2Fchanges%2F26%2F4226%2F1;hp=4bd7ffcf9c444db6c8e4280de203bf692f585996;hpb=84bd3343061399a2af06a34d384c56c34758d306;p=ric-plt%2Fe2.git diff --git a/RIC-E2-TERMINATION/sctpThread.cpp b/RIC-E2-TERMINATION/sctpThread.cpp index 4bd7ffc..987bbaf 100644 --- a/RIC-E2-TERMINATION/sctpThread.cpp +++ b/RIC-E2-TERMINATION/sctpThread.cpp @@ -20,16 +20,20 @@ +#include <3rdparty/oranE2/RANfunctions-List.h> #include "sctpThread.h" #include "BuildRunName.h" +//#include "3rdparty/oranE2SM/E2SM-gNB-NRT-RANfunction-Definition.h" +//#include "BuildXml.h" +//#include "pugixml/src/pugixml.hpp" + using namespace std; //using namespace std::placeholders; using namespace boost::filesystem; +using namespace prometheus; + -#ifdef __TRACING__ -using namespace opentracing; -#endif //#ifdef __cplusplus //extern "C" //{ @@ -81,13 +85,18 @@ std::atomic num_of_XAPP_messages{0}; static long transactionCounter = 0; int buildListeningPort(sctp_params_t &sctpParams) { - sctpParams.listenFD = socket (AF_INET6, SOCK_STREAM, IPPROTO_SCTP); + sctpParams.listenFD = socket(AF_INET6, SOCK_STREAM, IPPROTO_SCTP); + if (sctpParams.listenFD <= 0) { + mdclog_write(MDCLOG_ERR, "Error Opening socket, %s", strerror(errno)); + return -1; + } + struct sockaddr_in6 servaddr {}; servaddr.sin6_family = AF_INET6; servaddr.sin6_addr = in6addr_any; - servaddr.sin6_port = htons(SRC_PORT); + servaddr.sin6_port = htons(sctpParams.sctpPort); if (bind(sctpParams.listenFD, (SA *)&servaddr, sizeof(servaddr)) < 0 ) { - mdclog_write(MDCLOG_ERR, "Error binding. %s\n", strerror(errno)); + mdclog_write(MDCLOG_ERR, "Error binding port %d. %s", sctpParams.sctpPort, strerror(errno)); return -1; } if (setSocketNoBlocking(sctpParams.listenFD) == -1) { @@ -187,19 +196,19 @@ int buildConfiguration(sctp_params_t &sctpParams) { // define the file name in the tmp directory under the volume strcat(tmpLogFilespec,"/tmp/E2Term_%Y-%m-%d_%H-%M-%S.%N.tmpStr"); -// std::string localIP = conf.getStringValue("local-ip"); -// if (localIP.length() == 0) { -// mdclog_write(MDCLOG_ERR, "illigal local-ip. environment variable"); -// exit(-1); -// } - - //sctpParams.myIP.assign(getenv(localIP.c_str())); sctpParams.myIP = conf.getStringValue("local-ip"); if (sctpParams.myIP.length() == 0) { mdclog_write(MDCLOG_ERR, "illigal local-ip."); return -1; } + int sctpPort = conf.getIntValue("sctp-port"); + if (sctpPort == -1) { + mdclog_write(MDCLOG_ERR, "illigal SCTP port "); + return -1; + } + sctpParams.sctpPort = (uint16_t)sctpPort; + sctpParams.fqdn = conf.getStringValue("external-fqdn"); if (sctpParams.fqdn.length() == 0) { mdclog_write(MDCLOG_ERR, "illigal external-fqdn"); @@ -235,9 +244,35 @@ int buildConfiguration(sctp_params_t &sctpParams) { } jsonTrace = sctpParams.trace; - sctpParams.ka_message_length = snprintf(sctpParams.ka_message, 4096, "{\"address\": \"%s:%d\"," - "\"fqdn\": \"%s\"," - "\"pod_name\": \"%s\"}", + sctpParams.epollTimeOut = -1; + tmpStr = conf.getStringValue("prometheusMode"); + transform(tmpStr.begin(), tmpStr.end(), tmpStr.begin(), ::tolower); + sctpParams.prometheusMode = tmpStr; + if (tmpStr.length() != 0) { + if (tmpStr.compare("push") == 0) { + sctpParams.prometheusPushAddress = tmpStr; + auto timeout = conf.getIntValue("prometheusPushTimeOut"); + if (timeout >= 5 && timeout <= 300) { + sctpParams.epollTimeOut = timeout * 1000; + } else { + sctpParams.epollTimeOut = 10 * 1000; + } + } + } + + tmpStr = conf.getStringValue("prometheusPushAddr"); + if (tmpStr.length() != 0) { + sctpParams.prometheusPushAddress = tmpStr; + } + + tmpStr = conf.getStringValue("prometheusPort"); + if (tmpStr.length() != 0) { + sctpParams.prometheusPort = tmpStr; + } + + sctpParams.ka_message_length = snprintf(sctpParams.ka_message, KA_MESSAGE_SIZE, "{\"address\": \"%s:%d\"," + "\"fqdn\": \"%s\"," + "\"pod_name\": \"%s\"}", (const char *)sctpParams.myIP.c_str(), sctpParams.rmrPort, sctpParams.fqdn.c_str(), @@ -280,16 +315,53 @@ int buildConfiguration(sctp_params_t &sctpParams) { return 0; } +static std::string GetHostName() { + char hostname[1024]; + + if (::gethostname(hostname, sizeof(hostname))) { + return {}; + } + return hostname; +} + +void startPrometheus(sctp_params_t &sctpParams) { + sctpParams.prometheusFamily = &BuildCounter() + .Name("E2T") + .Help("E2T message counter") + .Labels({{"POD_NAME", sctpParams.podName}}) + .Register(*sctpParams.prometheusRegistry); + + if (strcmp(sctpParams.prometheusMode.c_str(),"pull") == 0) { + if (mdclog_level_get() >= MDCLOG_DEBUG) { + mdclog_write(MDCLOG_DEBUG, "Start Prometheus Pull mode on %s:%s", sctpParams.myIP.c_str(), sctpParams.prometheusPort.c_str()); + } + sctpParams.prometheusExposer = new Exposer(sctpParams.myIP + ":" + sctpParams.prometheusPort, 1); + sctpParams.prometheusExposer->RegisterCollectable(sctpParams.prometheusRegistry); + } else if (strcmp(sctpParams.prometheusMode.c_str(),"push") == 0) { + if (mdclog_level_get() >= MDCLOG_DEBUG) { + mdclog_write(MDCLOG_DEBUG, "Start Prometheus Push mode"); + } + const auto labels = Gateway::GetInstanceLabel(GetHostName()); + string address {}; + string port {}; + char ch = ':'; + auto found = sctpParams.prometheusPushAddress.find_last_of(ch); + // If string doesn't have + // character ch present in it + if (found != string::npos) { + address = sctpParams.prometheusPushAddress.substr(0,found); + port = sctpParams.prometheusPushAddress.substr(found + 1); + sctpParams.prometheusGateway = new Gateway(address, port, "E2T", labels); + sctpParams.prometheusGateway->RegisterCollectable(sctpParams.prometheusRegistry); + } else { + mdclog_write(MDCLOG_ERR, "failed to build Prometheus gateway no stats will be sent"); + } + } +} + int main(const int argc, char **argv) { sctp_params_t sctpParams; -#ifdef __TRACING__ - opentracing::Tracer::InitGlobal(tracelibcpp::createTracer("E2 Terminator")); - auto span = opentracing::Tracer::Global()->StartSpan(__FUNCTION__); -#else - otSpan span = 0; -#endif - { std::random_device device{}; std::mt19937 generator(device()); @@ -328,6 +400,13 @@ int main(const int argc, char **argv) { exit(-1); } + //auto registry = std::make_shared(); + sctpParams.prometheusRegistry = std::make_shared(); + + //sctpParams.promtheusFamily = new Family("E2T", "E2T message counter", {{"E", sctpParams.podName}}); + + startPrometheus(sctpParams); + // start epoll sctpParams.epoll_fd = epoll_create1(0); if (sctpParams.epoll_fd == -1) { @@ -335,7 +414,7 @@ int main(const int argc, char **argv) { exit(-1); } - getRmrContext(sctpParams, &span); + getRmrContext(sctpParams); if (sctpParams.rmrCtx == nullptr) { close(sctpParams.epoll_fd); exit(-1); @@ -373,6 +452,7 @@ int main(const int argc, char **argv) { } } + //loop over term_init until first message from xApp handleTermInit(sctpParams); @@ -380,9 +460,6 @@ int main(const int argc, char **argv) { t.join(); } -#ifdef __TRACING__ - opentracing::Tracer::Global()->Close(); -#endif return 0; } @@ -421,7 +498,7 @@ void sendTermInit(sctp_params_t &sctpParams) { rmr_bytes2xact(msg, tx, txLen); msg = rmr_send_msg(sctpParams.rmrCtx, msg); if (msg == nullptr) { - msg = rmr_alloc_msg(sctpParams.rmrCtx, sctpParams.myIP.length()); + msg = rmr_alloc_msg(sctpParams.rmrCtx, sctpParams.ka_message_length); } else if (msg->state == 0) { rmr_free_msg(msg); if (mdclog_level_get() >= MDCLOG_INFO) { @@ -430,7 +507,7 @@ void sendTermInit(sctp_params_t &sctpParams) { return; } else { if (count % 100 == 0) { - mdclog_write(MDCLOG_ERR, "Error sending E2_TERM_INIT cause : %d ", msg->state); + mdclog_write(MDCLOG_ERR, "Error sending E2_TERM_INIT cause : %s ", translateRmrErrorMessages(msg->state).c_str()); } sleep(1); } @@ -478,12 +555,12 @@ int buildInotify(sctp_params_t &sctpParams) { } sctpParams.inotifyWD = inotify_add_watch(sctpParams.inotifyFD, - (const char *)sctpParams.configFilePath.c_str(), + (const char *)sctpParams.configFilePath.c_str(), (unsigned)IN_OPEN | (unsigned)IN_CLOSE_WRITE | (unsigned)IN_CLOSE_NOWRITE); //IN_CLOSE = (IN_CLOSE_WRITE | IN_CLOSE_NOWRITE) if (sctpParams.inotifyWD == -1) { mdclog_write(MDCLOG_ERR, "Failed to add directory : %s to inotify (inotify_add_watch) %s", - sctpParams.configFilePath.c_str(), - strerror(errno)); + sctpParams.configFilePath.c_str(), + strerror(errno)); close(sctpParams.inotifyFD); return -1; } @@ -506,11 +583,6 @@ int buildInotify(sctp_params_t &sctpParams) { * @return */ void listener(sctp_params_t *params) { -#ifdef __TRACING__ - auto span = opentracing::Tracer::Global()->StartSpan(__FUNCTION__); -#else - otSpan span = 0; -#endif int num_of_SCTP_messages = 0; auto totalTime = 0.0; mdclog_mdc_clean(); @@ -535,7 +607,6 @@ void listener(sctp_params_t *params) { mdclog_write(MDCLOG_DEBUG, "started thread number %s", tid); } - RmrMessagesBuffer_t rmrMessageBuffer{}; //create and init RMR rmrMessageBuffer.rmrCtx = params->rmrCtx; @@ -557,26 +628,44 @@ void listener(sctp_params_t *params) { ReportingMessages_t message {}; - for (int i = 0; i < MAX_RMR_BUFF_ARRY; i++) { - rmrMessageBuffer.rcvBufferedMessages[i] = rmr_alloc_msg(rmrMessageBuffer.rmrCtx, RECEIVE_XAPP_BUFFER_SIZE); - rmrMessageBuffer.sendBufferedMessages[i] = rmr_alloc_msg(rmrMessageBuffer.rmrCtx, RECEIVE_XAPP_BUFFER_SIZE); - } +// for (int i = 0; i < MAX_RMR_BUFF_ARRY; i++) { +// rmrMessageBuffer.rcvBufferedMessages[i] = rmr_alloc_msg(rmrMessageBuffer.rmrCtx, RECEIVE_XAPP_BUFFER_SIZE); +// rmrMessageBuffer.sendBufferedMessages[i] = rmr_alloc_msg(rmrMessageBuffer.rmrCtx, RECEIVE_XAPP_BUFFER_SIZE); +// } + bool gatewayflag = false; while (true) { + future gateWay; + if (mdclog_level_get() >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "Start EPOLL Wait"); + mdclog_write(MDCLOG_DEBUG, "Start EPOLL Wait. Timeout = %d", params->epollTimeOut); } - auto numOfEvents = epoll_wait(params->epoll_fd, events, MAXEVENTS, -1); - if (numOfEvents < 0 && errno == EINTR) { - if (mdclog_level_get() >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "got EINTR : %s", strerror(errno)); + auto numOfEvents = epoll_wait(params->epoll_fd, events, MAXEVENTS, params->epollTimeOut); + if (numOfEvents == 0) { + if (params->prometheusGateway != nullptr) { + gateWay = params->prometheusGateway->AsyncPush(); + gatewayflag = true; } continue; - } - if (numOfEvents < 0) { + } else if (numOfEvents < 0) { + if (errno == EINTR) { + if (mdclog_level_get() >= MDCLOG_DEBUG) { + mdclog_write(MDCLOG_DEBUG, "got EINTR : %s", strerror(errno)); + } + continue; + } mdclog_write(MDCLOG_ERR, "Epoll wait failed, errno = %s", strerror(errno)); return; } + if (gatewayflag) { + gatewayflag = false; + auto rc = gateWay.get(); + if (rc != 200) { + mdclog_write(MDCLOG_ERR, "Async Send to Promethues faild with Return Code %d", rc); + } else if (mdclog_level_get() >= MDCLOG_DEBUG) { + mdclog_write(MDCLOG_DEBUG, "Stats sent to Prometheus"); + } + } for (auto i = 0; i < numOfEvents; i++) { if (mdclog_level_get() >= MDCLOG_DEBUG) { mdclog_write(MDCLOG_DEBUG, "handling epoll event %d out of %d", i + 1, numOfEvents); @@ -587,10 +676,13 @@ void listener(sctp_params_t *params) { if ((events[i].events & EPOLLERR) || (events[i].events & EPOLLHUP)) { - handlepoll_error(events[i], message, rmrMessageBuffer, params, &span); + handlepoll_error(events[i], message, rmrMessageBuffer, params); } else if (events[i].events & EPOLLOUT) { - handleEinprogressMessages(events[i], message, rmrMessageBuffer, params, &span); + handleEinprogressMessages(events[i], message, rmrMessageBuffer, params); } else if (params->listenFD == events[i].data.fd) { + if (mdclog_level_get() >= MDCLOG_INFO) { + mdclog_write(MDCLOG_INFO, "New connection request from sctp network\n"); + } // new connection is requested from RAN start build connection while (true) { struct sockaddr in_addr {}; @@ -617,8 +709,8 @@ void listener(sctp_params_t *params) { break; } auto ans = getnameinfo(&in_addr, in_len, - peerInfo->hostName, NI_MAXHOST, - peerInfo->portNumber, NI_MAXSERV, NI_NUMERICHOST | NI_NUMERICSERV); + peerInfo->hostName, NI_MAXHOST, + peerInfo->portNumber, NI_MAXSERV, (unsigned )((unsigned int)NI_NUMERICHOST | (unsigned int)NI_NUMERICSERV)); if (ans < 0) { mdclog_write(MDCLOG_ERR, "Failed to get info on connection request. %s\n", strerror(errno)); close(peerInfo->fileDescriptor); @@ -633,12 +725,10 @@ void listener(sctp_params_t *params) { peerInfo, (EPOLLIN | EPOLLET), params->sctpMap, nullptr, - 0, &span) != 0) { -#ifdef __TRACING__ - lspan->Finish(); -#endif + 0) != 0) { break; } + break; } } else if (params->rmrListenFd == events[i].data.fd) { // got message from XAPP @@ -647,11 +737,9 @@ void listener(sctp_params_t *params) { if (mdclog_level_get() >= MDCLOG_DEBUG) { mdclog_write(MDCLOG_DEBUG, "new message from RMR"); } - if (receiveXappMessages(params->epoll_fd, - params->sctpMap, + if (receiveXappMessages(params->sctpMap, rmrMessageBuffer, - message.message.time, - &span) != 0) { + message.message.time) != 0) { mdclog_write(MDCLOG_ERR, "Error handling Xapp message"); } } else if (params->inotifyFD == events[i].data.fd) { @@ -669,8 +757,7 @@ void listener(sctp_params_t *params) { params->sctpMap, num_of_SCTP_messages, rmrMessageBuffer, - message.message.time, - &span); + message.message.time); } clock_gettime(CLOCK_MONOTONIC, &end); @@ -685,11 +772,6 @@ void listener(sctp_params_t *params) { } } } -#ifdef __TRACING__ - span->Finish(); -#else - -#endif } /** @@ -728,12 +810,16 @@ void handleConfigChange(sctp_params_t *sctpParams) { // not the directory } if (event->len) { - if (!(sctpParams->configFileName.compare(event->name))) { + auto retVal = strcmp(sctpParams->configFileName.c_str(), event->name); + if (retVal != 0) { continue; } } // only the file we want if (event->mask & (uint32_t)IN_CLOSE_WRITE) { + if (mdclog_level_get() >= MDCLOG_INFO) { + mdclog_write(MDCLOG_INFO, "Configuration file changed"); + } if (exists(p)) { const int size = 2048; auto fileSize = file_size(p); @@ -797,6 +883,17 @@ void handleConfigChange(sctp_params_t *sctpParams) { sctpParams->trace = false; } jsonTrace = sctpParams->trace; + + if (strcmp(sctpParams->prometheusMode.c_str(), "push") == 0) { + auto timeout = conf.getIntValue("prometheusPushTimeOut"); + if (timeout >= 5 && timeout <= 300) { + sctpParams->epollTimeOut = timeout * 1000; + } else { + mdclog_write(MDCLOG_ERR, "prometheusPushTimeOut set wrong value %d, values are [5..300]", + timeout); + } + } + endlessLoop = false; } } @@ -809,19 +906,11 @@ void handleConfigChange(sctp_params_t *sctpParams) { * @param message * @param rmrMessageBuffer * @param params - * @param pSpan */ void handleEinprogressMessages(struct epoll_event &event, ReportingMessages_t &message, RmrMessagesBuffer_t &rmrMessageBuffer, - sctp_params_t *params, - otSpan *pSpan) { -#ifdef __TRACING__ - auto lspan = opentracing::Tracer::Global()->StartSpan( - __FUNCTION__, { opentracing::ChildOf(&pSpan->get()->context()) }); -#else - otSpan lspan = 0; -#endif + sctp_params_t *params) { auto *peerInfo = (ConnectedCU_t *)event.data.ptr; memcpy(message.message.enodbName, peerInfo->enodbName, sizeof(peerInfo->enodbName)); @@ -844,26 +933,20 @@ void handleEinprogressMessages(struct epoll_event &event, message.message.asnLength = rmrMessageBuffer.sendMessage->len; mdclog_write(MDCLOG_ERR, "%s", rmrMessageBuffer.sendMessage->payload); message.message.direction = 'N'; - if (sendRequestToXapp(message, RIC_SCTP_CONNECTION_FAILURE, rmrMessageBuffer, &lspan) != 0) { + if (sendRequestToXapp(message, RIC_SCTP_CONNECTION_FAILURE, rmrMessageBuffer) != 0) { mdclog_write(MDCLOG_ERR, "SCTP_CONNECTION_FAIL message failed to send to xAPP"); } memset(peerInfo->asnData, 0, peerInfo->asnLength); peerInfo->asnLength = 0; peerInfo->mtype = 0; -#ifdef __TRACING__ - lspan->Finish(); -#endif return; } peerInfo->isConnected = true; if (modifyToEpoll(params->epoll_fd, peerInfo, (EPOLLIN | EPOLLET), params->sctpMap, peerInfo->enodbName, - peerInfo->mtype, &lspan) != 0) { + peerInfo->mtype) != 0) { mdclog_write(MDCLOG_ERR, "epoll_ctl EPOLL_CTL_MOD"); -#ifdef __TRACING__ - lspan->Finish(); -#endif return; } @@ -876,36 +959,23 @@ void handleEinprogressMessages(struct epoll_event &event, mdclog_write(MDCLOG_DEBUG, "send the delayed SETUP/ENDC SETUP to sctp for %s", message.message.enodbName); } - if (sendSctpMsg(peerInfo, message, params->sctpMap, &lspan) != 0) { + if (sendSctpMsg(peerInfo, message, params->sctpMap) != 0) { if (mdclog_level_get() >= MDCLOG_DEBUG) { mdclog_write(MDCLOG_DEBUG, "Error write to SCTP %s %d", __func__, __LINE__); } -#ifdef __TRACING__ - lspan->Finish(); -#endif return; } memset(peerInfo->asnData, 0, peerInfo->asnLength); peerInfo->asnLength = 0; peerInfo->mtype = 0; -#ifdef __TRACING__ - lspan->Finish(); -#endif } void handlepoll_error(struct epoll_event &event, ReportingMessages_t &message, RmrMessagesBuffer_t &rmrMessageBuffer, - sctp_params_t *params, - otSpan *pSpan) { -#ifdef __TRACING__ - auto lspan = opentracing::Tracer::Global()->StartSpan( - __FUNCTION__, { opentracing::ChildOf(&pSpan->get()->context()) }); -#else - otSpan lspan = 0; -#endif + sctp_params_t *params) { if (event.data.fd != params->rmrListenFd) { auto *peerInfo = (ConnectedCU_t *)event.data.ptr; mdclog_write(MDCLOG_ERR, "epoll error, events %0x on fd %d, RAN NAME : %s", @@ -919,19 +989,16 @@ void handlepoll_error(struct epoll_event &event, memcpy(message.message.enodbName, peerInfo->enodbName, sizeof(peerInfo->enodbName)); message.message.direction = 'N'; - if (sendRequestToXapp(message, RIC_SCTP_CONNECTION_FAILURE, rmrMessageBuffer, &lspan) != 0) { + if (sendRequestToXapp(message, RIC_SCTP_CONNECTION_FAILURE, rmrMessageBuffer) != 0) { mdclog_write(MDCLOG_ERR, "SCTP_CONNECTION_FAIL message failed to send to xAPP"); } close(peerInfo->fileDescriptor); - cleanHashEntry((ConnectedCU_t *) event.data.ptr, params->sctpMap, &lspan); + params->sctpMap->erase(peerInfo->enodbName); + cleanHashEntry((ConnectedCU_t *) event.data.ptr, params->sctpMap); } else { mdclog_write(MDCLOG_ERR, "epoll error, events %0x on RMR FD", event.events); } -#ifdef __TRACING__ - lspan->Finish(); -#endif - } /** * @@ -963,15 +1030,8 @@ int setSocketNoBlocking(int socket) { * * @param val * @param m - * @param pSpan */ -void cleanHashEntry(ConnectedCU_t *val, Sctp_Map_t *m, otSpan *pSpan) { -#ifdef __TRACING__ - auto lspan = opentracing::Tracer::Global()->StartSpan( - __FUNCTION__, { opentracing::ChildOf(&pSpan->get()->context()) }); -#else -// otSpan lspan = 0; -#endif +void cleanHashEntry(ConnectedCU_t *val, Sctp_Map_t *m) { char *dummy; auto port = (uint16_t) strtol(val->portNumber, &dummy, 10); char searchBuff[2048]{}; @@ -981,9 +1041,6 @@ void cleanHashEntry(ConnectedCU_t *val, Sctp_Map_t *m, otSpan *pSpan) { m->erase(val->enodbName); free(val); -#ifdef __TRACING__ - lspan->Finish(); -#endif } /** @@ -996,13 +1053,7 @@ void cleanHashEntry(ConnectedCU_t *val, Sctp_Map_t *m, otSpan *pSpan) { * @param mtype message number * @return 0 success, anegative number on fail */ -int sendSctpMsg(ConnectedCU_t *peerInfo, ReportingMessages_t &message, Sctp_Map_t *m, otSpan *pSpan) { -#ifdef __TRACING__ - auto lspan = opentracing::Tracer::Global()->StartSpan( - __FUNCTION__, { opentracing::ChildOf(&pSpan->get()->context()) }); -#else - otSpan lspan = 0; -#endif +int sendSctpMsg(ConnectedCU_t *peerInfo, ReportingMessages_t &message, Sctp_Map_t *m) { auto loglevel = mdclog_level_get(); int fd = peerInfo->fileDescriptor; if (loglevel >= MDCLOG_DEBUG) { @@ -1018,12 +1069,9 @@ int sendSctpMsg(ConnectedCU_t *peerInfo, ReportingMessages_t &message, Sctp_Map_ mdclog_write(MDCLOG_ERR, "error writing to CU a message, %s ", strerror(errno)); if (!peerInfo->isConnected) { mdclog_write(MDCLOG_ERR, "connection to CU %s is still in progress.", message.message.enodbName); -#ifdef __TRACING__ - lspan->Finish(); -#endif return -1; } - cleanHashEntry(peerInfo, m, &lspan); + cleanHashEntry(peerInfo, m); close(fd); char key[MAX_ENODB_NAME_SIZE * 2]; snprintf(key, MAX_ENODB_NAME_SIZE * 2, "msg:%s|%d", message.message.enodbName, @@ -1031,14 +1079,11 @@ int sendSctpMsg(ConnectedCU_t *peerInfo, ReportingMessages_t &message, Sctp_Map_ if (loglevel >= MDCLOG_DEBUG) { mdclog_write(MDCLOG_DEBUG, "remove key = %s from %s at line %d", key, __FUNCTION__, __LINE__); } - auto tmp = m->find(key); + auto tmp = m->find(key); if (tmp) { free(tmp); } m->erase(key); -#ifdef __TRACING__ - lspan->Finish(); -#endif return -1; } message.message.direction = 'D'; @@ -1051,9 +1096,6 @@ int sendSctpMsg(ConnectedCU_t *peerInfo, ReportingMessages_t &message, Sctp_Map_ message.message.enodbName, __FUNCTION__); } -#ifdef __TRACING__ - lspan->Finish(); -#endif return 0; } } @@ -1062,17 +1104,8 @@ int sendSctpMsg(ConnectedCU_t *peerInfo, ReportingMessages_t &message, Sctp_Map_ * * @param message * @param rmrMessageBuffer - * @param pSpan */ -void getRequestMetaData(ReportingMessages_t &message, RmrMessagesBuffer_t &rmrMessageBuffer, otSpan *pSpan) { -#ifdef __TRACING__ - auto lspan = opentracing::Tracer::Global()->StartSpan( - __FUNCTION__, { opentracing::ChildOf(&pSpan->get()->context()) }); -#else -// otSpan lspan = 0; -#endif - rmr_get_meid(rmrMessageBuffer.rcvMessage, (unsigned char *)(message.message.enodbName)); - +void getRequestMetaData(ReportingMessages_t &message, RmrMessagesBuffer_t &rmrMessageBuffer) { message.message.asndata = rmrMessageBuffer.rcvMessage->payload; message.message.asnLength = rmrMessageBuffer.rcvMessage->len; @@ -1080,103 +1113,9 @@ void getRequestMetaData(ReportingMessages_t &message, RmrMessagesBuffer_t &rmrMe mdclog_write(MDCLOG_DEBUG, "Message from Xapp RAN name = %s message length = %ld", message.message.enodbName, (unsigned long) message.message.asnLength); } -#ifdef __TRACING__ - lspan->Finish(); -#endif - } -/** - * - * @param metaData all the data strip to structure - * @param data the data recived from xAPP - * @return 0 success all other values are fault - */ -int getSetupRequestMetaData(ReportingMessages_t &message, char *data, char *host, uint16_t &port, otSpan *pSpan) { -#ifdef __TRACING__ - auto lspan = opentracing::Tracer::Global()->StartSpan( - __FUNCTION__, { opentracing::ChildOf(&pSpan->get()->context()) }); -#else -// otSpan lspan = 0; -#endif - auto loglevel = mdclog_level_get(); - - char delimiter[4] {}; - memset(delimiter, 0, (size_t)4); - delimiter[0] = '|'; - char *tmp; - - char *val = strtok_r(data, delimiter, &tmp); - if (val != nullptr) { - if (mdclog_level_get() >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "SCTP ADDRESS parameter from message = %s", val); - } - memcpy(host, val, tmp - val ); - } else { - mdclog_write(MDCLOG_ERR, "wrong Host Name for setup request %s", data); -#ifdef __TRACING__ - lspan->Finish(); -#endif - return -1; - } - - val = strtok_r(nullptr, delimiter, &tmp); - if (val != nullptr) { - if (mdclog_level_get() >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "PORT parameter from message = %s", val); - } - char *dummy; - port = (uint16_t)strtol(val, &dummy, 10); - } else { - mdclog_write(MDCLOG_ERR, "wrong Port for setup request %s", data); -#ifdef __TRACING__ - lspan->Finish(); -#endif - return -2; - } - - val = strtok_r(nullptr, delimiter, &tmp); - if (val != nullptr) { - if (mdclog_level_get() >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "RAN NAME parameter from message = %s", val); - } - memcpy(message.message.enodbName, val, tmp - val); - } else { - mdclog_write(MDCLOG_ERR, "wrong gNb/Enodeb name for setup request %s", data); -#ifdef __TRACING__ - lspan->Finish(); -#endif - - return -3; - } - val = strtok_r(nullptr, delimiter, &tmp); - if (val != nullptr) { - if (mdclog_level_get() >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "ASN length parameter from message = %s", val); - } - char *dummy; - message.message.asnLength = (uint16_t) strtol(val, &dummy, 10); - } else { - mdclog_write(MDCLOG_ERR, "wrong ASN length for setup request %s", data); -#ifdef __TRACING__ - lspan->Finish(); -#endif - return -4; - } - - message.message.asndata = (unsigned char *)tmp; // tmp is local but point to the location in data - - if (loglevel >= MDCLOG_INFO) { - mdclog_write(MDCLOG_INFO, "Message from Xapp RAN name = %s host address = %s port = %d", - message.message.enodbName, host, port); - } -#ifdef __TRACING__ - lspan->Finish(); -#endif - - return 0; -} /** * @@ -1185,21 +1124,13 @@ int getSetupRequestMetaData(ReportingMessages_t &message, char *data, char *host * @param numOfMessages * @param rmrMessageBuffer * @param ts - * @param pSpan * @return */ int receiveDataFromSctp(struct epoll_event *events, Sctp_Map_t *sctpMap, int &numOfMessages, RmrMessagesBuffer_t &rmrMessageBuffer, - struct timespec &ts, - otSpan *pSpan) { -#ifdef __TRACING__ - auto lspan = opentracing::Tracer::Global()->StartSpan( - __FUNCTION__, { opentracing::ChildOf(&pSpan->get()->context()) }); -#else - otSpan lspan = 0; -#endif + struct timespec &ts) { /* We have data on the fd waiting to be read. Read and display it. * We must read whatever data is available completely, as we are running * in edge-triggered mode and won't get a notification again for the same data. */ @@ -1210,14 +1141,12 @@ int receiveDataFromSctp(struct epoll_event *events, // get the identity of the interface message.peerInfo = (ConnectedCU_t *)events->data.ptr; - struct timespec start{0, 0}; struct timespec decodestart{0, 0}; struct timespec end{0, 0}; E2AP_PDU_t *pdu = nullptr; - while (true) { if (loglevel >= MDCLOG_DEBUG) { mdclog_write(MDCLOG_DEBUG, "Start Read from SCTP %d fd", message.peerInfo->fileDescriptor); @@ -1230,8 +1159,9 @@ int receiveDataFromSctp(struct epoll_event *events, if (loglevel >= MDCLOG_DEBUG) { mdclog_write(MDCLOG_DEBUG, "Finish Read from SCTP %d fd message length = %ld", - message.peerInfo->fileDescriptor, message.message.asnLength); + message.peerInfo->fileDescriptor, message.message.asnLength); } + memcpy(message.message.enodbName, message.peerInfo->enodbName, sizeof(message.peerInfo->enodbName)); message.message.direction = 'U'; message.message.time.tv_nsec = ts.tv_nsec; @@ -1260,7 +1190,6 @@ int receiveDataFromSctp(struct epoll_event *events, break; } - asn_dec_rval_t rval; if (loglevel >= MDCLOG_DEBUG) { char printBuffer[4096]{}; char *tmp = printBuffer; @@ -1277,7 +1206,7 @@ int receiveDataFromSctp(struct epoll_event *events, clock_gettime(CLOCK_MONOTONIC, &decodestart); } - rval = asn_decode(nullptr, ATS_ALIGNED_BASIC_PER, &asn_DEF_E2AP_PDU, (void **) &pdu, + auto rval = asn_decode(nullptr, ATS_ALIGNED_BASIC_PER, &asn_DEF_E2AP_PDU, (void **) &pdu, message.message.asndata, message.message.asnLength); if (rval.code != RC_OK) { mdclog_write(MDCLOG_ERR, "Error %d Decoding (unpack) E2AP PDU from RAN : %s", rval.code, @@ -1299,15 +1228,15 @@ int receiveDataFromSctp(struct epoll_event *events, switch (pdu->present) { case E2AP_PDU_PR_initiatingMessage: {//initiating message - asnInitiatingRequest(pdu, message, rmrMessageBuffer, &lspan); + asnInitiatingRequest(pdu, sctpMap,message, rmrMessageBuffer); break; } case E2AP_PDU_PR_successfulOutcome: { //successful outcome - asnSuccsesfulMsg(pdu, message, sctpMap, rmrMessageBuffer, &lspan); + asnSuccsesfulMsg(pdu, sctpMap, message, rmrMessageBuffer); break; } case E2AP_PDU_PR_unsuccessfulOutcome: { //Unsuccessful Outcome - asnUnSuccsesfulMsg(pdu, message, sctpMap, rmrMessageBuffer, &lspan); + asnUnSuccsesfulMsg(pdu, sctpMap, message, rmrMessageBuffer); break; } default: @@ -1321,20 +1250,11 @@ int receiveDataFromSctp(struct epoll_event *events, message.peerInfo->enodbName, end.tv_sec - decodestart.tv_sec, end.tv_nsec - decodestart.tv_nsec); } numOfMessages++; - // remove the break for EAGAIN - //break; if (pdu != nullptr) { - //TODO need to test ASN_STRUCT_RESET(asn_DEF_E2AP_PDU, pdu); to get better performance - //ASN_STRUCT_RESET(asn_DEF_E2AP_PDU, pdu); - ASN_STRUCT_FREE(asn_DEF_E2AP_PDU, pdu); - pdu = nullptr; + ASN_STRUCT_RESET(asn_DEF_E2AP_PDU, pdu); + //ASN_STRUCT_FREE(asn_DEF_E2AP_PDU, pdu); + //pdu = nullptr; } - //clock_gettime(CLOCK_MONOTONIC, &start); - } - // in case of break to avoid memory leak - if (pdu != nullptr) { - ASN_STRUCT_FREE(asn_DEF_E2AP_PDU, pdu); - pdu = nullptr; } if (done) { @@ -1343,21 +1263,20 @@ int receiveDataFromSctp(struct epoll_event *events, } message.message.asnLength = rmrMessageBuffer.sendMessage->len = snprintf((char *)rmrMessageBuffer.sendMessage->payload, - 256, - "%s|CU disconnected unexpectedly", - message.peerInfo->enodbName); + 256, + "%s|CU disconnected unexpectedly", + message.peerInfo->enodbName); message.message.asndata = rmrMessageBuffer.sendMessage->payload; if (sendRequestToXapp(message, RIC_SCTP_CONNECTION_FAILURE, - rmrMessageBuffer, - &lspan) != 0) { + rmrMessageBuffer) != 0) { mdclog_write(MDCLOG_ERR, "SCTP_CONNECTION_FAIL message failed to send to xAPP"); } /* Closing descriptor make epoll remove it from the set of descriptors which are monitored. */ close(message.peerInfo->fileDescriptor); - cleanHashEntry((ConnectedCU_t *) events->data.ptr, sctpMap, &lspan); + cleanHashEntry((ConnectedCU_t *) events->data.ptr, sctpMap); } if (loglevel >= MDCLOG_DEBUG) { clock_gettime(CLOCK_MONOTONIC, &end); @@ -1365,106 +1284,369 @@ int receiveDataFromSctp(struct epoll_event *events, end.tv_sec - start.tv_sec, end.tv_nsec - start.tv_nsec); } -#ifdef __TRACING__ - lspan->Finish(); -#endif - return 0; } static void buildAndsendSetupRequest(ReportingMessages_t &message, - E2setupRequestIEs_t *ie, RmrMessagesBuffer_t &rmrMessageBuffer, - E2AP_PDU_t *pdu) { + E2AP_PDU_t *pdu/*, + string const &messageName, + string const &ieName, + vector &functionsToAdd_v, + vector &functionsToModified_v*/) { auto logLevel = mdclog_level_get(); + // now we can send the data to e2Mgr - - if (buildRanName(message.peerInfo->enodbName, ie) < 0) { - mdclog_write(MDCLOG_ERR, "Bad param in E2setupRequestIEs GlobalE2node_ID.\n"); - } else { - memcpy(message.message.enodbName, message.peerInfo->enodbName, strlen(message.peerInfo->enodbName)); + asn_enc_rval_t er; + auto buffer_size = RECEIVE_SCTP_BUFFER_SIZE * 2; + unsigned char buffer[RECEIVE_SCTP_BUFFER_SIZE * 2]; + while (true) { + er = asn_encode_to_buffer(nullptr, ATS_BASIC_XER, &asn_DEF_E2AP_PDU, pdu, buffer, buffer_size); + if (er.encoded == -1) { + mdclog_write(MDCLOG_ERR, "encoding of %s failed, %s", asn_DEF_E2AP_PDU.name, strerror(errno)); + return; + } else if (er.encoded > (ssize_t) buffer_size) { + buffer_size = er.encoded + 128; + mdclog_write(MDCLOG_WARN, "Buffer of size %d is to small for %s. Reallocate buffer of size %d", + (int) buffer_size, + asn_DEF_E2AP_PDU.name, buffer_size); + buffer_size = er.encoded + 128; +// free(buffer); + continue; + } + buffer[er.encoded] = '\0'; + break; } - // now we can send the data to e2Mgr - auto buffer_size = RECEIVE_SCTP_BUFFER_SIZE; + // encode to xml - auto *rmrMsg = rmr_alloc_msg(rmrMessageBuffer.rmrCtx, buffer_size); - // add addrees to message - auto j = snprintf((char *)rmrMsg->payload, 256, "%s:%d|", message.peerInfo->sctpParams->myIP.c_str(), message.peerInfo->sctpParams->rmrPort); + string res((char *)buffer); + res.erase(std::remove(res.begin(), res.end(), '\n'), res.end()); + res.erase(std::remove(res.begin(), res.end(), '\t'), res.end()); + res.erase(std::remove(res.begin(), res.end(), ' '), res.end()); +// string res {}; +// if (!functionsToAdd_v.empty() || !functionsToModified_v.empty()) { +// res = buildXmlData(messageName, ieName, functionsToAdd_v, functionsToModified_v, buffer, (size_t) er.encoded); +// } + rmr_mbuf_t *rmrMsg; +// if (res.length() == 0) { +// rmrMsg = rmr_alloc_msg(rmrMessageBuffer.rmrCtx, buffer_size + 256); +// rmrMsg->len = snprintf((char *) rmrMsg->payload, RECEIVE_SCTP_BUFFER_SIZE * 2, "%s:%d|%s", +// message.peerInfo->sctpParams->myIP.c_str(), +// message.peerInfo->sctpParams->rmrPort, +// buffer); +// } else { + rmrMsg = rmr_alloc_msg(rmrMessageBuffer.rmrCtx, (int)res.length() + 256); + rmrMsg->len = snprintf((char *) rmrMsg->payload, res.length() + 256, "%s:%d|%s", + message.peerInfo->sctpParams->myIP.c_str(), + message.peerInfo->sctpParams->rmrPort, + res.c_str()); +// } - unsigned char *buffer = &rmrMsg->payload[j]; - // encode to xml - asn_enc_rval_t er; - er = asn_encode_to_buffer(nullptr, ATS_BASIC_XER, &asn_DEF_E2AP_PDU, pdu, buffer, buffer_size - j); - if (er.encoded == -1) { - mdclog_write(MDCLOG_ERR, "encoding of %s failed, %s", asn_DEF_E2AP_PDU.name, strerror(errno)); - } else if (er.encoded > (ssize_t) buffer_size) { - mdclog_write(MDCLOG_ERR, "Buffer of size %d is to small for %s", - (int) buffer_size, - asn_DEF_E2AP_PDU.name); - } else { - if (logLevel >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "Buffer of size %d, data = %s", (int) er.encoded, buffer); - } - // TODO send to RMR - message.message.messageType = rmrMsg->mtype = RIC_X2_SETUP_REQ; - rmrMsg->state = 0; - rmr_bytes2meid(rmrMsg, (unsigned char *) message.message.enodbName, strlen(message.message.enodbName)); + if (logLevel >= MDCLOG_DEBUG) { + mdclog_write(MDCLOG_DEBUG, "Setup request of size %d :\n %s\n", rmrMsg->len, rmrMsg->payload); + } + // send to RMR + rmrMsg->mtype = message.message.messageType; + rmrMsg->state = 0; + rmr_bytes2meid(rmrMsg, (unsigned char *) message.message.enodbName, strlen(message.message.enodbName)); - static unsigned char tx[32]; - snprintf((char *) tx, sizeof tx, "%15ld", transactionCounter++); - rmr_bytes2xact(rmrMsg, tx, strlen((const char *) tx)); - - rmrMsg = rmr_send_msg(rmrMessageBuffer.rmrCtx, rmrMsg); - if (rmrMsg == nullptr) { - mdclog_write(MDCLOG_ERR, "RMR failed to send returned nullptr"); - } else if (rmrMsg->state != 0) { - char meid[RMR_MAX_MEID]{}; - if (rmrMsg->state == RMR_ERR_RETRY) { - usleep(5); - rmrMsg->state = 0; - mdclog_write(MDCLOG_INFO, "RETRY sending Message %d to Xapp from %s", - rmrMsg->mtype, rmr_get_meid(rmrMsg, (unsigned char *) meid)); - rmrMsg = rmr_send_msg(rmrMessageBuffer.rmrCtx, rmrMsg); - if (rmrMsg == nullptr) { - mdclog_write(MDCLOG_ERR, "RMR failed send returned nullptr"); - } else if (rmrMsg->state != 0) { - mdclog_write(MDCLOG_ERR, - "RMR Retry failed %s sending request %d to Xapp from %s", - translateRmrErrorMessages(rmrMsg->state).c_str(), - rmrMsg->mtype, - rmr_get_meid(rmrMsg, (unsigned char *) meid)); - } - } else { - mdclog_write(MDCLOG_ERR, "RMR failed: %s. sending request %d to Xapp from %s", + static unsigned char tx[32]; + snprintf((char *) tx, sizeof tx, "%15ld", transactionCounter++); + rmr_bytes2xact(rmrMsg, tx, strlen((const char *) tx)); + + rmrMsg = rmr_send_msg(rmrMessageBuffer.rmrCtx, rmrMsg); + if (rmrMsg == nullptr) { + mdclog_write(MDCLOG_ERR, "RMR failed to send returned nullptr"); + } else if (rmrMsg->state != 0) { + char meid[RMR_MAX_MEID]{}; + if (rmrMsg->state == RMR_ERR_RETRY) { + usleep(5); + rmrMsg->state = 0; + mdclog_write(MDCLOG_INFO, "RETRY sending Message %d to Xapp from %s", + rmrMsg->mtype, rmr_get_meid(rmrMsg, (unsigned char *) meid)); + rmrMsg = rmr_send_msg(rmrMessageBuffer.rmrCtx, rmrMsg); + if (rmrMsg == nullptr) { + mdclog_write(MDCLOG_ERR, "RMR failed send returned nullptr"); + } else if (rmrMsg->state != 0) { + mdclog_write(MDCLOG_ERR, + "RMR Retry failed %s sending request %d to Xapp from %s", translateRmrErrorMessages(rmrMsg->state).c_str(), rmrMsg->mtype, rmr_get_meid(rmrMsg, (unsigned char *) meid)); } + } else { + mdclog_write(MDCLOG_ERR, "RMR failed: %s. sending request %d to Xapp from %s", + translateRmrErrorMessages(rmrMsg->state).c_str(), + rmrMsg->mtype, + rmr_get_meid(rmrMsg, (unsigned char *) meid)); } - message.peerInfo->gotSetup = true; - buildJsonMessage(message); } + message.peerInfo->gotSetup = true; + buildJsonMessage(message); + if (rmrMsg != nullptr) { + rmr_free_msg(rmrMsg); + } +} + +#if 0 +int RAN_Function_list_To_Vector(RANfunctions_List_t& list, vector &runFunXML_v) { + auto index = 0; + runFunXML_v.clear(); + for (auto j = 0; j < list.list.count; j++) { + auto *raNfunctionItemIEs = (RANfunction_ItemIEs_t *)list.list.array[j]; + if (raNfunctionItemIEs->id == ProtocolIE_ID_id_RANfunction_Item && + (raNfunctionItemIEs->value.present == RANfunction_ItemIEs__value_PR_RANfunction_Item)) { + // encode to xml + E2SM_gNB_NRT_RANfunction_Definition_t *ranFunDef = nullptr; + auto rval = asn_decode(nullptr, ATS_ALIGNED_BASIC_PER, + &asn_DEF_E2SM_gNB_NRT_RANfunction_Definition, + (void **)&ranFunDef, + raNfunctionItemIEs->value.choice.RANfunction_Item.ranFunctionDefinition.buf, + raNfunctionItemIEs->value.choice.RANfunction_Item.ranFunctionDefinition.size); + if (rval.code != RC_OK) { + mdclog_write(MDCLOG_ERR, "Error %d Decoding (unpack) E2SM message from : %s", + rval.code, + asn_DEF_E2SM_gNB_NRT_RANfunction_Definition.name); + return -1; + } + + auto xml_buffer_size = RECEIVE_SCTP_BUFFER_SIZE * 2; + unsigned char xml_buffer[RECEIVE_SCTP_BUFFER_SIZE * 2]; + memset(xml_buffer, 0, RECEIVE_SCTP_BUFFER_SIZE * 2); + // encode to xml + auto er = asn_encode_to_buffer(nullptr, + ATS_BASIC_XER, + &asn_DEF_E2SM_gNB_NRT_RANfunction_Definition, + ranFunDef, + xml_buffer, + xml_buffer_size); + if (er.encoded == -1) { + mdclog_write(MDCLOG_ERR, "encoding of %s failed, %s", + asn_DEF_E2SM_gNB_NRT_RANfunction_Definition.name, + strerror(errno)); + } else if (er.encoded > (ssize_t)xml_buffer_size) { + mdclog_write(MDCLOG_ERR, "Buffer of size %d is to small for %s, at %s line %d", + (int) xml_buffer_size, + asn_DEF_E2SM_gNB_NRT_RANfunction_Definition.name, __func__, __LINE__); + } else { + if (mdclog_level_get() >= MDCLOG_DEBUG) { + mdclog_write(MDCLOG_DEBUG, "Encoding E2SM %s PDU number %d : %s", + asn_DEF_E2SM_gNB_NRT_RANfunction_Definition.name, + index++, + xml_buffer); + } + + string runFuncs = (char *)(xml_buffer); + runFunXML_v.emplace_back(runFuncs); + } + } + } + return 0; +} + +int collectServiceUpdate_RequestData(E2AP_PDU_t *pdu, + Sctp_Map_t *sctpMap, + ReportingMessages_t &message, + vector &RANfunctionsAdded_v, + vector &RANfunctionsModified_v) { + memset(message.peerInfo->enodbName, 0 , MAX_ENODB_NAME_SIZE); + for (auto i = 0; i < pdu->choice.initiatingMessage->value.choice.RICserviceUpdate.protocolIEs.list.count; i++) { + auto *ie = pdu->choice.initiatingMessage->value.choice.RICserviceUpdate.protocolIEs.list.array[i]; + if (ie->id == ProtocolIE_ID_id_RANfunctionsAdded) { + if (ie->value.present == RICserviceUpdate_IEs__value_PR_RANfunctionsID_List) { + if (mdclog_level_get() >= MDCLOG_DEBUG) { + mdclog_write(MDCLOG_DEBUG, "Run function list have %d entries", + ie->value.choice.RANfunctions_List.list.count); + } + if (RAN_Function_list_To_Vector(ie->value.choice.RANfunctions_List, RANfunctionsAdded_v) != 0 ) { + return -1; + } + } + } else if (ie->id == ProtocolIE_ID_id_RANfunctionsModified) { + if (ie->value.present == RICserviceUpdate_IEs__value_PR_RANfunctions_List) { + if (mdclog_level_get() >= MDCLOG_DEBUG) { + mdclog_write(MDCLOG_DEBUG, "Run function list have %d entries", + ie->value.choice.RANfunctions_List.list.count); + } + if (RAN_Function_list_To_Vector(ie->value.choice.RANfunctions_List, RANfunctionsModified_v) != 0 ) { + return -1; + } + } + } + } + if (mdclog_level_get() >= MDCLOG_DEBUG) { + mdclog_write(MDCLOG_DEBUG, "Run function vector have %ld entries", + RANfunctionsAdded_v.size()); + } + return 0; +} + +#endif + + +void buildPrometheuslist(ConnectedCU_t *peerInfo, Family *prometheusFamily) { + peerInfo->counters[IN_INITI][MSG_COUNTER][(ProcedureCode_id_E2setup)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"SetupRequest", "Messages"}}); + peerInfo->counters[IN_INITI][BYTES_COUNTER][(ProcedureCode_id_E2setup)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"SetupRequest", "Bytes"}}); + + peerInfo->counters[IN_INITI][MSG_COUNTER][(ProcedureCode_id_ErrorIndication)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"ErrorIndication", "Messages"}}); + peerInfo->counters[IN_INITI][BYTES_COUNTER][(ProcedureCode_id_ErrorIndication)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"ErrorIndication", "Bytes"}}); + + peerInfo->counters[IN_INITI][MSG_COUNTER][(ProcedureCode_id_RICindication)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICindication", "Messages"}}); + peerInfo->counters[IN_INITI][BYTES_COUNTER][(ProcedureCode_id_RICindication)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICindication", "Bytes"}}); + + peerInfo->counters[IN_INITI][MSG_COUNTER][(ProcedureCode_id_Reset)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"ResetRequest", "Messages"}}); + peerInfo->counters[IN_INITI][BYTES_COUNTER][(ProcedureCode_id_Reset)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"ResetRequest", "Bytes"}}); + + peerInfo->counters[IN_INITI][MSG_COUNTER][(ProcedureCode_id_RICserviceUpdate)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICserviceUpdate", "Messages"}}); + peerInfo->counters[IN_INITI][BYTES_COUNTER][(ProcedureCode_id_RICserviceUpdate)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICserviceUpdate", "Bytes"}}); + // --------------------------------------------- + peerInfo->counters[IN_SUCC][MSG_COUNTER][(ProcedureCode_id_Reset)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"ResetACK", "Messages"}}); + peerInfo->counters[IN_SUCC][BYTES_COUNTER][(ProcedureCode_id_Reset)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"ResetACK", "Bytes"}}); + + peerInfo->counters[IN_SUCC][MSG_COUNTER][(ProcedureCode_id_RICcontrol)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICcontrolACK", "Messages"}}); + peerInfo->counters[IN_SUCC][BYTES_COUNTER][(ProcedureCode_id_RICcontrol)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICcontrolACK", "Bytes"}}); + + peerInfo->counters[IN_SUCC][MSG_COUNTER][(ProcedureCode_id_RICsubscription)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICsubscriptionACK", "Messages"}}); + peerInfo->counters[IN_SUCC][BYTES_COUNTER][(ProcedureCode_id_RICsubscription)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICsubscriptionACK", "Bytes"}}); + + peerInfo->counters[IN_SUCC][MSG_COUNTER][(ProcedureCode_id_RICsubscriptionDelete)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICsubscriptionDeleteACK", "Messages"}}); + peerInfo->counters[IN_SUCC][BYTES_COUNTER][(ProcedureCode_id_RICsubscriptionDelete)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICsubscriptionDeleteACK", "Bytes"}}); + //------------------------------------------------------------- + + peerInfo->counters[IN_UN_SUCC][MSG_COUNTER][(ProcedureCode_id_RICcontrol)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICcontrolFailure", "Messages"}}); + peerInfo->counters[IN_UN_SUCC][BYTES_COUNTER][(ProcedureCode_id_RICcontrol)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICcontrolFailure", "Bytes"}}); + + peerInfo->counters[IN_UN_SUCC][MSG_COUNTER][(ProcedureCode_id_RICsubscription)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICsubscriptionFailure", "Messages"}}); + peerInfo->counters[IN_UN_SUCC][BYTES_COUNTER][(ProcedureCode_id_RICsubscription)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICsubscriptionFailure", "Bytes"}}); + + peerInfo->counters[IN_UN_SUCC][MSG_COUNTER][(ProcedureCode_id_RICsubscriptionDelete)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICsubscriptionDeleteFailure", "Messages"}}); + peerInfo->counters[IN_UN_SUCC][BYTES_COUNTER][(ProcedureCode_id_RICsubscriptionDelete)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICsubscriptionDeleteFailure", "Bytes"}}); + + //==================================================================================== + peerInfo->counters[OUT_INITI][MSG_COUNTER][(ProcedureCode_id_ErrorIndication)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"ErrorIndication", "Messages"}}); + peerInfo->counters[OUT_INITI][BYTES_COUNTER][(ProcedureCode_id_ErrorIndication)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"ErrorIndication", "Bytes"}}); + + peerInfo->counters[OUT_INITI][MSG_COUNTER][(ProcedureCode_id_Reset)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"ResetRequest", "Messages"}}); + peerInfo->counters[OUT_INITI][BYTES_COUNTER][(ProcedureCode_id_Reset)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"ResetRequest", "Bytes"}}); + + peerInfo->counters[OUT_INITI][MSG_COUNTER][(ProcedureCode_id_RICcontrol)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"RICcontrol", "Messages"}}); + peerInfo->counters[OUT_INITI][BYTES_COUNTER][(ProcedureCode_id_RICcontrol)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"RICcontrol", "Bytes"}}); + + peerInfo->counters[OUT_INITI][MSG_COUNTER][(ProcedureCode_id_RICserviceQuery)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"RICserviceQuery", "Messages"}}); + peerInfo->counters[OUT_INITI][BYTES_COUNTER][(ProcedureCode_id_RICserviceQuery)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"RICserviceQuery", "Bytes"}}); + + peerInfo->counters[OUT_INITI][MSG_COUNTER][(ProcedureCode_id_RICsubscription)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"RICsubscription", "Messages"}}); + peerInfo->counters[OUT_INITI][BYTES_COUNTER][(ProcedureCode_id_RICsubscription)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"RICsubscription", "Bytes"}}); + + peerInfo->counters[OUT_INITI][MSG_COUNTER][(ProcedureCode_id_RICsubscriptionDelete)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"RICsubscriptionDelete", "Messages"}}); + peerInfo->counters[OUT_INITI][BYTES_COUNTER][(ProcedureCode_id_RICsubscriptionDelete)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"RICsubscriptionDelete", "Bytes"}}); + //--------------------------------------------------------------------------------------------------------- + peerInfo->counters[OUT_SUCC][MSG_COUNTER][(ProcedureCode_id_E2setup)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"SetupResponse", "Messages"}}); + peerInfo->counters[OUT_SUCC][BYTES_COUNTER][(ProcedureCode_id_E2setup)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"SetupResponse", "Bytes"}}); + + peerInfo->counters[OUT_SUCC][MSG_COUNTER][(ProcedureCode_id_Reset)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"ResetACK", "Messages"}}); + peerInfo->counters[OUT_SUCC][BYTES_COUNTER][(ProcedureCode_id_Reset)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"ResetACK", "Bytes"}}); + + peerInfo->counters[OUT_SUCC][MSG_COUNTER][(ProcedureCode_id_RICserviceUpdate)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"RICserviceUpdateResponse", "Messages"}}); + peerInfo->counters[OUT_SUCC][BYTES_COUNTER][(ProcedureCode_id_RICserviceUpdate)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"RICserviceUpdateResponse", "Bytes"}}); + //---------------------------------------------------------------------------------------------------------------- + peerInfo->counters[OUT_UN_SUCC][MSG_COUNTER][(ProcedureCode_id_E2setup)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"SetupRequestFailure", "Messages"}}); + peerInfo->counters[OUT_UN_SUCC][BYTES_COUNTER][(ProcedureCode_id_E2setup)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"SetupRequestFailure", "Bytes"}}); + + peerInfo->counters[OUT_UN_SUCC][MSG_COUNTER][(ProcedureCode_id_RICserviceUpdate)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"RICserviceUpdateFailure", "Messages"}}); + peerInfo->counters[OUT_UN_SUCC][BYTES_COUNTER][(ProcedureCode_id_RICserviceUpdate)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"RICserviceUpdateFailure", "Bytes"}}); +} +/** + * + * @param pdu + * @param sctpMap + * @param message + * @param RANfunctionsAdded_v + * @return + */ +int collectSetupRequestData(E2AP_PDU_t *pdu, + Sctp_Map_t *sctpMap, + ReportingMessages_t &message /*, vector &RANfunctionsAdded_v*/) { + memset(message.peerInfo->enodbName, 0 , MAX_ENODB_NAME_SIZE); + for (auto i = 0; i < pdu->choice.initiatingMessage->value.choice.E2setupRequest.protocolIEs.list.count; i++) { + auto *ie = pdu->choice.initiatingMessage->value.choice.E2setupRequest.protocolIEs.list.array[i]; + if (ie->id == ProtocolIE_ID_id_GlobalE2node_ID) { + // get the ran name for meid + if (ie->value.present == E2setupRequestIEs__value_PR_GlobalE2node_ID) { + if (buildRanName(message.peerInfo->enodbName, ie) < 0) { + mdclog_write(MDCLOG_ERR, "Bad param in E2setupRequestIEs GlobalE2node_ID.\n"); + // no mesage will be sent + return -1; + } + + memcpy(message.message.enodbName, message.peerInfo->enodbName, strlen(message.peerInfo->enodbName)); + sctpMap->setkey(message.message.enodbName, message.peerInfo); + } + } /*else if (ie->id == ProtocolIE_ID_id_RANfunctionsAdded) { + if (ie->value.present == E2setupRequestIEs__value_PR_RANfunctions_List) { + if (mdclog_level_get() >= MDCLOG_DEBUG) { + mdclog_write(MDCLOG_DEBUG, "Run function list have %d entries", + ie->value.choice.RANfunctions_List.list.count); + } + if (RAN_Function_list_To_Vector(ie->value.choice.RANfunctions_List, RANfunctionsAdded_v) != 0 ) { + return -1; + } + } + } */ + } +// if (mdclog_level_get() >= MDCLOG_DEBUG) { +// mdclog_write(MDCLOG_DEBUG, "Run function vector have %ld entries", +// RANfunctionsAdded_v.size()); +// } + return 0; +} + +int XML_From_PER(ReportingMessages_t &message, RmrMessagesBuffer_t &rmrMessageBuffer) { + E2AP_PDU_t *pdu = nullptr; + + if (mdclog_level_get() >= MDCLOG_DEBUG) { + mdclog_write(MDCLOG_DEBUG, "got PER message of size %d is:%s", + rmrMessageBuffer.sendMessage->len, rmrMessageBuffer.sendMessage->payload); + } + auto rval = asn_decode(nullptr, ATS_ALIGNED_BASIC_PER, &asn_DEF_E2AP_PDU, (void **) &pdu, + rmrMessageBuffer.sendMessage->payload, rmrMessageBuffer.sendMessage->len); + if (rval.code != RC_OK) { + mdclog_write(MDCLOG_ERR, "Error %d Decoding (unpack) setup response from E2MGR : %s", + rval.code, + message.message.enodbName); + return -1; + } + + int buff_size = RECEIVE_XAPP_BUFFER_SIZE; + auto er = asn_encode_to_buffer(nullptr, ATS_BASIC_XER, &asn_DEF_E2AP_PDU, pdu, + rmrMessageBuffer.sendMessage->payload, buff_size); + if (er.encoded == -1) { + mdclog_write(MDCLOG_ERR, "encoding of %s failed, %s", asn_DEF_E2AP_PDU.name, strerror(errno)); + return -1; + } else if (er.encoded > (ssize_t)buff_size) { + mdclog_write(MDCLOG_ERR, "Buffer of size %d is to small for %s, at %s line %d", + (int)rmrMessageBuffer.sendMessage->len, + asn_DEF_E2AP_PDU.name, + __func__, + __LINE__); + return -1; + } + rmrMessageBuffer.sendMessage->len = er.encoded; + return 0; } + /** * * @param pdu * @param message * @param rmrMessageBuffer - * @param pSpan */ void asnInitiatingRequest(E2AP_PDU_t *pdu, + Sctp_Map_t *sctpMap, ReportingMessages_t &message, - RmrMessagesBuffer_t &rmrMessageBuffer, - otSpan *pSpan) { -#ifdef __TRACING__ - auto lspan = opentracing::Tracer::Global()->StartSpan( - __FUNCTION__, { opentracing::ChildOf(&pSpan->get()->context()) }); -#else - otSpan lspan = 0; -#endif - + RmrMessagesBuffer_t &rmrMessageBuffer) { auto logLevel = mdclog_level_get(); auto procedureCode = ((InitiatingMessage_t *) pdu->choice.initiatingMessage)->procedureCode; if (logLevel >= MDCLOG_DEBUG) { @@ -1473,26 +1655,56 @@ void asnInitiatingRequest(E2AP_PDU_t *pdu, switch (procedureCode) { case ProcedureCode_id_E2setup: { if (logLevel >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "Got E2setup\n"); + mdclog_write(MDCLOG_DEBUG, "Got E2setup"); } - memset(message.peerInfo->enodbName, 0 , MAX_ENODB_NAME_SIZE); - for (auto i = 0; i < pdu->choice.initiatingMessage->value.choice.E2setupRequest.protocolIEs.list.count; i++) { - auto *ie = pdu->choice.initiatingMessage->value.choice.E2setupRequest.protocolIEs.list.array[i]; - if (ie->id == ProtocolIE_ID_id_GlobalE2node_ID) { - if (ie->value.present == E2setupRequestIEs__value_PR_GlobalE2node_ID) { - buildAndsendSetupRequest(message, ie, rmrMessageBuffer, pdu); - break; - } - } +// vector RANfunctionsAdded_v; +// vector RANfunctionsModified_v; +// RANfunctionsAdded_v.clear(); +// RANfunctionsModified_v.clear(); + if (collectSetupRequestData(pdu, sctpMap, message) != 0) { + break; } + + buildPrometheuslist(message.peerInfo, message.peerInfo->sctpParams->prometheusFamily); + + string messageName("E2setupRequest"); + string ieName("E2setupRequestIEs"); + message.message.messageType = RIC_E2_SETUP_REQ; + message.peerInfo->counters[IN_INITI][MSG_COUNTER][ProcedureCode_id_E2setup]->Increment(); + message.peerInfo->counters[IN_INITI][BYTES_COUNTER][ProcedureCode_id_E2setup]->Increment((double)message.message.asnLength); + buildAndsendSetupRequest(message, rmrMessageBuffer, pdu); + break; + } + case ProcedureCode_id_RICserviceUpdate: { + if (logLevel >= MDCLOG_DEBUG) { + mdclog_write(MDCLOG_DEBUG, "Got RICserviceUpdate %s", message.message.enodbName); + } +// vector RANfunctionsAdded_v; +// vector RANfunctionsModified_v; +// RANfunctionsAdded_v.clear(); +// RANfunctionsModified_v.clear(); +// if (collectServiceUpdate_RequestData(pdu, sctpMap, message, +// RANfunctionsAdded_v, RANfunctionsModified_v) != 0) { +// break; +// } + + string messageName("RICserviceUpdate"); + string ieName("RICserviceUpdateIEs"); + message.message.messageType = RIC_SERVICE_UPDATE; + message.peerInfo->counters[IN_INITI][MSG_COUNTER][ProcedureCode_id_RICserviceUpdate]->Increment(); + message.peerInfo->counters[IN_INITI][BYTES_COUNTER][ProcedureCode_id_RICserviceUpdate]->Increment((double)message.message.asnLength); + + buildAndsendSetupRequest(message, rmrMessageBuffer, pdu); break; } case ProcedureCode_id_ErrorIndication: { if (logLevel >= MDCLOG_DEBUG) { mdclog_write(MDCLOG_DEBUG, "Got ErrorIndication %s", message.message.enodbName); } - if (sendRequestToXapp(message, RIC_ERROR_INDICATION, rmrMessageBuffer, &lspan) != 0) { + message.peerInfo->counters[IN_INITI][MSG_COUNTER][ProcedureCode_id_ErrorIndication]->Increment(); + message.peerInfo->counters[IN_INITI][BYTES_COUNTER][ProcedureCode_id_ErrorIndication]->Increment((double)message.message.asnLength); + if (sendRequestToXapp(message, RIC_ERROR_INDICATION, rmrMessageBuffer) != 0) { mdclog_write(MDCLOG_ERR, "RIC_ERROR_INDICATION failed to send to xAPP"); } break; @@ -1501,14 +1713,15 @@ void asnInitiatingRequest(E2AP_PDU_t *pdu, if (logLevel >= MDCLOG_DEBUG) { mdclog_write(MDCLOG_DEBUG, "Got Reset %s", message.message.enodbName); } - if (sendRequestToXapp(message, RIC_X2_RESET, rmrMessageBuffer, &lspan) != 0) { - mdclog_write(MDCLOG_ERR, "RIC_X2_RESET message failed to send to xAPP"); + + message.peerInfo->counters[IN_INITI][MSG_COUNTER][ProcedureCode_id_Reset]->Increment(); + message.peerInfo->counters[IN_INITI][BYTES_COUNTER][ProcedureCode_id_Reset]->Increment((double)message.message.asnLength); + if (XML_From_PER(message, rmrMessageBuffer) < 0) { + break; } - break; - } - case ProcedureCode_id_RICcontrol: { - if (logLevel >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "Got RICcontrol %s", message.message.enodbName); + + if (sendRequestToXapp(message, RIC_E2_RESET_REQ, rmrMessageBuffer) != 0) { + mdclog_write(MDCLOG_ERR, "RIC_E2_RESET_REQ message failed to send to xAPP"); } break; } @@ -1535,13 +1748,19 @@ void asnInitiatingRequest(E2AP_PDU_t *pdu, (unsigned char *)message.message.enodbName, strlen(message.message.enodbName)); rmrMessageBuffer.sendMessage->state = 0; - rmrMessageBuffer.sendMessage->sub_id = (int) ie->value.choice.RICrequestID.ricRequestorID; + rmrMessageBuffer.sendMessage->sub_id = (int)ie->value.choice.RICrequestID.ricInstanceID; + + //ie->value.choice.RICrequestID.ricInstanceID; if (mdclog_level_get() >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "RIC sub id = %d, message type = %d", + mdclog_write(MDCLOG_DEBUG, "sub id = %d, mtype = %d, ric instance id %ld, requestor id = %ld", rmrMessageBuffer.sendMessage->sub_id, - rmrMessageBuffer.sendMessage->mtype); + rmrMessageBuffer.sendMessage->mtype, + ie->value.choice.RICrequestID.ricInstanceID, + ie->value.choice.RICrequestID.ricRequestorID); } - sendRmrMessage(rmrMessageBuffer, message, &lspan); + message.peerInfo->counters[IN_INITI][MSG_COUNTER][ProcedureCode_id_RICindication]->Increment(); + message.peerInfo->counters[IN_INITI][BYTES_COUNTER][ProcedureCode_id_RICindication]->Increment((double)message.message.asnLength); + sendRmrMessage(rmrMessageBuffer, message); messageSent = true; } else { mdclog_write(MDCLOG_ERR, "RIC request id missing illigal request"); @@ -1553,33 +1772,6 @@ void asnInitiatingRequest(E2AP_PDU_t *pdu, } break; } - case ProcedureCode_id_RICserviceQuery: { - if (logLevel >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "Got RICserviceQuery %s", message.message.enodbName); - } - break; - } - case ProcedureCode_id_RICserviceUpdate: { - if (logLevel >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "Got RICserviceUpdate %s", message.message.enodbName); - } - if (sendRequestToXapp(message, RIC_SERVICE_UPDATE, rmrMessageBuffer, &lspan) != 0) { - mdclog_write(MDCLOG_ERR, "RIC_SERVICE_UPDATE message failed to send to xAPP"); - } - break; - } - case ProcedureCode_id_RICsubscription: { - if (logLevel >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "Got RICsubscription %s", message.message.enodbName); - } - break; - } - case ProcedureCode_id_RICsubscriptionDelete: { - if (logLevel >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "Got RICsubscriptionDelete %s", message.message.enodbName); - } - break; - } default: { mdclog_write(MDCLOG_ERR, "Undefined or not supported message = %ld", procedureCode); message.message.messageType = 0; // no RMR message type yet @@ -1589,55 +1781,35 @@ void asnInitiatingRequest(E2AP_PDU_t *pdu, break; } } -#ifdef __TRACING__ - lspan->Finish(); -#endif - } /** * * @param pdu * @param message - * @param sctpMap * @param rmrMessageBuffer - * @param pSpan */ -void asnSuccsesfulMsg(E2AP_PDU_t *pdu, ReportingMessages_t &message, Sctp_Map_t *sctpMap, - RmrMessagesBuffer_t &rmrMessageBuffer, otSpan *pSpan) { -#ifdef __TRACING__ - auto lspan = opentracing::Tracer::Global()->StartSpan( - __FUNCTION__, { opentracing::ChildOf(&pSpan->get()->context()) }); -#else - otSpan lspan = 0; -#endif +void asnSuccsesfulMsg(E2AP_PDU_t *pdu, + Sctp_Map_t *sctpMap, + ReportingMessages_t &message, + RmrMessagesBuffer_t &rmrMessageBuffer) { auto procedureCode = pdu->choice.successfulOutcome->procedureCode; auto logLevel = mdclog_level_get(); if (logLevel >= MDCLOG_INFO) { mdclog_write(MDCLOG_INFO, "Successful Outcome %ld", procedureCode); } switch (procedureCode) { - case ProcedureCode_id_E2setup: { - if (logLevel >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "Got E2setup\n"); - } - break; - } - case ProcedureCode_id_ErrorIndication: { - if (logLevel >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "Got ErrorIndication %s", message.message.enodbName); - } - if (sendRequestToXapp(message, RIC_ERROR_INDICATION, rmrMessageBuffer, &lspan) != 0) { - mdclog_write(MDCLOG_ERR, "RIC_ERROR_INDICATION failed to send to xAPP"); - } - break; - } case ProcedureCode_id_Reset: { if (logLevel >= MDCLOG_DEBUG) { mdclog_write(MDCLOG_DEBUG, "Got Reset %s", message.message.enodbName); } - if (sendRequestToXapp(message, RIC_X2_RESET, rmrMessageBuffer, &lspan) != 0) { - mdclog_write(MDCLOG_ERR, "RIC_X2_RESET message failed to send to xAPP"); + message.peerInfo->counters[IN_SUCC][MSG_COUNTER][ProcedureCode_id_Reset]->Increment(); + message.peerInfo->counters[IN_SUCC][BYTES_COUNTER][ProcedureCode_id_Reset]->Increment((double)message.message.asnLength); + if (XML_From_PER(message, rmrMessageBuffer) < 0) { + break; + } + if (sendRequestToXapp(message, RIC_E2_RESET_RESP, rmrMessageBuffer) != 0) { + mdclog_write(MDCLOG_ERR, "RIC_E2_RESET_RESP message failed to send to xAPP"); } break; } @@ -1659,7 +1831,9 @@ void asnSuccsesfulMsg(E2AP_PDU_t *pdu, ReportingMessages_t &message, Sctp_Map_t if (ie->value.present == RICcontrolAcknowledge_IEs__value_PR_RICrequestID) { message.message.messageType = rmrMessageBuffer.sendMessage->mtype = RIC_CONTROL_ACK; rmrMessageBuffer.sendMessage->state = 0; - rmrMessageBuffer.sendMessage->sub_id = (int) ie->value.choice.RICrequestID.ricRequestorID; +// rmrMessageBuffer.sendMessage->sub_id = (int) ie->value.choice.RICrequestID.ricRequestorID; + rmrMessageBuffer.sendMessage->sub_id = (int)ie->value.choice.RICrequestID.ricInstanceID; + static unsigned char tx[32]; snprintf((char *) tx, sizeof tx, "%15ld", transactionCounter++); rmr_bytes2xact(rmrMessageBuffer.sendMessage, tx, strlen((const char *) tx)); @@ -1667,7 +1841,9 @@ void asnSuccsesfulMsg(E2AP_PDU_t *pdu, ReportingMessages_t &message, Sctp_Map_t (unsigned char *)message.message.enodbName, strlen(message.message.enodbName)); - sendRmrMessage(rmrMessageBuffer, message, &lspan); + message.peerInfo->counters[IN_SUCC][MSG_COUNTER][ProcedureCode_id_RICcontrol]->Increment(); + message.peerInfo->counters[IN_SUCC][BYTES_COUNTER][ProcedureCode_id_RICcontrol]->Increment((double)message.message.asnLength); + sendRmrMessage(rmrMessageBuffer, message); messageSent = true; } else { mdclog_write(MDCLOG_ERR, "RIC request id missing illigal request"); @@ -1680,67 +1856,13 @@ void asnSuccsesfulMsg(E2AP_PDU_t *pdu, ReportingMessages_t &message, Sctp_Map_t break; } - case ProcedureCode_id_RICindication: { - if (logLevel >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "Got RICindication %s", message.message.enodbName); - } - for (auto i = 0; i < pdu->choice.initiatingMessage->value.choice.RICindication.protocolIEs.list.count; i++) { - auto messageSent = false; - RICindication_IEs_t *ie = pdu->choice.initiatingMessage->value.choice.RICindication.protocolIEs.list.array[i]; - if (logLevel >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "ie type (ProtocolIE_ID) = %ld", ie->id); - } - if (ie->id == ProtocolIE_ID_id_RICrequestID) { - if (logLevel >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "Got RIC requestId entry, ie type (ProtocolIE_ID) = %ld", ie->id); - } - if (ie->value.present == RICindication_IEs__value_PR_RICrequestID) { - static unsigned char tx[32]; - message.message.messageType = rmrMessageBuffer.sendMessage->mtype = RIC_INDICATION; - snprintf((char *) tx, sizeof tx, "%15ld", transactionCounter++); - rmr_bytes2xact(rmrMessageBuffer.sendMessage, tx, strlen((const char *) tx)); - rmr_bytes2meid(rmrMessageBuffer.sendMessage, - (unsigned char *)message.message.enodbName, - strlen(message.message.enodbName)); - rmrMessageBuffer.sendMessage->state = 0; - rmrMessageBuffer.sendMessage->sub_id = (int) ie->value.choice.RICrequestID.ricRequestorID; - if (mdclog_level_get() >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "RIC sub id = %d, message type = %d", - rmrMessageBuffer.sendMessage->sub_id, - rmrMessageBuffer.sendMessage->mtype); - } - sendRmrMessage(rmrMessageBuffer, message, &lspan); - messageSent = true; - } else { - mdclog_write(MDCLOG_ERR, "RIC request id missing illigal request"); - } - } - if (messageSent) { - break; - } - } - break; - } - case ProcedureCode_id_RICserviceQuery: { - if (logLevel >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "Got RICserviceQuery %s", message.message.enodbName); - } - break; - } - case ProcedureCode_id_RICserviceUpdate: { - if (logLevel >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "Got RICserviceUpdate %s", message.message.enodbName); - } - if (sendRequestToXapp(message, RIC_SERVICE_UPDATE, rmrMessageBuffer, &lspan) != 0) { - mdclog_write(MDCLOG_ERR, "RIC_SERVICE_UPDATE message failed to send to xAPP"); - } - break; - } case ProcedureCode_id_RICsubscription: { if (logLevel >= MDCLOG_DEBUG) { mdclog_write(MDCLOG_DEBUG, "Got RICsubscription %s", message.message.enodbName); } - if (sendRequestToXapp(message, RIC_SUB_RESP, rmrMessageBuffer, &lspan) != 0) { + message.peerInfo->counters[IN_SUCC][MSG_COUNTER][ProcedureCode_id_RICsubscription]->Increment(); + message.peerInfo->counters[IN_SUCC][BYTES_COUNTER][ProcedureCode_id_RICsubscription]->Increment((double)message.message.asnLength); + if (sendRequestToXapp(message, RIC_SUB_RESP, rmrMessageBuffer) != 0) { mdclog_write(MDCLOG_ERR, "Subscription successful message failed to send to xAPP"); } break; @@ -1749,7 +1871,9 @@ void asnSuccsesfulMsg(E2AP_PDU_t *pdu, ReportingMessages_t &message, Sctp_Map_t if (logLevel >= MDCLOG_DEBUG) { mdclog_write(MDCLOG_DEBUG, "Got RICsubscriptionDelete %s", message.message.enodbName); } - if (sendRequestToXapp(message, RIC_SUB_DEL_RESP, rmrMessageBuffer, &lspan) != 0) { + message.peerInfo->counters[IN_SUCC][MSG_COUNTER][ProcedureCode_id_RICsubscriptionDelete]->Increment(); + message.peerInfo->counters[IN_SUCC][BYTES_COUNTER][ProcedureCode_id_RICsubscriptionDelete]->Increment((double)message.message.asnLength); + if (sendRequestToXapp(message, RIC_SUB_DEL_RESP, rmrMessageBuffer) != 0) { mdclog_write(MDCLOG_ERR, "Subscription delete successful message failed to send to xAPP"); } break; @@ -1762,61 +1886,24 @@ void asnSuccsesfulMsg(E2AP_PDU_t *pdu, ReportingMessages_t &message, Sctp_Map_t break; } } -#ifdef __TRACING__ - lspan->Finish(); -#endif - } /** * * @param pdu * @param message - * @param sctpMap * @param rmrMessageBuffer - * @param pSpan */ void asnUnSuccsesfulMsg(E2AP_PDU_t *pdu, - ReportingMessages_t &message, Sctp_Map_t *sctpMap, - RmrMessagesBuffer_t &rmrMessageBuffer, - otSpan *pSpan) { -#ifdef __TRACING__ - auto lspan = opentracing::Tracer::Global()->StartSpan( - __FUNCTION__, { opentracing::ChildOf(&pSpan->get()->context()) }); -#else - otSpan lspan = 0; -#endif + ReportingMessages_t &message, + RmrMessagesBuffer_t &rmrMessageBuffer) { auto procedureCode = pdu->choice.unsuccessfulOutcome->procedureCode; auto logLevel = mdclog_level_get(); if (logLevel >= MDCLOG_INFO) { mdclog_write(MDCLOG_INFO, "Unsuccessful Outcome %ld", procedureCode); } switch (procedureCode) { - case ProcedureCode_id_E2setup: { - if (logLevel >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "Got E2setup\n"); - } - break; - } - case ProcedureCode_id_ErrorIndication: { - if (logLevel >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "Got ErrorIndication %s", message.message.enodbName); - } - if (sendRequestToXapp(message, RIC_ERROR_INDICATION, rmrMessageBuffer, &lspan) != 0) { - mdclog_write(MDCLOG_ERR, "RIC_ERROR_INDICATION failed to send to xAPP"); - } - break; - } - case ProcedureCode_id_Reset: { - if (logLevel >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "Got Reset %s", message.message.enodbName); - } - if (sendRequestToXapp(message, RIC_X2_RESET, rmrMessageBuffer, &lspan) != 0) { - mdclog_write(MDCLOG_ERR, "RIC_X2_RESET message failed to send to xAPP"); - } - break; - } case ProcedureCode_id_RICcontrol: { if (logLevel >= MDCLOG_DEBUG) { mdclog_write(MDCLOG_DEBUG, "Got RICcontrol %s", message.message.enodbName); @@ -1835,13 +1922,16 @@ void asnUnSuccsesfulMsg(E2AP_PDU_t *pdu, if (ie->value.present == RICcontrolFailure_IEs__value_PR_RICrequestID) { message.message.messageType = rmrMessageBuffer.sendMessage->mtype = RIC_CONTROL_FAILURE; rmrMessageBuffer.sendMessage->state = 0; - rmrMessageBuffer.sendMessage->sub_id = (int) ie->value.choice.RICrequestID.ricRequestorID; +// rmrMessageBuffer.sendMessage->sub_id = (int)ie->value.choice.RICrequestID.ricRequestorID; + rmrMessageBuffer.sendMessage->sub_id = (int)ie->value.choice.RICrequestID.ricInstanceID; static unsigned char tx[32]; snprintf((char *) tx, sizeof tx, "%15ld", transactionCounter++); rmr_bytes2xact(rmrMessageBuffer.sendMessage, tx, strlen((const char *) tx)); rmr_bytes2meid(rmrMessageBuffer.sendMessage, (unsigned char *) message.message.enodbName, strlen(message.message.enodbName)); - sendRmrMessage(rmrMessageBuffer, message, &lspan); + message.peerInfo->counters[IN_UN_SUCC][MSG_COUNTER][ProcedureCode_id_RICcontrol]->Increment(); + message.peerInfo->counters[IN_UN_SUCC][BYTES_COUNTER][ProcedureCode_id_RICcontrol]->Increment((double)message.message.asnLength); + sendRmrMessage(rmrMessageBuffer, message); messageSent = true; } else { mdclog_write(MDCLOG_ERR, "RIC request id missing illigal request"); @@ -1853,67 +1943,13 @@ void asnUnSuccsesfulMsg(E2AP_PDU_t *pdu, } break; } - case ProcedureCode_id_RICindication: { - if (logLevel >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "Got RICindication %s", message.message.enodbName); - } - for (auto i = 0; i < pdu->choice.initiatingMessage->value.choice.RICindication.protocolIEs.list.count; i++) { - auto messageSent = false; - RICindication_IEs_t *ie = pdu->choice.initiatingMessage->value.choice.RICindication.protocolIEs.list.array[i]; - if (logLevel >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "ie type (ProtocolIE_ID) = %ld", ie->id); - } - if (ie->id == ProtocolIE_ID_id_RICrequestID) { - if (logLevel >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "Got RIC requestId entry, ie type (ProtocolIE_ID) = %ld", ie->id); - } - if (ie->value.present == RICindication_IEs__value_PR_RICrequestID) { - static unsigned char tx[32]; - message.message.messageType = rmrMessageBuffer.sendMessage->mtype = RIC_INDICATION; - snprintf((char *) tx, sizeof tx, "%15ld", transactionCounter++); - rmr_bytes2xact(rmrMessageBuffer.sendMessage, tx, strlen((const char *) tx)); - rmr_bytes2meid(rmrMessageBuffer.sendMessage, - (unsigned char *)message.message.enodbName, - strlen(message.message.enodbName)); - rmrMessageBuffer.sendMessage->state = 0; - rmrMessageBuffer.sendMessage->sub_id = (int) ie->value.choice.RICrequestID.ricRequestorID; - if (mdclog_level_get() >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "RIC sub id = %d, message type = %d", - rmrMessageBuffer.sendMessage->sub_id, - rmrMessageBuffer.sendMessage->mtype); - } - sendRmrMessage(rmrMessageBuffer, message, &lspan); - messageSent = true; - } else { - mdclog_write(MDCLOG_ERR, "RIC request id missing illigal request"); - } - } - if (messageSent) { - break; - } - } - break; - } - case ProcedureCode_id_RICserviceQuery: { - if (logLevel >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "Got RICserviceQuery %s", message.message.enodbName); - } - break; - } - case ProcedureCode_id_RICserviceUpdate: { - if (logLevel >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "Got RICserviceUpdate %s", message.message.enodbName); - } - if (sendRequestToXapp(message, RIC_SERVICE_UPDATE, rmrMessageBuffer, &lspan) != 0) { - mdclog_write(MDCLOG_ERR, "RIC_SERVICE_UPDATE message failed to send to xAPP"); - } - break; - } case ProcedureCode_id_RICsubscription: { if (logLevel >= MDCLOG_DEBUG) { mdclog_write(MDCLOG_DEBUG, "Got RICsubscription %s", message.message.enodbName); } - if (sendRequestToXapp(message, RIC_SUB_FAILURE, rmrMessageBuffer, &lspan) != 0) { + message.peerInfo->counters[IN_UN_SUCC][MSG_COUNTER][ProcedureCode_id_RICsubscription]->Increment(); + message.peerInfo->counters[IN_UN_SUCC][BYTES_COUNTER][ProcedureCode_id_RICsubscription]->Increment((double)message.message.asnLength); + if (sendRequestToXapp(message, RIC_SUB_FAILURE, rmrMessageBuffer) != 0) { mdclog_write(MDCLOG_ERR, "Subscription unsuccessful message failed to send to xAPP"); } break; @@ -1922,7 +1958,9 @@ void asnUnSuccsesfulMsg(E2AP_PDU_t *pdu, if (logLevel >= MDCLOG_DEBUG) { mdclog_write(MDCLOG_DEBUG, "Got RICsubscriptionDelete %s", message.message.enodbName); } - if (sendRequestToXapp(message, RIC_SUB_DEL_FAILURE, rmrMessageBuffer, &lspan) != 0) { + message.peerInfo->counters[IN_UN_SUCC][MSG_COUNTER][ProcedureCode_id_RICsubscriptionDelete]->Increment(); + message.peerInfo->counters[IN_UN_SUCC][BYTES_COUNTER][ProcedureCode_id_RICsubscriptionDelete]->Increment((double)message.message.asnLength); + if (sendRequestToXapp(message, RIC_SUB_FAILURE, rmrMessageBuffer) != 0) { mdclog_write(MDCLOG_ERR, "Subscription Delete unsuccessful message failed to send to xAPP"); } break; @@ -1936,10 +1974,6 @@ void asnUnSuccsesfulMsg(E2AP_PDU_t *pdu, break; } } -#ifdef __TRACING__ - lspan->Finish(); -#endif - } /** @@ -1947,19 +1981,11 @@ void asnUnSuccsesfulMsg(E2AP_PDU_t *pdu, * @param message * @param requestId * @param rmrMmessageBuffer - * @param pSpan * @return */ int sendRequestToXapp(ReportingMessages_t &message, int requestId, - RmrMessagesBuffer_t &rmrMmessageBuffer, - otSpan *pSpan) { -#ifdef __TRACING__ - auto lspan = opentracing::Tracer::Global()->StartSpan( - __FUNCTION__, { opentracing::ChildOf(&pSpan->get()->context()) }); -#else - otSpan lspan = 0; -#endif + RmrMessagesBuffer_t &rmrMmessageBuffer) { rmr_bytes2meid(rmrMmessageBuffer.sendMessage, (unsigned char *)message.message.enodbName, strlen(message.message.enodbName)); @@ -1969,29 +1995,19 @@ int sendRequestToXapp(ReportingMessages_t &message, snprintf((char *) tx, sizeof tx, "%15ld", transactionCounter++); rmr_bytes2xact(rmrMmessageBuffer.sendMessage, tx, strlen((const char *) tx)); - auto rc = sendRmrMessage(rmrMmessageBuffer, message, &lspan); -#ifdef __TRACING__ - lspan->Finish(); -#endif - + auto rc = sendRmrMessage(rmrMmessageBuffer, message); return rc; } - -void getRmrContext(sctp_params_t &pSctpParams, otSpan *pSpan) { -#ifdef __TRACING__ - auto lspan = opentracing::Tracer::Global()->StartSpan( - __FUNCTION__, { opentracing::ChildOf(&pSpan->get()->context()) }); -#else -// otSpan lspan = 0; -#endif +/** + * + * @param pSctpParams + */ +void getRmrContext(sctp_params_t &pSctpParams) { pSctpParams.rmrCtx = nullptr; - pSctpParams.rmrCtx = rmr_init(pSctpParams.rmrAddress, RMR_MAX_RCV_BYTES, RMRFL_NONE); + pSctpParams.rmrCtx = rmr_init(pSctpParams.rmrAddress, RECEIVE_XAPP_BUFFER_SIZE, RMRFL_NONE); if (pSctpParams.rmrCtx == nullptr) { mdclog_write(MDCLOG_ERR, "Failed to initialize RMR"); -#ifdef __TRACING__ - lspan->Finish(); -#endif return; } @@ -2014,9 +2030,6 @@ void getRmrContext(sctp_params_t &pSctpParams, otSpan *pSpan) { if (mdclog_level_get() >= MDCLOG_INFO) { mdclog_write(MDCLOG_INFO, "RMR running"); } -#ifdef __TRACING__ - lspan->Finish(); -#endif rmr_init_trace(pSctpParams.rmrCtx, 200); // get the RMR fd for the epoll pSctpParams.rmrListenFd = rmr_get_rcvfd(pSctpParams.rmrCtx); @@ -2035,31 +2048,57 @@ void getRmrContext(sctp_params_t &pSctpParams, otSpan *pSpan) { /** * - * @param epoll_fd + * @param message + * @param rmrMessageBuffer + * @return + */ +int PER_FromXML(ReportingMessages_t &message, RmrMessagesBuffer_t &rmrMessageBuffer) { + E2AP_PDU_t *pdu = nullptr; + + if (mdclog_level_get() >= MDCLOG_DEBUG) { + mdclog_write(MDCLOG_DEBUG, "got xml Format data from xApp of size %d is:%s", + rmrMessageBuffer.rcvMessage->len, rmrMessageBuffer.rcvMessage->payload); + } + auto rval = asn_decode(nullptr, ATS_BASIC_XER, &asn_DEF_E2AP_PDU, (void **) &pdu, + rmrMessageBuffer.rcvMessage->payload, rmrMessageBuffer.rcvMessage->len); + if (rval.code != RC_OK) { + mdclog_write(MDCLOG_ERR, "Error %d Decoding (unpack) setup response from E2MGR : %s", + rval.code, + message.message.enodbName); + return -1; + } + + int buff_size = RECEIVE_XAPP_BUFFER_SIZE; + auto er = asn_encode_to_buffer(nullptr, ATS_ALIGNED_BASIC_PER, &asn_DEF_E2AP_PDU, pdu, + rmrMessageBuffer.rcvMessage->payload, buff_size); + if (er.encoded == -1) { + mdclog_write(MDCLOG_ERR, "encoding of %s failed, %s", asn_DEF_E2AP_PDU.name, strerror(errno)); + return -1; + } else if (er.encoded > (ssize_t)buff_size) { + mdclog_write(MDCLOG_ERR, "Buffer of size %d is to small for %s, at %s line %d", + (int)rmrMessageBuffer.rcvMessage->len, + asn_DEF_E2AP_PDU.name, + __func__, + __LINE__); + return -1; + } + rmrMessageBuffer.rcvMessage->len = er.encoded; + return 0; +} + +/** + * * @param sctpMap * @param rmrMessageBuffer * @param ts - * @param pSpan * @return */ -int receiveXappMessages(int epoll_fd, - Sctp_Map_t *sctpMap, +int receiveXappMessages(Sctp_Map_t *sctpMap, RmrMessagesBuffer_t &rmrMessageBuffer, - struct timespec &ts, - otSpan *pSpan) { -#ifdef __TRACING__ - auto lspan = opentracing::Tracer::Global()->StartSpan( - __FUNCTION__, { opentracing::ChildOf(&pSpan->get()->context()) }); -#else - otSpan lspan = 0; -#endif + struct timespec &ts) { if (rmrMessageBuffer.rcvMessage == nullptr) { //we have error mdclog_write(MDCLOG_ERR, "RMR Allocation message, %s", strerror(errno)); -#ifdef __TRACING__ - lspan->Finish(); -#endif - return -1; } @@ -2070,10 +2109,6 @@ int receiveXappMessages(int epoll_fd, if (rmrMessageBuffer.rcvMessage == nullptr) { mdclog_write(MDCLOG_ERR, "RMR Receving message with null pointer, Realloc rmr mesage buffer"); rmrMessageBuffer.rcvMessage = rmr_alloc_msg(rmrMessageBuffer.rmrCtx, RECEIVE_XAPP_BUFFER_SIZE); -#ifdef __TRACING__ - lspan->Finish(); -#endif - return -2; } ReportingMessages_t message; @@ -2085,212 +2120,140 @@ int receiveXappMessages(int epoll_fd, //auto msgData = msg->payload; if (rmrMessageBuffer.rcvMessage->state != 0) { mdclog_write(MDCLOG_ERR, "RMR Receving message with stat = %d", rmrMessageBuffer.rcvMessage->state); -#ifdef __TRACING__ - lspan->Finish(); -#endif - return -1; } - switch (rmrMessageBuffer.rcvMessage->mtype) { - case RIC_X2_SETUP_REQ: { - if (connectToCUandSetUp(rmrMessageBuffer, message, epoll_fd, sctpMap, &lspan) != 0) { - mdclog_write(MDCLOG_ERR, "ERROR in connectToCUandSetUp on RIC_X2_SETUP_REQ"); - message.message.messageType = rmrMessageBuffer.sendMessage->mtype = RIC_SCTP_CONNECTION_FAILURE; - message.message.direction = 'N'; - message.message.asnLength = rmrMessageBuffer.sendMessage->len = - snprintf((char *)rmrMessageBuffer.sendMessage->payload, - 256, - "ERROR in connectToCUandSetUp on RIC_X2_SETUP_REQ"); - rmrMessageBuffer.sendMessage->state = 0; - message.message.asndata = rmrMessageBuffer.sendMessage->payload; - - if (mdclog_level_get() >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "start writing to rmr buffer"); - } - rmr_bytes2xact(rmrMessageBuffer.sendMessage, rmrMessageBuffer.rcvMessage->xaction, RMR_MAX_XID); - rmr_str2meid(rmrMessageBuffer.sendMessage, (unsigned char *)message.message.enodbName); - - sendRmrMessage(rmrMessageBuffer, message, &lspan); -#ifdef __TRACING__ - lspan->Finish(); -#endif - return -3; - } - break; - } - case RIC_ENDC_X2_SETUP_REQ: { - if (connectToCUandSetUp(rmrMessageBuffer, message, epoll_fd, sctpMap, &lspan) != 0) { - mdclog_write(MDCLOG_ERR, "ERROR in connectToCUandSetUp on RIC_ENDC_X2_SETUP_REQ"); - message.message.messageType = rmrMessageBuffer.sendMessage->mtype = RIC_SCTP_CONNECTION_FAILURE; - message.message.direction = 'N'; - message.message.asnLength = rmrMessageBuffer.sendMessage->len = - snprintf((char *)rmrMessageBuffer.sendMessage->payload, 256, - "ERROR in connectToCUandSetUp on RIC_ENDC_X2_SETUP_REQ"); - rmrMessageBuffer.sendMessage->state = 0; - message.message.asndata = rmrMessageBuffer.sendMessage->payload; - - if (mdclog_level_get() >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "start writing to rmr buffer"); - } - - rmr_bytes2xact(rmrMessageBuffer.sendMessage, rmrMessageBuffer.rcvMessage->xaction, RMR_MAX_XID); - rmr_str2meid(rmrMessageBuffer.sendMessage, (unsigned char *) message.message.enodbName); - - sendRmrMessage(rmrMessageBuffer, message, &lspan); -#ifdef __TRACING__ - lspan->Finish(); -#endif - return -3; - } - break; - } - case RIC_ENDC_CONF_UPDATE: { - if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap, &lspan) != 0) { - mdclog_write(MDCLOG_ERR, "Failed to send RIC_ENDC_CONF_UPDATE"); -#ifdef __TRACING__ - lspan->Finish(); -#endif - return -4; - } - break; - } - case RIC_ENDC_CONF_UPDATE_ACK: { - if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap, &lspan) != 0) { - mdclog_write(MDCLOG_ERR, "Failed to send RIC_ENDC_CONF_UPDATE_ACK"); -#ifdef __TRACING__ - lspan->Finish(); -#endif - return -4; - } - break; + rmr_get_meid(rmrMessageBuffer.rcvMessage, (unsigned char *)message.message.enodbName); + message.peerInfo = (ConnectedCU_t *) sctpMap->find(message.message.enodbName); + if (message.peerInfo == nullptr) { + auto type = rmrMessageBuffer.rcvMessage->mtype; + switch (type) { + case RIC_SCTP_CLEAR_ALL: + case E2_TERM_KEEP_ALIVE_REQ: + case RIC_HEALTH_CHECK_REQ: + break; + default: + mdclog_write(MDCLOG_ERR, "Failed to send message no CU entry %s", message.message.enodbName); + return -1; } - case RIC_ENDC_CONF_UPDATE_FAILURE: { - if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap, &lspan) != 0) { - mdclog_write(MDCLOG_ERR, "Failed to send RIC_ENDC_CONF_UPDATE_FAILURE"); -#ifdef __TRACING__ - lspan->Finish(); -#endif + } - return -4; + switch (rmrMessageBuffer.rcvMessage->mtype) { + case RIC_E2_SETUP_RESP : { + if (PER_FromXML(message, rmrMessageBuffer) != 0) { + break; } - break; - } - case RIC_ENB_CONF_UPDATE: { - if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap, &lspan) != 0) { - mdclog_write(MDCLOG_ERR, "Failed to send RIC_ENDC_CONF_UPDATE"); -#ifdef __TRACING__ - lspan->Finish(); -#endif - return -4; + message.peerInfo->counters[OUT_SUCC][MSG_COUNTER][ProcedureCode_id_E2setup]->Increment(); + message.peerInfo->counters[OUT_SUCC][BYTES_COUNTER][ProcedureCode_id_E2setup]->Increment(rmrMessageBuffer.rcvMessage->len); + if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap) != 0) { + mdclog_write(MDCLOG_ERR, "Failed to send RIC_E2_SETUP_RESP"); + return -6; } break; } - case RIC_ENB_CONF_UPDATE_ACK: { - if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap, &lspan) != 0) { - mdclog_write(MDCLOG_ERR, "Failed to send RIC_ENB_CONF_UPDATE_ACK"); -#ifdef __TRACING__ - lspan->Finish(); -#endif - return -4; + case RIC_E2_SETUP_FAILURE : { + if (PER_FromXML(message, rmrMessageBuffer) != 0) { + break; } - break; - } - case RIC_ENB_CONF_UPDATE_FAILURE: { - if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap, &lspan) != 0) { - mdclog_write(MDCLOG_ERR, "Failed to send RIC_ENB_CONF_UPDATE_FAILURE"); -#ifdef __TRACING__ - lspan->Finish(); -#endif - return -4; + message.peerInfo->counters[OUT_UN_SUCC][MSG_COUNTER][ProcedureCode_id_E2setup]->Increment(); + message.peerInfo->counters[OUT_UN_SUCC][BYTES_COUNTER][ProcedureCode_id_E2setup]->Increment(rmrMessageBuffer.rcvMessage->len); + if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap) != 0) { + mdclog_write(MDCLOG_ERR, "Failed to send RIC_E2_SETUP_FAILURE"); + return -6; } break; } - case RIC_RES_STATUS_REQ: { - if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap, &lspan) != 0) { - mdclog_write(MDCLOG_ERR, "Failed to send RIC_RES_STATUS_REQ"); -#ifdef __TRACING__ - lspan->Finish(); -#endif + case RIC_ERROR_INDICATION: { + message.peerInfo->counters[OUT_INITI][MSG_COUNTER][ProcedureCode_id_ErrorIndication]->Increment(); + message.peerInfo->counters[OUT_INITI][BYTES_COUNTER][ProcedureCode_id_ErrorIndication]->Increment(rmrMessageBuffer.rcvMessage->len); + if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap) != 0) { + mdclog_write(MDCLOG_ERR, "Failed to send RIC_ERROR_INDICATION"); return -6; } break; } case RIC_SUB_REQ: { - if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap, &lspan) != 0) { + message.peerInfo->counters[OUT_INITI][MSG_COUNTER][ProcedureCode_id_RICsubscription]->Increment(); + message.peerInfo->counters[OUT_INITI][BYTES_COUNTER][ProcedureCode_id_RICsubscription]->Increment(rmrMessageBuffer.rcvMessage->len); + if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap) != 0) { mdclog_write(MDCLOG_ERR, "Failed to send RIC_SUB_REQ"); -#ifdef __TRACING__ - lspan->Finish(); -#endif return -6; } break; } case RIC_SUB_DEL_REQ: { - if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap, &lspan) != 0) { + message.peerInfo->counters[OUT_INITI][MSG_COUNTER][ProcedureCode_id_RICsubscriptionDelete]->Increment(); + message.peerInfo->counters[OUT_INITI][BYTES_COUNTER][ProcedureCode_id_RICsubscriptionDelete]->Increment(rmrMessageBuffer.rcvMessage->len); + if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap) != 0) { mdclog_write(MDCLOG_ERR, "Failed to send RIC_SUB_DEL_REQ"); -#ifdef __TRACING__ - lspan->Finish(); -#endif return -6; } break; } case RIC_CONTROL_REQ: { - if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap, &lspan) != 0) { + message.peerInfo->counters[OUT_INITI][MSG_COUNTER][ProcedureCode_id_RICcontrol]->Increment(); + message.peerInfo->counters[OUT_INITI][BYTES_COUNTER][ProcedureCode_id_RICcontrol]->Increment(rmrMessageBuffer.rcvMessage->len); + if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap) != 0) { mdclog_write(MDCLOG_ERR, "Failed to send RIC_CONTROL_REQ"); -#ifdef __TRACING__ - lspan->Finish(); -#endif return -6; } break; } case RIC_SERVICE_QUERY: { - if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap, &lspan) != 0) { + if (PER_FromXML(message, rmrMessageBuffer) != 0) { + break; + } + message.peerInfo->counters[OUT_INITI][MSG_COUNTER][ProcedureCode_id_RICserviceQuery]->Increment(); + message.peerInfo->counters[OUT_INITI][BYTES_COUNTER][ProcedureCode_id_RICserviceQuery]->Increment(rmrMessageBuffer.rcvMessage->len); + if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap) != 0) { mdclog_write(MDCLOG_ERR, "Failed to send RIC_SERVICE_QUERY"); -#ifdef __TRACING__ - lspan->Finish(); -#endif return -6; } break; } case RIC_SERVICE_UPDATE_ACK: { - if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap, &lspan) != 0) { + if (PER_FromXML(message, rmrMessageBuffer) != 0) { + break; + } + message.peerInfo->counters[OUT_SUCC][MSG_COUNTER][ProcedureCode_id_RICserviceUpdate]->Increment(); + message.peerInfo->counters[OUT_SUCC][BYTES_COUNTER][ProcedureCode_id_RICserviceQuery]->Increment(rmrMessageBuffer.rcvMessage->len); + if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap) != 0) { mdclog_write(MDCLOG_ERR, "Failed to send RIC_SERVICE_UPDATE_ACK"); -#ifdef __TRACING__ - lspan->Finish(); -#endif return -6; } break; } case RIC_SERVICE_UPDATE_FAILURE: { - if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap, &lspan) != 0) { + if (PER_FromXML(message, rmrMessageBuffer) != 0) { + break; + } + message.peerInfo->counters[OUT_UN_SUCC][MSG_COUNTER][ProcedureCode_id_RICserviceUpdate]->Increment(); + message.peerInfo->counters[OUT_UN_SUCC][BYTES_COUNTER][ProcedureCode_id_RICserviceQuery]->Increment(rmrMessageBuffer.rcvMessage->len); + if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap) != 0) { mdclog_write(MDCLOG_ERR, "Failed to send RIC_SERVICE_UPDATE_FAILURE"); -#ifdef __TRACING__ - lspan->Finish(); -#endif return -6; } break; } - case RIC_X2_RESET: { - if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap, &lspan) != 0) { - mdclog_write(MDCLOG_ERR, "Failed to send RIC_X2_RESET"); -#ifdef __TRACING__ - lspan->Finish(); -#endif + case RIC_E2_RESET_REQ: { + if (PER_FromXML(message, rmrMessageBuffer) != 0) { + break; + } + message.peerInfo->counters[OUT_INITI][MSG_COUNTER][ProcedureCode_id_Reset]->Increment(); + message.peerInfo->counters[OUT_INITI][BYTES_COUNTER][ProcedureCode_id_Reset]->Increment(rmrMessageBuffer.rcvMessage->len); + if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap) != 0) { + mdclog_write(MDCLOG_ERR, "Failed to send RIC_E2_RESET"); return -6; } break; } - case RIC_X2_RESET_RESP: { - if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap, &lspan) != 0) { - mdclog_write(MDCLOG_ERR, "Failed to send RIC_X2_RESET_RESP"); -#ifdef __TRACING__ - lspan->Finish(); -#endif + case RIC_E2_RESET_RESP: { + if (PER_FromXML(message, rmrMessageBuffer) != 0) { + break; + } + message.peerInfo->counters[OUT_SUCC][MSG_COUNTER][ProcedureCode_id_Reset]->Increment(); + message.peerInfo->counters[OUT_SUCC][BYTES_COUNTER][ProcedureCode_id_Reset]->Increment(rmrMessageBuffer.rcvMessage->len); + if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap) != 0) { + mdclog_write(MDCLOG_ERR, "Failed to send RIC_E2_RESET_RESP"); return -6; } break; @@ -2314,13 +2277,12 @@ int receiveXappMessages(int epoll_fd, message.message.asnLength = rmrMessageBuffer.sendMessage->len = snprintf((char *)rmrMessageBuffer.sendMessage->payload, - 256, - "%s|RIC_SCTP_CLEAR_ALL", - peerInfo->enodbName); + 256, + "%s|RIC_SCTP_CLEAR_ALL", + peerInfo->enodbName); message.message.asndata = rmrMessageBuffer.sendMessage->payload; mdclog_write(MDCLOG_INFO, "%s", message.message.asndata); - if (sendRequestToXapp(message, - RIC_SCTP_CONNECTION_FAILURE, rmrMessageBuffer, &lspan) != 0) { + if (sendRequestToXapp(message, RIC_SCTP_CONNECTION_FAILURE, rmrMessageBuffer) != 0) { mdclog_write(MDCLOG_ERR, "SCTP_CONNECTION_FAIL message failed to send to xAPP"); } free(peerInfo); @@ -2334,8 +2296,8 @@ int receiveXappMessages(int epoll_fd, case E2_TERM_KEEP_ALIVE_REQ: { // send message back rmr_bytes2payload(rmrMessageBuffer.sendMessage, - (unsigned char *)rmrMessageBuffer.ka_message, - rmrMessageBuffer.ka_message_len); + (unsigned char *)rmrMessageBuffer.ka_message, + rmrMessageBuffer.ka_message_len); rmrMessageBuffer.sendMessage->mtype = E2_TERM_KEEP_ALIVE_RESP; rmrMessageBuffer.sendMessage->state = 0; static unsigned char tx[32]; @@ -2347,13 +2309,37 @@ int receiveXappMessages(int epoll_fd, mdclog_write(MDCLOG_ERR, "Failed to send E2_TERM_KEEP_ALIVE_RESP RMR message returned NULL"); } else if (rmrMessageBuffer.sendMessage->state != 0) { mdclog_write(MDCLOG_ERR, "Failed to send E2_TERM_KEEP_ALIVE_RESP, on RMR state = %d ( %s)", - rmrMessageBuffer.sendMessage->state, translateRmrErrorMessages(rmrMessageBuffer.sendMessage->state).c_str()); - } else if (mdclog_level_get() >= MDCLOG_INFO) { - mdclog_write(MDCLOG_INFO, "Got Keep Alive Request send : %s", rmrMessageBuffer.ka_message); + rmrMessageBuffer.sendMessage->state, translateRmrErrorMessages(rmrMessageBuffer.sendMessage->state).c_str()); + } else if (mdclog_level_get() >= MDCLOG_DEBUG) { + mdclog_write(MDCLOG_DEBUG, "Got Keep Alive Request send : %s", rmrMessageBuffer.ka_message); + } + + break; + } + case RIC_HEALTH_CHECK_REQ: { + // send message back + rmr_bytes2payload(rmrMessageBuffer.sendMessage, + (unsigned char *)"OK", + 2); + rmrMessageBuffer.sendMessage->mtype = RIC_HEALTH_CHECK_RESP; + rmrMessageBuffer.sendMessage->state = 0; + static unsigned char tx[32]; + auto txLen = snprintf((char *) tx, sizeof tx, "%15ld", transactionCounter++); + rmr_bytes2xact(rmrMessageBuffer.sendMessage, tx, txLen); + rmrMessageBuffer.sendMessage = rmr_send_msg(rmrMessageBuffer.rmrCtx, rmrMessageBuffer.sendMessage); + if (rmrMessageBuffer.sendMessage == nullptr) { + rmrMessageBuffer.sendMessage = rmr_alloc_msg(rmrMessageBuffer.rmrCtx, RECEIVE_XAPP_BUFFER_SIZE); + mdclog_write(MDCLOG_ERR, "Failed to send RIC_HEALTH_CHECK_RESP RMR message returned NULL"); + } else if (rmrMessageBuffer.sendMessage->state != 0) { + mdclog_write(MDCLOG_ERR, "Failed to send RIC_HEALTH_CHECK_RESP, on RMR state = %d ( %s)", + rmrMessageBuffer.sendMessage->state, translateRmrErrorMessages(rmrMessageBuffer.sendMessage->state).c_str()); + } else if (mdclog_level_get() >= MDCLOG_DEBUG) { + mdclog_write(MDCLOG_DEBUG, "Got RIC_HEALTH_CHECK_REQ Request send : OK"); } break; } + default: mdclog_write(MDCLOG_WARN, "Message Type : %d is not seported", rmrMessageBuffer.rcvMessage->mtype); message.message.asndata = rmrMessageBuffer.rcvMessage->payload; @@ -2365,17 +2351,11 @@ int receiveXappMessages(int epoll_fd, buildJsonMessage(message); -#ifdef __TRACING__ - lspan->Finish(); -#endif return -7; } if (mdclog_level_get() >= MDCLOG_DEBUG) { mdclog_write(MDCLOG_DEBUG, "EXIT OK from %s", __FUNCTION__); } -#ifdef __TRACING__ - lspan->Finish(); -#endif return 0; } @@ -2385,31 +2365,19 @@ int receiveXappMessages(int epoll_fd, * @param message * @param failedMsgId * @param sctpMap - * @param pSpan * @return */ int sendDirectionalSctpMsg(RmrMessagesBuffer_t &messageBuffer, ReportingMessages_t &message, int failedMsgId, - Sctp_Map_t *sctpMap, - otSpan *pSpan) { -#ifdef __TRACING__ - auto lspan = opentracing::Tracer::Global()->StartSpan( - __FUNCTION__, { opentracing::ChildOf(&pSpan->get()->context()) }); -#else - otSpan lspan = 0; -#endif + Sctp_Map_t *sctpMap) { - getRequestMetaData(message, messageBuffer, &lspan); + getRequestMetaData(message, messageBuffer); if (mdclog_level_get() >= MDCLOG_INFO) { mdclog_write(MDCLOG_INFO, "send message to %s address", message.message.enodbName); } - auto rc = sendMessagetoCu(sctpMap, messageBuffer, message, failedMsgId, &lspan); -#ifdef __TRACING__ - lspan->Finish(); -#endif - + auto rc = sendMessagetoCu(sctpMap, messageBuffer, message, failedMsgId); return rc; } @@ -2419,41 +2387,15 @@ int sendDirectionalSctpMsg(RmrMessagesBuffer_t &messageBuffer, * @param messageBuffer * @param message * @param failedMesgId - * @param pSpan * @return */ int sendMessagetoCu(Sctp_Map_t *sctpMap, RmrMessagesBuffer_t &messageBuffer, ReportingMessages_t &message, - int failedMesgId, - otSpan *pSpan) { -#ifdef __TRACING__ - auto lspan = opentracing::Tracer::Global()->StartSpan( - __FUNCTION__, { opentracing::ChildOf(&pSpan->get()->context()) }); -#else - otSpan lspan = 0; -#endif - auto *peerInfo = (ConnectedCU_t *) sctpMap->find(message.message.enodbName); - if (peerInfo == nullptr) { - if (failedMesgId != 0) { - sendFailedSendingMessagetoXapp(messageBuffer, message, failedMesgId, &lspan); - } else { - mdclog_write(MDCLOG_ERR, "Failed to send message no CU entry %s", message.message.enodbName); - } -#ifdef __TRACING__ - lspan->Finish(); -#endif - - return -1; - } - + int failedMesgId) { // get the FD message.message.messageType = messageBuffer.rcvMessage->mtype; - auto rc = sendSctpMsg(peerInfo, message, sctpMap, &lspan); -#ifdef __TRACING__ - lspan->Finish(); -#endif - + auto rc = sendSctpMsg(message.peerInfo, message, sctpMap); return rc; } @@ -2465,14 +2407,7 @@ int sendMessagetoCu(Sctp_Map_t *sctpMap, * @param failedMesgId the return message type error */ void -sendFailedSendingMessagetoXapp(RmrMessagesBuffer_t &rmrMessageBuffer, ReportingMessages_t &message, int failedMesgId, - otSpan *pSpan) { -#ifdef __TRACING__ - auto lspan = opentracing::Tracer::Global()->StartSpan( - __FUNCTION__, { opentracing::ChildOf(&pSpan->get()->context()) }); -#else - otSpan lspan = 0; -#endif +sendFailedSendingMessagetoXapp(RmrMessagesBuffer_t &rmrMessageBuffer, ReportingMessages_t &message, int failedMesgId) { rmr_mbuf_t *msg = rmrMessageBuffer.sendMessage; msg->len = snprintf((char *) msg->payload, 200, "the gNb/eNode name %s not found", message.message.enodbName); @@ -2486,356 +2421,10 @@ sendFailedSendingMessagetoXapp(RmrMessagesBuffer_t &rmrMessageBuffer, ReportingM snprintf((char *) tx, sizeof tx, "%15ld", transactionCounter++); rmr_bytes2xact(msg, tx, strlen((const char *) tx)); - sendRmrMessage(rmrMessageBuffer, message, &lspan); -#ifdef __TRACING__ - lspan->Finish();pLogSink -#endif - + sendRmrMessage(rmrMessageBuffer, message); } -/** - * Send Response back to xApp, message is used only when there was a request from the xApp - * - * @param enodbName the name of the gNb/eNodeB - * @param msgType the value of the message to the xApp - * @param requestType The request that was sent by the xAPP - * @param rmrCtx the rmr identifier - * @param sctpMap hash map holds data on the requestrs - * @param buf the buffer to send to xAPP - * @param size size of the buffer to send - * @return - */ -/* -int sendResponseToXapp(ReportingMessages_t &message, - int msgType, - int requestType, - RmrMessagesBuffer_t &rmrMessageBuffer, - Sctp_Map_t *sctpMap, - otSpan *pSpan) { -#ifdef __TRACING__ - auto lspan = opentracing::Tracer::Global()->StartSpan( - __FUNCTION__, { opentracing::ChildOf(&pSpan->get()->context()) }); -#else - otSpan lspan = 0; -#endif - char key[MAX_ENODB_NAME_SIZE * 2]; - snprintf(key, MAX_ENODB_NAME_SIZE * 2, "msg:%s|%d", message.message.enodbName, requestType); - - auto xact = sctpMap->find(key); - if (xact == nullptr) { - mdclog_write(MDCLOG_ERR, "NO Request %s found for this response from CU: %s", key, - message.message.enodbName); -#ifdef __TRACING__ - lspan->Finish(); -#endif - - return -1; - } - sctpMap->erase(key); - - message.message.messageType = rmrMessageBuffer.sendMessage->mtype = msgType; //SETUP_RESPONSE_MESSAGE_TYPE; - rmr_bytes2payload(rmrMessageBuffer.sendMessage, (unsigned char *) message.message.asndata, - message.message.asnLength); - rmr_bytes2xact(rmrMessageBuffer.sendMessage, (const unsigned char *)xact, strlen((const char *)xact)); - rmr_str2meid(rmrMessageBuffer.sendMessage, (unsigned char *) message.message.enodbName); - rmrMessageBuffer.sendMessage->state = 0; - - if (mdclog_level_get() >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "remove key = %s from %s at line %d", key, __FUNCTION__, __LINE__); - } - free(xact); - - auto rc = sendRmrMessage(rmrMessageBuffer, message, &lspan); -#ifdef __TRACING__ - lspan->Finish(); -#endif - return rc; -} -*/ - -/** - * build the SCTP connection to eNodB or gNb - * @param rmrMessageBuffer - * @param message - * @param epoll_fd - * @param sctpMap - * @param pSpan - * @return - */ -int connectToCUandSetUp(RmrMessagesBuffer_t &rmrMessageBuffer, - ReportingMessages_t &message, - int epoll_fd, - Sctp_Map_t *sctpMap, - otSpan *pSpan) { -#ifdef __TRACING__ - auto lspan = opentracing::Tracer::Global()->StartSpan( - __FUNCTION__, { opentracing::ChildOf(&pSpan->get()->context()) }); -#else - otSpan lspan = 0; -#endif - struct sockaddr_in6 servaddr{}; - struct addrinfo hints{}, *result; - auto msgData = rmrMessageBuffer.rcvMessage->payload; - unsigned char meid[RMR_MAX_MEID]{}; - char host[256]{}; - uint16_t port = 0; - - message.message.messageType = rmrMessageBuffer.rcvMessage->mtype; - rmr_mbuf_t *msg = rmrMessageBuffer.rcvMessage; - rmr_get_meid(msg, meid); - - if (mdclog_level_get() >= MDCLOG_INFO) { - mdclog_write(MDCLOG_INFO, "message %d Received for MEID :%s. SETUP/EN-DC Setup Request from xApp, Message = %s", - msg->mtype, meid, msgData); - } - if (getSetupRequestMetaData(message, (char *)msgData, host, port, &lspan) < 0) { - if (mdclog_level_get() >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "Error in setup parameters %s, %d", __func__, __LINE__); - } -#ifdef __TRACING__ - lspan->Finish(); -#endif - return -1; - } - - //// message asndata points to the start of the asndata of the message and not to start of payload - // search if the same host:port but not the same enodbname - char searchBuff[256]{}; - snprintf(searchBuff, sizeof searchBuff, "host:%s:%d", host, port); - auto e = (char *)sctpMap->find(searchBuff); - if (e != nullptr) { - // found one compare if not the same - if (strcmp(message.message.enodbName, e) != 0) { - mdclog_write(MDCLOG_ERR, - "Try to connect CU %s to Host %s but %s already connected", - message.message.enodbName, host, e); -#ifdef __TRACING__ - lspan->Finish(); -#endif - return -1; - } - } - - // check if not connected. if connected send the request and return - auto *peerInfo = (ConnectedCU_t *)sctpMap->find(message.message.enodbName); - if (peerInfo != nullptr) { - if (mdclog_level_get() >= MDCLOG_INFO) { - mdclog_write(MDCLOG_INFO, "Device already connected to %s", - message.message.enodbName); - } - message.message.messageType = msg->mtype; - auto rc = sendSctpMsg(peerInfo, message, sctpMap, &lspan); - if (rc != 0) { - mdclog_write(MDCLOG_ERR, "failed write to SCTP %s, %d", __func__, __LINE__); -#ifdef __TRACING__ - lspan->Finish(); -#endif - return -1; - } - - char key[MAX_ENODB_NAME_SIZE * 2]; - snprintf(key, MAX_ENODB_NAME_SIZE * 2, "msg:%s|%d", message.message.enodbName, msg->mtype); - int xaction_len = strlen((const char *) msg->xaction); - auto *xaction = (unsigned char *) calloc(1, xaction_len); - memcpy(xaction, msg->xaction, xaction_len); - sctpMap->setkey(key, xaction); - if (mdclog_level_get() >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "set key = %s from %s at line %d", key, __FUNCTION__, __LINE__); - } -#ifdef __TRACING__ - lspan->Finish(); -#endif - return 0; - } - - peerInfo = (ConnectedCU_t *) calloc(1, sizeof(ConnectedCU_t)); - memcpy(peerInfo->enodbName, message.message.enodbName, sizeof(message.message.enodbName)); - - // new connection - if ((peerInfo->fileDescriptor = socket(AF_INET6, SOCK_STREAM, IPPROTO_SCTP)) < 0) { - mdclog_write(MDCLOG_ERR, "Socket Error, %s %s, %d", strerror(errno), __func__, __LINE__); -#ifdef __TRACING__ - lspan->Finish(); -#endif - return -1; - } - - auto optval = 1; - if (setsockopt(peerInfo->fileDescriptor, SOL_SOCKET, SO_REUSEPORT, &optval, sizeof optval) != 0) { - mdclog_write(MDCLOG_ERR, "setsockopt SO_REUSEPORT Error, %s %s, %d", strerror(errno), __func__, __LINE__); -#ifdef __TRACING__ - lspan->Finish(); -#endif - return -1; - } - optval = 1; - if (setsockopt(peerInfo->fileDescriptor, SOL_SOCKET, SO_REUSEADDR, &optval, sizeof optval) != 0) { - mdclog_write(MDCLOG_ERR, "setsockopt SO_REUSEADDR Error, %s %s, %d", strerror(errno), __func__, __LINE__); -#ifdef __TRACING__ - lspan->Finish(); -#endif - return -1; - } - servaddr.sin6_family = AF_INET6; - - struct sockaddr_in6 localAddr {}; - localAddr.sin6_family = AF_INET6; - localAddr.sin6_addr = in6addr_any; - localAddr.sin6_port = htons(SRC_PORT); - - if (bind(peerInfo->fileDescriptor, (struct sockaddr*)&localAddr , sizeof(struct sockaddr_in6)) < 0) { - mdclog_write(MDCLOG_ERR, "bind Socket Error, %s %s, %d", strerror(errno), __func__, __LINE__); -#ifdef __TRACING__ - lspan->Finish(); -#endif - return -1; - }//Ends the binding. - - memset(&hints, 0, sizeof hints); - hints.ai_flags = AI_NUMERICHOST; - if (getaddrinfo(host, nullptr, &hints, &result) < 0) { - close(peerInfo->fileDescriptor); - mdclog_write(MDCLOG_ERR, "getaddrinfo error for %s, Error = %s", host, strerror(errno)); -#ifdef __TRACING__ - lspan->Finish(); -#endif - return -1; - } - memcpy(&servaddr, result->ai_addr, sizeof(struct sockaddr_in6)); - freeaddrinfo(result); - - servaddr.sin6_port = htons(port); /* daytime server */ - if (mdclog_level_get() >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "Send Connect FD = %d host : %s port %d", - peerInfo->fileDescriptor, - host, - port); - } - - // Add to Epol - if (addToEpoll(epoll_fd, peerInfo, (EPOLLOUT | EPOLLIN | EPOLLET), sctpMap, message.message.enodbName, - msg->mtype, &lspan) != 0) { -#ifdef __TRACING__ - lspan->Finish(); -#endif - return -1; - } - - char hostBuff[NI_MAXHOST]; - char portBuff[NI_MAXHOST]; - - if (getnameinfo((SA *) &servaddr, sizeof(servaddr), - hostBuff, sizeof(hostBuff), - portBuff, sizeof(portBuff), - (uint) (NI_NUMERICHOST) | (uint) (NI_NUMERICSERV)) != 0) { - mdclog_write(MDCLOG_ERR, "getnameinfo() Error, %s %s %d", strerror(errno), __func__, __LINE__); -#ifdef __TRACING__ - lspan->Finish(); -#endif - return -1; - } - - if (setSocketNoBlocking(peerInfo->fileDescriptor) != 0) { - mdclog_write(MDCLOG_ERR, "setSocketNoBlocking failed to set new connection %s on sctpPort %s", hostBuff, - portBuff); - close(peerInfo->fileDescriptor); -#ifdef __TRACING__ - lspan->Finish(); -#endif - return -1; - } - - memcpy(peerInfo->hostName, hostBuff, strlen(hostBuff)); - peerInfo->hostName[strlen(hostBuff)] = 0; - memcpy(peerInfo->portNumber, portBuff, strlen(portBuff)); - peerInfo->portNumber[strlen(portBuff)] = 0; - - // map by enoodb/gnb name - sctpMap->setkey(message.message.enodbName, peerInfo); - //map host and port to enodeb - sctpMap->setkey(searchBuff, message.message.enodbName); - - // save message for the return values - char key[MAX_ENODB_NAME_SIZE * 2]; - snprintf(key, MAX_ENODB_NAME_SIZE * 2, "msg:%s|%d", message.message.enodbName, msg->mtype); - int xaction_len = strlen((const char *) msg->xaction); - auto *xaction = (unsigned char *) calloc(1, xaction_len); - memcpy(xaction, msg->xaction, xaction_len); - sctpMap->setkey(key, xaction); - if (mdclog_level_get() >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "End building peerinfo: %s for CU %s", key, message.message.enodbName); - } - - if (mdclog_level_get() >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "Send connect to FD %d, %s, %d", - peerInfo->fileDescriptor, __func__, __LINE__); - } - if (connect(peerInfo->fileDescriptor, (SA *) &servaddr, sizeof(servaddr)) < 0) { - if (errno != EINPROGRESS) { - mdclog_write(MDCLOG_ERR, "connect FD %d to host : %s port %d, %s", - peerInfo->fileDescriptor, host, port, strerror(errno)); - close(peerInfo->fileDescriptor); -#ifdef __TRACING__ - lspan->Finish(); -#endif - return -1; - } - if (mdclog_level_get() >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, - "Connect to FD %d returned with EINPROGRESS : %s", - peerInfo->fileDescriptor, strerror(errno)); - } - // since message.message.asndata is pointing to the asndata in the rmr message payload we copy it like this - memcpy(peerInfo->asnData, message.message.asndata, message.message.asnLength); - peerInfo->asnLength = message.message.asnLength; - peerInfo->mtype = msg->mtype; -#ifdef __TRACING__ - lspan->Finish(); -#endif - return 0; - } - - if (mdclog_level_get() >= MDCLOG_INFO) { - mdclog_write(MDCLOG_INFO, "Connect to FD %d returned OK without EINPROGRESS", peerInfo->fileDescriptor); - } - peerInfo->isConnected = true; - - if (modifyToEpoll(epoll_fd, peerInfo, (EPOLLIN | EPOLLET), sctpMap, message.message.enodbName, msg->mtype, - &lspan) != 0) { -#ifdef __TRACING__ - lspan->Finish(); -#endif - return -1; - } - - if (mdclog_level_get() >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "Connected to host : %s port %d", host, port); - } - - message.message.messageType = msg->mtype; - if (mdclog_level_get() >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "Send SCTP message to FD %d", peerInfo->fileDescriptor); - } - if (sendSctpMsg(peerInfo, message, sctpMap, &lspan) != 0) { - mdclog_write(MDCLOG_ERR, "Error write to SCTP %s %d", __func__, __LINE__); -#ifdef __TRACING__ - lspan->Finish(); -#endif - return -1; - } - memset(peerInfo->asnData, 0, message.message.asnLength); - peerInfo->asnLength = 0; - peerInfo->mtype = 0; - - if (mdclog_level_get() >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "Sent message to SCTP for %s", message.message.enodbName); - } -#ifdef __TRACING__ - lspan->Finish(); -#endif - return 0; -} /** * @@ -2845,7 +2434,6 @@ int connectToCUandSetUp(RmrMessagesBuffer_t &rmrMessageBuffer, * @param sctpMap * @param enodbName * @param msgType - * @param pSpan * @return */ int addToEpoll(int epoll_fd, @@ -2853,14 +2441,7 @@ int addToEpoll(int epoll_fd, uint32_t events, Sctp_Map_t *sctpMap, char *enodbName, - int msgType, - otSpan *pSpan) { -#ifdef __TRACING__ - auto lspan = opentracing::Tracer::Global()->StartSpan( - __FUNCTION__, { opentracing::ChildOf(&pSpan->get()->context()) }); -#else - otSpan lspan = 0; -#endif + int msgType) { // Add to Epol struct epoll_event event{}; event.data.ptr = peerInfo; @@ -2872,7 +2453,7 @@ int addToEpoll(int epoll_fd, } close(peerInfo->fileDescriptor); if (enodbName != nullptr) { - cleanHashEntry(peerInfo, sctpMap, &lspan); + cleanHashEntry(peerInfo, sctpMap); char key[MAX_ENODB_NAME_SIZE * 2]; snprintf(key, MAX_ENODB_NAME_SIZE * 2, "msg:%s|%d", enodbName, msgType); if (mdclog_level_get() >= MDCLOG_DEBUG) { @@ -2887,14 +2468,8 @@ int addToEpoll(int epoll_fd, peerInfo->enodbName[0] = 0; } mdclog_write(MDCLOG_ERR, "epoll_ctl EPOLL_CTL_ADD (may chack not to quit here)"); -#ifdef __TRACING__ - lspan->Finish(); -#endif return -1; } -#ifdef __TRACING__ - lspan->Finish(); -#endif return 0; } @@ -2906,7 +2481,6 @@ int addToEpoll(int epoll_fd, * @param sctpMap * @param enodbName * @param msgType - * @param pSpan * @return */ int modifyToEpoll(int epoll_fd, @@ -2914,14 +2488,7 @@ int modifyToEpoll(int epoll_fd, uint32_t events, Sctp_Map_t *sctpMap, char *enodbName, - int msgType, - otSpan *pSpan) { -#ifdef __TRACING__ - auto lspan = opentracing::Tracer::Global()->StartSpan( - __FUNCTION__, { opentracing::ChildOf(&pSpan->get()->context()) }); -#else - otSpan lspan = 0; -#endif + int msgType) { // Add to Epol struct epoll_event event{}; event.data.ptr = peerInfo; @@ -2932,54 +2499,25 @@ int modifyToEpoll(int epoll_fd, strerror(errno), __func__, __LINE__); } close(peerInfo->fileDescriptor); - cleanHashEntry(peerInfo, sctpMap, &lspan); + cleanHashEntry(peerInfo, sctpMap); char key[MAX_ENODB_NAME_SIZE * 2]; snprintf(key, MAX_ENODB_NAME_SIZE * 2, "msg:%s|%d", enodbName, msgType); if (mdclog_level_get() >= MDCLOG_DEBUG) { mdclog_write(MDCLOG_DEBUG, "remove key = %s from %s at line %d", key, __FUNCTION__, __LINE__); } - auto tmp = sctpMap->find(key); + auto tmp = sctpMap->find(key); if (tmp) { free(tmp); } sctpMap->erase(key); mdclog_write(MDCLOG_ERR, "epoll_ctl EPOLL_CTL_ADD (may chack not to quit here)"); -#ifdef __TRACING__ - lspan->Finish(); -#endif return -1; } -#ifdef __TRACING__ - lspan->Finish(); -#endif return 0; } -int sendRmrMessage(RmrMessagesBuffer_t &rmrMessageBuffer, ReportingMessages_t &message, otSpan *pSpan) { -#ifdef __TRACING__ - auto lspan = opentracing::Tracer::Global()->StartSpan( - __FUNCTION__, { opentracing::ChildOf(&pSpan->get()->context()) }); -#else -// otSpan lspan = 0; -#endif - //serialize the span -#ifdef __TRACING__ - std::unordered_map data; - RICCarrierWriter carrier(data); - opentracing::Tracer::Global()->Inject((lspan.get())->context(), carrier); - nlohmann::json j = data; - std::string str = j.dump(); - static auto maxTraceLength = 0; - - maxTraceLength = str.length() > maxTraceLength ? str.length() : maxTraceLength; - // serialized context can be put to RMR message using function: - if (mdclog_level_get() >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "max trace length is %d trace data length = %ld data = %s", maxTraceLength, - str.length(), str.c_str()); - } - rmr_set_trace(rmrMessageBuffer.sendMessage, (const unsigned char *) str.c_str(), str.length()); -#endif +int sendRmrMessage(RmrMessagesBuffer_t &rmrMessageBuffer, ReportingMessages_t &message) { buildJsonMessage(message); rmrMessageBuffer.sendMessage = rmr_send_msg(rmrMessageBuffer.rmrCtx, rmrMessageBuffer.sendMessage); @@ -2987,9 +2525,6 @@ int sendRmrMessage(RmrMessagesBuffer_t &rmrMessageBuffer, ReportingMessages_t &m if (rmrMessageBuffer.sendMessage == nullptr) { rmrMessageBuffer.sendMessage = rmr_alloc_msg(rmrMessageBuffer.rmrCtx, RECEIVE_XAPP_BUFFER_SIZE); mdclog_write(MDCLOG_ERR, "RMR failed send message returned with NULL pointer"); -#ifdef __TRACING__ - lspan->Finish(); -#endif return -1; } @@ -3005,9 +2540,6 @@ int sendRmrMessage(RmrMessagesBuffer_t &rmrMessageBuffer, ReportingMessages_t &m if (rmrMessageBuffer.sendMessage == nullptr) { mdclog_write(MDCLOG_ERR, "RMR failed send message returned with NULL pointer"); rmrMessageBuffer.sendMessage = rmr_alloc_msg(rmrMessageBuffer.rmrCtx, RECEIVE_XAPP_BUFFER_SIZE); -#ifdef __TRACING__ - lspan->Finish(); -#endif return -1; } else if (rmrMessageBuffer.sendMessage->state != 0) { mdclog_write(MDCLOG_ERR, @@ -3016,9 +2548,6 @@ int sendRmrMessage(RmrMessagesBuffer_t &rmrMessageBuffer, ReportingMessages_t &m rmrMessageBuffer.sendMessage->mtype, rmr_get_meid(rmrMessageBuffer.sendMessage, (unsigned char *)meid)); auto rc = rmrMessageBuffer.sendMessage->state; -#ifdef __TRACING__ - lspan->Finish(); -#endif return rc; } } else { @@ -3026,9 +2555,6 @@ int sendRmrMessage(RmrMessagesBuffer_t &rmrMessageBuffer, ReportingMessages_t &m translateRmrErrorMessages(rmrMessageBuffer.sendMessage->state).c_str(), rmrMessageBuffer.sendMessage->mtype, rmr_get_meid(rmrMessageBuffer.sendMessage, (unsigned char *)meid)); -#ifdef __TRACING__ - lspan->Finish(); -#endif return rmrMessageBuffer.sendMessage->state; } } @@ -3043,25 +2569,25 @@ void buildJsonMessage(ReportingMessages_t &message) { message.base64Data, message.outLen); if (mdclog_level_get() >= MDCLOG_DEBUG) { - mdclog_write(MDCLOG_DEBUG, "asn data length = %d, base64 message length = %d ", + mdclog_write(MDCLOG_DEBUG, "Tracing: ASN length = %d, base64 message length = %d ", (int) message.message.asnLength, (int) message.outLen); } snprintf(message.buffer, sizeof(message.buffer), - "{\"header\": {\"ts\": \"%ld.%09ld\"," - "\"ranName\": \"%s\"," - "\"messageType\": %d," - "\"direction\": \"%c\"}," - "\"base64Length\": %d," - "\"asnBase64\": \"%s\"}", - message.message.time.tv_sec, - message.message.time.tv_nsec, - message.message.enodbName, - message.message.messageType, - message.message.direction, - (int) message.outLen, - message.base64Data); + "{\"header\": {\"ts\": \"%ld.%09ld\"," + "\"ranName\": \"%s\"," + "\"messageType\": %d," + "\"direction\": \"%c\"}," + "\"base64Length\": %d," + "\"asnBase64\": \"%s\"}", + message.message.time.tv_sec, + message.message.time.tv_nsec, + message.message.enodbName, + message.message.messageType, + message.message.direction, + (int) message.outLen, + message.base64Data); static src::logger_mt &lg = my_logger::get(); BOOST_LOG(lg) << message.buffer;