1 // Copyright 2019 AT&T Intellectual Property
2 // Copyright 2019 Nokia
4 // Licensed under the Apache License, Version 2.0 (the "License");
5 // you may not use this file except in compliance with the License.
6 // You may obtain a copy of the License at
8 // http://www.apache.org/licenses/LICENSE-2.0
10 // Unless required by applicable law or agreed to in writing, software
11 // distributed under the License is distributed on an "AS IS" BASIS,
12 // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 // See the License for the specific language governing permissions and
14 // limitations under the License.
16 // This source code is part of the near-RT RIC (RAN Intelligent Controller)
17 // platform project (RICP).
19 // TODO: High-level file comment.
23 #include <3rdparty/oranE2/RANfunctions-List.h>
24 #include "sctpThread.h"
25 #include "BuildRunName.h"
27 //#include "3rdparty/oranE2SM/E2SM-gNB-NRT-RANfunction-Definition.h"
28 //#include "BuildXml.h"
29 //#include "pugixml/src/pugixml.hpp"
32 #include <sys/inotify.h>
37 //using namespace std::placeholders;
38 using namespace boost::filesystem;
39 using namespace prometheus;
47 // need to expose without the include of gcov
48 extern "C" void __gcov_flush(void);
49 #define LOG_FILE_CONFIG_MAP "CONFIG_MAP_NAME"
51 static void catch_function(int signal) {
57 BOOST_LOG_INLINE_GLOBAL_LOGGER_DEFAULT(my_logger, src::logger_mt)
59 boost::shared_ptr<sinks::synchronous_sink<sinks::text_file_backend>> boostLogger;
60 double cpuClock = 0.0;
61 bool jsonTrace = false;
64 static int enable_log_change_notify(const char* fileName)
68 if ( lstat(fileName,&fileInfo) == 0 )
70 ret = register_log_change_notify(fileName);
76 static int register_log_change_notify(const char *fileName)
78 pthread_attr_t cb_attr;
80 pthread_attr_init(&cb_attr);
81 pthread_attr_setdetachstate(&cb_attr,PTHREAD_CREATE_DETACHED);
82 return pthread_create(&tid, &cb_attr,&monitor_loglevel_change_handler,(void *)strdup(fileName));
86 static void * monitor_loglevel_change_handler(void* arg)
88 char *fileName = (char*) arg;
89 int ifd; // the inotify file des
90 int wfd; // the watched file des
92 char rbuf[4096]; // large read buffer as the event is var len
95 struct timeval timeout;
96 char* dname=NULL; // directory name
97 char* bname = NULL; // basename
101 dname = strdup( fileName); // defrock the file name into dir and basename
102 if( (tok = strrchr( dname, '/' )) != NULL ) {
104 bname = strdup( tok+1 );
108 ifd = inotify_init1( 0 ); // initialise watcher setting blocking read (no option)
110 fprintf( stderr, "### ERR ### unable to initialise file watch %s\n", strerror( errno ) );
112 wfd = inotify_add_watch( ifd, dname, IN_MOVED_TO | IN_CLOSE_WRITE ); // we only care about close write changes
115 fprintf( stderr, "### ERR ### unable to add watch on config file %s: %s\n", fileName, strerror( errno ) );
119 memset( &timeout, 0, sizeof(timeout) );
124 res = select (ifd + 1, &fds, NULL, NULL, &timeout);
127 n = read( ifd, rbuf, sizeof( rbuf ) ); // read the event
129 if( errno == EAGAIN ) {
131 fprintf( stderr, "### CRIT ### config listener read err: %s\n", strerror( errno ) );
136 //Retrieving Log Level from configmap by parsing configmap file
137 log_level = parse_file(fileName);
138 update_mdc_log_level_severity(log_level); //setting log level
142 inotify_rm_watch(ifd,wfd);
152 void update_mdc_log_level_severity(char* log_level)
154 mdclog_severity_t level = MDCLOG_ERR;
156 if(log_level == NULL)
158 printf("### ERR ### Invalid Log-Level Configuration in ConfigMap, Default Log-Level Applied: %d\n",level);
160 else if(strcasecmp(log_level,"1")==0)
164 else if(strcasecmp(log_level,"2")==0)
168 else if(strcasecmp(log_level,"3")==0)
172 else if(strcasecmp(log_level,"4")==0)
174 level = MDCLOG_DEBUG;
177 mdclog_level_set(level);
179 static char* parse_file(char* filename)
183 char *string_match = "log-level";
185 FILE *file = fopen ( filename, "r" );
189 while ( fgets ( line, sizeof line, file ) != NULL )
191 token = strtok(line, search);
192 if(strcmp(token,string_match)==0)
195 token = strtok(NULL, search);
196 token = strtok(token, "\n");//removing newline if any
203 return(strdup(token));
208 char *read_env_param(const char*envkey)
212 char *value = getenv(envkey);
214 return strdup(value);
219 void dynamic_log_level_change()
221 char *logFile_Name = read_env_param(LOG_FILE_CONFIG_MAP);
222 char* log_level_init=NULL;
225 log_level_init = parse_file(logFile_Name);
226 update_mdc_log_level_severity(log_level_init); //setting log level
227 free(log_level_init);
230 enable_log_change_notify(logFile_Name);
236 int log_change_monitor = 0;
238 mdclog_attr_init(&attr);
239 mdclog_attr_set_ident(attr, "E2Terminator");
241 if(mdclog_format_initialize(log_change_monitor)!=0)
242 mdclog_write(MDCLOG_ERR, "Failed to intialize MDC log format !!!");
243 dynamic_log_level_change();
244 mdclog_attr_destroy(attr);
246 auto start_time = std::chrono::high_resolution_clock::now();
247 typedef std::chrono::duration<double, std::ratio<1,1>> seconds_t;
250 return seconds_t(std::chrono::high_resolution_clock::now() - start_time).count();
253 double approx_CPU_MHz(unsigned sleepTime) {
254 using namespace std::chrono_literals;
256 uint64_t cycles_start = rdtscp(aux);
257 double time_start = age();
258 std::this_thread::sleep_for(sleepTime * 1ms);
259 uint64_t elapsed_cycles = rdtscp(aux) - cycles_start;
260 double elapsed_time = age() - time_start;
261 return elapsed_cycles / elapsed_time;
264 //std::atomic<int64_t> rmrCounter{0};
265 std::atomic<int64_t> num_of_messages{0};
266 std::atomic<int64_t> num_of_XAPP_messages{0};
267 static long transactionCounter = 0;
269 int buildListeningPort(sctp_params_t &sctpParams) {
270 sctpParams.listenFD = socket(AF_INET6, SOCK_STREAM, IPPROTO_SCTP);
271 if (sctpParams.listenFD <= 0) {
272 mdclog_write(MDCLOG_ERR, "Error Opening socket, %s", strerror(errno));
276 struct sockaddr_in6 serverAddress {};
277 serverAddress.sin6_family = AF_INET6;
278 serverAddress.sin6_addr = in6addr_any;
279 serverAddress.sin6_port = htons(sctpParams.sctpPort);
280 if (bind(sctpParams.listenFD, (SA *)&serverAddress, sizeof(serverAddress)) < 0 ) {
281 mdclog_write(MDCLOG_ERR, "Error binding port %d. %s", sctpParams.sctpPort, strerror(errno));
284 if (setSocketNoBlocking(sctpParams.listenFD) == -1) {
285 //mdclog_write(MDCLOG_ERR, "Error binding. %s", strerror(errno));
288 if (mdclog_level_get() >= MDCLOG_DEBUG) {
289 struct sockaddr_in6 clientAddress {};
290 socklen_t len = sizeof(clientAddress);
291 getsockname(sctpParams.listenFD, (SA *)&clientAddress, &len);
293 inet_ntop(AF_INET6, &clientAddress.sin6_addr, buff, sizeof(buff));
294 mdclog_write(MDCLOG_DEBUG, "My address: %s, port %d\n", buff, htons(clientAddress.sin6_port));
297 if (listen(sctpParams.listenFD, SOMAXCONN) < 0) {
298 mdclog_write(MDCLOG_ERR, "Error listening. %s\n", strerror(errno));
301 struct epoll_event event {};
302 event.events = EPOLLIN | EPOLLET;
303 event.data.fd = sctpParams.listenFD;
305 // add listening port to epoll
306 if (epoll_ctl(sctpParams.epoll_fd, EPOLL_CTL_ADD, sctpParams.listenFD, &event)) {
307 printf("Failed to add descriptor to epoll\n");
308 mdclog_write(MDCLOG_ERR, "Failed to add descriptor to epoll. %s\n", strerror(errno));
315 int buildConfiguration(sctp_params_t &sctpParams) {
316 path p = (sctpParams.configFilePath + "/" + sctpParams.configFileName).c_str();
318 const int size = 2048;
319 auto fileSize = file_size(p);
320 if (fileSize > size) {
321 mdclog_write(MDCLOG_ERR, "File %s larger than %d", p.string().c_str(), size);
325 mdclog_write(MDCLOG_ERR, "Configuration File %s not exists", p.string().c_str());
330 if (conf.openConfigFile(p.string()) == -1) {
331 mdclog_write(MDCLOG_ERR, "Filed to open config file %s, %s",
332 p.string().c_str(), strerror(errno));
335 int rmrPort = conf.getIntValue("nano");
337 mdclog_write(MDCLOG_ERR, "illegal RMR port ");
340 sctpParams.rmrPort = (uint16_t)rmrPort;
341 snprintf(sctpParams.rmrAddress, sizeof(sctpParams.rmrAddress), "%d", (int) (sctpParams.rmrPort));
342 auto tmpStr = conf.getStringValue("volume");
343 if (tmpStr.length() == 0) {
344 mdclog_write(MDCLOG_ERR, "illegal volume.");
348 char tmpLogFilespec[VOLUME_URL_SIZE];
349 tmpLogFilespec[0] = 0;
350 sctpParams.volume[0] = 0;
351 snprintf(sctpParams.volume, VOLUME_URL_SIZE, "%s", tmpStr.c_str());
352 // copy the name to temp file as well
353 snprintf(tmpLogFilespec, VOLUME_URL_SIZE, "%s", tmpStr.c_str());
356 // define the file name in the tmp directory under the volume
357 strcat(tmpLogFilespec,"/tmp/E2Term_%Y-%m-%d_%H-%M-%S.%N.tmpStr");
359 sctpParams.myIP = conf.getStringValue("local-ip");
360 if (sctpParams.myIP.length() == 0) {
361 mdclog_write(MDCLOG_ERR, "illegal local-ip.");
365 int sctpPort = conf.getIntValue("sctp-port");
366 if (sctpPort == -1) {
367 mdclog_write(MDCLOG_ERR, "illegal SCTP port ");
370 sctpParams.sctpPort = (uint16_t)sctpPort;
372 sctpParams.fqdn = conf.getStringValue("external-fqdn");
373 if (sctpParams.fqdn.length() == 0) {
374 mdclog_write(MDCLOG_ERR, "illegal external-fqdn");
378 std::string pod = conf.getStringValue("pod_name");
380 if (pod.length() == 0) {
381 mdclog_write(MDCLOG_ERR, "illegal pod_name in config file");
384 auto *podName = getenv(pod.c_str());
385 if (podName == nullptr) {
386 mdclog_write(MDCLOG_ERR, "illegal pod_name or environment variable not exists : %s", pod.c_str());
390 sctpParams.podName.assign(podName);
391 if (sctpParams.podName.length() == 0) {
392 mdclog_write(MDCLOG_ERR, "illegal pod_name");
397 tmpStr = conf.getStringValue("trace");
398 transform(tmpStr.begin(), tmpStr.end(), tmpStr.begin(), ::tolower);
399 if ((tmpStr.compare("start")) == 0) {
400 mdclog_write(MDCLOG_INFO, "Trace set to: start");
401 sctpParams.trace = true;
402 } else if ((tmpStr.compare("stop")) == 0) {
403 mdclog_write(MDCLOG_INFO, "Trace set to: stop");
404 sctpParams.trace = false;
406 mdclog_write(MDCLOG_ERR, "Trace was set to wrong value %s, set to stop", tmpStr.c_str());
407 sctpParams.trace = false;
409 jsonTrace = sctpParams.trace;
411 sctpParams.epollTimeOut = -1;
413 tmpStr = conf.getStringValue("prometheusPort");
414 if (tmpStr.length() != 0) {
415 sctpParams.prometheusPort = tmpStr;
418 sctpParams.ka_message_length = snprintf(sctpParams.ka_message, KA_MESSAGE_SIZE, "{\"address\": \"%s:%d\","
420 "\"pod_name\": \"%s\"}",
421 (const char *)sctpParams.myIP.c_str(),
423 sctpParams.fqdn.c_str(),
424 sctpParams.podName.c_str());
426 if (mdclog_level_get() >= MDCLOG_INFO) {
427 mdclog_write(MDCLOG_DEBUG,"RMR Port: %s", to_string(sctpParams.rmrPort).c_str());
428 mdclog_write(MDCLOG_DEBUG,"LogLevel: %s", to_string(sctpParams.logLevel).c_str());
429 mdclog_write(MDCLOG_DEBUG,"volume: %s", sctpParams.volume);
430 mdclog_write(MDCLOG_DEBUG,"tmpLogFilespec: %s", tmpLogFilespec);
431 mdclog_write(MDCLOG_DEBUG,"my ip: %s", sctpParams.myIP.c_str());
432 mdclog_write(MDCLOG_DEBUG,"pod name: %s", sctpParams.podName.c_str());
434 mdclog_write(MDCLOG_INFO, "running parameters for instance : %s", sctpParams.ka_message);
437 // Files written to the current working directory
438 boostLogger = logging::add_file_log(
439 keywords::file_name = tmpLogFilespec, // to temp directory
440 keywords::rotation_size = 10 * 1024 * 1024,
441 keywords::time_based_rotation = sinks::file::rotation_at_time_interval(posix_time::hours(1)),
442 keywords::format = "%Message%"
443 //keywords::format = "[%TimeStamp%]: %Message%" // use each tmpStr with time stamp
446 // Setup a destination folder for collecting rotated (closed) files --since the same volume can use rename()
447 boostLogger->locked_backend()->set_file_collector(sinks::file::make_collector(
448 keywords::target = sctpParams.volume
451 // Upon restart, scan the directory for files matching the file_name pattern
452 boostLogger->locked_backend()->scan_for_files();
454 // Enable auto-flushing after each tmpStr record written
455 if (mdclog_level_get() >= MDCLOG_DEBUG) {
456 boostLogger->locked_backend()->auto_flush(true);
462 void startPrometheus(sctp_params_t &sctpParams) {
463 sctpParams.prometheusFamily = &BuildCounter()
465 .Help("E2T message counter")
466 .Labels({{"POD_NAME", sctpParams.podName}})
467 .Register(*sctpParams.prometheusRegistry);
469 string prometheusPath = sctpParams.prometheusPort + "," + "[::]:" + sctpParams.prometheusPort;
470 if (mdclog_level_get() >= MDCLOG_DEBUG) {
471 mdclog_write(MDCLOG_DEBUG, "Start Prometheus Pull mode on %s", prometheusPath.c_str());
473 sctpParams.prometheusExposer = new Exposer(prometheusPath, 1);
474 sctpParams.prometheusExposer->RegisterCollectable(sctpParams.prometheusRegistry);
478 int main(const int argc, char **argv) {
479 sctp_params_t sctpParams;
481 std::random_device device{};
482 std::mt19937 generator(device());
483 std::uniform_int_distribution<long> distribution(1, (long) 1e12);
484 transactionCounter = distribution(generator);
488 // uint32_t aux1 = 0;
489 // st = rdtscp(aux1);
491 unsigned num_cpus = std::thread::hardware_concurrency();
493 if (std::signal(SIGINT, catch_function) == SIG_ERR) {
494 mdclog_write(MDCLOG_ERR, "Error initializing SIGINT");
497 if (std::signal(SIGABRT, catch_function)== SIG_ERR) {
498 mdclog_write(MDCLOG_ERR, "Error initializing SIGABRT");
501 if (std::signal(SIGTERM, catch_function)== SIG_ERR) {
502 mdclog_write(MDCLOG_ERR, "Error initializing SIGTERM");
506 cpuClock = approx_CPU_MHz(100);
508 mdclog_write(MDCLOG_DEBUG, "CPU speed %11.11f", cpuClock);
510 auto result = parse(argc, argv, sctpParams);
512 if (buildConfiguration(sctpParams) != 0) {
516 //auto registry = std::make_shared<Registry>();
517 sctpParams.prometheusRegistry = std::make_shared<Registry>();
519 //sctpParams.prometheusFamily = new Family<Counter>("E2T", "E2T message counter", {{"E", sctpParams.podName}});
521 startPrometheus(sctpParams);
524 sctpParams.epoll_fd = epoll_create1(0);
525 if (sctpParams.epoll_fd == -1) {
526 mdclog_write(MDCLOG_ERR, "failed to open epoll descriptor");
529 getRmrContext(sctpParams);
530 if (sctpParams.rmrCtx == nullptr) {
531 close(sctpParams.epoll_fd);
535 if (buildInotify(sctpParams) == -1) {
536 close(sctpParams.rmrListenFd);
537 rmr_close(sctpParams.rmrCtx);
538 close(sctpParams.epoll_fd);
542 if (buildListeningPort(sctpParams) != 0) {
543 close(sctpParams.rmrListenFd);
544 rmr_close(sctpParams.rmrCtx);
545 close(sctpParams.epoll_fd);
549 sctpParams.sctpMap = new mapWrapper();
551 std::vector<std::thread> threads(num_cpus);
552 // std::vector<std::thread> threads;
555 for (unsigned int i = 0; i < num_cpus; i++) {
556 threads[i] = std::thread(listener, &sctpParams);
561 int rc = pthread_setaffinity_np(threads[i].native_handle(), sizeof(cpu_set_t), &cpuset);
563 mdclog_write(MDCLOG_ERR, "Error calling pthread_setaffinity_np: %d", rc);
568 //loop over term_init until first message from xApp
569 handleTermInit(sctpParams);
571 for (auto &t : threads) {
578 void handleTermInit(sctp_params_t &sctpParams) {
579 sendTermInit(sctpParams);
580 //send to e2 manager init of e2 term
585 auto xappMessages = num_of_XAPP_messages.load(std::memory_order_acquire);
586 if (xappMessages > 0) {
587 if (mdclog_level_get() >= MDCLOG_INFO) {
588 mdclog_write(MDCLOG_INFO, "Got a message from some application, stop sending E2_TERM_INIT");
594 if (count % 1000 == 0) {
595 mdclog_write(MDCLOG_ERR, "GOT No messages from any xApp");
596 sendTermInit(sctpParams);
601 void sendTermInit(sctp_params_t &sctpParams) {
602 rmr_mbuf_t *msg = rmr_alloc_msg(sctpParams.rmrCtx, sctpParams.ka_message_length);
605 msg->mtype = E2_TERM_INIT;
607 rmr_bytes2payload(msg, (unsigned char *)sctpParams.ka_message, sctpParams.ka_message_length);
608 static unsigned char tx[32];
609 auto txLen = snprintf((char *) tx, sizeof tx, "%15ld", transactionCounter++);
610 rmr_bytes2xact(msg, tx, txLen);
611 msg = rmr_send_msg(sctpParams.rmrCtx, msg);
612 if (msg == nullptr) {
613 msg = rmr_alloc_msg(sctpParams.rmrCtx, sctpParams.ka_message_length);
614 } else if (msg->state == 0) {
616 if (mdclog_level_get() >= MDCLOG_INFO) {
617 mdclog_write(MDCLOG_INFO, "E2_TERM_INIT successfully sent ");
621 if (count % 100 == 0) {
622 mdclog_write(MDCLOG_ERR, "Error sending E2_TERM_INIT cause : %s ", translateRmrErrorMessages(msg->state).c_str());
637 cxxopts::ParseResult parse(int argc, char *argv[], sctp_params_t &sctpParams) {
638 cxxopts::Options options(argv[0], "e2 term help");
639 options.positional_help("[optional args]").show_positional_help();
640 options.allow_unrecognised_options().add_options()
641 ("p,path", "config file path", cxxopts::value<std::string>(sctpParams.configFilePath)->default_value("config"))
642 ("f,file", "config file name", cxxopts::value<std::string>(sctpParams.configFileName)->default_value("config.conf"))
643 ("h,help", "Print help");
645 auto result = options.parse(argc, (const char **&)argv);
647 if (result.count("help")) {
648 std::cout << options.help({""}) << std::endl;
657 * @return -1 failed 0 success
659 int buildInotify(sctp_params_t &sctpParams) {
660 sctpParams.inotifyFD = inotify_init1(IN_NONBLOCK);
661 if (sctpParams.inotifyFD == -1) {
662 mdclog_write(MDCLOG_ERR, "Failed to init inotify (inotify_init1) %s", strerror(errno));
666 sctpParams.inotifyWD = inotify_add_watch(sctpParams.inotifyFD,
667 (const char *)sctpParams.configFilePath.c_str(),
668 (unsigned)IN_OPEN | (unsigned)IN_CLOSE_WRITE | (unsigned)IN_CLOSE_NOWRITE); //IN_CLOSE = (IN_CLOSE_WRITE | IN_CLOSE_NOWRITE)
669 if (sctpParams.inotifyWD == -1) {
670 mdclog_write(MDCLOG_ERR, "Failed to add directory : %s to inotify (inotify_add_watch) %s",
671 sctpParams.configFilePath.c_str(),
673 close(sctpParams.inotifyFD);
677 struct epoll_event event{};
678 event.events = (EPOLLIN);
679 event.data.fd = sctpParams.inotifyFD;
680 // add listening RMR FD to epoll
681 if (epoll_ctl(sctpParams.epoll_fd, EPOLL_CTL_ADD, sctpParams.inotifyFD, &event)) {
682 mdclog_write(MDCLOG_ERR, "Failed to add inotify FD to epoll");
683 close(sctpParams.inotifyFD);
694 void listener(sctp_params_t *params) {
695 int num_of_SCTP_messages = 0;
696 auto totalTime = 0.0;
697 std::thread::id this_id = std::this_thread::get_id();
699 auto pod_name = std::getenv("POD_NAME");
700 auto container_name = std::getenv("CONTAINER_NAME");
701 auto service_name = std::getenv("SERVICE_NAME");
702 auto host_name = std::getenv("HOST_NAME");
703 auto system_name = std::getenv("SYSTEM_NAME");
704 auto pid = std::to_string(getpid()).c_str();
705 streambuf *oldCout = cout.rdbuf();
706 ostringstream memCout;
708 cout.rdbuf(memCout.rdbuf());
710 //return to the normal cout
714 memcpy(tid, memCout.str().c_str(), memCout.str().length() < 32 ? memCout.str().length() : 31);
715 tid[memCout.str().length()] = 0;
716 mdclog_mdc_add("SYSTEM_NAME", system_name);
717 mdclog_mdc_add("HOST_NAME", host_name);
718 mdclog_mdc_add("SERVICE_NAME", service_name);
719 mdclog_mdc_add("CONTAINER_NAME", container_name);
720 mdclog_mdc_add("POD_NAME", pod_name);
721 mdclog_mdc_add("PID", pid);
723 if (mdclog_level_get() >= MDCLOG_DEBUG) {
724 mdclog_write(MDCLOG_DEBUG, "started thread number %s", tid);
727 RmrMessagesBuffer_t rmrMessageBuffer{};
728 //create and init RMR
729 rmrMessageBuffer.rmrCtx = params->rmrCtx;
731 auto *events = (struct epoll_event *) calloc(MAXEVENTS, sizeof(struct epoll_event));
732 struct timespec end{0, 0};
733 struct timespec start{0, 0};
735 rmrMessageBuffer.rcvMessage = rmr_alloc_msg(rmrMessageBuffer.rmrCtx, RECEIVE_XAPP_BUFFER_SIZE);
736 rmrMessageBuffer.sendMessage = rmr_alloc_msg(rmrMessageBuffer.rmrCtx, RECEIVE_XAPP_BUFFER_SIZE);
738 memcpy(rmrMessageBuffer.ka_message, params->ka_message, params->ka_message_length);
739 rmrMessageBuffer.ka_message_len = params->ka_message_length;
740 rmrMessageBuffer.ka_message[rmrMessageBuffer.ka_message_len] = 0;
742 if (mdclog_level_get() >= MDCLOG_DEBUG) {
743 mdclog_write(MDCLOG_DEBUG, "keep alive message is : %s", rmrMessageBuffer.ka_message);
746 ReportingMessages_t message {};
748 // for (int i = 0; i < MAX_RMR_BUFF_ARRAY; i++) {
749 // rmrMessageBuffer.rcvBufferedMessages[i] = rmr_alloc_msg(rmrMessageBuffer.rmrCtx, RECEIVE_XAPP_BUFFER_SIZE);
750 // rmrMessageBuffer.sendBufferedMessages[i] = rmr_alloc_msg(rmrMessageBuffer.rmrCtx, RECEIVE_XAPP_BUFFER_SIZE);
754 if (mdclog_level_get() >= MDCLOG_DEBUG) {
755 mdclog_write(MDCLOG_DEBUG, "Start EPOLL Wait. Timeout = %d", params->epollTimeOut);
758 auto numOfEvents = epoll_wait(params->epoll_fd, events, MAXEVENTS, params->epollTimeOut);
760 auto numOfEvents = 1;
762 if (numOfEvents == 0) { // time out
763 if (mdclog_level_get() >= MDCLOG_DEBUG) {
764 mdclog_write(MDCLOG_DEBUG, "got epoll timeout");
767 } else if (numOfEvents < 0) {
768 if (errno == EINTR) {
769 if (mdclog_level_get() >= MDCLOG_DEBUG) {
770 mdclog_write(MDCLOG_DEBUG, "got EINTR : %s", strerror(errno));
774 mdclog_write(MDCLOG_ERR, "Epoll wait failed, errno = %s", strerror(errno));
777 for (auto i = 0; i < numOfEvents; i++) {
778 if (mdclog_level_get() >= MDCLOG_DEBUG) {
779 mdclog_write(MDCLOG_DEBUG, "handling epoll event %d out of %d", i + 1, numOfEvents);
781 clock_gettime(CLOCK_MONOTONIC, &message.message.time);
782 start.tv_sec = message.message.time.tv_sec;
783 start.tv_nsec = message.message.time.tv_nsec;
786 if ((events[i].events & EPOLLERR) || (events[i].events & EPOLLHUP)) {
787 handlepoll_error(events[i], message, rmrMessageBuffer, params);
788 } else if (events[i].events & EPOLLOUT) {
789 handleEinprogressMessages(events[i], message, rmrMessageBuffer, params);
790 } else if (params->listenFD == events[i].data.fd) {
791 if (mdclog_level_get() >= MDCLOG_INFO) {
792 mdclog_write(MDCLOG_INFO, "New connection request from sctp network\n");
794 // new connection is requested from RAN start build connection
796 struct sockaddr in_addr {};
798 char hostBuff[NI_MAXHOST];
799 char portBuff[NI_MAXSERV];
801 in_len = sizeof(in_addr);
802 auto *peerInfo = (ConnectedCU_t *)calloc(1, sizeof(ConnectedCU_t));
803 if(peerInfo == nullptr){
804 mdclog_write(MDCLOG_ERR, "calloc failed");
807 peerInfo->sctpParams = params;
808 peerInfo->fileDescriptor = accept(params->listenFD, &in_addr, &in_len);
809 if (peerInfo->fileDescriptor == -1) {
810 if ((errno == EAGAIN) || (errno == EWOULDBLOCK)) {
811 /* We have processed all incoming connections. */
814 mdclog_write(MDCLOG_ERR, "Accept error, errno = %s", strerror(errno));
818 if (setSocketNoBlocking(peerInfo->fileDescriptor) == -1) {
819 mdclog_write(MDCLOG_ERR, "setSocketNoBlocking failed to set new connection %s on port %s\n", hostBuff, portBuff);
820 close(peerInfo->fileDescriptor);
823 auto ans = getnameinfo(&in_addr, in_len,
824 peerInfo->hostName, NI_MAXHOST,
825 peerInfo->portNumber, NI_MAXSERV, (unsigned )((unsigned int)NI_NUMERICHOST | (unsigned int)NI_NUMERICSERV));
827 mdclog_write(MDCLOG_ERR, "Failed to get info on connection request. %s\n", strerror(errno));
828 close(peerInfo->fileDescriptor);
831 if (mdclog_level_get() >= MDCLOG_DEBUG) {
832 mdclog_write(MDCLOG_DEBUG, "Accepted connection on descriptor %d (host=%s, port=%s)\n", peerInfo->fileDescriptor, peerInfo->hostName, peerInfo->portNumber);
834 peerInfo->isConnected = false;
835 peerInfo->gotSetup = false;
836 if (addToEpoll(params->epoll_fd,
839 params->sctpMap, nullptr,
845 } else if (params->rmrListenFd == events[i].data.fd) {
846 // got message from XAPP
847 //num_of_XAPP_messages.fetch_add(1, std::memory_order_release);
848 num_of_messages.fetch_add(1, std::memory_order_release);
849 if (mdclog_level_get() >= MDCLOG_DEBUG) {
850 mdclog_write(MDCLOG_DEBUG, "new RMR message");
852 if (receiveXappMessages(params->sctpMap,
854 message.message.time) != 0) {
855 mdclog_write(MDCLOG_ERR, "Error handling Xapp message");
857 } else if (params->inotifyFD == events[i].data.fd) {
858 mdclog_write(MDCLOG_INFO, "Got event from inotify (configuration update)");
859 handleConfigChange(params);
861 /* We RMR_ERR_RETRY have data on the fd waiting to be read. Read and display it.
862 * We must read whatever data is available completely, as we are running
863 * in edge-triggered mode and won't get a notification again for the same data. */
864 num_of_messages.fetch_add(1, std::memory_order_release);
865 if (mdclog_level_get() >= MDCLOG_DEBUG) {
866 mdclog_write(MDCLOG_DEBUG, "new message from SCTP, epoll flags are : %0x", events[i].events);
868 receiveDataFromSctp(&events[i],
870 num_of_SCTP_messages,
872 message.message.time);
875 clock_gettime(CLOCK_MONOTONIC, &end);
876 if (mdclog_level_get() >= MDCLOG_INFO) {
877 totalTime += ((end.tv_sec + 1.0e-9 * end.tv_nsec) -
878 ((double) start.tv_sec + 1.0e-9 * start.tv_nsec));
880 if (mdclog_level_get() >= MDCLOG_DEBUG) {
881 mdclog_write(MDCLOG_DEBUG, "message handling is %ld seconds %ld nanoseconds",
882 end.tv_sec - start.tv_sec,
883 end.tv_nsec - start.tv_nsec);
896 void handleConfigChange(sctp_params_t *sctpParams) {
897 char buf[4096] __attribute__ ((aligned(__alignof__(struct inotify_event))));
898 const struct inotify_event *event;
901 struct inotify_event tmpEvent;
903 path p = (sctpParams->configFilePath + "/" + sctpParams->configFileName).c_str();
904 auto endlessLoop = true;
905 while (endlessLoop) {
907 auto len = read(sctpParams->inotifyFD, buf, sizeof buf);
912 if (errno != EAGAIN) {
913 mdclog_write(MDCLOG_ERR, "read %s ", strerror(errno));
923 for (ptr = buf; ptr < buf + len; ptr += sizeof(struct inotify_event) + event->len) {
925 event = (const struct inotify_event *)ptr;
927 tmpEvent.mask = (uint32_t)IN_CLOSE_WRITE;
930 if (event->mask & (uint32_t)IN_ISDIR) {
934 // the directory name
935 if (sctpParams->inotifyWD == event->wd) {
939 auto retVal = strcmp(sctpParams->configFileName.c_str(), event->name);
944 // only the file we want
945 if (event->mask & (uint32_t)IN_CLOSE_WRITE) {
946 if (mdclog_level_get() >= MDCLOG_INFO) {
947 mdclog_write(MDCLOG_INFO, "Configuration file changed");
950 const int size = 2048;
951 auto fileSize = file_size(p);
952 if (fileSize > size) {
953 mdclog_write(MDCLOG_ERR, "File %s larger than %d", p.string().c_str(), size);
957 mdclog_write(MDCLOG_ERR, "Configuration File %s not exists", p.string().c_str());
962 if (conf.openConfigFile(p.string()) == -1) {
963 mdclog_write(MDCLOG_ERR, "Filed to open config file %s, %s",
964 p.string().c_str(), strerror(errno));
967 auto tmpStr = conf.getStringValue("loglevel");
968 if (tmpStr.length() == 0) {
969 mdclog_write(MDCLOG_ERR, "illegal loglevel. Set loglevel to MDCLOG_INFO");
972 transform(tmpStr.begin(), tmpStr.end(), tmpStr.begin(), ::tolower);
974 if ((tmpStr.compare("debug")) == 0) {
975 mdclog_write(MDCLOG_INFO, "Log level set to MDCLOG_DEBUG");
976 sctpParams->logLevel = MDCLOG_DEBUG;
977 } else if ((tmpStr.compare("info")) == 0) {
978 mdclog_write(MDCLOG_INFO, "Log level set to MDCLOG_INFO");
979 sctpParams->logLevel = MDCLOG_INFO;
980 } else if ((tmpStr.compare("warning")) == 0) {
981 mdclog_write(MDCLOG_INFO, "Log level set to MDCLOG_WARN");
982 sctpParams->logLevel = MDCLOG_WARN;
983 } else if ((tmpStr.compare("error")) == 0) {
984 mdclog_write(MDCLOG_INFO, "Log level set to MDCLOG_ERR");
985 sctpParams->logLevel = MDCLOG_ERR;
987 mdclog_write(MDCLOG_ERR, "illegal loglevel = %s. Set loglevel to MDCLOG_INFO", tmpStr.c_str());
988 sctpParams->logLevel = MDCLOG_INFO;
990 mdclog_level_set(sctpParams->logLevel);
991 tmpStr = conf.getStringValue("trace");
992 if (tmpStr.length() == 0) {
993 mdclog_write(MDCLOG_ERR, "illegal trace. Set trace to stop");
997 transform(tmpStr.begin(), tmpStr.end(), tmpStr.begin(), ::tolower);
998 if ((tmpStr.compare("start")) == 0) {
999 mdclog_write(MDCLOG_INFO, "Trace set to: start");
1000 sctpParams->trace = true;
1001 } else if ((tmpStr.compare("stop")) == 0) {
1002 mdclog_write(MDCLOG_INFO, "Trace set to: stop");
1003 sctpParams->trace = false;
1005 mdclog_write(MDCLOG_ERR, "Trace was set to wrong value %s, set to stop", tmpStr.c_str());
1006 sctpParams->trace = false;
1008 jsonTrace = sctpParams->trace;
1011 endlessLoop = false;
1024 * @param rmrMessageBuffer
1027 void handleEinprogressMessages(struct epoll_event &event,
1028 ReportingMessages_t &message,
1029 RmrMessagesBuffer_t &rmrMessageBuffer,
1030 sctp_params_t *params) {
1031 auto *peerInfo = (ConnectedCU_t *)event.data.ptr;
1032 memcpy(message.message.enodbName, peerInfo->enodbName, sizeof(peerInfo->enodbName));
1034 mdclog_write(MDCLOG_INFO, "file descriptor %d got EPOLLOUT", peerInfo->fileDescriptor);
1036 socklen_t retValLen = 0;
1037 auto rc = getsockopt(peerInfo->fileDescriptor, SOL_SOCKET, SO_ERROR, &retVal, &retValLen);
1038 if (rc != 0 || retVal != 0) {
1041 rmrMessageBuffer.sendMessage->len = snprintf((char *)rmrMessageBuffer.sendMessage->payload, 256,
1042 "%s|Failed SCTP Connection, after EINPROGRESS the getsockopt%s",
1043 peerInfo->enodbName, strerror(errno));
1044 } else if (retVal != 0) {
1045 rmrMessageBuffer.sendMessage->len = snprintf((char *)rmrMessageBuffer.sendMessage->payload, 256,
1046 "%s|Failed SCTP Connection after EINPROGRESS, SO_ERROR",
1047 peerInfo->enodbName);
1050 message.message.asndata = rmrMessageBuffer.sendMessage->payload;
1051 message.message.asnLength = rmrMessageBuffer.sendMessage->len;
1052 mdclog_write(MDCLOG_ERR, "%s", rmrMessageBuffer.sendMessage->payload);
1053 message.message.direction = 'N';
1054 if (sendRequestToXapp(message, RIC_SCTP_CONNECTION_FAILURE, rmrMessageBuffer) != 0) {
1055 mdclog_write(MDCLOG_ERR, "SCTP_CONNECTION_FAIL message failed to send to xAPP");
1058 memset(peerInfo->asnData, 0, peerInfo->asnLength);
1059 peerInfo->asnLength = 0;
1060 peerInfo->mtype = 0;
1064 peerInfo->isConnected = true;
1066 if (modifyToEpoll(params->epoll_fd, peerInfo, (EPOLLIN | EPOLLET), params->sctpMap, peerInfo->enodbName,
1067 peerInfo->mtype) != 0) {
1068 mdclog_write(MDCLOG_ERR, "epoll_ctl EPOLL_CTL_MOD");
1072 message.message.asndata = (unsigned char *)peerInfo->asnData;
1073 message.message.asnLength = peerInfo->asnLength;
1074 message.message.messageType = peerInfo->mtype;
1075 memcpy(message.message.enodbName, peerInfo->enodbName, sizeof(peerInfo->enodbName));
1076 num_of_messages.fetch_add(1, std::memory_order_release);
1077 if (mdclog_level_get() >= MDCLOG_DEBUG) {
1078 mdclog_write(MDCLOG_DEBUG, "send the delayed SETUP/ENDC SETUP to sctp for %s",
1079 message.message.enodbName);
1081 if (sendSctpMsg(peerInfo, message, params->sctpMap) != 0) {
1082 if (mdclog_level_get() >= MDCLOG_DEBUG) {
1083 mdclog_write(MDCLOG_DEBUG, "Error write to SCTP %s %d", __func__, __LINE__);
1088 memset(peerInfo->asnData, 0, peerInfo->asnLength);
1089 peerInfo->asnLength = 0;
1090 peerInfo->mtype = 0;
1094 void handlepoll_error(struct epoll_event &event,
1095 ReportingMessages_t &message,
1096 RmrMessagesBuffer_t &rmrMessageBuffer,
1097 sctp_params_t *params) {
1098 if (event.data.fd != params->rmrListenFd) {
1099 auto *peerInfo = (ConnectedCU_t *)event.data.ptr;
1100 mdclog_write(MDCLOG_ERR, "epoll error, events %0x on fd %d, RAN NAME : %s",
1101 event.events, peerInfo->fileDescriptor, peerInfo->enodbName);
1104 rmrMessageBuffer.sendMessage->len = snprintf((char *)rmrMessageBuffer.sendMessage->payload, 256,
1105 "%s|Failed SCTP Connection",
1106 peerInfo->enodbName);
1107 message.message.asndata = rmrMessageBuffer.sendMessage->payload;
1108 message.message.asnLength = rmrMessageBuffer.sendMessage->len;
1110 memcpy(message.message.enodbName, peerInfo->enodbName, sizeof(peerInfo->enodbName));
1111 message.message.direction = 'N';
1112 if (sendRequestToXapp(message, RIC_SCTP_CONNECTION_FAILURE, rmrMessageBuffer) != 0) {
1113 mdclog_write(MDCLOG_ERR, "SCTP_CONNECTION_FAIL message failed to send to xAPP");
1116 close(peerInfo->fileDescriptor);
1117 params->sctpMap->erase(peerInfo->enodbName);
1118 cleanHashEntry((ConnectedCU_t *) event.data.ptr, params->sctpMap);
1120 mdclog_write(MDCLOG_ERR, "epoll error, events %0x on RMR FD", event.events);
1128 int setSocketNoBlocking(int socket) {
1129 auto flags = fcntl(socket, F_GETFL, 0);
1132 mdclog_write(MDCLOG_ERR, "%s, %s", __FUNCTION__, strerror(errno));
1136 flags = (unsigned) flags | (unsigned) O_NONBLOCK;
1137 if (fcntl(socket, F_SETFL, flags) == -1) {
1138 mdclog_write(MDCLOG_ERR, "%s, %s", __FUNCTION__, strerror(errno));
1150 void cleanHashEntry(ConnectedCU_t *val, Sctp_Map_t *m) {
1152 auto port = (uint16_t) strtol(val->portNumber, &dummy, 10);
1153 char searchBuff[2048]{};
1155 snprintf(searchBuff, sizeof searchBuff, "host:%s:%d", val->hostName, port);
1156 m->erase(searchBuff);
1158 m->erase(val->enodbName);
1166 * @param fd file descriptor
1167 * @param data the asn data to send
1168 * @param len length of the data
1169 * @param enodbName the enodbName as in the map for printing purpose
1170 * @param m map host information
1171 * @param mtype message number
1172 * @return 0 success, a negative number on fail
1174 int sendSctpMsg(ConnectedCU_t *peerInfo, ReportingMessages_t &message, Sctp_Map_t *m) {
1175 auto loglevel = mdclog_level_get();
1177 int fd = peerInfo->fileDescriptor;
1179 int fd = FILE_DESCRIPTOR;
1181 if (loglevel >= MDCLOG_DEBUG) {
1182 mdclog_write(MDCLOG_DEBUG, "Send SCTP message for CU %s, %s",
1183 message.message.enodbName, __FUNCTION__);
1187 if (send(fd,message.message.asndata, message.message.asnLength,MSG_NOSIGNAL) < 0) {
1188 if (errno == EINTR) {
1191 mdclog_write(MDCLOG_ERR, "error writing to CU a message, %s ", strerror(errno));
1193 if (!peerInfo->isConnected) {
1194 mdclog_write(MDCLOG_ERR, "connection to CU %s is still in progress.", message.message.enodbName);
1197 cleanHashEntry(peerInfo, m);
1200 char key[MAX_ENODB_NAME_SIZE * 2];
1201 snprintf(key, MAX_ENODB_NAME_SIZE * 2, "msg:%s|%d", message.message.enodbName,
1202 message.message.messageType);
1203 if (loglevel >= MDCLOG_DEBUG) {
1204 mdclog_write(MDCLOG_DEBUG, "remove key = %s from %s at line %d", key, __FUNCTION__, __LINE__);
1206 auto tmp = m->find(key);
1215 message.message.direction = 'D';
1216 // send report.buffer of size
1217 buildJsonMessage(message);
1219 if (loglevel >= MDCLOG_DEBUG) {
1220 mdclog_write(MDCLOG_DEBUG,
1221 "SCTP message for CU %s sent from %s",
1222 message.message.enodbName,
1232 * @param rmrMessageBuffer
1234 void getRequestMetaData(ReportingMessages_t &message, RmrMessagesBuffer_t &rmrMessageBuffer) {
1235 message.message.asndata = rmrMessageBuffer.rcvMessage->payload;
1236 message.message.asnLength = rmrMessageBuffer.rcvMessage->len;
1238 if (mdclog_level_get() >= MDCLOG_DEBUG) {
1239 mdclog_write(MDCLOG_DEBUG, "Message from Xapp RAN name = %s message length = %ld",
1240 message.message.enodbName, (unsigned long) message.message.asnLength);
1250 * @param numOfMessages
1251 * @param rmrMessageBuffer
1255 int receiveDataFromSctp(struct epoll_event *events,
1256 Sctp_Map_t *sctpMap,
1258 RmrMessagesBuffer_t &rmrMessageBuffer,
1259 struct timespec &ts) {
1260 /* We have data on the fd waiting to be read. Read and display it.
1261 * We must read whatever data is available completely, as we are running
1262 * in edge-triggered mode and won't get a notification again for the same data. */
1263 ReportingMessages_t message {};
1265 auto loglevel = mdclog_level_get();
1267 // get the identity of the interface
1268 message.peerInfo = (ConnectedCU_t *)events->data.ptr;
1270 struct timespec start{0, 0};
1271 struct timespec decodeStart{0, 0};
1272 struct timespec end{0, 0};
1274 E2AP_PDU_t *pdu = nullptr;
1277 if (loglevel >= MDCLOG_DEBUG) {
1278 mdclog_write(MDCLOG_DEBUG, "Start Read from SCTP %d fd", message.peerInfo->fileDescriptor);
1279 clock_gettime(CLOCK_MONOTONIC, &start);
1281 // read the buffer directly to rmr payload
1282 message.message.asndata = rmrMessageBuffer.sendMessage->payload;
1284 message.message.asnLength = rmrMessageBuffer.sendMessage->len =
1285 read(message.peerInfo->fileDescriptor, rmrMessageBuffer.sendMessage->payload, RECEIVE_SCTP_BUFFER_SIZE);
1287 message.message.asnLength = rmrMessageBuffer.sendMessage->len;
1290 if (loglevel >= MDCLOG_DEBUG) {
1291 mdclog_write(MDCLOG_DEBUG, "Finish Read from SCTP %d fd message length = %ld",
1292 message.peerInfo->fileDescriptor, message.message.asnLength);
1295 memcpy(message.message.enodbName, message.peerInfo->enodbName, sizeof(message.peerInfo->enodbName));
1296 message.message.direction = 'U';
1297 message.message.time.tv_nsec = ts.tv_nsec;
1298 message.message.time.tv_sec = ts.tv_sec;
1300 if (message.message.asnLength < 0) {
1301 if (errno == EINTR) {
1304 /* If errno == EAGAIN, that means we have read all
1305 data. So goReportingMessages_t back to the main loop. */
1306 if (errno != EAGAIN) {
1307 mdclog_write(MDCLOG_ERR, "Read error, %s ", strerror(errno));
1309 } else if (loglevel >= MDCLOG_DEBUG) {
1310 mdclog_write(MDCLOG_DEBUG, "EAGAIN - descriptor = %d", message.peerInfo->fileDescriptor);
1313 } else if (message.message.asnLength == 0) {
1314 /* End of file. The remote has closed the connection. */
1315 if (loglevel >= MDCLOG_INFO) {
1316 mdclog_write(MDCLOG_INFO, "END of File Closed connection - descriptor = %d",
1317 message.peerInfo->fileDescriptor);
1323 if (loglevel >= MDCLOG_DEBUG) {
1324 char printBuffer[RECEIVE_SCTP_BUFFER_SIZE]{};
1325 char *tmp = printBuffer;
1326 for (size_t i = 0; i < (size_t)message.message.asnLength; ++i) {
1327 snprintf(tmp, 3, "%02x", message.message.asndata[i]);
1330 printBuffer[message.message.asnLength] = 0;
1331 clock_gettime(CLOCK_MONOTONIC, &end);
1332 mdclog_write(MDCLOG_DEBUG, "Before Encoding E2AP PDU for : %s, Read time is : %ld seconds, %ld nanoseconds",
1333 message.peerInfo->enodbName, end.tv_sec - start.tv_sec, end.tv_nsec - start.tv_nsec);
1334 mdclog_write(MDCLOG_DEBUG, "PDU buffer length = %ld, data = : %s", message.message.asnLength,
1336 clock_gettime(CLOCK_MONOTONIC, &decodeStart);
1339 auto rval = asn_decode(nullptr, ATS_ALIGNED_BASIC_PER, &asn_DEF_E2AP_PDU, (void **) &pdu,
1340 message.message.asndata, message.message.asnLength);
1342 asn_dec_rval_t rval = {RC_OK, 0};
1343 pdu = (E2AP_PDU_t*)rmrMessageBuffer.sendMessage->tp_buf;
1345 if (rval.code != RC_OK) {
1346 mdclog_write(MDCLOG_ERR, "Error %d Decoding (unpack) E2AP PDU from RAN : %s", rval.code,
1347 message.peerInfo->enodbName);
1351 if (loglevel >= MDCLOG_DEBUG) {
1352 clock_gettime(CLOCK_MONOTONIC, &end);
1353 mdclog_write(MDCLOG_DEBUG, "After Encoding E2AP PDU for : %s, Read time is : %ld seconds, %ld nanoseconds",
1354 message.peerInfo->enodbName, end.tv_sec - decodeStart.tv_sec, end.tv_nsec - decodeStart.tv_nsec);
1357 FILE *stream = open_memstream(&printBuffer, &size);
1358 asn_fprint(stream, &asn_DEF_E2AP_PDU, pdu);
1359 mdclog_write(MDCLOG_DEBUG, "Encoding E2AP PDU past : %s", printBuffer);
1360 clock_gettime(CLOCK_MONOTONIC, &decodeStart);
1366 switch (pdu->present) {
1367 case E2AP_PDU_PR_initiatingMessage: {//initiating message
1368 asnInitiatingRequest(pdu, sctpMap,message, rmrMessageBuffer);
1371 case E2AP_PDU_PR_successfulOutcome: { //successful outcome
1372 asnSuccessfulMsg(pdu, sctpMap, message, rmrMessageBuffer);
1375 case E2AP_PDU_PR_unsuccessfulOutcome: { //Unsuccessful Outcome
1376 asnUnSuccsesfulMsg(pdu, sctpMap, message, rmrMessageBuffer);
1380 mdclog_write(MDCLOG_ERR, "Unknown index %d in E2AP PDU", pdu->present);
1383 if (loglevel >= MDCLOG_DEBUG) {
1384 clock_gettime(CLOCK_MONOTONIC, &end);
1385 mdclog_write(MDCLOG_DEBUG,
1386 "After processing message and sent to rmr for : %s, Read time is : %ld seconds, %ld nanoseconds",
1387 message.peerInfo->enodbName, end.tv_sec - decodeStart.tv_sec, end.tv_nsec - decodeStart.tv_nsec);
1390 if (pdu != nullptr) {
1391 ASN_STRUCT_RESET(asn_DEF_E2AP_PDU, pdu);
1392 //ASN_STRUCT_FREE(asn_DEF_E2AP_PDU, pdu);
1402 if (loglevel >= MDCLOG_INFO) {
1403 mdclog_write(MDCLOG_INFO, "Closed connection - descriptor = %d", message.peerInfo->fileDescriptor);
1405 message.message.asnLength = rmrMessageBuffer.sendMessage->len =
1406 snprintf((char *)rmrMessageBuffer.sendMessage->payload,
1408 "%s|CU disconnected unexpectedly",
1409 message.peerInfo->enodbName);
1410 message.message.asndata = rmrMessageBuffer.sendMessage->payload;
1412 if (sendRequestToXapp(message,
1413 RIC_SCTP_CONNECTION_FAILURE,
1414 rmrMessageBuffer) != 0) {
1415 mdclog_write(MDCLOG_ERR, "SCTP_CONNECTION_FAIL message failed to send to xAPP");
1419 /* Closing descriptor make epoll remove it from the set of descriptors which are monitored. */
1420 close(message.peerInfo->fileDescriptor);
1421 cleanHashEntry((ConnectedCU_t *) events->data.ptr, sctpMap);
1423 if (loglevel >= MDCLOG_DEBUG) {
1424 clock_gettime(CLOCK_MONOTONIC, &end);
1425 mdclog_write(MDCLOG_DEBUG, "from receive SCTP to send RMR time is %ld seconds and %ld nanoseconds",
1426 end.tv_sec - start.tv_sec, end.tv_nsec - start.tv_nsec);
1432 static void buildAndSendSetupRequest(ReportingMessages_t &message,
1433 RmrMessagesBuffer_t &rmrMessageBuffer,
1435 string const &messageName,
1436 string const &ieName,
1437 vector<string> &functionsToAdd_v,
1438 vector<string> &functionsToModified_v*/) {
1439 auto logLevel = mdclog_level_get();
1440 // now we can send the data to e2Mgr
1443 auto buffer_size = RECEIVE_SCTP_BUFFER_SIZE * 2;
1444 unsigned char *buffer = nullptr;
1445 buffer = (unsigned char *) calloc(buffer_size, sizeof(unsigned char));
1448 mdclog_write(MDCLOG_ERR, "Allocating buffer for %s failed, %s", asn_DEF_E2AP_PDU.name, strerror(errno));
1452 er = asn_encode_to_buffer(nullptr, ATS_BASIC_XER, &asn_DEF_E2AP_PDU, pdu, buffer, buffer_size);
1453 if (er.encoded == -1) {
1454 mdclog_write(MDCLOG_ERR, "encoding of %s failed, %s", asn_DEF_E2AP_PDU.name, strerror(errno));
1456 } else if (er.encoded > (ssize_t) buffer_size) {
1457 buffer_size = er.encoded + 128;
1458 mdclog_write(MDCLOG_WARN, "Buffer of size %d is to small for %s. Reallocate buffer of size %d",
1460 asn_DEF_E2AP_PDU.name, buffer_size);
1461 buffer_size = er.encoded + 128;
1463 unsigned char *newBuffer = nullptr;
1464 newBuffer = (unsigned char *) realloc(buffer, buffer_size);
1468 mdclog_write(MDCLOG_ERR, "Reallocating buffer for %s failed, %s", asn_DEF_E2AP_PDU.name, strerror(errno));
1475 buffer[er.encoded] = '\0';
1480 string res((char *)buffer);
1481 res.erase(std::remove(res.begin(), res.end(), '\n'), res.end());
1482 res.erase(std::remove(res.begin(), res.end(), '\t'), res.end());
1483 res.erase(std::remove(res.begin(), res.end(), ' '), res.end());
1486 // if (!functionsToAdd_v.empty() || !functionsToModified_v.empty()) {
1487 // res = buildXmlData(messageName, ieName, functionsToAdd_v, functionsToModified_v, buffer, (size_t) er.encoded);
1490 // if (res.length() == 0) {
1491 // rmrMsg = rmr_alloc_msg(rmrMessageBuffer.rmrCtx, buffer_size + 256);
1492 // rmrMsg->len = snprintf((char *) rmrMsg->payload, RECEIVE_SCTP_BUFFER_SIZE * 2, "%s:%d|%s",
1493 // message.peerInfo->sctpParams->myIP.c_str(),
1494 // message.peerInfo->sctpParams->rmrPort,
1497 rmrMsg = rmr_alloc_msg(rmrMessageBuffer.rmrCtx, (int)res.length() + 256);
1498 rmrMsg->len = snprintf((char *) rmrMsg->payload, res.length() + 256, "%s:%d|%s",
1499 message.peerInfo->sctpParams->myIP.c_str(),
1500 message.peerInfo->sctpParams->rmrPort,
1504 if (logLevel >= MDCLOG_DEBUG) {
1505 mdclog_write(MDCLOG_DEBUG, "Setup request of size %d :\n %s\n", rmrMsg->len, rmrMsg->payload);
1508 rmrMsg->mtype = message.message.messageType;
1510 rmr_bytes2meid(rmrMsg, (unsigned char *) message.message.enodbName, strlen(message.message.enodbName));
1512 static unsigned char tx[32];
1513 snprintf((char *) tx, sizeof tx, "%15ld", transactionCounter++);
1514 rmr_bytes2xact(rmrMsg, tx, strlen((const char *) tx));
1516 rmrMsg = rmr_send_msg(rmrMessageBuffer.rmrCtx, rmrMsg);
1518 if (rmrMsg == nullptr) {
1519 mdclog_write(MDCLOG_ERR, "RMR failed to send returned nullptr");
1520 } else if (rmrMsg->state != 0) {
1521 char meid[RMR_MAX_MEID]{};
1522 if (rmrMsg->state == RMR_ERR_RETRY) {
1525 mdclog_write(MDCLOG_INFO, "RETRY sending Message %d to Xapp from %s",
1526 rmrMsg->mtype, rmr_get_meid(rmrMsg, (unsigned char *) meid));
1528 rmrMsg = rmr_send_msg(rmrMessageBuffer.rmrCtx, rmrMsg);
1530 if (rmrMsg == nullptr) {
1531 mdclog_write(MDCLOG_ERR, "RMR failed send returned nullptr");
1532 } else if (rmrMsg->state != 0) {
1533 mdclog_write(MDCLOG_ERR,
1534 "RMR Retry failed %s sending request %d to Xapp from %s",
1535 translateRmrErrorMessages(rmrMsg->state).c_str(),
1537 rmr_get_meid(rmrMsg, (unsigned char *) meid));
1540 mdclog_write(MDCLOG_ERR, "RMR failed: %s. sending request %d to Xapp from %s",
1541 translateRmrErrorMessages(rmrMsg->state).c_str(),
1543 rmr_get_meid(rmrMsg, (unsigned char *) meid));
1546 message.peerInfo->gotSetup = true;
1547 buildJsonMessage(message);
1549 if (rmrMsg != nullptr) {
1550 rmr_free_msg(rmrMsg);
1558 int RAN_Function_list_To_Vector(RANfunctions_List_t& list, vector <string> &runFunXML_v) {
1560 runFunXML_v.clear();
1561 for (auto j = 0; j < list.list.count; j++) {
1562 auto *raNfunctionItemIEs = (RANfunction_ItemIEs_t *)list.list.array[j];
1563 if (raNfunctionItemIEs->id == ProtocolIE_ID_id_RANfunction_Item &&
1564 (raNfunctionItemIEs->value.present == RANfunction_ItemIEs__value_PR_RANfunction_Item)) {
1566 E2SM_gNB_NRT_RANfunction_Definition_t *ranFunDef = nullptr;
1567 auto rval = asn_decode(nullptr, ATS_ALIGNED_BASIC_PER,
1568 &asn_DEF_E2SM_gNB_NRT_RANfunction_Definition,
1569 (void **)&ranFunDef,
1570 raNfunctionItemIEs->value.choice.RANfunction_Item.ranFunctionDefinition.buf,
1571 raNfunctionItemIEs->value.choice.RANfunction_Item.ranFunctionDefinition.size);
1572 if (rval.code != RC_OK) {
1573 mdclog_write(MDCLOG_ERR, "Error %d Decoding (unpack) E2SM message from : %s",
1575 asn_DEF_E2SM_gNB_NRT_RANfunction_Definition.name);
1579 auto xml_buffer_size = RECEIVE_SCTP_BUFFER_SIZE * 2;
1580 unsigned char xml_buffer[RECEIVE_SCTP_BUFFER_SIZE * 2];
1581 memset(xml_buffer, 0, RECEIVE_SCTP_BUFFER_SIZE * 2);
1583 auto er = asn_encode_to_buffer(nullptr,
1585 &asn_DEF_E2SM_gNB_NRT_RANfunction_Definition,
1589 if (er.encoded == -1) {
1590 mdclog_write(MDCLOG_ERR, "encoding of %s failed, %s",
1591 asn_DEF_E2SM_gNB_NRT_RANfunction_Definition.name,
1593 } else if (er.encoded > (ssize_t)xml_buffer_size) {
1594 mdclog_write(MDCLOG_ERR, "Buffer of size %d is to small for %s, at %s line %d",
1595 (int) xml_buffer_size,
1596 asn_DEF_E2SM_gNB_NRT_RANfunction_Definition.name, __func__, __LINE__);
1598 if (mdclog_level_get() >= MDCLOG_DEBUG) {
1599 mdclog_write(MDCLOG_DEBUG, "Encoding E2SM %s PDU number %d : %s",
1600 asn_DEF_E2SM_gNB_NRT_RANfunction_Definition.name,
1605 string runFuncs = (char *)(xml_buffer);
1606 runFunXML_v.emplace_back(runFuncs);
1613 int collectServiceUpdate_RequestData(E2AP_PDU_t *pdu,
1614 Sctp_Map_t *sctpMap,
1615 ReportingMessages_t &message,
1616 vector <string> &RANfunctionsAdded_v,
1617 vector <string> &RANfunctionsModified_v) {
1618 memset(message.peerInfo->enodbName, 0 , MAX_ENODB_NAME_SIZE);
1619 for (auto i = 0; i < pdu->choice.initiatingMessage->value.choice.RICserviceUpdate.protocolIEs.list.count; i++) {
1620 auto *ie = pdu->choice.initiatingMessage->value.choice.RICserviceUpdate.protocolIEs.list.array[i];
1621 if (ie->id == ProtocolIE_ID_id_RANfunctionsAdded) {
1622 if (ie->value.present == RICserviceUpdate_IEs__value_PR_RANfunctionsID_List) {
1623 if (mdclog_level_get() >= MDCLOG_DEBUG) {
1624 mdclog_write(MDCLOG_DEBUG, "Run function list have %d entries",
1625 ie->value.choice.RANfunctions_List.list.count);
1627 if (RAN_Function_list_To_Vector(ie->value.choice.RANfunctions_List, RANfunctionsAdded_v) != 0 ) {
1631 } else if (ie->id == ProtocolIE_ID_id_RANfunctionsModified) {
1632 if (ie->value.present == RICserviceUpdate_IEs__value_PR_RANfunctions_List) {
1633 if (mdclog_level_get() >= MDCLOG_DEBUG) {
1634 mdclog_write(MDCLOG_DEBUG, "Run function list have %d entries",
1635 ie->value.choice.RANfunctions_List.list.count);
1637 if (RAN_Function_list_To_Vector(ie->value.choice.RANfunctions_List, RANfunctionsModified_v) != 0 ) {
1643 if (mdclog_level_get() >= MDCLOG_DEBUG) {
1644 mdclog_write(MDCLOG_DEBUG, "Run function vector have %ld entries",
1645 RANfunctionsAdded_v.size());
1653 void buildPrometheusList(ConnectedCU_t *peerInfo, Family<Counter> *prometheusFamily) {
1654 peerInfo->counters[IN_INITI][MSG_COUNTER][(ProcedureCode_id_E2setup)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"SetupRequest", "Messages"}});
1655 peerInfo->counters[IN_INITI][BYTES_COUNTER][(ProcedureCode_id_E2setup)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"SetupRequest", "Bytes"}});
1657 peerInfo->counters[IN_INITI][MSG_COUNTER][(ProcedureCode_id_ErrorIndication)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"ErrorIndication", "Messages"}});
1658 peerInfo->counters[IN_INITI][BYTES_COUNTER][(ProcedureCode_id_ErrorIndication)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"ErrorIndication", "Bytes"}});
1660 peerInfo->counters[IN_INITI][MSG_COUNTER][(ProcedureCode_id_RICindication)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICindication", "Messages"}});
1661 peerInfo->counters[IN_INITI][BYTES_COUNTER][(ProcedureCode_id_RICindication)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICindication", "Bytes"}});
1663 peerInfo->counters[IN_INITI][MSG_COUNTER][(ProcedureCode_id_Reset)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"ResetRequest", "Messages"}});
1664 peerInfo->counters[IN_INITI][BYTES_COUNTER][(ProcedureCode_id_Reset)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"ResetRequest", "Bytes"}});
1666 peerInfo->counters[IN_INITI][MSG_COUNTER][(ProcedureCode_id_RICserviceUpdate)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICserviceUpdate", "Messages"}});
1667 peerInfo->counters[IN_INITI][BYTES_COUNTER][(ProcedureCode_id_RICserviceUpdate)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICserviceUpdate", "Bytes"}});
1668 // ---------------------------------------------
1669 peerInfo->counters[IN_SUCC][MSG_COUNTER][(ProcedureCode_id_Reset)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"ResetACK", "Messages"}});
1670 peerInfo->counters[IN_SUCC][BYTES_COUNTER][(ProcedureCode_id_Reset)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"ResetACK", "Bytes"}});
1672 peerInfo->counters[IN_SUCC][MSG_COUNTER][(ProcedureCode_id_RICcontrol)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICcontrolACK", "Messages"}});
1673 peerInfo->counters[IN_SUCC][BYTES_COUNTER][(ProcedureCode_id_RICcontrol)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICcontrolACK", "Bytes"}});
1675 peerInfo->counters[IN_SUCC][MSG_COUNTER][(ProcedureCode_id_RICsubscription)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICsubscriptionACK", "Messages"}});
1676 peerInfo->counters[IN_SUCC][BYTES_COUNTER][(ProcedureCode_id_RICsubscription)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICsubscriptionACK", "Bytes"}});
1678 peerInfo->counters[IN_SUCC][MSG_COUNTER][(ProcedureCode_id_RICsubscriptionDelete)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICsubscriptionDeleteACK", "Messages"}});
1679 peerInfo->counters[IN_SUCC][BYTES_COUNTER][(ProcedureCode_id_RICsubscriptionDelete)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICsubscriptionDeleteACK", "Bytes"}});
1680 //-------------------------------------------------------------
1682 peerInfo->counters[IN_UN_SUCC][MSG_COUNTER][(ProcedureCode_id_RICcontrol)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICcontrolFailure", "Messages"}});
1683 peerInfo->counters[IN_UN_SUCC][BYTES_COUNTER][(ProcedureCode_id_RICcontrol)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICcontrolFailure", "Bytes"}});
1685 peerInfo->counters[IN_UN_SUCC][MSG_COUNTER][(ProcedureCode_id_RICsubscription)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICsubscriptionFailure", "Messages"}});
1686 peerInfo->counters[IN_UN_SUCC][BYTES_COUNTER][(ProcedureCode_id_RICsubscription)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICsubscriptionFailure", "Bytes"}});
1688 peerInfo->counters[IN_UN_SUCC][MSG_COUNTER][(ProcedureCode_id_RICsubscriptionDelete)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICsubscriptionDeleteFailure", "Messages"}});
1689 peerInfo->counters[IN_UN_SUCC][BYTES_COUNTER][(ProcedureCode_id_RICsubscriptionDelete)] = &prometheusFamily->Add({{peerInfo->enodbName, "IN"}, {"RICsubscriptionDeleteFailure", "Bytes"}});
1691 //====================================================================================
1692 peerInfo->counters[OUT_INITI][MSG_COUNTER][(ProcedureCode_id_ErrorIndication)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"ErrorIndication", "Messages"}});
1693 peerInfo->counters[OUT_INITI][BYTES_COUNTER][(ProcedureCode_id_ErrorIndication)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"ErrorIndication", "Bytes"}});
1695 peerInfo->counters[OUT_INITI][MSG_COUNTER][(ProcedureCode_id_Reset)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"ResetRequest", "Messages"}});
1696 peerInfo->counters[OUT_INITI][BYTES_COUNTER][(ProcedureCode_id_Reset)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"ResetRequest", "Bytes"}});
1698 peerInfo->counters[OUT_INITI][MSG_COUNTER][(ProcedureCode_id_RICcontrol)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"RICcontrol", "Messages"}});
1699 peerInfo->counters[OUT_INITI][BYTES_COUNTER][(ProcedureCode_id_RICcontrol)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"RICcontrol", "Bytes"}});
1701 peerInfo->counters[OUT_INITI][MSG_COUNTER][(ProcedureCode_id_RICserviceQuery)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"RICserviceQuery", "Messages"}});
1702 peerInfo->counters[OUT_INITI][BYTES_COUNTER][(ProcedureCode_id_RICserviceQuery)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"RICserviceQuery", "Bytes"}});
1704 peerInfo->counters[OUT_INITI][MSG_COUNTER][(ProcedureCode_id_RICsubscription)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"RICsubscription", "Messages"}});
1705 peerInfo->counters[OUT_INITI][BYTES_COUNTER][(ProcedureCode_id_RICsubscription)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"RICsubscription", "Bytes"}});
1707 peerInfo->counters[OUT_INITI][MSG_COUNTER][(ProcedureCode_id_RICsubscriptionDelete)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"RICsubscriptionDelete", "Messages"}});
1708 peerInfo->counters[OUT_INITI][BYTES_COUNTER][(ProcedureCode_id_RICsubscriptionDelete)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"RICsubscriptionDelete", "Bytes"}});
1709 //---------------------------------------------------------------------------------------------------------
1710 peerInfo->counters[OUT_SUCC][MSG_COUNTER][(ProcedureCode_id_E2setup)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"SetupResponse", "Messages"}});
1711 peerInfo->counters[OUT_SUCC][BYTES_COUNTER][(ProcedureCode_id_E2setup)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"SetupResponse", "Bytes"}});
1713 peerInfo->counters[OUT_SUCC][MSG_COUNTER][(ProcedureCode_id_Reset)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"ResetACK", "Messages"}});
1714 peerInfo->counters[OUT_SUCC][BYTES_COUNTER][(ProcedureCode_id_Reset)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"ResetACK", "Bytes"}});
1716 peerInfo->counters[OUT_SUCC][MSG_COUNTER][(ProcedureCode_id_RICserviceUpdate)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"RICserviceUpdateResponse", "Messages"}});
1717 peerInfo->counters[OUT_SUCC][BYTES_COUNTER][(ProcedureCode_id_RICserviceUpdate)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"RICserviceUpdateResponse", "Bytes"}});
1718 //----------------------------------------------------------------------------------------------------------------
1719 peerInfo->counters[OUT_UN_SUCC][MSG_COUNTER][(ProcedureCode_id_E2setup)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"SetupRequestFailure", "Messages"}});
1720 peerInfo->counters[OUT_UN_SUCC][BYTES_COUNTER][(ProcedureCode_id_E2setup)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"SetupRequestFailure", "Bytes"}});
1722 peerInfo->counters[OUT_UN_SUCC][MSG_COUNTER][(ProcedureCode_id_RICserviceUpdate)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"RICserviceUpdateFailure", "Messages"}});
1723 peerInfo->counters[OUT_UN_SUCC][BYTES_COUNTER][(ProcedureCode_id_RICserviceUpdate)] = &prometheusFamily->Add({{peerInfo->enodbName, "OUT"}, {"RICserviceUpdateFailure", "Bytes"}});
1730 * @param RANfunctionsAdded_v
1733 int collectSetupRequestData(E2AP_PDU_t *pdu,
1734 Sctp_Map_t *sctpMap,
1735 ReportingMessages_t &message /*, vector <string> &RANfunctionsAdded_v*/) {
1736 memset(message.peerInfo->enodbName, 0 , MAX_ENODB_NAME_SIZE);
1737 for (auto i = 0; i < pdu->choice.initiatingMessage->value.choice.E2setupRequest.protocolIEs.list.count; i++) {
1738 auto *ie = pdu->choice.initiatingMessage->value.choice.E2setupRequest.protocolIEs.list.array[i];
1739 if (ie->id == ProtocolIE_ID_id_GlobalE2node_ID) {
1740 // get the ran name for meid
1741 if (ie->value.present == E2setupRequestIEs__value_PR_GlobalE2node_ID) {
1742 if (buildRanName(message.peerInfo->enodbName, ie) < 0) {
1743 mdclog_write(MDCLOG_ERR, "Bad param in E2setupRequestIEs GlobalE2node_ID.\n");
1744 // no message will be sent
1748 memcpy(message.message.enodbName, message.peerInfo->enodbName, strlen(message.peerInfo->enodbName));
1749 sctpMap->setkey(message.message.enodbName, message.peerInfo);
1751 } /*else if (ie->id == ProtocolIE_ID_id_RANfunctionsAdded) {
1752 if (ie->value.present == E2setupRequestIEs__value_PR_RANfunctions_List) {
1753 if (mdclog_level_get() >= MDCLOG_DEBUG) {
1754 mdclog_write(MDCLOG_DEBUG, "Run function list have %d entries",
1755 ie->value.choice.RANfunctions_List.list.count);
1757 if (RAN_Function_list_To_Vector(ie->value.choice.RANfunctions_List, RANfunctionsAdded_v) != 0 ) {
1763 // if (mdclog_level_get() >= MDCLOG_DEBUG) {
1764 // mdclog_write(MDCLOG_DEBUG, "Run function vector have %ld entries",
1765 // RANfunctionsAdded_v.size());
1770 int XML_From_PER(ReportingMessages_t &message, RmrMessagesBuffer_t &rmrMessageBuffer) {
1771 E2AP_PDU_t *pdu = nullptr;
1773 if (mdclog_level_get() >= MDCLOG_DEBUG) {
1774 mdclog_write(MDCLOG_DEBUG, "got PER message of size %d is:%s",
1775 rmrMessageBuffer.sendMessage->len, rmrMessageBuffer.sendMessage->payload);
1777 auto rval = asn_decode(nullptr, ATS_ALIGNED_BASIC_PER, &asn_DEF_E2AP_PDU, (void **) &pdu,
1778 rmrMessageBuffer.sendMessage->payload, rmrMessageBuffer.sendMessage->len);
1779 if (rval.code != RC_OK) {
1780 mdclog_write(MDCLOG_ERR, "Error %d Decoding (unpack) setup response from E2MGR : %s",
1782 message.message.enodbName);
1786 int buff_size = RECEIVE_XAPP_BUFFER_SIZE;
1787 auto er = asn_encode_to_buffer(nullptr, ATS_BASIC_XER, &asn_DEF_E2AP_PDU, pdu,
1788 rmrMessageBuffer.sendMessage->payload, buff_size);
1789 if (er.encoded == -1) {
1790 mdclog_write(MDCLOG_ERR, "encoding of %s failed, %s", asn_DEF_E2AP_PDU.name, strerror(errno));
1792 } else if (er.encoded > (ssize_t)buff_size) {
1793 mdclog_write(MDCLOG_ERR, "Buffer of size %d is to small for %s, at %s line %d",
1794 (int)rmrMessageBuffer.sendMessage->len,
1795 asn_DEF_E2AP_PDU.name,
1800 rmrMessageBuffer.sendMessage->len = er.encoded;
1809 * @param rmrMessageBuffer
1811 void asnInitiatingRequest(E2AP_PDU_t *pdu,
1812 Sctp_Map_t *sctpMap,
1813 ReportingMessages_t &message,
1814 RmrMessagesBuffer_t &rmrMessageBuffer) {
1815 auto logLevel = mdclog_level_get();
1816 auto procedureCode = ((InitiatingMessage_t *) pdu->choice.initiatingMessage)->procedureCode;
1817 if (logLevel >= MDCLOG_DEBUG) {
1818 mdclog_write(MDCLOG_DEBUG, "Initiating message %ld\n", procedureCode);
1820 switch (procedureCode) {
1821 case ProcedureCode_id_E2setup: {
1822 if (logLevel >= MDCLOG_DEBUG) {
1823 mdclog_write(MDCLOG_DEBUG, "Got E2setup");
1826 // vector <string> RANfunctionsAdded_v;
1827 // vector <string> RANfunctionsModified_v;
1828 // RANfunctionsAdded_v.clear();
1829 // RANfunctionsModified_v.clear();
1830 if (collectSetupRequestData(pdu, sctpMap, message) != 0) {
1834 buildPrometheusList(message.peerInfo, message.peerInfo->sctpParams->prometheusFamily);
1836 string messageName("E2setupRequest");
1837 string ieName("E2setupRequestIEs");
1838 message.message.messageType = RIC_E2_SETUP_REQ;
1839 message.peerInfo->counters[IN_INITI][MSG_COUNTER][ProcedureCode_id_E2setup]->Increment();
1840 message.peerInfo->counters[IN_INITI][BYTES_COUNTER][ProcedureCode_id_E2setup]->Increment((double)message.message.asnLength);
1841 buildAndSendSetupRequest(message, rmrMessageBuffer, pdu);
1844 case ProcedureCode_id_RICserviceUpdate: {
1845 if (logLevel >= MDCLOG_DEBUG) {
1846 mdclog_write(MDCLOG_DEBUG, "Got RICserviceUpdate %s", message.message.enodbName);
1848 // vector <string> RANfunctionsAdded_v;
1849 // vector <string> RANfunctionsModified_v;
1850 // RANfunctionsAdded_v.clear();
1851 // RANfunctionsModified_v.clear();
1852 // if (collectServiceUpdate_RequestData(pdu, sctpMap, message,
1853 // RANfunctionsAdded_v, RANfunctionsModified_v) != 0) {
1857 string messageName("RICserviceUpdate");
1858 string ieName("RICserviceUpdateIEs");
1859 message.message.messageType = RIC_SERVICE_UPDATE;
1861 message.peerInfo->counters[IN_INITI][MSG_COUNTER][ProcedureCode_id_RICserviceUpdate]->Increment();
1862 message.peerInfo->counters[IN_INITI][BYTES_COUNTER][ProcedureCode_id_RICserviceUpdate]->Increment((double)message.message.asnLength);
1864 buildAndSendSetupRequest(message, rmrMessageBuffer, pdu);
1867 case ProcedureCode_id_ErrorIndication: {
1868 if (logLevel >= MDCLOG_DEBUG) {
1869 mdclog_write(MDCLOG_DEBUG, "Got ErrorIndication %s", message.message.enodbName);
1872 message.peerInfo->counters[IN_INITI][MSG_COUNTER][ProcedureCode_id_ErrorIndication]->Increment();
1873 message.peerInfo->counters[IN_INITI][BYTES_COUNTER][ProcedureCode_id_ErrorIndication]->Increment((double)message.message.asnLength);
1875 if (sendRequestToXapp(message, RIC_ERROR_INDICATION, rmrMessageBuffer) != 0) {
1876 mdclog_write(MDCLOG_ERR, "RIC_ERROR_INDICATION failed to send to xAPP");
1880 case ProcedureCode_id_Reset: {
1881 if (logLevel >= MDCLOG_DEBUG) {
1882 mdclog_write(MDCLOG_DEBUG, "Got Reset %s", message.message.enodbName);
1885 message.peerInfo->counters[IN_INITI][MSG_COUNTER][ProcedureCode_id_Reset]->Increment();
1886 message.peerInfo->counters[IN_INITI][BYTES_COUNTER][ProcedureCode_id_Reset]->Increment((double)message.message.asnLength);
1888 if (XML_From_PER(message, rmrMessageBuffer) < 0) {
1892 if (sendRequestToXapp(message, RIC_E2_RESET_REQ, rmrMessageBuffer) != 0) {
1893 mdclog_write(MDCLOG_ERR, "RIC_E2_RESET_REQ message failed to send to xAPP");
1897 case ProcedureCode_id_RICindication: {
1898 if (logLevel >= MDCLOG_DEBUG) {
1899 mdclog_write(MDCLOG_DEBUG, "Got RICindication %s", message.message.enodbName);
1901 for (auto i = 0; i < pdu->choice.initiatingMessage->value.choice.RICindication.protocolIEs.list.count; i++) {
1902 auto messageSent = false;
1903 RICindication_IEs_t *ie = pdu->choice.initiatingMessage->value.choice.RICindication.protocolIEs.list.array[i];
1904 if (logLevel >= MDCLOG_DEBUG) {
1905 mdclog_write(MDCLOG_DEBUG, "ie type (ProtocolIE_ID) = %ld", ie->id);
1907 if (ie->id == ProtocolIE_ID_id_RICrequestID) {
1908 if (logLevel >= MDCLOG_DEBUG) {
1909 mdclog_write(MDCLOG_DEBUG, "Got RIC requestId entry, ie type (ProtocolIE_ID) = %ld", ie->id);
1911 if (ie->value.present == RICindication_IEs__value_PR_RICrequestID) {
1912 static unsigned char tx[32];
1913 message.message.messageType = rmrMessageBuffer.sendMessage->mtype = RIC_INDICATION;
1914 snprintf((char *) tx, sizeof tx, "%15ld", transactionCounter++);
1915 rmr_bytes2xact(rmrMessageBuffer.sendMessage, tx, strlen((const char *) tx));
1916 rmr_bytes2meid(rmrMessageBuffer.sendMessage,
1917 (unsigned char *)message.message.enodbName,
1918 strlen(message.message.enodbName));
1919 rmrMessageBuffer.sendMessage->state = 0;
1920 rmrMessageBuffer.sendMessage->sub_id = (int)ie->value.choice.RICrequestID.ricInstanceID;
1922 //ie->value.choice.RICrequestID.ricInstanceID;
1923 if (mdclog_level_get() >= MDCLOG_DEBUG) {
1924 mdclog_write(MDCLOG_DEBUG, "sub id = %d, mtype = %d, ric instance id %ld, requestor id = %ld",
1925 rmrMessageBuffer.sendMessage->sub_id,
1926 rmrMessageBuffer.sendMessage->mtype,
1927 ie->value.choice.RICrequestID.ricInstanceID,
1928 ie->value.choice.RICrequestID.ricRequestorID);
1931 message.peerInfo->counters[IN_INITI][MSG_COUNTER][ProcedureCode_id_RICindication]->Increment();
1932 message.peerInfo->counters[IN_INITI][BYTES_COUNTER][ProcedureCode_id_RICindication]->Increment((double)message.message.asnLength);
1934 sendRmrMessage(rmrMessageBuffer, message);
1937 mdclog_write(MDCLOG_ERR, "RIC request id missing illegal request");
1947 mdclog_write(MDCLOG_ERR, "Undefined or not supported message = %ld", procedureCode);
1948 message.message.messageType = 0; // no RMR message type yet
1950 buildJsonMessage(message);
1961 * @param rmrMessageBuffer
1963 void asnSuccessfulMsg(E2AP_PDU_t *pdu,
1964 Sctp_Map_t *sctpMap,
1965 ReportingMessages_t &message,
1966 RmrMessagesBuffer_t &rmrMessageBuffer) {
1967 auto procedureCode = pdu->choice.successfulOutcome->procedureCode;
1968 auto logLevel = mdclog_level_get();
1969 if (logLevel >= MDCLOG_INFO) {
1970 mdclog_write(MDCLOG_INFO, "Successful Outcome %ld", procedureCode);
1972 switch (procedureCode) {
1973 case ProcedureCode_id_Reset: {
1974 if (logLevel >= MDCLOG_DEBUG) {
1975 mdclog_write(MDCLOG_DEBUG, "Got Reset %s", message.message.enodbName);
1978 message.peerInfo->counters[IN_SUCC][MSG_COUNTER][ProcedureCode_id_Reset]->Increment();
1979 message.peerInfo->counters[IN_SUCC][BYTES_COUNTER][ProcedureCode_id_Reset]->Increment((double)message.message.asnLength);
1981 if (XML_From_PER(message, rmrMessageBuffer) < 0) {
1984 if (sendRequestToXapp(message, RIC_E2_RESET_RESP, rmrMessageBuffer) != 0) {
1985 mdclog_write(MDCLOG_ERR, "RIC_E2_RESET_RESP message failed to send to xAPP");
1989 case ProcedureCode_id_RICcontrol: {
1990 if (logLevel >= MDCLOG_DEBUG) {
1991 mdclog_write(MDCLOG_DEBUG, "Got RICcontrol %s", message.message.enodbName);
1994 i < pdu->choice.successfulOutcome->value.choice.RICcontrolAcknowledge.protocolIEs.list.count; i++) {
1995 auto messageSent = false;
1996 RICcontrolAcknowledge_IEs_t *ie = pdu->choice.successfulOutcome->value.choice.RICcontrolAcknowledge.protocolIEs.list.array[i];
1997 if (mdclog_level_get() >= MDCLOG_DEBUG) {
1998 mdclog_write(MDCLOG_DEBUG, "ie type (ProtocolIE_ID) = %ld", ie->id);
2000 if (ie->id == ProtocolIE_ID_id_RICrequestID) {
2001 if (mdclog_level_get() >= MDCLOG_DEBUG) {
2002 mdclog_write(MDCLOG_DEBUG, "Got RIC requestId entry, ie type (ProtocolIE_ID) = %ld", ie->id);
2004 if (ie->value.present == RICcontrolAcknowledge_IEs__value_PR_RICrequestID) {
2005 message.message.messageType = rmrMessageBuffer.sendMessage->mtype = RIC_CONTROL_ACK;
2006 rmrMessageBuffer.sendMessage->state = 0;
2007 // rmrMessageBuffer.sendMessage->sub_id = (int) ie->value.choice.RICrequestID.ricRequestorID;
2008 rmrMessageBuffer.sendMessage->sub_id = (int)ie->value.choice.RICrequestID.ricInstanceID;
2010 static unsigned char tx[32];
2011 snprintf((char *) tx, sizeof tx, "%15ld", transactionCounter++);
2012 rmr_bytes2xact(rmrMessageBuffer.sendMessage, tx, strlen((const char *) tx));
2013 rmr_bytes2meid(rmrMessageBuffer.sendMessage,
2014 (unsigned char *)message.message.enodbName,
2015 strlen(message.message.enodbName));
2017 message.peerInfo->counters[IN_SUCC][MSG_COUNTER][ProcedureCode_id_RICcontrol]->Increment();
2018 message.peerInfo->counters[IN_SUCC][BYTES_COUNTER][ProcedureCode_id_RICcontrol]->Increment((double)message.message.asnLength);
2020 sendRmrMessage(rmrMessageBuffer, message);
2023 mdclog_write(MDCLOG_ERR, "RIC request id missing illegal request");
2033 case ProcedureCode_id_RICsubscription: {
2034 if (logLevel >= MDCLOG_DEBUG) {
2035 mdclog_write(MDCLOG_DEBUG, "Got RICsubscription %s", message.message.enodbName);
2038 message.peerInfo->counters[IN_SUCC][MSG_COUNTER][ProcedureCode_id_RICsubscription]->Increment();
2039 message.peerInfo->counters[IN_SUCC][BYTES_COUNTER][ProcedureCode_id_RICsubscription]->Increment((double)message.message.asnLength);
2041 if (sendRequestToXapp(message, RIC_SUB_RESP, rmrMessageBuffer) != 0) {
2042 mdclog_write(MDCLOG_ERR, "Subscription successful message failed to send to xAPP");
2046 case ProcedureCode_id_RICsubscriptionDelete: {
2047 if (logLevel >= MDCLOG_DEBUG) {
2048 mdclog_write(MDCLOG_DEBUG, "Got RICsubscriptionDelete %s", message.message.enodbName);
2051 message.peerInfo->counters[IN_SUCC][MSG_COUNTER][ProcedureCode_id_RICsubscriptionDelete]->Increment();
2052 message.peerInfo->counters[IN_SUCC][BYTES_COUNTER][ProcedureCode_id_RICsubscriptionDelete]->Increment((double)message.message.asnLength);
2054 if (sendRequestToXapp(message, RIC_SUB_DEL_RESP, rmrMessageBuffer) != 0) {
2055 mdclog_write(MDCLOG_ERR, "Subscription delete successful message failed to send to xAPP");
2060 mdclog_write(MDCLOG_WARN, "Undefined or not supported message = %ld", procedureCode);
2061 message.message.messageType = 0; // no RMR message type yet
2062 buildJsonMessage(message);
2073 * @param rmrMessageBuffer
2075 void asnUnSuccsesfulMsg(E2AP_PDU_t *pdu,
2076 Sctp_Map_t *sctpMap,
2077 ReportingMessages_t &message,
2078 RmrMessagesBuffer_t &rmrMessageBuffer) {
2079 auto procedureCode = pdu->choice.unsuccessfulOutcome->procedureCode;
2080 auto logLevel = mdclog_level_get();
2081 if (logLevel >= MDCLOG_INFO) {
2082 mdclog_write(MDCLOG_INFO, "Unsuccessful Outcome %ld", procedureCode);
2084 switch (procedureCode) {
2085 case ProcedureCode_id_RICcontrol: {
2086 if (logLevel >= MDCLOG_DEBUG) {
2087 mdclog_write(MDCLOG_DEBUG, "Got RICcontrol %s", message.message.enodbName);
2090 i < pdu->choice.unsuccessfulOutcome->value.choice.RICcontrolFailure.protocolIEs.list.count; i++) {
2091 auto messageSent = false;
2092 RICcontrolFailure_IEs_t *ie = pdu->choice.unsuccessfulOutcome->value.choice.RICcontrolFailure.protocolIEs.list.array[i];
2093 if (logLevel >= MDCLOG_DEBUG) {
2094 mdclog_write(MDCLOG_DEBUG, "ie type (ProtocolIE_ID) = %ld", ie->id);
2096 if (ie->id == ProtocolIE_ID_id_RICrequestID) {
2097 if (logLevel >= MDCLOG_DEBUG) {
2098 mdclog_write(MDCLOG_DEBUG, "Got RIC requestId entry, ie type (ProtocolIE_ID) = %ld", ie->id);
2100 if (ie->value.present == RICcontrolFailure_IEs__value_PR_RICrequestID) {
2101 message.message.messageType = rmrMessageBuffer.sendMessage->mtype = RIC_CONTROL_FAILURE;
2102 rmrMessageBuffer.sendMessage->state = 0;
2103 // rmrMessageBuffer.sendMessage->sub_id = (int)ie->value.choice.RICrequestID.ricRequestorID;
2104 rmrMessageBuffer.sendMessage->sub_id = (int)ie->value.choice.RICrequestID.ricInstanceID;
2105 static unsigned char tx[32];
2106 snprintf((char *) tx, sizeof tx, "%15ld", transactionCounter++);
2107 rmr_bytes2xact(rmrMessageBuffer.sendMessage, tx, strlen((const char *) tx));
2108 rmr_bytes2meid(rmrMessageBuffer.sendMessage, (unsigned char *) message.message.enodbName,
2109 strlen(message.message.enodbName));
2111 message.peerInfo->counters[IN_UN_SUCC][MSG_COUNTER][ProcedureCode_id_RICcontrol]->Increment();
2112 message.peerInfo->counters[IN_UN_SUCC][BYTES_COUNTER][ProcedureCode_id_RICcontrol]->Increment((double)message.message.asnLength);
2114 sendRmrMessage(rmrMessageBuffer, message);
2117 mdclog_write(MDCLOG_ERR, "RIC request id missing illegal request");
2126 case ProcedureCode_id_RICsubscription: {
2127 if (logLevel >= MDCLOG_DEBUG) {
2128 mdclog_write(MDCLOG_DEBUG, "Got RICsubscription %s", message.message.enodbName);
2131 message.peerInfo->counters[IN_UN_SUCC][MSG_COUNTER][ProcedureCode_id_RICsubscription]->Increment();
2132 message.peerInfo->counters[IN_UN_SUCC][BYTES_COUNTER][ProcedureCode_id_RICsubscription]->Increment((double)message.message.asnLength);
2134 if (sendRequestToXapp(message, RIC_SUB_FAILURE, rmrMessageBuffer) != 0) {
2135 mdclog_write(MDCLOG_ERR, "Subscription unsuccessful message failed to send to xAPP");
2139 case ProcedureCode_id_RICsubscriptionDelete: {
2140 if (logLevel >= MDCLOG_DEBUG) {
2141 mdclog_write(MDCLOG_DEBUG, "Got RICsubscriptionDelete %s", message.message.enodbName);
2144 message.peerInfo->counters[IN_UN_SUCC][MSG_COUNTER][ProcedureCode_id_RICsubscriptionDelete]->Increment();
2145 message.peerInfo->counters[IN_UN_SUCC][BYTES_COUNTER][ProcedureCode_id_RICsubscriptionDelete]->Increment((double)message.message.asnLength);
2147 if (sendRequestToXapp(message, RIC_SUB_FAILURE, rmrMessageBuffer) != 0) {
2148 mdclog_write(MDCLOG_ERR, "Subscription Delete unsuccessful message failed to send to xAPP");
2153 mdclog_write(MDCLOG_WARN, "Undefined or not supported message = %ld", procedureCode);
2154 message.message.messageType = 0; // no RMR message type yet
2156 buildJsonMessage(message);
2167 * @param rmrMmessageBuffer
2170 int sendRequestToXapp(ReportingMessages_t &message,
2172 RmrMessagesBuffer_t &rmrMmessageBuffer) {
2173 rmr_bytes2meid(rmrMmessageBuffer.sendMessage,
2174 (unsigned char *)message.message.enodbName,
2175 strlen(message.message.enodbName));
2176 message.message.messageType = rmrMmessageBuffer.sendMessage->mtype = requestId;
2177 rmrMmessageBuffer.sendMessage->state = 0;
2178 static unsigned char tx[32];
2179 snprintf((char *) tx, sizeof tx, "%15ld", transactionCounter++);
2180 rmr_bytes2xact(rmrMmessageBuffer.sendMessage, tx, strlen((const char *) tx));
2182 auto rc = sendRmrMessage(rmrMmessageBuffer, message);
2188 * @param pSctpParams
2190 void getRmrContext(sctp_params_t &pSctpParams) {
2191 pSctpParams.rmrCtx = nullptr;
2192 pSctpParams.rmrCtx = rmr_init(pSctpParams.rmrAddress, RECEIVE_XAPP_BUFFER_SIZE, RMRFL_NONE);
2193 if (pSctpParams.rmrCtx == nullptr) {
2194 mdclog_write(MDCLOG_ERR, "Failed to initialize RMR");
2198 rmr_set_stimeout(pSctpParams.rmrCtx, 0); // disable retries for any send operation
2199 // we need to find that routing table exist and we can run
2200 if (mdclog_level_get() >= MDCLOG_INFO) {
2201 mdclog_write(MDCLOG_INFO, "We are after RMR INIT wait for RMR_Ready");
2206 if ((rmrReady = rmr_ready(pSctpParams.rmrCtx)) == 0) {
2210 if (count % 60 == 0) {
2211 mdclog_write(MDCLOG_INFO, "waiting to RMR ready state for %d seconds", count);
2214 if (mdclog_level_get() >= MDCLOG_INFO) {
2215 mdclog_write(MDCLOG_INFO, "RMR running");
2217 rmr_init_trace(pSctpParams.rmrCtx, 200);
2218 // get the RMR fd for the epoll
2219 pSctpParams.rmrListenFd = rmr_get_rcvfd(pSctpParams.rmrCtx);
2220 struct epoll_event event{};
2221 // add RMR fd to epoll
2222 event.events = (EPOLLIN);
2223 event.data.fd = pSctpParams.rmrListenFd;
2224 // add listening RMR FD to epoll
2225 if (epoll_ctl(pSctpParams.epoll_fd, EPOLL_CTL_ADD, pSctpParams.rmrListenFd, &event)) {
2226 mdclog_write(MDCLOG_ERR, "Failed to add RMR descriptor to epoll");
2227 close(pSctpParams.rmrListenFd);
2228 rmr_close(pSctpParams.rmrCtx);
2229 pSctpParams.rmrCtx = nullptr;
2236 * @param rmrMessageBuffer
2239 int PER_FromXML(ReportingMessages_t &message, RmrMessagesBuffer_t &rmrMessageBuffer) {
2240 E2AP_PDU_t *pdu = nullptr;
2242 if (mdclog_level_get() >= MDCLOG_DEBUG) {
2243 mdclog_write(MDCLOG_DEBUG, "got xml Format data from xApp of size %d is:%s",
2244 rmrMessageBuffer.rcvMessage->len, rmrMessageBuffer.rcvMessage->payload);
2246 auto rval = asn_decode(nullptr, ATS_BASIC_XER, &asn_DEF_E2AP_PDU, (void **) &pdu,
2247 rmrMessageBuffer.rcvMessage->payload, rmrMessageBuffer.rcvMessage->len);
2248 if (mdclog_level_get() >= MDCLOG_DEBUG) {
2249 mdclog_write(MDCLOG_DEBUG, "%s After decoding the XML to PDU", __func__ );
2251 if (rval.code != RC_OK) {
2252 mdclog_write(MDCLOG_ERR, "Error %d Decoding (unpack) setup response from E2MGR : %s",
2254 message.message.enodbName);
2258 int buff_size = RECEIVE_XAPP_BUFFER_SIZE;
2259 auto er = asn_encode_to_buffer(nullptr, ATS_ALIGNED_BASIC_PER, &asn_DEF_E2AP_PDU, pdu,
2260 rmrMessageBuffer.rcvMessage->payload, buff_size);
2261 if (mdclog_level_get() >= MDCLOG_DEBUG) {
2262 mdclog_write(MDCLOG_DEBUG, "%s After encoding PDU to PER", __func__ );
2264 if (er.encoded == -1) {
2265 mdclog_write(MDCLOG_ERR, "encoding of %s failed, %s", asn_DEF_E2AP_PDU.name, strerror(errno));
2267 } else if (er.encoded > (ssize_t)buff_size) {
2268 mdclog_write(MDCLOG_ERR, "Buffer of size %d is to small for %s, at %s line %d",
2269 (int)rmrMessageBuffer.rcvMessage->len,
2270 asn_DEF_E2AP_PDU.name,
2275 rmrMessageBuffer.rcvMessage->len = er.encoded;
2282 * @param rmrMessageBuffer
2286 int receiveXappMessages(Sctp_Map_t *sctpMap,
2287 RmrMessagesBuffer_t &rmrMessageBuffer,
2288 struct timespec &ts) {
2289 int loglevel = mdclog_level_get();
2290 if (rmrMessageBuffer.rcvMessage == nullptr) {
2292 mdclog_write(MDCLOG_ERR, "RMR Allocation message, %s", strerror(errno));
2296 // if (loglevel >= MDCLOG_DEBUG) {
2297 // mdclog_write(MDCLOG_DEBUG, "Call to rmr_rcv_msg");
2299 rmrMessageBuffer.rcvMessage = rmr_rcv_msg(rmrMessageBuffer.rmrCtx, rmrMessageBuffer.rcvMessage);
2300 if (rmrMessageBuffer.rcvMessage == nullptr) {
2301 mdclog_write(MDCLOG_ERR, "RMR Receiving message with null pointer, Reallocated rmr message buffer");
2302 rmrMessageBuffer.rcvMessage = rmr_alloc_msg(rmrMessageBuffer.rmrCtx, RECEIVE_XAPP_BUFFER_SIZE);
2305 ReportingMessages_t message;
2306 message.message.direction = 'D';
2307 message.message.time.tv_nsec = ts.tv_nsec;
2308 message.message.time.tv_sec = ts.tv_sec;
2310 // get message payload
2311 //auto msgData = msg->payload;
2313 rmrMessageBuffer.rcvMessage->state = 0;
2315 if (rmrMessageBuffer.rcvMessage->state != 0) {
2316 mdclog_write(MDCLOG_ERR, "RMR Receiving message with stat = %d", rmrMessageBuffer.rcvMessage->state);
2319 rmr_get_meid(rmrMessageBuffer.rcvMessage, (unsigned char *)message.message.enodbName);
2320 message.peerInfo = (ConnectedCU_t *) sctpMap->find(message.message.enodbName);
2321 if (message.peerInfo == nullptr) {
2322 auto type = rmrMessageBuffer.rcvMessage->mtype;
2324 case RIC_SCTP_CLEAR_ALL:
2325 case E2_TERM_KEEP_ALIVE_REQ:
2326 case RIC_HEALTH_CHECK_REQ:
2332 mdclog_write(MDCLOG_ERR, "Failed to send message no CU entry %s", message.message.enodbName);
2337 if (rmrMessageBuffer.rcvMessage->mtype != RIC_HEALTH_CHECK_REQ) {
2338 num_of_XAPP_messages.fetch_add(1, std::memory_order_release);
2341 switch (rmrMessageBuffer.rcvMessage->mtype) {
2342 case RIC_E2_SETUP_RESP : {
2343 if (loglevel >= MDCLOG_DEBUG) {
2344 mdclog_write(MDCLOG_DEBUG, "RIC_E2_SETUP_RESP");
2346 if (PER_FromXML(message, rmrMessageBuffer) != 0) {
2350 message.peerInfo->counters[OUT_SUCC][MSG_COUNTER][ProcedureCode_id_E2setup]->Increment();
2351 message.peerInfo->counters[OUT_SUCC][BYTES_COUNTER][ProcedureCode_id_E2setup]->Increment(rmrMessageBuffer.rcvMessage->len);
2353 if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap) != 0) {
2354 mdclog_write(MDCLOG_ERR, "Failed to send RIC_E2_SETUP_RESP");
2359 case RIC_E2_SETUP_FAILURE : {
2360 if (loglevel >= MDCLOG_DEBUG) {
2361 mdclog_write(MDCLOG_DEBUG, "RIC_E2_SETUP_FAILURE");
2363 if (PER_FromXML(message, rmrMessageBuffer) != 0) {
2367 message.peerInfo->counters[OUT_UN_SUCC][MSG_COUNTER][ProcedureCode_id_E2setup]->Increment();
2368 message.peerInfo->counters[OUT_UN_SUCC][BYTES_COUNTER][ProcedureCode_id_E2setup]->Increment(rmrMessageBuffer.rcvMessage->len);
2370 if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap) != 0) {
2371 mdclog_write(MDCLOG_ERR, "Failed to send RIC_E2_SETUP_FAILURE");
2376 case RIC_ERROR_INDICATION: {
2377 if (loglevel >= MDCLOG_DEBUG) {
2378 mdclog_write(MDCLOG_DEBUG, "RIC_ERROR_INDICATION");
2381 message.peerInfo->counters[OUT_INITI][MSG_COUNTER][ProcedureCode_id_ErrorIndication]->Increment();
2382 message.peerInfo->counters[OUT_INITI][BYTES_COUNTER][ProcedureCode_id_ErrorIndication]->Increment(rmrMessageBuffer.rcvMessage->len);
2384 if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap) != 0) {
2385 mdclog_write(MDCLOG_ERR, "Failed to send RIC_ERROR_INDICATION");
2391 if (loglevel >= MDCLOG_DEBUG) {
2392 mdclog_write(MDCLOG_DEBUG, "RIC_SUB_REQ");
2395 message.peerInfo->counters[OUT_INITI][MSG_COUNTER][ProcedureCode_id_RICsubscription]->Increment();
2396 message.peerInfo->counters[OUT_INITI][BYTES_COUNTER][ProcedureCode_id_RICsubscription]->Increment(rmrMessageBuffer.rcvMessage->len);
2398 if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap) != 0) {
2399 mdclog_write(MDCLOG_ERR, "Failed to send RIC_SUB_REQ");
2404 case RIC_SUB_DEL_REQ: {
2405 if (loglevel >= MDCLOG_DEBUG) {
2406 mdclog_write(MDCLOG_DEBUG, "RIC_SUB_DEL_REQ");
2409 message.peerInfo->counters[OUT_INITI][MSG_COUNTER][ProcedureCode_id_RICsubscriptionDelete]->Increment();
2410 message.peerInfo->counters[OUT_INITI][BYTES_COUNTER][ProcedureCode_id_RICsubscriptionDelete]->Increment(rmrMessageBuffer.rcvMessage->len);
2412 if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap) != 0) {
2413 mdclog_write(MDCLOG_ERR, "Failed to send RIC_SUB_DEL_REQ");
2418 case RIC_CONTROL_REQ: {
2419 if (loglevel >= MDCLOG_DEBUG) {
2420 mdclog_write(MDCLOG_DEBUG, "RIC_CONTROL_REQ");
2423 message.peerInfo->counters[OUT_INITI][MSG_COUNTER][ProcedureCode_id_RICcontrol]->Increment();
2424 message.peerInfo->counters[OUT_INITI][BYTES_COUNTER][ProcedureCode_id_RICcontrol]->Increment(rmrMessageBuffer.rcvMessage->len);
2426 if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap) != 0) {
2427 mdclog_write(MDCLOG_ERR, "Failed to send RIC_CONTROL_REQ");
2432 case RIC_SERVICE_QUERY: {
2433 if (loglevel >= MDCLOG_DEBUG) {
2434 mdclog_write(MDCLOG_DEBUG, "RIC_SERVICE_QUERY");
2436 if (PER_FromXML(message, rmrMessageBuffer) != 0) {
2440 message.peerInfo->counters[OUT_INITI][MSG_COUNTER][ProcedureCode_id_RICserviceQuery]->Increment();
2441 message.peerInfo->counters[OUT_INITI][BYTES_COUNTER][ProcedureCode_id_RICserviceQuery]->Increment(rmrMessageBuffer.rcvMessage->len);
2443 if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap) != 0) {
2444 mdclog_write(MDCLOG_ERR, "Failed to send RIC_SERVICE_QUERY");
2449 case RIC_SERVICE_UPDATE_ACK: {
2450 if (loglevel >= MDCLOG_DEBUG) {
2451 mdclog_write(MDCLOG_DEBUG, "RIC_SERVICE_UPDATE_ACK");
2453 if (PER_FromXML(message, rmrMessageBuffer) != 0) {
2454 mdclog_write(MDCLOG_ERR, "error in PER_FromXML");
2458 message.peerInfo->counters[OUT_SUCC][MSG_COUNTER][ProcedureCode_id_RICserviceUpdate]->Increment();
2459 message.peerInfo->counters[OUT_SUCC][BYTES_COUNTER][ProcedureCode_id_RICserviceUpdate]->Increment(rmrMessageBuffer.rcvMessage->len);
2461 if (loglevel >= MDCLOG_DEBUG) {
2462 mdclog_write(MDCLOG_DEBUG, "Before sending to CU");
2464 if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap) != 0) {
2465 mdclog_write(MDCLOG_ERR, "Failed to send RIC_SERVICE_UPDATE_ACK");
2470 case RIC_SERVICE_UPDATE_FAILURE: {
2471 if (loglevel >= MDCLOG_DEBUG) {
2472 mdclog_write(MDCLOG_DEBUG, "RIC_SERVICE_UPDATE_FAILURE");
2474 if (PER_FromXML(message, rmrMessageBuffer) != 0) {
2478 message.peerInfo->counters[OUT_UN_SUCC][MSG_COUNTER][ProcedureCode_id_RICserviceUpdate]->Increment();
2479 message.peerInfo->counters[OUT_UN_SUCC][BYTES_COUNTER][ProcedureCode_id_RICserviceUpdate]->Increment(rmrMessageBuffer.rcvMessage->len);
2481 if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap) != 0) {
2482 mdclog_write(MDCLOG_ERR, "Failed to send RIC_SERVICE_UPDATE_FAILURE");
2487 case RIC_E2_RESET_REQ: {
2488 if (loglevel >= MDCLOG_DEBUG) {
2489 mdclog_write(MDCLOG_DEBUG, "RIC_E2_RESET_REQ");
2491 if (PER_FromXML(message, rmrMessageBuffer) != 0) {
2495 message.peerInfo->counters[OUT_INITI][MSG_COUNTER][ProcedureCode_id_Reset]->Increment();
2496 message.peerInfo->counters[OUT_INITI][BYTES_COUNTER][ProcedureCode_id_Reset]->Increment(rmrMessageBuffer.rcvMessage->len);
2498 if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap) != 0) {
2499 mdclog_write(MDCLOG_ERR, "Failed to send RIC_E2_RESET");
2504 case RIC_E2_RESET_RESP: {
2505 if (loglevel >= MDCLOG_DEBUG) {
2506 mdclog_write(MDCLOG_DEBUG, "RIC_E2_RESET_RESP");
2508 if (PER_FromXML(message, rmrMessageBuffer) != 0) {
2512 message.peerInfo->counters[OUT_SUCC][MSG_COUNTER][ProcedureCode_id_Reset]->Increment();
2513 message.peerInfo->counters[OUT_SUCC][BYTES_COUNTER][ProcedureCode_id_Reset]->Increment(rmrMessageBuffer.rcvMessage->len);
2515 if (sendDirectionalSctpMsg(rmrMessageBuffer, message, 0, sctpMap) != 0) {
2516 mdclog_write(MDCLOG_ERR, "Failed to send RIC_E2_RESET_RESP");
2521 case RIC_SCTP_CLEAR_ALL: {
2522 mdclog_write(MDCLOG_INFO, "RIC_SCTP_CLEAR_ALL");
2523 // loop on all keys and close socket and then erase all map.
2525 sctpMap->getKeys(v);
2526 for (auto const &iter : v) { //}; iter != sctpMap.end(); iter++) {
2527 if (!boost::starts_with((string) (iter), "host:") && !boost::starts_with((string) (iter), "msg:")) {
2528 auto *peerInfo = (ConnectedCU_t *) sctpMap->find(iter);
2529 if (peerInfo == nullptr) {
2532 close(peerInfo->fileDescriptor);
2533 memcpy(message.message.enodbName, peerInfo->enodbName, sizeof(peerInfo->enodbName));
2534 message.message.direction = 'D';
2535 message.message.time.tv_nsec = ts.tv_nsec;
2536 message.message.time.tv_sec = ts.tv_sec;
2538 message.message.asnLength = rmrMessageBuffer.sendMessage->len =
2539 snprintf((char *)rmrMessageBuffer.sendMessage->payload,
2541 "%s|RIC_SCTP_CLEAR_ALL",
2542 peerInfo->enodbName);
2543 message.message.asndata = rmrMessageBuffer.sendMessage->payload;
2544 mdclog_write(MDCLOG_INFO, "%s", message.message.asndata);
2545 if (sendRequestToXapp(message, RIC_SCTP_CONNECTION_FAILURE, rmrMessageBuffer) != 0) {
2546 mdclog_write(MDCLOG_ERR, "SCTP_CONNECTION_FAIL message failed to send to xAPP");
2556 case E2_TERM_KEEP_ALIVE_REQ: {
2557 // send message back
2558 rmr_bytes2payload(rmrMessageBuffer.sendMessage,
2559 (unsigned char *)rmrMessageBuffer.ka_message,
2560 rmrMessageBuffer.ka_message_len);
2561 rmrMessageBuffer.sendMessage->mtype = E2_TERM_KEEP_ALIVE_RESP;
2562 rmrMessageBuffer.sendMessage->state = 0;
2563 static unsigned char tx[32];
2564 auto txLen = snprintf((char *) tx, sizeof tx, "%15ld", transactionCounter++);
2565 rmr_bytes2xact(rmrMessageBuffer.sendMessage, tx, txLen);
2567 rmrMessageBuffer.sendMessage = rmr_send_msg(rmrMessageBuffer.rmrCtx, rmrMessageBuffer.sendMessage);
2569 if (rmrMessageBuffer.sendMessage == nullptr) {
2570 rmrMessageBuffer.sendMessage = rmr_alloc_msg(rmrMessageBuffer.rmrCtx, RECEIVE_XAPP_BUFFER_SIZE);
2571 mdclog_write(MDCLOG_ERR, "Failed to send E2_TERM_KEEP_ALIVE_RESP RMR message returned NULL");
2572 } else if (rmrMessageBuffer.sendMessage->state != 0) {
2573 mdclog_write(MDCLOG_ERR, "Failed to send E2_TERM_KEEP_ALIVE_RESP, on RMR state = %d ( %s)",
2574 rmrMessageBuffer.sendMessage->state, translateRmrErrorMessages(rmrMessageBuffer.sendMessage->state).c_str());
2575 } else if (loglevel >= MDCLOG_DEBUG) {
2576 mdclog_write(MDCLOG_DEBUG, "Got Keep Alive Request send : %s", rmrMessageBuffer.ka_message);
2581 case RIC_HEALTH_CHECK_REQ: {
2582 static int counter = 0;
2583 // send message back
2584 rmr_bytes2payload(rmrMessageBuffer.rcvMessage,
2585 (unsigned char *)"OK",
2587 rmrMessageBuffer.rcvMessage->mtype = RIC_HEALTH_CHECK_RESP;
2588 rmrMessageBuffer.rcvMessage->state = 0;
2589 static unsigned char tx[32];
2590 auto txLen = snprintf((char *) tx, sizeof tx, "%15ld", transactionCounter++);
2591 rmr_bytes2xact(rmrMessageBuffer.rcvMessage, tx, txLen);
2592 rmrMessageBuffer.rcvMessage = rmr_rts_msg(rmrMessageBuffer.rmrCtx, rmrMessageBuffer.rcvMessage);
2593 //rmrMessageBuffer.sendMessage = rmr_send_msg(rmrMessageBuffer.rmrCtx, rmrMessageBuffer.sendMessage);
2594 if (rmrMessageBuffer.rcvMessage == nullptr) {
2595 rmrMessageBuffer.rcvMessage = rmr_alloc_msg(rmrMessageBuffer.rmrCtx, RECEIVE_XAPP_BUFFER_SIZE);
2596 mdclog_write(MDCLOG_ERR, "Failed to send RIC_HEALTH_CHECK_RESP RMR message returned NULL");
2597 } else if (rmrMessageBuffer.rcvMessage->state != 0) {
2598 mdclog_write(MDCLOG_ERR, "Failed to send RIC_HEALTH_CHECK_RESP, on RMR state = %d ( %s)",
2599 rmrMessageBuffer.rcvMessage->state, translateRmrErrorMessages(rmrMessageBuffer.rcvMessage->state).c_str());
2600 } else if (loglevel >= MDCLOG_DEBUG && ++counter % 100 == 0) {
2601 mdclog_write(MDCLOG_DEBUG, "Got %d RIC_HEALTH_CHECK_REQ Request send : OK", counter);
2608 mdclog_write(MDCLOG_WARN, "Message Type : %d is not supported", rmrMessageBuffer.rcvMessage->mtype);
2609 message.message.asndata = rmrMessageBuffer.rcvMessage->payload;
2610 message.message.asnLength = rmrMessageBuffer.rcvMessage->len;
2611 message.message.time.tv_nsec = ts.tv_nsec;
2612 message.message.time.tv_sec = ts.tv_sec;
2613 message.message.messageType = rmrMessageBuffer.rcvMessage->mtype;
2615 buildJsonMessage(message);
2620 if (mdclog_level_get() >= MDCLOG_DEBUG) {
2621 mdclog_write(MDCLOG_DEBUG, "EXIT OK from %s", __FUNCTION__);
2627 * Send message to the CU that is not expecting for successful or unsuccessful results
2628 * @param messageBuffer
2630 * @param failedMsgId
2634 int sendDirectionalSctpMsg(RmrMessagesBuffer_t &messageBuffer,
2635 ReportingMessages_t &message,
2637 Sctp_Map_t *sctpMap) {
2638 if (mdclog_level_get() >= MDCLOG_DEBUG) {
2639 mdclog_write(MDCLOG_DEBUG, "send message: %d to %s address", message.message.messageType, message.message.enodbName);
2642 getRequestMetaData(message, messageBuffer);
2643 if (mdclog_level_get() >= MDCLOG_INFO) {
2644 mdclog_write(MDCLOG_INFO, "send message to %s address", message.message.enodbName);
2647 auto rc = sendMessagetoCu(sctpMap, messageBuffer, message, failedMsgId);
2654 * @param messageBuffer
2656 * @param failedMesgId
2659 int sendMessagetoCu(Sctp_Map_t *sctpMap,
2660 RmrMessagesBuffer_t &messageBuffer,
2661 ReportingMessages_t &message,
2664 message.message.messageType = messageBuffer.rcvMessage->mtype;
2665 auto rc = sendSctpMsg(message.peerInfo, message, sctpMap);
2680 int addToEpoll(int epoll_fd,
2681 ConnectedCU_t *peerInfo,
2683 Sctp_Map_t *sctpMap,
2687 struct epoll_event event{};
2688 event.data.ptr = peerInfo;
2689 event.events = events;
2690 if (epoll_ctl(epoll_fd, EPOLL_CTL_ADD, peerInfo->fileDescriptor, &event) < 0) {
2691 if (mdclog_level_get() >= MDCLOG_DEBUG) {
2692 mdclog_write(MDCLOG_DEBUG, "epoll_ctl EPOLL_CTL_ADD (may check not to quit here), %s, %s %d",
2693 strerror(errno), __func__, __LINE__);
2695 close(peerInfo->fileDescriptor);
2696 if (enodbName != nullptr) {
2697 cleanHashEntry(peerInfo, sctpMap);
2698 char key[MAX_ENODB_NAME_SIZE * 2];
2699 snprintf(key, MAX_ENODB_NAME_SIZE * 2, "msg:%s|%d", enodbName, msgType);
2700 if (mdclog_level_get() >= MDCLOG_DEBUG) {
2701 mdclog_write(MDCLOG_DEBUG, "remove key = %s from %s at line %d", key, __FUNCTION__, __LINE__);
2703 auto tmp = sctpMap->find(key);
2706 sctpMap->erase(key);
2709 peerInfo->enodbName[0] = 0;
2711 mdclog_write(MDCLOG_ERR, "epoll_ctl EPOLL_CTL_ADD (may check not to quit here)");
2727 int modifyToEpoll(int epoll_fd,
2728 ConnectedCU_t *peerInfo,
2730 Sctp_Map_t *sctpMap,
2734 struct epoll_event event{};
2735 event.data.ptr = peerInfo;
2736 event.events = events;
2737 if (epoll_ctl(epoll_fd, EPOLL_CTL_MOD, peerInfo->fileDescriptor, &event) < 0) {
2738 if (mdclog_level_get() >= MDCLOG_DEBUG) {
2739 mdclog_write(MDCLOG_DEBUG, "epoll_ctl EPOLL_CTL_MOD (may check not to quit here), %s, %s %d",
2740 strerror(errno), __func__, __LINE__);
2742 close(peerInfo->fileDescriptor);
2743 cleanHashEntry(peerInfo, sctpMap);
2744 char key[MAX_ENODB_NAME_SIZE * 2];
2745 snprintf(key, MAX_ENODB_NAME_SIZE * 2, "msg:%s|%d", enodbName, msgType);
2746 if (mdclog_level_get() >= MDCLOG_DEBUG) {
2747 mdclog_write(MDCLOG_DEBUG, "remove key = %s from %s at line %d", key, __FUNCTION__, __LINE__);
2749 auto tmp = sctpMap->find(key);
2753 sctpMap->erase(key);
2754 mdclog_write(MDCLOG_ERR, "epoll_ctl EPOLL_CTL_ADD (may check not to quit here)");
2761 int sendRmrMessage(RmrMessagesBuffer_t &rmrMessageBuffer, ReportingMessages_t &message) {
2762 buildJsonMessage(message);
2764 rmrMessageBuffer.sendMessage = rmr_send_msg(rmrMessageBuffer.rmrCtx, rmrMessageBuffer.sendMessage);
2766 rmrMessageBuffer.sendMessage->state = RMR_ERR_RETRY;
2768 if (rmrMessageBuffer.sendMessage == nullptr) {
2769 rmrMessageBuffer.sendMessage = rmr_alloc_msg(rmrMessageBuffer.rmrCtx, RECEIVE_XAPP_BUFFER_SIZE);
2770 mdclog_write(MDCLOG_ERR, "RMR failed send message returned with NULL pointer");
2774 if (rmrMessageBuffer.sendMessage->state != 0) {
2775 char meid[RMR_MAX_MEID]{};
2776 if (rmrMessageBuffer.sendMessage->state == RMR_ERR_RETRY) {
2778 rmrMessageBuffer.sendMessage->state = 0;
2779 mdclog_write(MDCLOG_INFO, "RETRY sending Message type %d to Xapp from %s",
2780 rmrMessageBuffer.sendMessage->mtype,
2781 rmr_get_meid(rmrMessageBuffer.sendMessage, (unsigned char *)meid));
2783 rmrMessageBuffer.sendMessage = rmr_send_msg(rmrMessageBuffer.rmrCtx, rmrMessageBuffer.sendMessage);
2785 if (rmrMessageBuffer.sendMessage == nullptr) {
2786 mdclog_write(MDCLOG_ERR, "RMR failed send message returned with NULL pointer");
2787 rmrMessageBuffer.sendMessage = rmr_alloc_msg(rmrMessageBuffer.rmrCtx, RECEIVE_XAPP_BUFFER_SIZE);
2789 } else if (rmrMessageBuffer.sendMessage->state != 0) {
2790 mdclog_write(MDCLOG_ERR,
2791 "Message state %s while sending request %d to Xapp from %s after retry of 10 microseconds",
2792 translateRmrErrorMessages(rmrMessageBuffer.sendMessage->state).c_str(),
2793 rmrMessageBuffer.sendMessage->mtype,
2794 rmr_get_meid(rmrMessageBuffer.sendMessage, (unsigned char *)meid));
2795 auto rc = rmrMessageBuffer.sendMessage->state;
2799 mdclog_write(MDCLOG_ERR, "Message state %s while sending request %d to Xapp from %s",
2800 translateRmrErrorMessages(rmrMessageBuffer.sendMessage->state).c_str(),
2801 rmrMessageBuffer.sendMessage->mtype,
2802 rmr_get_meid(rmrMessageBuffer.sendMessage, (unsigned char *)meid));
2803 return rmrMessageBuffer.sendMessage->state;
2809 void buildJsonMessage(ReportingMessages_t &message) {
2814 message.outLen = sizeof(message.base64Data);
2815 base64::encode((const unsigned char *) message.message.asndata,
2816 (const int) message.message.asnLength,
2819 if (mdclog_level_get() >= MDCLOG_DEBUG) {
2820 mdclog_write(MDCLOG_DEBUG, "Tracing: ASN length = %d, base64 message length = %d ",
2821 (int) message.message.asnLength,
2822 (int) message.outLen);
2825 snprintf(message.buffer, sizeof(message.buffer),
2826 "{\"header\": {\"ts\": \"%ld.%09ld\","
2827 "\"ranName\": \"%s\","
2828 "\"messageType\": %d,"
2829 "\"direction\": \"%c\"},"
2830 "\"base64Length\": %d,"
2831 "\"asnBase64\": \"%s\"}",
2832 message.message.time.tv_sec,
2833 message.message.time.tv_nsec,
2834 message.message.enodbName,
2835 message.message.messageType,
2836 message.message.direction,
2837 (int) message.outLen,
2838 message.base64Data);
2839 static src::logger_mt &lg = my_logger::get();
2841 BOOST_LOG(lg) << message.buffer;
2847 * take RMR error code to string
2851 string translateRmrErrorMessages(int state) {
2855 str = "RMR_OK - state is good";
2857 case RMR_ERR_BADARG:
2858 str = "RMR_ERR_BADARG - argument passed to function was unusable";
2860 case RMR_ERR_NOENDPT:
2861 str = "RMR_ERR_NOENDPT - send//call could not find an endpoint based on msg type";
2864 str = "RMR_ERR_EMPTY - msg received had no payload; attempt to send an empty message";
2867 str = "RMR_ERR_NOHDR - message didn't contain a valid header";
2869 case RMR_ERR_SENDFAILED:
2870 str = "RMR_ERR_SENDFAILED - send failed; errno has nano reason";
2872 case RMR_ERR_CALLFAILED:
2873 str = "RMR_ERR_CALLFAILED - unable to send call() message";
2875 case RMR_ERR_NOWHOPEN:
2876 str = "RMR_ERR_NOWHOPEN - no wormholes are open";
2879 str = "RMR_ERR_WHID - wormhole id was invalid";
2881 case RMR_ERR_OVERFLOW:
2882 str = "RMR_ERR_OVERFLOW - operation would have busted through a buffer/field size";
2885 str = "RMR_ERR_RETRY - request (send/call/rts) failed, but caller should retry (EAGAIN for wrappers)";
2887 case RMR_ERR_RCVFAILED:
2888 str = "RMR_ERR_RCVFAILED - receive failed (hard error)";
2890 case RMR_ERR_TIMEOUT:
2891 str = "RMR_ERR_TIMEOUT - message processing call timed out";
2894 str = "RMR_ERR_UNSET - the message hasn't been populated with a transport buffer";
2897 str = "RMR_ERR_TRUNC - received message likely truncated";
2899 case RMR_ERR_INITFAILED:
2900 str = "RMR_ERR_INITFAILED - initialisation of something (probably message) failed";
2902 case RMR_ERR_NOTSUPP:
2903 str = "RMR_ERR_NOTSUPP - the request is not supported, or RMr was not initialised for the request";
2907 snprintf(buf, sizeof buf, "UNDOCUMENTED RMR_ERR : %d", state);