X-Git-Url: https://gerrit.o-ran-sc.org/r/gitweb?a=blobdiff_plain;f=src%2Frmr%2Fnng%2Fsrc%2Fsr_nng_static.c;h=dfe3d0477072ef4a2c7f3ca3b5c80e877389d042;hb=f23c7756b6c75f571e28346c6ee1466f108932c8;hp=7c175894dc2ebd5485313b8466a372730767e7a8;hpb=68c1ab2191d9959fde0bd275a560f7c9cf6df485;p=ric-plt%2Flib%2Frmr.git diff --git a/src/rmr/nng/src/sr_nng_static.c b/src/rmr/nng/src/sr_nng_static.c index 7c17589..dfe3d04 100644 --- a/src/rmr/nng/src/sr_nng_static.c +++ b/src/rmr/nng/src/sr_nng_static.c @@ -1,4 +1,4 @@ -// : vi ts=4 sw=4 noet : +// : vi ts=4 sw=4 noet 2 /* ================================================================================== Copyright (c) 2019 Nokia @@ -142,8 +142,8 @@ static rmr_mbuf_t* alloc_zcmsg( uta_ctx_t* ctx, rmr_mbuf_t* msg, int size, int s hdr->sub_id = htonl( UNSET_SUBID ); SET_HDR_LEN( hdr ); // ensure these are converted to net byte order SET_HDR_TR_LEN( hdr, ctx->trace_data_len ); - //SET_HDR_D1_LEN( hdr, ctx->d1_len ); // no need until we start using them - //SET_HDR_D2_LEN( hdr, ctx->d2_len ); + SET_HDR_D1_LEN( hdr, ctx->d1_len ); + //SET_HDR_D2_LEN( hdr, ctx->d2_len ); // future } msg->len = 0; // length of data in the payload msg->alloc_len = mlen; // length of allocated transport buffer @@ -153,7 +153,8 @@ static rmr_mbuf_t* alloc_zcmsg( uta_ctx_t* ctx, rmr_mbuf_t* msg, int size, int s msg->xaction = ((uta_mhdr_t *)msg->header)->xid; // point at transaction id in header area msg->state = state; // fill in caller's state (likely the state of the last operation) msg->flags |= MFL_ZEROCOPY; // this is a zerocopy sendable message - strncpy( (char *) ((uta_mhdr_t *)msg->header)->src, ctx->my_name, RMR_MAX_SID ); + strncpy( (char *) ((uta_mhdr_t *)msg->header)->src, ctx->my_name, RMR_MAX_SRC ); + strncpy( (char *) ((uta_mhdr_t *)msg->header)->srcip, ctx->my_ip, RMR_MAX_SRC ); if( DEBUG > 1 ) fprintf( stderr, "[DBUG] alloc_zcmsg mlen=%ld size=%d mpl=%d flags=%02x\n", (long) mlen, size, ctx->max_plen, msg->flags ); @@ -206,7 +207,7 @@ static rmr_mbuf_t* alloc_mbuf( uta_ctx_t* ctx, int state ) { is somewhat based on header version. */ static void ref_tpbuf( rmr_mbuf_t* msg, size_t alen ) { - uta_mhdr_t* hdr; // current header + uta_mhdr_t* hdr = NULL; // current header uta_v1mhdr_t* v1hdr; // version 1 header int ver; int hlen; // header len to use for a truncation check @@ -249,7 +250,7 @@ static void ref_tpbuf( rmr_mbuf_t* msg, size_t alen ) { break; } - if( msg->len > (msg->alloc_len - hlen ) ) { // more than we should have had room for; error + if( msg->len > (msg->alloc_len - hlen ) ) { msg->state = RMR_ERR_TRUNC; msg->len = msg->alloc_len - hlen; // adjust len down so user app doesn't overrun } else { @@ -322,7 +323,6 @@ static inline rmr_mbuf_t* realloc_msg( rmr_mbuf_t* old_msg, int tr_len ) { uta_mhdr_t* hdr; uta_v1mhdr_t* v1hdr; int tr_old_len; // tr size in new buffer - int coffset; // an offset to something in the header for copy nm = (rmr_mbuf_t *) malloc( sizeof *nm ); if( nm == NULL ) { @@ -349,22 +349,19 @@ static inline rmr_mbuf_t* realloc_msg( rmr_mbuf_t* old_msg, int tr_len ) { nm->payload = (void *) v1hdr + sizeof( *v1hdr ); break; - default: // current message always caught here + default: // current message version always caught here hdr = nm->header; - memcpy( hdr, old_msg->header, sizeof( uta_mhdr_t ) ); // ONLY copy the header portion; trace and data might have changed - if( RMR_D1_LEN( hdr ) ) { - coffset = DATA1_OFFSET( hdr ); // offset to d1 - memcpy( hdr + coffset, old_msg->header + coffset, RMR_D1_LEN( hdr ) ); // copy data1 and data2 if necessary + memcpy( hdr, old_msg->header, sizeof( uta_mhdr_t ) ); // ONLY copy the header portion; trace and data offsets might have changed + SET_HDR_TR_LEN( hdr, tr_len ); // must adjust trace len in new message before copy + if( RMR_D1_LEN( hdr ) ) { + memcpy( DATA1_ADDR( hdr ), DATA1_ADDR( old_msg->header), RMR_D1_LEN( hdr ) ); // copy data1 and data2 if necessary } if( RMR_D2_LEN( hdr ) ) { - coffset = DATA2_OFFSET( hdr ); // offset to d2 - memcpy( hdr + coffset, old_msg->header + coffset, RMR_D2_LEN( hdr ) ); // copy data2 and data2 if necessary + memcpy( DATA2_ADDR( hdr ), DATA2_ADDR( old_msg->header), RMR_D2_LEN( hdr ) ); } - SET_HDR_TR_LEN( hdr, tr_len ); // MUST set before pointing payload nm->payload = PAYLOAD_ADDR( hdr ); // directly at the payload - SET_HDR_TR_LEN( hdr, tr_len ); // do NOT copy old trace data, just set the new header break; } @@ -405,6 +402,9 @@ static inline rmr_mbuf_t* realloc_msg( rmr_mbuf_t* old_msg, int tr_len ) { reuse. They have their reasons I guess. Thus, we will free the old transport buffer if user passes the message in; at least our mbuf will be reused. + + When msg->state is not ok, this function must set tp_state in the message as some API + fucntions return the message directly and do not propigate errno into the message. */ static rmr_mbuf_t* rcv_msg( uta_ctx_t* ctx, rmr_mbuf_t* old_msg ) { int state; @@ -427,14 +427,18 @@ static rmr_mbuf_t* rcv_msg( uta_ctx_t* ctx, rmr_mbuf_t* old_msg ) { msg->len = 0; msg->payload = NULL; msg->xaction = NULL; + msg->tp_buf = NULL; msg->state = nng_recvmsg( ctx->nn_sock, (nng_msg **) &msg->tp_buf, NO_FLAGS ); // blocks hard until received if( (msg->state = xlate_nng_state( msg->state, RMR_ERR_RCVFAILED )) != RMR_OK ) { + msg->tp_state = errno; return msg; } + msg->tp_state = 0; if( msg->tp_buf == NULL ) { // if state is good this _should_ not be nil, but parninoia says check anyway msg->state = RMR_ERR_EMPTY; + msg->tp_state = 0; return msg; } @@ -448,6 +452,7 @@ static rmr_mbuf_t* rcv_msg( uta_ctx_t* ctx, rmr_mbuf_t* old_msg ) { msg->mtype, msg->state, msg->len, msg->payload - (unsigned char *) msg->header ); } else { msg->state = RMR_ERR_EMPTY; + msg->tp_state = 0; msg->len = 0; msg->alloc_len = rsize; msg->payload = NULL; @@ -511,12 +516,15 @@ static void* rcv_payload( uta_ctx_t* ctx, rmr_mbuf_t* old_msg ) { Called by rmr_send_msg() and rmr_rts_msg(), etc. and thus we assume that all pointer validation has been done prior. + + When msg->state is not ok, this function must set tp_state in the message as some API + fucntions return the message directly and do not propigate errno into the message. */ static rmr_mbuf_t* send_msg( uta_ctx_t* ctx, rmr_mbuf_t* msg, nng_socket nn_sock, int retries ) { int state; uta_mhdr_t* hdr; int nng_flags = NNG_FLAG_NONBLOCK; // if we need to set any nng flags (zc buffer) add it to this - int spin_retries = 1000; // if eagain/timeout we'll spin this many times before giving up the CPU + int spin_retries = 1000; // if eagain/timeout we'll spin, at max, this many times before giving up the CPU int tr_len; // trace len in sending message so we alloc new message with same trace size // future: ensure that application did not overrun the XID buffer; last byte must be 0 @@ -528,7 +536,8 @@ static rmr_mbuf_t* send_msg( uta_ctx_t* ctx, rmr_mbuf_t* msg, nng_socket nn_sock tr_len = RMR_TR_LEN( hdr ); // snarf trace len before sending as hdr is invalid after send if( msg->flags & MFL_ADDSRC ) { // buffer was allocated as a receive buffer; must add our source - strncpy( (char *) ((uta_mhdr_t *)msg->header)->src, ctx->my_name, RMR_MAX_SID ); // must overlay the source to be ours + strncpy( (char *) ((uta_mhdr_t *)msg->header)->src, ctx->my_name, RMR_MAX_SRC ); // must overlay the source to be ours + strncpy( (char *) ((uta_mhdr_t *)msg->header)->srcip, ctx->my_ip, RMR_MAX_SRC ); } errno = 0; @@ -540,7 +549,9 @@ static rmr_mbuf_t* send_msg( uta_ctx_t* ctx, rmr_mbuf_t* msg, nng_socket nn_sock if( retries > 0 && (state == NNG_EAGAIN || state == NNG_ETIMEDOUT) ) { if( --spin_retries <= 0 ) { // don't give up the processor if we don't have to retries--; - usleep( 1 ); // sigh, give up the cpu and hope it's just 1 miscrosec + if( retries > 0 ) { // only if we'll loop through again + usleep( 1 ); // sigh, give up the cpu and hope it's just 1 miscrosec + } spin_retries = 1000; } } else { @@ -559,6 +570,7 @@ static rmr_mbuf_t* send_msg( uta_ctx_t* ctx, rmr_mbuf_t* msg, nng_socket nn_sock // future: this should not happen as all buffers we deal with are zc buffers; might make sense to remove the test and else msg->state = RMR_ERR_SENDFAILED; errno = ENOTSUP; + msg->tp_state = errno; return msg; /* NOT SUPPORTED @@ -590,6 +602,142 @@ static rmr_mbuf_t* send_msg( uta_ctx_t* ctx, rmr_mbuf_t* msg, nng_socket nn_sock return msg; } +/* + send message with maximum timeout. + Accept a message and send it to an endpoint based on message type. + If NNG reports that the send attempt timed out, or should be retried, + RMr will retry for approximately max_to microseconds; rounded to the next + higher value of 10. + + Allocates a new message buffer for the next send. If a message type has + more than one group of endpoints defined, then the message will be sent + in round robin fashion to one endpoint in each group. + + An endpoint will be looked up in the route table using the message type and + the subscription id. If the subscription id is "UNSET_SUBID", then only the + message type is used. If the initial lookup, with a subid, fails, then a + second lookup using just the mtype is tried. + + When msg->state is not OK, this function must set tp_state in the message as + some API fucntions return the message directly and do not propigate errno into + the message. + + CAUTION: this is a non-blocking send. If the message cannot be sent, then + it will return with an error and errno set to eagain. If the send is + a limited fanout, then the returned status is the status of the last + send attempt. + +*/ +static rmr_mbuf_t* mtosend_msg( void* vctx, rmr_mbuf_t* msg, int max_to ) { + rtable_ent_t* rte; // the route table entry which matches the message key + nng_socket nn_sock; // endpoint socket for send + uta_ctx_t* ctx; + int group; // selected group to get socket for + int send_again; // true if the message must be sent again + rmr_mbuf_t* clone_m; // cloned message for an nth send + int sock_ok; // got a valid socket from round robin select + char* d1; + int ok_sends = 0; // track number of ok sends + + if( (ctx = (uta_ctx_t *) vctx) == NULL || msg == NULL ) { // bad stuff, bail fast + errno = EINVAL; // if msg is null, this is their clue + if( msg != NULL ) { + msg->state = RMR_ERR_BADARG; + errno = EINVAL; // must ensure it's not eagain + msg->tp_state = errno; + } + return msg; + } + + errno = 0; // clear; nano might set, but ensure it's not left over if it doesn't + if( msg->header == NULL ) { + fprintf( stderr, "rmr_send_msg: ERROR: message had no header\n" ); + msg->state = RMR_ERR_NOHDR; + errno = EBADMSG; // must ensure it's not eagain + msg->tp_state = errno; + return msg; + } + + if( max_to < 0 ) { + max_to = ctx->send_retries; // convert to retries + } + + if( (rte = uta_get_rte( ctx->rtable, msg->sub_id, msg->mtype, TRUE )) == NULL ) { // find the entry which matches subid/type allow fallback to type only key + if( ctx->flags & CTXFL_WARN ) { + fprintf( stderr, "[WARN] no endpoint for mtype=%d sub_id=%d\n", msg->mtype, msg->sub_id ); + } + msg->state = RMR_ERR_NOENDPT; + errno = ENXIO; // must ensure it's not eagain + msg->tp_state = errno; + return msg; // caller can resend (maybe) or free + } + + send_again = 1; // force loop entry + group = 0; // always start with group 0 + while( send_again ) { + sock_ok = uta_epsock_rr( rte, group, &send_again, &nn_sock ); // select endpt from rr group and set again if more groups + + if( DEBUG ) fprintf( stderr, "[DBUG] mtosend_msg: flgs=0x%04x type=%d again=%d group=%d len=%d sock_ok=%d\n", + msg->flags, msg->mtype, send_again, group, msg->len, sock_ok ); + + group++; + + if( sock_ok ) { // with an rte we _should_ always have a socket, but don't bet on it + if( send_again ) { + clone_m = clone_msg( msg ); // must make a copy as once we send this message is not available + if( clone_m == NULL ) { + msg->state = RMR_ERR_SENDFAILED; + errno = ENOMEM; + msg->tp_state = errno; + if( ctx->flags & CTXFL_WARN ) { + fprintf( stderr, "[WARN] unable to clone message for multiple rr-group send\n" ); + } + return msg; + } + + if( DEBUG ) fprintf( stderr, "[DBUG] msg cloned: type=%d len=%d\n", msg->mtype, msg->len ); + msg->flags |= MFL_NOALLOC; // keep send from allocating a new message; we have a clone to use + msg = send_msg( ctx, msg, nn_sock, max_to ); // do the hard work, msg should be nil on success + + if( msg != NULL ) { // returned message indicates send error of some sort + rmr_free_msg( msg ); // must ditchone; pick msg so we don't have to unfiddle flags + msg = clone_m; + } else { + ok_sends++; + msg = clone_m; // clone will be the next to send + } + } else { + msg = send_msg( ctx, msg, nn_sock, max_to ); // send the last, and allocate a new buffer; drops the clone if it was + if( DEBUG ) { + if( msg == NULL ) { + fprintf( stderr, "[DBUG] mtosend_msg: send returned nil message!\n" ); + } + } + } + } else { + if( ctx->flags & CTXFL_WARN ) { + fprintf( stderr, "[WARN] invalid socket for rte, setting no endpoint err: mtype=%d sub_id=%d\n", msg->mtype, msg->sub_id ); + } + msg->state = RMR_ERR_NOENDPT; + errno = ENXIO; + } + } + + if( msg ) { // call functions don't get a buffer back, so a nil check is required + msg->flags &= ~MFL_NOALLOC; // must return with this flag off + if( ok_sends ) { // multiple rr-groups and one was successful; report ok + msg->state = RMR_OK; + } + + if( DEBUG ) fprintf( stderr, "[DBUG] final send stats: ok=%d group=%d state=%d\n\n", ok_sends, group, msg->state ); + + msg->tp_state = errno; + } + + return msg; // last message caries the status of last/only send attempt +} + + /* A generic wrapper to the real send to keep wormhole stuff agnostic. We assume the wormhole function vetted the buffer so we don't have to.