Code Review
/
ric-plt
/
lib
/
rmr.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
review
|
tree
raw
|
inline
| side by side
Beef up unit tests for SI95 code
[ric-plt/lib/rmr.git]
/
test
/
rmr_nng_test.c
diff --git
a/test/rmr_nng_test.c
b/test/rmr_nng_test.c
index
a3aeb40
..
bef4a24
100644
(file)
--- a/
test/rmr_nng_test.c
+++ b/
test/rmr_nng_test.c
@@
-62,6
+62,7
@@
#include <nng/protocol/pipeline0/push.h>
#include <nng/protocol/pipeline0/pull.h>
#include <nng/protocol/pipeline0/push.h>
#include <nng/protocol/pipeline0/pull.h>
+#define NNG_UNDER_TEST 1
#define EMULATE_NNG
#include "test_nng_em.c" // nng/nn emulation (before including things under test)
#define EMULATE_NNG
#include "test_nng_em.c" // nng/nn emulation (before including things under test)
@@
-70,7
+71,6
@@
#include "rmr_symtab.h"
#include "rmr_logging.h"
#include "rmr_agnostic.h" // transport agnostic header
#include "rmr_symtab.h"
#include "rmr_logging.h"
#include "rmr_agnostic.h" // transport agnostic header
-#include "rmr_nng_private.h" // transport specific
#include "symtab.c"
#include "logging.c"
#include "symtab.c"
#include "logging.c"
@@
-81,6
+81,7
@@
static void gen_rt( uta_ctx_t* ctx ); // defined in sr_nng_static_test, but use
// specific test tools in this directory
#include "test_support.c" // things like fail_if()
// specific test tools in this directory
#include "test_support.c" // things like fail_if()
+#include "test_ctx_support.c" // specifically geared to creating dummy contex structs
#include "test_gen_rt.c"
// and finally....
#include "tools_static_test.c" // local test functions pulled directly because of static nature of things
#include "test_gen_rt.c"
// and finally....
#include "tools_static_test.c" // local test functions pulled directly because of static nature of things
@@
-135,6
+136,7
@@
int main() {
errors += mbuf_api_test( );
fprintf( stderr, "<INFO> error count: %d\n", errors );
errors += mbuf_api_test( );
fprintf( stderr, "<INFO> error count: %d\n", errors );
+ test_summary( errors, "tool tests" );
if( errors == 0 ) {
fprintf( stderr, "<PASS> all tests were OK\n\n" );
} else {
if( errors == 0 ) {
fprintf( stderr, "<PASS> all tests were OK\n\n" );
} else {