X-Git-Url: https://gerrit.o-ran-sc.org/r/gitweb?a=blobdiff_plain;f=src%2Fftacmp%2Ftranslate_fta.cc;h=b6855ff8aea4bc965bf699f1003ccf88e820662d;hb=eb761d89890df8d74532dd4faad118db18fd3b7d;hp=4a1263fd08a0863cffee8bf188eb8621fd4d71ac;hpb=b39da83c4c942a0fb301471a0bc127ae3c32bce2;p=com%2Fgs-lite.git diff --git a/src/ftacmp/translate_fta.cc b/src/ftacmp/translate_fta.cc index 4a1263f..b6855ff 100644 --- a/src/ftacmp/translate_fta.cc +++ b/src/ftacmp/translate_fta.cc @@ -164,7 +164,7 @@ int main(int argc, char **argv){ set::iterator si; - vector query_names; // for lfta.c registration + vector registration_query_names; // for lfta.c registration map > mach_query_names; // list queries of machine vector snap_lengths; // for lfta.c registration vector interface_names; // for lfta.c registration @@ -555,11 +555,37 @@ int main(int argc, char **argv){ } Schema = fta_parse_result->tables; +// Ensure that all schema_ids, if set, are distinct. +// Obsolete? There is code elsewhere to ensure that schema IDs are +// distinct on a per-interface basis. +/* + set found_ids; + set dup_ids; + for(int t=0;tsize();++t){ + int sch_id = Schema->get_table(t)->get_schema_id(); + if(sch_id> 0){ + if(found_ids.find(sch_id) != found_ids.end()){ + dup_ids.insert(sch_id); + }else{ + found_ids.insert(sch_id); + } + } + if(dup_ids.size()>0){ + fprintf(stderr, "Error, the schema has duplicate schema_ids:"); + for(auto dit=dup_ids.begin();dit!=dup_ids.end(); ++dit) + fprintf(stderr," %d",(*dit)); + fprintf(stderr,"\n"); + exit(1); + } + } +*/ + + // Process schema field inheritance int retval; retval = Schema->unroll_tables(err_str); if(retval){ - fprintf(stderr,"Error processing schema filed inheritance:\n %s\n", err_str.c_str() ); + fprintf(stderr,"Error processing schema field inheritance:\n %s\n", err_str.c_str() ); exit(1); } }else{ @@ -607,24 +633,31 @@ int main(int argc, char **argv){ map lfta_prefilter_val; string lfta_header = -"#include \n\n" +"#include \n" "#include \"rts.h\"\n" "#include \"fta.h\"\n" "#include \"lapp.h\"\n" -"#include \"rts_udaf.h\"\n\n" +"#include \"rts_udaf.h\"\n" +"#include\n" +"#include \n" +"#include \"rdtsc.h\"\n" +"#include \"watchlist.h\"\n\n" + ; // Get any locally defined parsing headers glob_t glob_result; memset(&glob_result, 0, sizeof(glob_result)); - // do the glob operation + // do the glob operation TODO should be from GSROOT int return_value = glob("../../include/lfta/local/*h", GLOB_TILDE, NULL, &glob_result); if(return_value == 0){ + lfta_header += "\n"; for(size_t i = 0; i < glob_result.gl_pathc; ++i) { char *flds[1000]; int nflds = split_string(glob_result.gl_pathv[i],'/',flds,1000); - lfta_header += "#include \"local/"+string(flds[nflds-1])+"\"\n\n"; + lfta_header += "#include \"local/"+string(flds[nflds-1])+"\"\n"; } + lfta_header += "\n"; }else{ fprintf(stderr,"Warning, glob on ../../include/lfta/local/*h failed.\n"); } @@ -1625,11 +1658,11 @@ for(q=0;q lfta_list; - stream_query *rootq; - int qi,qj; + map > schema_of_schemaid; // ensure that schema IDs are unique to an + for(qi=hfta_topsort.size()-1;qi>=0;--qi){ int hfta_id = hfta_topsort[qi]; @@ -1862,21 +1895,158 @@ for(q=0;qquery_name) == 0){ // Grab the lfta for global optimization. vector tvec = split_queries[l]->query_plan[0]->get_input_tbls(); - string liface = tvec[0]->get_interface(); - string lmach = tvec[0]->get_machine(); - if (lmach == "") - lmach = hostname; - interface_names.push_back(liface); - machine_names.push_back(lmach); + string liface = "_local_"; +// string lmach = ""; + string lmach = hostname; + if(tvec.size()>0){ + liface = tvec[0]->get_interface(); // iface queries have been resolved + lmach = tvec[0]->get_machine(); + } // else{ + interface_names.push_back(liface); + machine_names.push_back(lmach); +// } + + vector schemaid_preds; + for(int irv=0;irvget_schema_name(); + string rvar_name = tvec[irv]->get_var_name(); + int schema_ref = tvec[irv]->get_schema_ref(); + if (lmach == "") + lmach = hostname; +// interface_names.push_back(liface); +// machine_names.push_back(lmach); + //printf("Machine is %s\n",lmach.c_str()); +// Check if a schemaId constraint needs to be inserted. + if(schema_ref<0){ // can result from some kinds of splits + schema_ref = Schema->get_table_ref(schema_name); + } + int schema_id = Schema->get_schema_id(schema_ref); // id associated with PROTOCOL + int errnum = 0; + string if_error; + iface_t *iface = ifaces_db-> get_interface(lmach, liface, errnum, if_error); + if(iface==NULL){ + fprintf(stderr,"Error looking up the interface %s on host %s: %s\n",liface.c_str(), lmach.c_str(), if_error.c_str()); + exit(1); + } + + + if(tvec[irv]->get_interface() != "_local_"){ + if(iface->has_multiple_schemas()){ + if(schema_id<0){ // invalid schema_id + fprintf(stderr,"Error, schema %s has no schema_id, but is sourced from multi-schema interface %s (ref'd in query %s)\n", schema_name.c_str(), liface.c_str(), split_queries[0]->query_name.c_str()); + exit(1); + } + vector iface_schemas = iface->get_property("Schemas"); + if(iface_schemas.size()>0 && find(iface_schemas.begin(), iface_schemas.end(), schema_name) == iface_schemas.end()){ + fprintf(stderr,"Error, query %s references schema %s from interface %s, but this schema is not in this interface's Schemas list\n", split_queries[0]->query_name.c_str(), schema_name.c_str(), liface.c_str()); + exit(1); + } +// Ensure that in liface, schema_id is used for only one schema + if(schema_of_schemaid.count(liface)==0){ + map empty_map; + schema_of_schemaid[liface] = empty_map; + } + if(schema_of_schemaid[liface].count(schema_id)==0){ + schema_of_schemaid[liface][schema_id] = schema_name; + }else{ + if(schema_of_schemaid[liface][schema_id] != schema_name){ + fprintf(stderr, "ERROR, For interface %s, schema id %d is used in schemas %s and %s\n", liface.c_str(), schema_id, schema_name.c_str(), schema_of_schemaid[liface][schema_id].c_str()); + exit(1); + } + } + }else{ // single-schema interface + schema_id = -1; // don't generate schema_id predicate + vector iface_schemas = iface->get_property("Schemas"); + if(iface_schemas.size()>0 && find(iface_schemas.begin(), iface_schemas.end(), schema_name) == iface_schemas.end()){ + fprintf(stderr,"Error, query %s references schema %s from interface %s, but this schema is not in this interface's Schemas list\n", split_queries[0]->query_name.c_str(), schema_name.c_str(), liface.c_str()); + exit(1); + } + if(iface_schemas.size()>1){ + fprintf(stderr, "ERROR, interface %s is a single-schema interface, but has %d elements in its Schemas list.\n", liface.c_str(), (int)iface_schemas.size()); + exit(1); + } + } + }else{ + schema_id = -1; + } + +// If we need to check the schema_id, insert a predicate into the lfta. +// TODO not just schema_id, the full all_schema_ids set. + if(schema_id>=0){ + colref_t *schid_cr = new colref_t("schemaId"); + schid_cr->schema_ref = schema_ref; + schid_cr->table_name = rvar_name; + schid_cr->tablevar_ref = 0; + schid_cr->default_table = false; + scalarexp_t *schid_se = new scalarexp_t(schid_cr); + data_type *schid_dt = new data_type("uint"); + schid_se->dt = schid_dt; + + string schid_str = int_to_string(schema_id); + literal_t *schid_lit = new literal_t(schid_str.c_str(), LITERAL_INT); + scalarexp_t *lit_se = new scalarexp_t(schid_lit); + lit_se->dt = schid_dt; + + predicate_t *schid_pr = new predicate_t(schid_se, "=", lit_se); + vector clist; + make_cnf_from_pr(schid_pr, clist); + analyze_cnf(clist[0]); + clist[0]->cost = 1; // cheap one comparison +// cnf built, now insert it. + split_queries[l]->query_plan[0]->append_to_where(clist[0]); + +// Specialized processing +// filter join, get two schemaid preds + string node_type = split_queries[l]->query_plan[0]->node_type(); + if(node_type == "filter_join"){ + filter_join_qpn *fj = (filter_join_qpn *)split_queries[l]->query_plan[0]; + if(irv==0){ + fj->pred_t0.push_back(clist[0]); + }else{ + fj->pred_t1.push_back(clist[0]); + } + schemaid_preds.push_back(schid_pr); + } +// watchlist join, get the first schemaid pred + if(node_type == "watch_join"){ + watch_join_qpn *fj = (watch_join_qpn *)split_queries[l]->query_plan[0]; + if(irv==0){ + fj->pred_t0.push_back(clist[0]); + schemaid_preds.push_back(schid_pr); + } + } + } + } +// Specialized processing, currently filter join. + if(schemaid_preds.size()>1){ + string node_type = split_queries[l]->query_plan[0]->node_type(); + if(node_type == "filter_join"){ + filter_join_qpn *fj = (filter_join_qpn *)split_queries[l]->query_plan[0]; + predicate_t *filter_pr = new predicate_t("OR", schemaid_preds[0], schemaid_preds[1]); + vector clist; + make_cnf_from_pr(filter_pr, clist); + analyze_cnf(clist[0]); + clist[0]->cost = 1; // cheap one comparison + fj->shared_pred.push_back(clist[0]); + } + } + + + + + + + // Set the ht size from the recommendation, if there is one in the rec file if(lfta_htsize.count(split_queries[l]->query_name)>0){ split_queries[l]->query_plan[0]->set_definition("aggregate_slots",int_to_string(lfta_htsize[split_queries[l]->query_name])); @@ -1891,7 +2061,9 @@ for(q=0;qquery_plan[0]->bind_to_schema(Schema); + // split_queries[l]->query_plan[0]->definitions = split_queries[l]->defines; /* @@ -1915,8 +2087,12 @@ for(q=0;qquery_plan[0], Schema)); - query_names.push_back(split_queries[l]->query_name); - mach_query_names[lmach].push_back(query_names.size()-1); + +// TODO NOTE : I'd like it to be the case that registration_query_names +// are the queries to be registered for subsciption. +// but there is some complex bookkeeping here. + registration_query_names.push_back(split_queries[l]->query_name); + mach_query_names[lmach].push_back(registration_query_names.size()-1); // NOTE: I will assume a 1-1 correspondance between // mach_query_names[lmach] and lfta_mach_lists[lmach] // where mach_query_names[lmach][i] contains the index into @@ -2050,7 +2226,7 @@ if (partitioned_mode) { } -print_hftas("preopt_hfta_info.txt", hfta_list, lfta_names, query_names, interface_names); +print_hftas("preopt_hfta_info.txt", hfta_list, lfta_names, registration_query_names, interface_names); int num_hfta = hfta_list.size(); for(i=0; i < hfta_list.size(); ++i){ @@ -2063,7 +2239,7 @@ for(i=num_hfta; i < hfta_list.size(); ++i){ Schema->append_table(td); } -print_hftas("postopt_hfta_info.txt", hfta_list, lfta_names, query_names, interface_names); +print_hftas("postopt_hfta_info.txt", hfta_list, lfta_names, registration_query_names, interface_names); @@ -2121,17 +2297,26 @@ for(i=0;idefines["namespace"]; if(hfta_list[i]->defines.count("Namespace")>0) hfta_namespace = hfta_list[i]->defines["Namespace"]; - if(hfta_list[i]->defines.count("Namespace")>0) - hfta_namespace = hfta_list[i]->defines["Namespace"]; + if(hfta_list[i]->defines.count("NAMESPACE")>0) + hfta_namespace = hfta_list[i]->defines["NAMESPACE"]; if(field_verifier != NULL){ string warning_str; if(hfta_comment == "") warning_str += "\tcomment not found.\n"; - if(hfta_title == "") - warning_str += "\ttitle not found.\n"; - if(hfta_namespace == "") - warning_str += "\tnamespace not found.\n"; + +// Obsolete stuff that Carsten wanted +// if(hfta_title == "") +// warning_str += "\ttitle not found.\n"; +// if(hfta_namespace == "") +// warning_str += "\tnamespace not found.\n"; + +// STOPPED HERE +// There is a get_tbl_keys method implemented for qp_nodes, +// integrate it into steam_query, then call it to find keys, +// and annotate feidls with their key-ness. +// If there is a "keys" proprty in the defines block, override anything returned +// from the automated analysis vector flds = hfta_list[i]->get_output_tabledef()->get_fields(); int fi; @@ -2143,6 +2328,62 @@ for(i=0;iget_output_tabledef()->get_tbl_name().c_str(),warning_str.c_str()); } +// Get the fields in this query + vector flds = hfta_list[i]->get_output_tabledef()->get_fields(); + +// do key processing + string hfta_keys_s = ""; + if(hfta_list[i]->defines.count("keys")>0) + hfta_keys_s = hfta_list[i]->defines["keys"]; + if(hfta_list[i]->defines.count("Keys")>0) + hfta_keys_s = hfta_list[i]->defines["Keys"]; + if(hfta_list[i]->defines.count("KEYS")>0) + hfta_keys_s = hfta_list[i]->defines["KEYS"]; + string xtra_keys_s = ""; + if(hfta_list[i]->defines.count("extra_keys")>0) + xtra_keys_s = hfta_list[i]->defines["extra_keys"]; + if(hfta_list[i]->defines.count("Extra_Keys")>0) + xtra_keys_s = hfta_list[i]->defines["Extra_Keys"]; + if(hfta_list[i]->defines.count("EXTRA_KEYS")>0) + xtra_keys_s = hfta_list[i]->defines["EXTRA_KEYS"]; +// get the keys + vector hfta_keys; + vector partial_keys; + vector xtra_keys; + if(hfta_keys_s==""){ + hfta_keys = hfta_list[i]->get_tbl_keys(partial_keys); + if(xtra_keys_s.size()>0){ + xtra_keys = split_string(xtra_keys_s, ','); + } + for(int xi=0;xi missing_keys; + for(int ki=0;kiget_name()) + break; + } + if(fi==flds.size()) + missing_keys.push_back(hfta_keys[ki]); + } + if(missing_keys.size()>0){ + fprintf(stderr, "Error, HFTA stream %s has keys defined which aren't in the output:", hfta_list[i]->get_output_tabledef()->get_tbl_name().c_str()); + for(int hi=0; hi\n",hfta_list[i]->get_output_tabledef()->get_tbl_name().c_str()); if(hfta_comment != "") fprintf(qtree_output,"\t\t\n",hfta_comment.c_str()); @@ -2154,15 +2395,25 @@ for(i=0;i\n"); // write info about fields to qtree.xml - vector flds = hfta_list[i]->get_output_tabledef()->get_fields(); int fi; for(fi=0;figet_name().c_str(), fi, flds[fi]->get_type().c_str()); if(flds[fi]->get_modifier_list()->size()){ - fprintf(qtree_output,"mods='%s' ", flds[fi]->get_modifier_list()->to_string().c_str()); + fprintf(qtree_output,"mods='%s'", flds[fi]->get_modifier_list()->to_string().c_str()); } fprintf(qtree_output," />\n"); } +// info about keys + for(int hi=0;hi\n",hfta_keys[hi].c_str()); + } + for(int hi=0;hi\n",partial_keys[hi].c_str()); + } + for(int hi=0;hi\n",xtra_keys[hi].c_str()); + } + // extract liveness timeout from query definition int liveness_timeout = atoi(hfta_list[i]->query_plan[hfta_list[i]->qhead]->get_val_of_def("liveness_timeout").c_str()); @@ -2247,7 +2498,10 @@ for(ssi_el=extra_external_libs.begin();ssi_el!=extra_external_libs.end();++ssi_e int li; for(li=0;li tvec = mach_lftas[li]->query_plan[0]->get_input_tbls(); - string lfta_iface = tvec[0]->get_interface(); + string lfta_iface = "_local_"; + if(tvec.size()>0){ + string lfta_iface = tvec[0]->get_interface(); + } lfta_iface_lists[lfta_iface].push_back(mach_lftas[li]); } @@ -2308,7 +2562,10 @@ for(ssi_el=extra_external_libs.begin();ssi_el!=extra_external_libs.end();++ssi_e // for fta_init for(li=0;li tvec = mach_lftas[li]->query_plan[0]->get_input_tbls(); - string lfta_iface = tvec[0]->get_interface(); + string lfta_iface = "_local_"; + if(tvec.size()>0){ + lfta_iface = tvec[0]->get_interface(); + } lfta_iface_lists[lfta_iface].push_back(mach_lftas[li]); lfta_iface_qname_ix[lfta_iface].push_back(mach_query_names[lmach][li]); } @@ -2319,11 +2576,14 @@ for(ssi_el=extra_external_libs.begin();ssi_el!=extra_external_libs.end();++ssi_e // But this is defunct code for gs-lite for(li=0;li tvec = mach_lftas[li]->query_plan[0]->get_input_tbls(); - string liface = tvec[0]->get_interface(); + string liface = "_local_"; + if(tvec.size()>0){ + liface = tvec[0]->get_interface(); + } vector iface_codegen_type = ifaces_db->get_iface_vals(lmach,liface,"iface_codegen_type",erri,err_str); if(iface_codegen_type.size()){ if(npdb->get_nic_prop_val(iface_codegen_type[0],"Return",erri) == "Packed"){ - packed_return = true; + packed_return = true; } } } @@ -2454,7 +2714,10 @@ for(ssi_el=extra_external_libs.begin();ssi_el!=extra_external_libs.end();++ssi_e lfta_val[lmach] += "\t} else if (!strcmp(iface_name, \"" + *sir + "\")) {\n"; // iterate through interface properties - vector iface_properties = ifaces_db->get_iface_properties(lmach,*sir,erri,err_str); + vector iface_properties; + if(*sir!="_local_"){ // dummy watchlist interface, don't process. + iface_properties = ifaces_db->get_iface_properties(lmach,*sir,erri,err_str); + } if (erri) { fprintf(stderr,"ERROR cannot retrieve interface properties for %s.%s, %s\n",lmach.c_str(), sir->c_str(), err_str.c_str()); exit(1); @@ -2470,8 +2733,9 @@ for(ssi_el=extra_external_libs.begin();ssi_el!=extra_external_libs.end();++ssi_e // combine all values for the interface property using comma separator vector vals = ifaces_db->get_iface_vals(lmach,*sir,iface_properties[i],erri,err_str); + lfta_val[lmach] += "\t\t\treturn \""; for (int j = 0; j < vals.size(); ++j) { - lfta_val[lmach] += "\t\t\treturn \"" + vals[j]; + lfta_val[lmach] += vals[j]; if (j != vals.size()-1) lfta_val[lmach] += ","; } @@ -2490,11 +2754,25 @@ for(ssi_el=extra_external_libs.begin();ssi_el!=extra_external_libs.end();++ssi_e lfta_val[lmach] += "// list of FTAs clearinghouse expects to register themselves\n"; lfta_val[lmach] += "const gs_sp_t fta_names[] = {"; - for (i = 0; i < query_names.size(); ++i) { - if (i) - lfta_val[lmach] += ", "; - lfta_val[lmach] += "\"" + query_names[i] + "\""; + bool first = true; + for(auto mvsi=lfta_iface_lists.begin(); mvsi!=lfta_iface_lists.end(); ++mvsi){ + string liface = (*mvsi).first; + if(liface != "_local_"){ // these don't register themselves + vector lfta_list = (*mvsi).second; + for(i=0;iquery_name + "\""; } @@ -2506,6 +2784,7 @@ for(ssi_el=extra_external_libs.begin();ssi_el!=extra_external_libs.end();++ssi_e // set the prefilter function accordingly. // see the example in demo/err2 lfta_val[lmach] += "void fta_init(gs_sp_t device){\n"; + lfta_val[lmach] += "// note: the last parameter in fta_register is the prefilter signature\n"; // for(i=0;i flds = lfta_sq->get_output_tabledef()->get_fields(); int fi; @@ -2611,26 +2898,30 @@ for(ssi_el=extra_external_libs.begin();ssi_el!=extra_external_libs.end();++ssi_e } if(warning_str != "") fprintf(stderr,"Warning, in LFTA stream %s:\n%s", - query_names[mi].c_str(),warning_str.c_str()); + registration_query_names[mi].c_str(),warning_str.c_str()); } // Create qtree output - fprintf(qtree_output,"\t\n",query_names[mi].c_str()); - if(lfta_comment != "") - fprintf(qtree_output,"\t\t\n",lfta_comment.c_str()); - if(lfta_title != "") - fprintf(qtree_output,"\t\t\n",lfta_title.c_str()); - if(lfta_namespace != "") - fprintf(qtree_output,"\t\t<Namespace value='%s' />\n",lfta_namespace.c_str()); - if(lfta_ht_size != "") - fprintf(qtree_output,"\t\t<HtSize value='%s' />\n",lfta_ht_size.c_str()); + fprintf(qtree_output,"\t<LFTA name='%s' >\n",registration_query_names[mi].c_str()); + if(lfta_comment != "") + fprintf(qtree_output,"\t\t<Description value='%s' />\n",lfta_comment.c_str()); + if(lfta_title != "") + fprintf(qtree_output,"\t\t<Title value='%s' />\n",lfta_title.c_str()); + if(lfta_namespace != "") + fprintf(qtree_output,"\t\t<Namespace value='%s' />\n",lfta_namespace.c_str()); + if(lfta_ht_size != "") + fprintf(qtree_output,"\t\t<HtSize value='%s' />\n",lfta_ht_size.c_str()); if(lmach != "") fprintf(qtree_output,"\t\t<Host value='%s' />\n",lmach.c_str()); else fprintf(qtree_output,"\t\t<Host value='%s' />\n",hostname.c_str()); fprintf(qtree_output,"\t\t<Interface value='%s' />\n",interface_names[mi].c_str()); - fprintf(qtree_output,"\t\t<ReadsFrom value='%s' />\n",interface_names[mi].c_str()); + std::vector<tablevar_t *> itbls = lfta_sq->get_input_tables(); + for(int t=0;t<itbls.size();++t){ + fprintf(qtree_output,"\t\t<ReadsFrom value='%s' />\n",itbls[t]->get_schema_name().c_str()); + } +// fprintf(qtree_output,"\t\t<ReadsFrom value='%s' />\n",interface_names[mi].c_str()); fprintf(qtree_output,"\t\t<Rate value='100' />\n"); fprintf(qtree_output,"\t\t<LivenessTimeout value='%d' />\n", lfta_liveness_timeouts[mi]); // write info about fields to qtree.xml @@ -2774,7 +3065,12 @@ void generate_makefile(vector<string> &input_file_names, int nfiles, } // Do we need to include protobuf libraries? +// TODO Move to the interface library: get the libraries to include +// for an interface type + bool use_proto = false; + bool use_bsa = false; + bool use_kafka = false; int erri; string err_str; for(ssi=ifaces.begin();ssi!=ifaces.end();++ssi){ @@ -2785,6 +3081,18 @@ void generate_makefile(vector<string> &input_file_names, int nfiles, use_proto = true; } } + ift = ifdb->get_iface_vals(ifmachines[ifnm],ifnm, "BSA", erri, err_str); + for(int ift_i=0;ift_i<ift.size();ift_i++){ + if(ift[ift_i] == "TRUE" || ift[ift_i] == "True" || ift[ift_i] == "true"){ + use_bsa = true; + } + } + ift = ifdb->get_iface_vals(ifmachines[ifnm],ifnm, "KAFKA", erri, err_str); + for(int ift_i=0;ift_i<ift.size();ift_i++){ + if(ift[ift_i] == "TRUE" || ift[ift_i] == "True" || ift[ift_i] == "true"){ + use_kafka = true; + } + } } fprintf(outfl, @@ -2814,6 +3122,10 @@ void generate_makefile(vector<string> &input_file_names, int nfiles, fprintf(outfl, " -ldll -ldl "); if(use_proto) fprintf(outfl, " -L/usr/local/lib/ -lprotobuf-c "); + if(use_bsa) + fprintf(outfl, " -lbsa_stream "); + if(use_kafka) + fprintf(outfl, " -lrdkafka "); fprintf(outfl," -lgscpaux"); #ifdef GCOV fprintf(outfl," -fprofile-arcs"); @@ -2896,6 +3208,9 @@ void generate_makefile(vector<string> &input_file_names, int nfiles, // string err_str; for(ssi=ifaces.begin();ssi!=ifaces.end();++ssi){ string ifnm = (*ssi); + // suppress internal _local_ interface + if (ifnm == "_local_") + continue; fprintf(outfl, "%s ",ifnm.c_str()); vector<string> ifv = ifdb->get_iface_vals(ifmachines[ifnm],ifnm, "Command", erri, err_str); for(j=0;j<ifv.size();++j)