X-Git-Url: https://gerrit.o-ran-sc.org/r/gitweb?a=blobdiff_plain;f=src%2Fftacmp%2Ftranslate_fta.cc;h=19ff6349ee040838f2820da8ad8b33dffd7e732e;hb=dec9c93423775db0f4783a93145f016d5d780ffd;hp=4a1263fd08a0863cffee8bf188eb8621fd4d71ac;hpb=e981e864b812c938d3df8b555b6bb98bb89273e7;p=com%2Fgs-lite.git diff --git a/src/ftacmp/translate_fta.cc b/src/ftacmp/translate_fta.cc index 4a1263f..19ff634 100644 --- a/src/ftacmp/translate_fta.cc +++ b/src/ftacmp/translate_fta.cc @@ -55,6 +55,8 @@ Copyright 2014 AT&T Intellectual Property #include"xml_t.h" #include"field_list.h" +#include "gsconfig.h" + extern int xmlParserparse(void); extern FILE *xmlParserin; extern int xmlParserdebug; @@ -164,9 +166,10 @@ int main(int argc, char **argv){ set::iterator si; - vector query_names; // for lfta.c registration + vector registration_query_names; // for lfta.c registration map > mach_query_names; // list queries of machine vector snap_lengths; // for lfta.c registration + vector snap_position; // for lfta.c registration vector interface_names; // for lfta.c registration vector machine_names; // machine of interface vector lfta_reuse_options; // for lfta.c registration @@ -395,24 +398,26 @@ int main(int argc, char **argv){ while(fgets(tmpstr,TMPSTRLEN,osp_in)){ o_lineno++; int nflds = split_string(tmpstr,',',flds,MAXFLDS); - if(nflds == 7){ + if(tmpstr[0]!='\n' && tmpstr[0]!='\r' && tmpstr[0]!='\0' && tmpstr[0]!='#'){ + if(nflds == 7){ // make operator type lowercase - char *tmpc; - for(tmpc=flds[1];*tmpc!='\0';++tmpc) - *tmpc = tolower(*tmpc); - - ospec_str *tmp_ospec = new ospec_str(); - tmp_ospec->query = flds[0]; - tmp_ospec->operator_type = flds[1]; - tmp_ospec->operator_param = flds[2]; - tmp_ospec->output_directory = flds[3]; - tmp_ospec->bucketwidth = atoi(flds[4]); - tmp_ospec->partitioning_flds = flds[5]; - tmp_ospec->n_partitions = atoi(flds[6]); - qname_to_ospec.insert(pair(tmp_ospec->query,output_specs.size())); - output_specs.push_back(tmp_ospec); - }else{ - fprintf(stderr,"Warning, line %d corrupted in output_spec.cfg, has %d fields.\n",o_lineno,nflds); + char *tmpc; + for(tmpc=flds[1];*tmpc!='\0';++tmpc) + *tmpc = tolower(*tmpc); + + ospec_str *tmp_ospec = new ospec_str(); + tmp_ospec->query = flds[0]; + tmp_ospec->operator_type = flds[1]; + tmp_ospec->operator_param = flds[2]; + tmp_ospec->output_directory = flds[3]; + tmp_ospec->bucketwidth = atoi(flds[4]); + tmp_ospec->partitioning_flds = flds[5]; + tmp_ospec->n_partitions = atoi(flds[6]); + qname_to_ospec.insert(pair(tmp_ospec->query,output_specs.size())); + output_specs.push_back(tmp_ospec); + }else{ + fprintf(stderr,"Warning, line %d corrupted in output_spec.cfg, has %d fields.\n",o_lineno,nflds); + } } } fclose(osp_in); @@ -555,11 +560,37 @@ int main(int argc, char **argv){ } Schema = fta_parse_result->tables; +// Ensure that all schema_ids, if set, are distinct. +// Obsolete? There is code elsewhere to ensure that schema IDs are +// distinct on a per-interface basis. +/* + set found_ids; + set dup_ids; + for(int t=0;tsize();++t){ + int sch_id = Schema->get_table(t)->get_schema_id(); + if(sch_id> 0){ + if(found_ids.find(sch_id) != found_ids.end()){ + dup_ids.insert(sch_id); + }else{ + found_ids.insert(sch_id); + } + } + if(dup_ids.size()>0){ + fprintf(stderr, "Error, the schema has duplicate schema_ids:"); + for(auto dit=dup_ids.begin();dit!=dup_ids.end(); ++dit) + fprintf(stderr," %d",(*dit)); + fprintf(stderr,"\n"); + exit(1); + } + } +*/ + + // Process schema field inheritance int retval; retval = Schema->unroll_tables(err_str); if(retval){ - fprintf(stderr,"Error processing schema filed inheritance:\n %s\n", err_str.c_str() ); + fprintf(stderr,"Error processing schema field inheritance:\n %s\n", err_str.c_str() ); exit(1); } }else{ @@ -607,24 +638,31 @@ int main(int argc, char **argv){ map lfta_prefilter_val; string lfta_header = -"#include \n\n" +"#include \n" "#include \"rts.h\"\n" "#include \"fta.h\"\n" "#include \"lapp.h\"\n" -"#include \"rts_udaf.h\"\n\n" +"#include \"rts_udaf.h\"\n" +"#include\n" +"#include \n" +"#include \"rdtsc.h\"\n" +"#include \"watchlist.h\"\n\n" + ; // Get any locally defined parsing headers glob_t glob_result; memset(&glob_result, 0, sizeof(glob_result)); - // do the glob operation + // do the glob operation TODO should be from GSROOT int return_value = glob("../../include/lfta/local/*h", GLOB_TILDE, NULL, &glob_result); if(return_value == 0){ + lfta_header += "\n"; for(size_t i = 0; i < glob_result.gl_pathc; ++i) { char *flds[1000]; int nflds = split_string(glob_result.gl_pathv[i],'/',flds,1000); - lfta_header += "#include \"local/"+string(flds[nflds-1])+"\"\n\n"; + lfta_header += "#include \"local/"+string(flds[nflds-1])+"\"\n"; } + lfta_header += "\n"; }else{ fprintf(stderr,"Warning, glob on ../../include/lfta/local/*h failed.\n"); } @@ -1318,6 +1356,7 @@ fprintf(stderr,"Parsing file %s\n",qpathname.c_str()); } if(dangling_ospecs!=""){ fprintf(stderr,"WARNING, the following entries in output_spec.cfg don't have a matching query: %s\n",dangling_ospecs.c_str()); + exit(1); } string dangling_par = ""; @@ -1625,11 +1664,11 @@ for(q=0;q lfta_list; - stream_query *rootq; - int qi,qj; + map > schema_of_schemaid; // ensure that schema IDs are unique to an + for(qi=hfta_topsort.size()-1;qi>=0;--qi){ int hfta_id = hfta_topsort[qi]; @@ -1862,21 +1901,162 @@ for(q=0;qquery_name) == 0){ // Grab the lfta for global optimization. vector tvec = split_queries[l]->query_plan[0]->get_input_tbls(); - string liface = tvec[0]->get_interface(); - string lmach = tvec[0]->get_machine(); - if (lmach == "") - lmach = hostname; - interface_names.push_back(liface); - machine_names.push_back(lmach); + string liface = "_local_"; +// string lmach = ""; + string lmach = hostname; + if(tvec.size()>0){ + liface = tvec[0]->get_interface(); // iface queries have been resolved + if(tvec[0]->get_machine() != ""){ + lmach = tvec[0]->get_machine(); + }else{ + fprintf(stderr,"WARNING, lfta %s has empty machine name, using %s\n", split_queries[l]->query_plan[0]->node_name.c_str(), hostname.c_str()); + } + } // else{ + interface_names.push_back(liface); + machine_names.push_back(lmach); +// } + + vector schemaid_preds; + for(int irv=0;irvget_schema_name(); + string rvar_name = tvec[irv]->get_var_name(); + int schema_ref = tvec[irv]->get_schema_ref(); + if (lmach == "") + lmach = hostname; +// interface_names.push_back(liface); +// machine_names.push_back(lmach); + //printf("Machine is %s\n",lmach.c_str()); +// Check if a schemaId constraint needs to be inserted. + if(schema_ref<0){ // can result from some kinds of splits + schema_ref = Schema->get_table_ref(schema_name); + } + int schema_id = Schema->get_schema_id(schema_ref); // id associated with PROTOCOL + int errnum = 0; + string if_error; + iface_t *iface = ifaces_db-> get_interface(lmach, liface, errnum, if_error); + if(iface==NULL){ + fprintf(stderr,"Error looking up the interface %s on host %s: %s\n",liface.c_str(), lmach.c_str(), if_error.c_str()); + exit(1); + } + + + if(tvec[irv]->get_interface() != "_local_"){ + if(iface->has_multiple_schemas()){ + if(schema_id<0){ // invalid schema_id + fprintf(stderr,"Error, schema %s has no schema_id, but is sourced from multi-schema interface %s (ref'd in query %s)\n", schema_name.c_str(), liface.c_str(), split_queries[0]->query_name.c_str()); + exit(1); + } + vector iface_schemas = iface->get_property("Schemas"); + if(iface_schemas.size()>0 && find(iface_schemas.begin(), iface_schemas.end(), schema_name) == iface_schemas.end()){ + fprintf(stderr,"Error, query %s references schema %s from interface %s, but this schema is not in this interface's Schemas list\n", split_queries[0]->query_name.c_str(), schema_name.c_str(), liface.c_str()); + exit(1); + } +// Ensure that in liface, schema_id is used for only one schema + if(schema_of_schemaid.count(liface)==0){ + map empty_map; + schema_of_schemaid[liface] = empty_map; + } + if(schema_of_schemaid[liface].count(schema_id)==0){ + schema_of_schemaid[liface][schema_id] = schema_name; + }else{ + if(schema_of_schemaid[liface][schema_id] != schema_name){ + fprintf(stderr, "ERROR, For interface %s, schema id %d is used in schemas %s and %s\n", liface.c_str(), schema_id, schema_name.c_str(), schema_of_schemaid[liface][schema_id].c_str()); + exit(1); + } + } + }else{ // single-schema interface + schema_id = -1; // don't generate schema_id predicate + vector iface_schemas = iface->get_property("Schemas"); + if(iface_schemas.size()>0 && find(iface_schemas.begin(), iface_schemas.end(), schema_name) == iface_schemas.end()){ + fprintf(stderr,"Error, query %s references schema %s from interface %s, but this schema is not in this interface's Schemas list\n", split_queries[0]->query_name.c_str(), schema_name.c_str(), liface.c_str()); + exit(1); + } + if(iface_schemas.size()>1){ + fprintf(stderr, "ERROR, interface %s is a single-schema interface, but has %d elements in its Schemas list.\n", liface.c_str(), (int)iface_schemas.size()); + exit(1); + } + } + }else{ + schema_id = -1; + } + +// If we need to check the schema_id, insert a predicate into the lfta. +// TODO not just schema_id, the full all_schema_ids set. + if(schema_id>=0){ + colref_t *schid_cr = new colref_t("schemaId"); + schid_cr->schema_ref = schema_ref; + schid_cr->table_name = rvar_name; + schid_cr->tablevar_ref = 0; + schid_cr->default_table = false; + scalarexp_t *schid_se = new scalarexp_t(schid_cr); + data_type *schid_dt = new data_type("uint"); + schid_se->dt = schid_dt; + + string schid_str = int_to_string(schema_id); + literal_t *schid_lit = new literal_t(schid_str.c_str(), LITERAL_INT); + scalarexp_t *lit_se = new scalarexp_t(schid_lit); + lit_se->dt = schid_dt; + + predicate_t *schid_pr = new predicate_t(schid_se, "=", lit_se); + vector clist; + make_cnf_from_pr(schid_pr, clist); + analyze_cnf(clist[0]); + clist[0]->cost = 1; // cheap one comparison +// cnf built, now insert it. + split_queries[l]->query_plan[0]->append_to_where(clist[0]); + +// Specialized processing +// filter join, get two schemaid preds + string node_type = split_queries[l]->query_plan[0]->node_type(); + if(node_type == "filter_join"){ + filter_join_qpn *fj = (filter_join_qpn *)split_queries[l]->query_plan[0]; + if(irv==0){ + fj->pred_t0.push_back(clist[0]); + }else{ + fj->pred_t1.push_back(clist[0]); + } + schemaid_preds.push_back(schid_pr); + } +// watchlist join, get the first schemaid pred + if(node_type == "watch_join"){ + watch_join_qpn *fj = (watch_join_qpn *)split_queries[l]->query_plan[0]; + if(irv==0){ + fj->pred_t0.push_back(clist[0]); + schemaid_preds.push_back(schid_pr); + } + } + } + } +// Specialized processing, currently filter join. + if(schemaid_preds.size()>1){ + string node_type = split_queries[l]->query_plan[0]->node_type(); + if(node_type == "filter_join"){ + filter_join_qpn *fj = (filter_join_qpn *)split_queries[l]->query_plan[0]; + predicate_t *filter_pr = new predicate_t("OR", schemaid_preds[0], schemaid_preds[1]); + vector clist; + make_cnf_from_pr(filter_pr, clist); + analyze_cnf(clist[0]); + clist[0]->cost = 1; // cheap one comparison + fj->shared_pred.push_back(clist[0]); + } + } + + + + + + + // Set the ht size from the recommendation, if there is one in the rec file if(lfta_htsize.count(split_queries[l]->query_name)>0){ split_queries[l]->query_plan[0]->set_definition("aggregate_slots",int_to_string(lfta_htsize[split_queries[l]->query_name])); @@ -1891,7 +2071,9 @@ for(q=0;qquery_plan[0]->bind_to_schema(Schema); + // split_queries[l]->query_plan[0]->definitions = split_queries[l]->defines; /* @@ -1914,9 +2096,17 @@ for(q=0;qquery_plan[0], Schema, split_queries[l]->get_gid(), Ext_fcns, lfta_schema_embed, ifaces_db,nicprop); */ - snap_lengths.push_back(compute_snap_len(split_queries[l]->query_plan[0], Schema)); - query_names.push_back(split_queries[l]->query_name); - mach_query_names[lmach].push_back(query_names.size()-1); + snap_lengths.push_back(compute_snap_len(split_queries[l]->query_plan[0], Schema, "snap")); + snap_position.push_back(compute_snap_len(split_queries[l]->query_plan[0], Schema, "index")); + +// STOPPED HERE need to figure out how to generate the code that Vlad needs +// from snap_postion + +// TODO NOTE : I'd like it to be the case that registration_query_names +// are the queries to be registered for subsciption. +// but there is some complex bookkeeping here. + registration_query_names.push_back(split_queries[l]->query_name); + mach_query_names[lmach].push_back(registration_query_names.size()-1); // NOTE: I will assume a 1-1 correspondance between // mach_query_names[lmach] and lfta_mach_lists[lmach] // where mach_query_names[lmach][i] contains the index into @@ -2050,7 +2240,7 @@ if (partitioned_mode) { } -print_hftas("preopt_hfta_info.txt", hfta_list, lfta_names, query_names, interface_names); +print_hftas("preopt_hfta_info.txt", hfta_list, lfta_names, registration_query_names, interface_names); int num_hfta = hfta_list.size(); for(i=0; i < hfta_list.size(); ++i){ @@ -2063,7 +2253,7 @@ for(i=num_hfta; i < hfta_list.size(); ++i){ Schema->append_table(td); } -print_hftas("postopt_hfta_info.txt", hfta_list, lfta_names, query_names, interface_names); +print_hftas("postopt_hfta_info.txt", hfta_list, lfta_names, registration_query_names, interface_names); @@ -2121,17 +2311,25 @@ for(i=0;idefines["namespace"]; if(hfta_list[i]->defines.count("Namespace")>0) hfta_namespace = hfta_list[i]->defines["Namespace"]; - if(hfta_list[i]->defines.count("Namespace")>0) - hfta_namespace = hfta_list[i]->defines["Namespace"]; + if(hfta_list[i]->defines.count("NAMESPACE")>0) + hfta_namespace = hfta_list[i]->defines["NAMESPACE"]; if(field_verifier != NULL){ string warning_str; if(hfta_comment == "") warning_str += "\tcomment not found.\n"; - if(hfta_title == "") - warning_str += "\ttitle not found.\n"; - if(hfta_namespace == "") - warning_str += "\tnamespace not found.\n"; + +// Obsolete stuff that Carsten wanted +// if(hfta_title == "") +// warning_str += "\ttitle not found.\n"; +// if(hfta_namespace == "") +// warning_str += "\tnamespace not found.\n"; + +// There is a get_tbl_keys method implemented for qp_nodes, +// integrate it into steam_query, then call it to find keys, +// and annotate feidls with their key-ness. +// If there is a "keys" proprty in the defines block, override anything returned +// from the automated analysis vector flds = hfta_list[i]->get_output_tabledef()->get_fields(); int fi; @@ -2143,6 +2341,62 @@ for(i=0;iget_output_tabledef()->get_tbl_name().c_str(),warning_str.c_str()); } +// Get the fields in this query + vector flds = hfta_list[i]->get_output_tabledef()->get_fields(); + +// do key processing + string hfta_keys_s = ""; + if(hfta_list[i]->defines.count("keys")>0) + hfta_keys_s = hfta_list[i]->defines["keys"]; + if(hfta_list[i]->defines.count("Keys")>0) + hfta_keys_s = hfta_list[i]->defines["Keys"]; + if(hfta_list[i]->defines.count("KEYS")>0) + hfta_keys_s = hfta_list[i]->defines["KEYS"]; + string xtra_keys_s = ""; + if(hfta_list[i]->defines.count("extra_keys")>0) + xtra_keys_s = hfta_list[i]->defines["extra_keys"]; + if(hfta_list[i]->defines.count("Extra_Keys")>0) + xtra_keys_s = hfta_list[i]->defines["Extra_Keys"]; + if(hfta_list[i]->defines.count("EXTRA_KEYS")>0) + xtra_keys_s = hfta_list[i]->defines["EXTRA_KEYS"]; +// get the keys + vector hfta_keys; + vector partial_keys; + vector xtra_keys; + if(hfta_keys_s==""){ + hfta_keys = hfta_list[i]->get_tbl_keys(partial_keys); + if(xtra_keys_s.size()>0){ + xtra_keys = split_string(xtra_keys_s, ','); + } + for(int xi=0;xi missing_keys; + for(int ki=0;kiget_name()) + break; + } + if(fi==flds.size()) + missing_keys.push_back(hfta_keys[ki]); + } + if(missing_keys.size()>0){ + fprintf(stderr, "Error, HFTA stream %s has keys defined which aren't in the output:", hfta_list[i]->get_output_tabledef()->get_tbl_name().c_str()); + for(int hi=0; hi\n",hfta_list[i]->get_output_tabledef()->get_tbl_name().c_str()); if(hfta_comment != "") fprintf(qtree_output,"\t\t\n",hfta_comment.c_str()); @@ -2154,15 +2408,25 @@ for(i=0;i\n"); // write info about fields to qtree.xml - vector flds = hfta_list[i]->get_output_tabledef()->get_fields(); int fi; for(fi=0;figet_name().c_str(), fi, flds[fi]->get_type().c_str()); if(flds[fi]->get_modifier_list()->size()){ - fprintf(qtree_output,"mods='%s' ", flds[fi]->get_modifier_list()->to_string().c_str()); + fprintf(qtree_output,"mods='%s'", flds[fi]->get_modifier_list()->to_string().c_str()); } fprintf(qtree_output," />\n"); } +// info about keys + for(int hi=0;hi\n",hfta_keys[hi].c_str()); + } + for(int hi=0;hi\n",partial_keys[hi].c_str()); + } + for(int hi=0;hi\n",xtra_keys[hi].c_str()); + } + // extract liveness timeout from query definition int liveness_timeout = atoi(hfta_list[i]->query_plan[hfta_list[i]->qhead]->get_val_of_def("liveness_timeout").c_str()); @@ -2247,7 +2511,10 @@ for(ssi_el=extra_external_libs.begin();ssi_el!=extra_external_libs.end();++ssi_e int li; for(li=0;li tvec = mach_lftas[li]->query_plan[0]->get_input_tbls(); - string lfta_iface = tvec[0]->get_interface(); + string lfta_iface = "_local_"; + if(tvec.size()>0){ + string lfta_iface = tvec[0]->get_interface(); + } lfta_iface_lists[lfta_iface].push_back(mach_lftas[li]); } @@ -2290,7 +2557,10 @@ for(ssi_el=extra_external_libs.begin();ssi_el!=extra_external_libs.end();++ssi_e // Find common filter predicates in the LFTAs. -// in addition generate structs to store the temporal attributes unpacked by prefilter +// in addition generate structs to store the +// temporal attributes unpacked by prefilter +// compute & provide interface for per-interface +// record extraction properties map >::iterator ssqi; for(ssqi=lfta_mach_lists.begin(); ssqi!=lfta_mach_lists.end(); ++ssqi){ @@ -2308,7 +2578,10 @@ for(ssi_el=extra_external_libs.begin();ssi_el!=extra_external_libs.end();++ssi_e // for fta_init for(li=0;li tvec = mach_lftas[li]->query_plan[0]->get_input_tbls(); - string lfta_iface = tvec[0]->get_interface(); + string lfta_iface = "_local_"; + if(tvec.size()>0){ + lfta_iface = tvec[0]->get_interface(); + } lfta_iface_lists[lfta_iface].push_back(mach_lftas[li]); lfta_iface_qname_ix[lfta_iface].push_back(mach_query_names[lmach][li]); } @@ -2319,11 +2592,14 @@ for(ssi_el=extra_external_libs.begin();ssi_el!=extra_external_libs.end();++ssi_e // But this is defunct code for gs-lite for(li=0;li tvec = mach_lftas[li]->query_plan[0]->get_input_tbls(); - string liface = tvec[0]->get_interface(); + string liface = "_local_"; + if(tvec.size()>0){ + liface = tvec[0]->get_interface(); + } vector iface_codegen_type = ifaces_db->get_iface_vals(lmach,liface,"iface_codegen_type",erri,err_str); if(iface_codegen_type.size()){ if(npdb->get_nic_prop_val(iface_codegen_type[0],"Return",erri) == "Packed"){ - packed_return = true; + packed_return = true; } } } @@ -2387,10 +2663,13 @@ for(ssi_el=extra_external_libs.begin();ssi_el!=extra_external_libs.end();++ssi_e lfta_prefilter_val[lmach] += "#define PREFILTER_DEFINED 1;\n\n"; #endif map > lfta_sigs; // used again later + map lfta_snap_pos; // optimize csv parsing + // compute now, use in get_iface_properties for(auto mvsi=lfta_iface_lists.begin(); mvsi!=lfta_iface_lists.end(); ++mvsi){ string liface = (*mvsi).first; vector empty_list; lfta_sigs[liface] = empty_list; + lfta_snap_pos[liface] = -1; vector lfta_cols; vector lfta_snap_length; @@ -2404,7 +2683,10 @@ for(ssi_el=extra_external_libs.begin();ssi_el!=extra_external_libs.end();++ssi_e } lfta_sigs[liface].push_back(mask); lfta_cols.push_back(lfta_iface_lists[liface][li]->query_plan[0]->get_colrefs(true,Schema)); - lfta_snap_length.push_back(compute_snap_len(lfta_iface_lists[liface][li]->query_plan[0], Schema)); + lfta_snap_length.push_back(compute_snap_len(lfta_iface_lists[liface][li]->query_plan[0], Schema, "snap")); + int this_snap_pos = compute_snap_len(lfta_iface_lists[liface][li]->query_plan[0], Schema, "index"); + if(this_snap_pos > lfta_snap_pos[liface]) + lfta_snap_pos[liface] = this_snap_pos; } //for(li=0;li iface_properties = ifaces_db->get_iface_properties(lmach,*sir,erri,err_str); + vector iface_properties; + if(*sir!="_local_"){ // dummy watchlist interface, don't process. + iface_properties = ifaces_db->get_iface_properties(lmach,*sir,erri,err_str); + } if (erri) { fprintf(stderr,"ERROR cannot retrieve interface properties for %s.%s, %s\n",lmach.c_str(), sir->c_str(), err_str.c_str()); exit(1); @@ -2470,13 +2755,16 @@ for(ssi_el=extra_external_libs.begin();ssi_el!=extra_external_libs.end();++ssi_e // combine all values for the interface property using comma separator vector vals = ifaces_db->get_iface_vals(lmach,*sir,iface_properties[i],erri,err_str); + lfta_val[lmach] += "\t\t\treturn \""; for (int j = 0; j < vals.size(); ++j) { - lfta_val[lmach] += "\t\t\treturn \"" + vals[j]; + lfta_val[lmach] += vals[j]; if (j != vals.size()-1) lfta_val[lmach] += ","; } lfta_val[lmach] += "\";\n"; } + lfta_val[lmach] += "\t\t}else if(!strcmp(property_name, \"_max_csv_pos\")){\n"; + lfta_val[lmach] += "\t\t\treturn \""+int_to_string(lfta_snap_pos[(*sir)])+"\";\n"; lfta_val[lmach] += "\t\t} else\n"; lfta_val[lmach] += "\t\t\treturn NULL;\n"; } @@ -2490,11 +2778,25 @@ for(ssi_el=extra_external_libs.begin();ssi_el!=extra_external_libs.end();++ssi_e lfta_val[lmach] += "// list of FTAs clearinghouse expects to register themselves\n"; lfta_val[lmach] += "const gs_sp_t fta_names[] = {"; - for (i = 0; i < query_names.size(); ++i) { - if (i) - lfta_val[lmach] += ", "; - lfta_val[lmach] += "\"" + query_names[i] + "\""; + bool first = true; + for(auto mvsi=lfta_iface_lists.begin(); mvsi!=lfta_iface_lists.end(); ++mvsi){ + string liface = (*mvsi).first; + if(liface != "_local_"){ // these don't register themselves + vector lfta_list = (*mvsi).second; + for(i=0;iquery_name + "\""; } @@ -2506,6 +2808,7 @@ for(ssi_el=extra_external_libs.begin();ssi_el!=extra_external_libs.end();++ssi_e // set the prefilter function accordingly. // see the example in demo/err2 lfta_val[lmach] += "void fta_init(gs_sp_t device){\n"; + lfta_val[lmach] += "// note: the last parameter in fta_register is the prefilter signature\n"; // for(i=0;i flds = lfta_sq->get_output_tabledef()->get_fields(); int fi; @@ -2611,26 +2922,30 @@ for(ssi_el=extra_external_libs.begin();ssi_el!=extra_external_libs.end();++ssi_e } if(warning_str != "") fprintf(stderr,"Warning, in LFTA stream %s:\n%s", - query_names[mi].c_str(),warning_str.c_str()); + registration_query_names[mi].c_str(),warning_str.c_str()); } // Create qtree output - fprintf(qtree_output,"\t\n",query_names[mi].c_str()); - if(lfta_comment != "") - fprintf(qtree_output,"\t\t\n",lfta_comment.c_str()); - if(lfta_title != "") - fprintf(qtree_output,"\t\t\n",lfta_title.c_str()); - if(lfta_namespace != "") - fprintf(qtree_output,"\t\t<Namespace value='%s' />\n",lfta_namespace.c_str()); - if(lfta_ht_size != "") - fprintf(qtree_output,"\t\t<HtSize value='%s' />\n",lfta_ht_size.c_str()); + fprintf(qtree_output,"\t<LFTA name='%s' >\n",registration_query_names[mi].c_str()); + if(lfta_comment != "") + fprintf(qtree_output,"\t\t<Description value='%s' />\n",lfta_comment.c_str()); + if(lfta_title != "") + fprintf(qtree_output,"\t\t<Title value='%s' />\n",lfta_title.c_str()); + if(lfta_namespace != "") + fprintf(qtree_output,"\t\t<Namespace value='%s' />\n",lfta_namespace.c_str()); + if(lfta_ht_size != "") + fprintf(qtree_output,"\t\t<HtSize value='%s' />\n",lfta_ht_size.c_str()); if(lmach != "") fprintf(qtree_output,"\t\t<Host value='%s' />\n",lmach.c_str()); else fprintf(qtree_output,"\t\t<Host value='%s' />\n",hostname.c_str()); fprintf(qtree_output,"\t\t<Interface value='%s' />\n",interface_names[mi].c_str()); - fprintf(qtree_output,"\t\t<ReadsFrom value='%s' />\n",interface_names[mi].c_str()); + std::vector<tablevar_t *> itbls = lfta_sq->get_input_tables(); + for(int t=0;t<itbls.size();++t){ + fprintf(qtree_output,"\t\t<ReadsFrom value='%s' />\n",itbls[t]->get_schema_name().c_str()); + } +// fprintf(qtree_output,"\t\t<ReadsFrom value='%s' />\n",interface_names[mi].c_str()); fprintf(qtree_output,"\t\t<Rate value='100' />\n"); fprintf(qtree_output,"\t\t<LivenessTimeout value='%d' />\n", lfta_liveness_timeouts[mi]); // write info about fields to qtree.xml @@ -2774,7 +3089,13 @@ void generate_makefile(vector<string> &input_file_names, int nfiles, } // Do we need to include protobuf libraries? +// TODO Move to the interface library: get the libraries to include +// for an interface type + bool use_proto = false; + bool use_bsa = false; + bool use_kafka = false; + bool use_ssl = false; int erri; string err_str; for(ssi=ifaces.begin();ssi!=ifaces.end();++ssi){ @@ -2782,9 +3103,47 @@ void generate_makefile(vector<string> &input_file_names, int nfiles, vector<string> ift = ifdb->get_iface_vals(ifmachines[ifnm],ifnm, "InterfaceType", erri, err_str); for(int ift_i=0;ift_i<ift.size();ift_i++){ if(ift[ift_i]=="PROTO"){ +#ifdef PROTO_ENABLED use_proto = true; +#else + fprintf(stderr,"Runtime libraries built without PROTO support. Rebuild with PROTO_ENABLED defined in gsoptions.h\n"); + exit(0); +#endif + } + } + ift = ifdb->get_iface_vals(ifmachines[ifnm],ifnm, "BSA", erri, err_str); + for(int ift_i=0;ift_i<ift.size();ift_i++){ + if(ift[ift_i] == "TRUE" || ift[ift_i] == "True" || ift[ift_i] == "true"){ +#ifdef BSA_ENABLED + use_bsa = true; +#else + fprintf(stderr,"Runtime libraries built without BSA support. Rebuild with BSA_ENABLED defined in gsoptions.h\n"); + exit(0); +#endif + } + } + ift = ifdb->get_iface_vals(ifmachines[ifnm],ifnm, "KAFKA", erri, err_str); + for(int ift_i=0;ift_i<ift.size();ift_i++){ + if(ift[ift_i] == "TRUE" || ift[ift_i] == "True" || ift[ift_i] == "true"){ +#ifdef KAFKA_ENABLED + use_kafka = true; +#else + fprintf(stderr,"Runtime libraries built without KAFKA support. Rebuild with KAFKA_ENABLED defined in gsoptions.h\n"); + exit(0); +#endif } } + ift = ifdb->get_iface_vals(ifmachines[ifnm],ifnm, "ENCRYPTED", erri, err_str); + for(int ift_i=0;ift_i<ift.size();ift_i++){ + if(ift[ift_i] == "TRUE" || ift[ift_i] == "True" || ift[ift_i] == "true"){ +#ifdef SSL_ENABLED + use_ssl = true; +#else + fprintf(stderr,"Runtime libraries built without SSL support. Rebuild with SSL_ENABLED defined in gsoptions.h\n"); + exit(0); +#endif + } + } } fprintf(outfl, @@ -2805,15 +3164,26 @@ void generate_makefile(vector<string> &input_file_names, int nfiles, if(use_pads) fprintf(outfl,"-lgscppads -lpads "); fprintf(outfl, -"-lgscprts -lgscphost -lm -lgscpaux -lgscplftaaux -lclearinghouse -lresolv -lpthread -lgscpinterface -lz"); +"-lgscprts -lgscphost -lm -lgscpaux -lgscplftaaux -lclearinghouse -lresolv -lpthread -lgscpinterface -lz -lrt"); if(use_pads) fprintf(outfl, " -lpz -lz -lbz "); if(libz_exists && libast_exists) fprintf(outfl," -last "); if(use_pads) fprintf(outfl, " -ldll -ldl "); - if(use_proto) - fprintf(outfl, " -L/usr/local/lib/ -lprotobuf-c "); + +#ifdef PROTO_ENABLED + fprintf(outfl, " -L/usr/local/lib/ -lprotobuf-c "); +#endif +#ifdef BSA_ENABLED + fprintf(outfl, " -lbsa_stream "); +#endif +#ifdef KAFKA_ENABLED + fprintf(outfl, " -lrdkafka "); +#endif +#ifdef SSL_ENABLED + fprintf(outfl, " -lssl -lcrypto "); +#endif fprintf(outfl," -lgscpaux"); #ifdef GCOV fprintf(outfl," -fprofile-arcs"); @@ -2896,6 +3266,9 @@ void generate_makefile(vector<string> &input_file_names, int nfiles, // string err_str; for(ssi=ifaces.begin();ssi!=ifaces.end();++ssi){ string ifnm = (*ssi); + // suppress internal _local_ interface + if (ifnm == "_local_") + continue; fprintf(outfl, "%s ",ifnm.c_str()); vector<string> ifv = ifdb->get_iface_vals(ifmachines[ifnm],ifnm, "Command", erri, err_str); for(j=0;j<ifv.size();++j)