X-Git-Url: https://gerrit.o-ran-sc.org/r/gitweb?a=blobdiff_plain;f=test%2Fcommon%2Fagent_api_functions.sh;h=767abb9a924ba102af1a0de05a5e35f0bb53e627;hb=a893ee0669b7820fdb30af24b213ea69956377e1;hp=cb48d7888b3140c9057358817fc49a9bf64e113c;hpb=a69cd90b798f1a25820d742710bcd13dfd1c4c60;p=nonrtric.git diff --git a/test/common/agent_api_functions.sh b/test/common/agent_api_functions.sh index cb48d788..767abb9a 100644 --- a/test/common/agent_api_functions.sh +++ b/test/common/agent_api_functions.sh @@ -183,7 +183,8 @@ start_policy_agent() { if [ $retcode_p -eq 0 ]; then echo -e " Using existing $POLICY_AGENT_APP_NAME deployment and service" echo " Setting $POLICY_AGENT_APP_NAME replicas=1" - __kube_scale deployment $POLICY_AGENT_APP_NAME $KUBE_NONRTRIC_NAMESPACE 1 + res_type=$(__kube_get_resource_type $POLICY_AGENT_APP_NAME $KUBE_NONRTRIC_NAMESPACE) + __kube_scale $res_type $POLICY_AGENT_APP_NAME $KUBE_NONRTRIC_NAMESPACE 1 fi if [ $retcode_i -eq 0 ]; then @@ -270,9 +271,13 @@ start_policy_agent() { fi # Keep the initial worker node in case the pod need to be "restarted" - must be made to the same node due to a volume mounted on the host - __PA_WORKER_NODE=$(kubectl get pod -l "autotest=PA" -n $KUBE_NONRTRIC_NAMESPACE -o jsonpath='{.items[*].spec.nodeName}') - if [ -z "$__PA_WORKER_NODE" ]; then - echo -e $YELLOW" Cannot find worker node for pod for $POLICY_AGENT_APP_NAME, persistency may not work"$EYELLOW + if [ $retcode_i -eq 0 ]; then + __PA_WORKER_NODE=$(kubectl get pod -l "autotest=PA" -n $KUBE_NONRTRIC_NAMESPACE -o jsonpath='{.items[*].spec.nodeName}') + if [ -z "$__PA_WORKER_NODE" ]; then + echo -e $YELLOW" Cannot find worker node for pod for $POLICY_AGENT_APP_NAME, persistency may not work"$EYELLOW + fi + else + echo -e $YELLOW" Persistency may not work for app $POLICY_AGENT_APP_NAME in multi-worker node config when running it as a prestarted app"$EYELLOW fi echo " Retrieving host and ports for service..." @@ -374,6 +379,14 @@ stop_policy_agent() { echo -e $BOLD"Stopping $POLICY_AGENT_DISPLAY_NAME"$EBOLD if [ $RUNMODE == "KUBE" ]; then + + __check_prestarted_image "PA" + if [ $? -eq 0 ]; then + echo -e $YELLOW" Persistency may not work for app $POLICY_AGENT_APP_NAME in multi-worker node config when running it as a prestarted app"$EYELLOW + res_type=$(__kube_get_resource_type $POLICY_AGENT_APP_NAME $KUBE_NONRTRIC_NAMESPACE) + __kube_scale $res_type $POLICY_AGENT_APP_NAME $KUBE_NONRTRIC_NAMESPACE 0 + return 0 + fi __kube_scale_all_resources $KUBE_NONRTRIC_NAMESPACE autotest PA echo " Deleting the replica set - a new will be started when the app is started" tmp=$(kubectl delete rs -n $KUBE_NONRTRIC_NAMESPACE -l "autotest=PA") @@ -404,6 +417,15 @@ start_stopped_policy_agent() { if [ $RUNMODE == "KUBE" ]; then + __check_prestarted_image "PA" + if [ $? -eq 0 ]; then + echo -e $YELLOW" Persistency may not work for app $POLICY_AGENT_APP_NAME in multi-worker node config when running it as a prestarted app"$EYELLOW + res_type=$(__kube_get_resource_type $POLICY_AGENT_APP_NAME $KUBE_NONRTRIC_NAMESPACE) + __kube_scale $res_type $POLICY_AGENT_APP_NAME $KUBE_NONRTRIC_NAMESPACE 1 + __check_service_start $POLICY_AGENT_APP_NAME $PA_PATH$POLICY_AGENT_ALIVE_URL + return 0 + fi + # Tie the PMS to the same worker node it was initially started on # A PVC of type hostPath is mounted to PMS, for persistent storage, so the PMS must always be on the node which mounted the volume if [ -z "$__PA_WORKER_NODE" ]; then @@ -419,7 +441,6 @@ start_stopped_policy_agent() { fi __kube_scale deployment $POLICY_AGENT_APP_NAME $KUBE_NONRTRIC_NAMESPACE 1 fi - else docker start $POLICY_AGENT_APP_NAME &> ./tmp/.dockererr if [ $? -ne 0 ]; then @@ -2106,7 +2127,7 @@ api_put_configuration() { return 1 fi if [ ! -f $2 ]; then - _log_test_fail_general "Config file "$2", does not exist" + __log_test_fail_general "Config file "$2", does not exist" return 1 fi inputJson=$(< $2) @@ -2142,7 +2163,7 @@ api_get_configuration() { return 1 fi if [ ! -f $2 ]; then - _log_test_fail_general "Config file "$2" for comparison, does not exist" + __log_test_fail_general "Config file "$2" for comparison, does not exist" return 1 fi @@ -2170,6 +2191,30 @@ api_get_configuration() { fi fi + __log_test_pass + return 0 +} + +########################################## +#### Reset types and instances #### +########################################## + +# Admin reset to remove all policies and services +# All types and instances etc are removed - types and instances in a1 sims need to be removed separately +# NOTE - only works in kubernetes and the pod should not be running +# args: - +# (Function for test scripts) + +pms_kube_pvc_reset() { + __log_test_start $@ + + pvc_name=$(kubectl get pvc -n nonrtric --no-headers -o custom-columns=":metadata.name" | grep policy) + if [ -z "$pvc_name" ]; then + pvc_name=policymanagementservice-vardata-pvc + fi + echo " Trying to reset pvc: "$pvc_name + __kube_clean_pvc $POLICY_AGENT_APP_NAME nonrtric $pvc_name /var/policy-management-service/database + __log_test_pass return 0 } \ No newline at end of file