#!/bin/bash
# ============LICENSE_START===============================================
-# Copyright (C) 2020 Nordix Foundation. All rights reserved.
+# Copyright (C) 2020-22023 Nordix Foundation. All rights reserved.
# ========================================================================
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
__print_args() {
echo "Args: remote|remote-remove docker|kube --env-file <environment-filename> [release] [auto-clean] [--stop-at-error] "
echo " [--ricsim-prefix <prefix> ] [--use-local-image <app-nam>+] [--use-snapshot-image <app-nam>+]"
- echo " [--use-staging-image <app-nam>+] [--use-release-image <app-nam>+] [--image-repo <repo-address]"
+ echo " [--use-staging-image <app-nam>+] [--use-release-image <app-nam>+] [--use-external-image <app-nam>+] [--image-repo <repo-address>]"
echo " [--repo-policy local|remote] [--cluster-timeout <timeout-in seconds>] [--print-stats]"
echo " [--override <override-environment-filename>] [--pre-clean] [--gen-stats] [--delete-namespaces]"
- echo " [--delete-containers] [--endpoint-stats]"
+ echo " [--delete-containers] [--endpoint-stats] [--kubeconfig <config-file>] [--host-path-dir <local-host-dir>]"
+ echo " [--kubecontext <context-name>] [--docker-host <docker-host-url>] [--docker-proxy <host-or-ip>]"
}
if [ $# -eq 1 ] && [ "$1" == "help" ]; then
echo "--use-snapshot-image - The script will use images from the nexus snapshot repo for the supplied apps, space separated list of app short names"
echo "--use-staging-image - The script will use images from the nexus staging repo for the supplied apps, space separated list of app short names"
echo "--use-release-image - The script will use images from the nexus release repo for the supplied apps, space separated list of app short names"
+ echo "--use-external-image - The script will use images from the external (non oran/onap) repo for the supplied apps, space separated list of app short names"
echo "--image-repo - Url to optional image repo. Only locally built images will be re-tagged and pushed to this repo"
echo "--repo-policy - Policy controlling which images to re-tag and push if param --image-repo is set. Default is 'local'"
echo "--cluster-timeout - Optional timeout for cluster where it takes time to obtain external ip/host-name. Timeout in seconds. "
echo "--delete-namespaces - Delete kubernetes namespaces before starting tests - but only those created by the test scripts. Kube mode only. Ignored if running with prestarted apps."
echo "--delete-containers - Delete docker containers before starting tests - but only those created by the test scripts. Docker mode only."
echo "--endpoint-stats - Collect endpoint statistics"
+ echo "--kubeconfig - Configure kubectl to use cluster specific cluster config file"
+ echo "--host-path-dir - (Base-)path on local-hostmounted to all VMs (nodes), for hostpath volumes in kube"
+ echo "--kubecontext - Configure kubectl to use a certain context, e.g 'minikube'"
+ echo "--docker-host - Configure docker to use docker in e.g. a VM"
+ echo "--docker-proxy - Configure ip/host to docker when docker is running in a VM"
echo ""
echo "List of app short names supported: "$APP_SHORT_NAMES
exit 0
echo "Test case started as: ${BASH_SOURCE[$i+1]} "$@
-#Localhost constants
-LOCALHOST_NAME="localhost"
-LOCALHOST_HTTP="http://localhost"
-LOCALHOST_HTTPS="https://localhost"
-
# Var to hold 'auto' in case containers shall be stopped when test case ends
AUTO_CLEAN=""
# Var to hold the app names to use remote release images for
USE_RELEASE_IMAGES=""
+# Var to hold the app names to use external release images for
+USE_EXTERNAL_IMAGES=""
# Use this var (STOP_AT_ERROR=1 in the test script) for debugging/trouble shooting to take all logs and exit at first FAIL test case
STOP_AT_ERROR=0
#Var to indicate docker-compose version, V1 or V2
#V1 names replicated containers <proj-name>_<service-name>_<index>
#V2 names replicated containers <proj-name>-<service-name>-<index>
-DOCKER_COMPOSE_VERION="V1"
+DOCKER_COMPOSE_VERSION="V1"
# Function to indent cmd output with one space
indent1() { sed 's/^/ /'; }
TCLOG=$TESTLOGS/$ATC/TC.log
exec &> >(tee ${TCLOG})
+echo $(date) > $TESTLOGS/$ATC/endpoint_tc_start.log
+echo "$TC_ONELINE_DESCR" > $TESTLOGS/$ATC/endpoint_tc_slogan.log
+echo "Test failed" > $TESTLOGS/$ATC/endpoint_tc_end.log # Will be overritten if test is ok
+
#Variables for counting tests as well as passed and failed tests
RES_TEST=0
RES_PASS=0
#Var to control if containers shall be delete before test setup
DELETE_CONTAINERS=0
+#Var to configure kubectl from a config file or context
+KUBECONF=""
+
+#Localhost, may be set to another host/ip by cmd parameter
+LOCALHOST_NAME="localhost"
+
+#Reseting vars related to token/keys used by kubeproxy when istio is enabled
+#The vars are populated if istio is used in the testcase
+KUBE_PROXY_CURL_JWT=""
+KUBE_PROXY_ISTIO_JWKS_KEYS=""
+
+#Var pointing to dir mounted to each kubernetes node (master and workers)
+#Persistent volumes using "hostpath" are allocated beneath the point.
+#Typically it is a dir on local host mounted to each VM running the master and worker.
+#So the intention is make this dir available so the PODs can be restarted on any
+#node and still access the persistent data
+#If not set from cmd line, the path is defaults to "/tmp"
+HOST_PATH_BASE_DIR=""
+
#File to keep deviation messages
DEVIATION_FILE=".tmp_deviations"
rm $DEVIATION_FILE &> /dev/null
# Function to log the start of a test case
__log_test_start() {
TIMESTAMP=$(date "+%Y-%m-%d %H:%M:%S")
- echo -e $BOLD"TEST $TEST_SEQUENCE_NR (${BASH_LINENO[1]}): ${FUNCNAME[1]}" $@ $EBOLD
- echo "TEST $TEST_SEQUENCE_NR - ${TIMESTAMP}: (${BASH_LINENO[1]}): ${FUNCNAME[1]}" $@ >> $HTTPLOG
+ echo -e $BOLD"TEST $TEST_SEQUENCE_NR - (${BASH_LINENO[1]}) - ${TIMESTAMP}: ${FUNCNAME[1]}" $@ $EBOLD
+ echo "TEST $TEST_SEQUENCE_NR - (${BASH_LINENO[1]}) - ${TIMESTAMP}: ${FUNCNAME[1]}" $@ >> $HTTPLOG
((RES_TEST++))
((TEST_SEQUENCE_NR++))
}
__check_stop_at_error
}
+# Function to log a test case that is not supported but will not fail
+__log_test_info_not_supported() {
+ echo -e $YELLOW" INFO, function not supported"$YELLOW
+ __print_current_stats
+}
+
# General function to log a passed test case
__log_test_pass() {
if [ $# -gt 0 ]; then
# Function to log the start of a configuration setup
__log_conf_start() {
TIMESTAMP=$(date "+%Y-%m-%d %H:%M:%S")
- echo -e $BOLD"CONF $CONF_SEQUENCE_NR (${BASH_LINENO[1]}): "${FUNCNAME[1]} $@ $EBOLD
- echo "CONF $CONF_SEQUENCE_NR - ${TIMESTAMP}: (${BASH_LINENO[1]}): "${FUNCNAME[1]} $@ >> $HTTPLOG
+ echo -e $BOLD"CONF $CONF_SEQUENCE_NR - (${BASH_LINENO[1]}) - ${TIMESTAMP}: "${FUNCNAME[1]} $@ $EBOLD
+ echo "CONF $CONF_SEQUENCE_NR - (${BASH_LINENO[1]}) - ${TIMESTAMP}: "${FUNCNAME[1]} $@ >> $HTTPLOG
((CONF_SEQUENCE_NR++))
}
return
fi
ENDPOINT_STAT_FILE=$TESTLOGS/$ATC/imageinfo_$ATC_$1".log"
- echo $POLICY_AGENT_IMAGE > $ENDPOINT_STAT_FILE
+ echo $A1PMS_IMAGE > $ENDPOINT_STAT_FILE
}
#Var for measuring execution time
fi
fi
fi
+ if [ $paramerror -eq 0 ]; then
+ if [ "$1" == "--use-external-image" ]; then
+ USE_EXTERNAL_IMAGES=""
+ shift
+ while [ $# -gt 0 ] && [[ "$1" != "--"* ]]; do
+ USE_EXTERNAL_IMAGES=$USE_EXTERNAL_IMAGES" "$1
+ if [[ "$AVAILABLE_IMAGES_OVERRIDE" != *"$1"* ]]; then
+ paramerror=1
+ if [ -z "$paramerror_str" ]; then
+ paramerror_str="App name $1 is not available for release override for flag: '--use-external-image'"
+ fi
+ fi
+ shift;
+ done
+ foundparm=0
+ if [ -z "$USE_EXTERNAL_IMAGES" ]; then
+ paramerror=1
+ if [ -z "$paramerror_str" ]; then
+ paramerror_str="No app name found for flag: '--use-use-external-image'"
+ fi
+ else
+ echo "Option set - Overriding with external images for app(s):"$USE_EXTERNAL_IMAGES
+ fi
+ fi
+ fi
if [ $paramerror -eq 0 ]; then
if [ "$1" == "--image-repo" ]; then
shift;
foundparm=0
fi
fi
-
+ if [ $paramerror -eq 0 ]; then
+ if [ "$1" == "--kubeconfig" ]; then
+ shift;
+ if [ -z "$1" ]; then
+ paramerror=1
+ if [ -z "$paramerror_str" ]; then
+ paramerror_str="No path found for : '--kubeconfig'"
+ fi
+ else
+ if [ -f $1 ]; then
+ if [ ! -z "$KUBECONF" ]; then
+ paramerror=1
+ if [ -z "$paramerror_str" ]; then
+ paramerror_str="Only one of --kubeconfig/--kubecontext can be set"
+ fi
+ else
+ KUBECONF="--kubeconfig $1"
+ echo "Option set - Kubeconfig path: "$1
+ shift;
+ foundparm=0
+ fi
+ else
+ paramerror=1
+ if [ -z "$paramerror_str" ]; then
+ paramerror_str="File $1 for --kubeconfig not found"
+ fi
+ fi
+ fi
+ fi
+ fi
+ if [ $paramerror -eq 0 ]; then
+ if [ "$1" == "--kubecontext" ]; then
+ shift;
+ if [ -z "$1" ]; then
+ paramerror=1
+ if [ -z "$paramerror_str" ]; then
+ paramerror_str="No context-name found for : '--kubecontext'"
+ fi
+ else
+ if [ ! -z "$KUBECONF" ]; then
+ paramerror=1
+ if [ -z "$paramerror_str" ]; then
+ paramerror_str="Only one of --kubeconfig or --kubecontext can be set"
+ fi
+ else
+ KUBECONF="--context $1"
+ echo "Option set - Kubecontext name: "$1
+ shift;
+ foundparm=0
+ fi
+ fi
+ fi
+ fi
+ if [ $paramerror -eq 0 ]; then
+ if [ "$1" == "--host-path-dir" ]; then
+ shift;
+ if [ -z "$1" ]; then
+ paramerror=1
+ if [ -z "$paramerror_str" ]; then
+ paramerror_str="No path found for : '--host-path-dir'"
+ fi
+ else
+ HOST_PATH_BASE_DIR=$1
+ echo "Option set - Host path for kube set to: "$1
+ shift
+ foundparm=0
+ fi
+ fi
+ fi
+ if [ $paramerror -eq 0 ]; then
+ if [ "$1" == "--docker-host" ]; then
+ shift;
+ if [ -z "$1" ]; then
+ paramerror=1
+ if [ -z "$paramerror_str" ]; then
+ paramerror_str="No url found for : '--docker-host'"
+ fi
+ else
+ export DOCKER_HOST="$1"
+ echo "Option set - DOCKER_HOST set to: "$1
+ shift
+ foundparm=0
+ fi
+ fi
+ fi
+ if [ $paramerror -eq 0 ]; then
+ if [ "$1" == "--docker-host" ]; then
+ shift;
+ if [ -z "$1" ]; then
+ paramerror=1
+ if [ -z "$paramerror_str" ]; then
+ paramerror_str="No url found for : '--docker-host'"
+ fi
+ else
+ export DOCKER_HOST="$1"
+ echo "Option set - DOCKER_HOST set to: "$1
+ shift
+ foundparm=0
+ fi
+ fi
+ fi
+ if [ $paramerror -eq 0 ]; then
+ if [ "$1" == "--docker-proxy" ]; then
+ shift;
+ if [ -z "$1" ]; then
+ paramerror=1
+ if [ -z "$paramerror_str" ]; then
+ paramerror_str="No ip/host found for : '--docker-proxy'"
+ fi
+ else
+ export LOCALHOST_NAME=$1
+ echo "Option set - docker proxy set to: "$1
+ shift
+ foundparm=0
+ fi
+ fi
+ fi
done
echo ""
exit 1
fi
+LOCALHOST_HTTP="http://$LOCALHOST_NAME"
+LOCALHOST_HTTPS="https://$LOCALHOST_NAME"
+
# sourcing the selected env variables for the test case
if [ -f "$TEST_ENV_VAR_FILE" ]; then
echo -e $BOLD"Sourcing env vars from: "$TEST_ENV_VAR_FILE$EBOLD
echo " docker is installed and using versions:"
echo " $(docker version --format 'Client version {{.Client.Version}} Server version {{.Server.Version}}')"
-tmp=$(which docker-compose)
-if [ $? -ne 0 ] || [ -z "$tmp" ]; then
- if [ $RUNMODE == "DOCKER" ]; then
+if [ $RUNMODE == "DOCKER" ]; then
+ tmp=$(which docker-compose)
+ if [ $? -ne 0 ] || [ -z "$tmp" ]; then
echo -e $RED"docker-compose is required to run the test environment, pls install"$ERED
exit 1
+ else
+ tmp=$(docker-compose version)
+ echo " docker-compose installed and using version $tmp"
+ if [[ "$tmp" == *'v2'* ]]; then
+ DOCKER_COMPOSE_VERSION="V2"
+ fi
fi
fi
-tmp=$(docker-compose version --short)
-echo " docker-compose installed and using version $tmp"
-if [[ "$tmp" == *'v2'* ]]; then
- DOCKER_COMPOSE_VERION="V2"
-fi
-
-tmp=$(which kubectl)
-if [ $? -ne 0 ] || [ -z tmp ]; then
- if [ $RUNMODE == "KUBE" ]; then
+if [ $RUNMODE == "KUBE" ]; then
+ tmp=$(which kubectl)
+ if [ $? -ne 0 ] || [ -z tmp ]; then
echo -e $RED"kubectl is required to run the test environment in kubernetes mode, pls install"$ERED
exit 1
- fi
-else
- if [ $RUNMODE == "KUBE" ]; then
+ else
echo " kubectl is installed and using versions:"
- echo $(kubectl version --short=true) | indent2
- res=$(kubectl cluster-info 2>&1)
+ echo $(kubectl $KUBECONF version --short=true) | indent2
+ res=$(kubectl $KUBECONF cluster-info 2>&1)
if [ $? -ne 0 ]; then
echo -e "$BOLD$RED############################################# $ERED$EBOLD"
- echo -e $BOLD$RED"Command 'kubectl cluster-info' returned error $ERED$EBOLD"
+ echo -e $BOLD$RED"Command 'kubectl '$KUBECONF' cluster-info' returned error $ERED$EBOLD"
echo -e "$BOLD$RED############################################# $ERED$EBOLD"
echo " "
echo "kubectl response:"
echo $res
echo " "
echo "This script may have been started with user with no permission to run kubectl"
- echo "Try running with 'sudo' or set 'KUBECONFIG'"
+ echo "Try running with 'sudo', set env KUBECONFIG or set '--kubeconfig' parameter"
echo "Do either 1, 2 or 3 "
echo " "
echo "1"
echo -e $BOLD"sudo -E <test-script-and-parameters>"$EBOLD
echo " "
echo "3"
- echo "Set KUBECONFIG inline (replace user)"
- echo -e $BOLD"sudo KUBECONFIG='/home/<user>/.kube/config' <test-script-and-parameters>"$EBOLD
+ echo "Set KUBECONFIG via script parameter"
+ echo -e $BOLD"sudo ... --kubeconfig /home/<user>/.kube/<config-file> ...."$EBOLD
+ echo "The config file need to downloaded from the cluster"
exit 1
fi
echo " Node(s) and container runtime config"
- kubectl get nodes -o wide | indent2
+ kubectl $KUBECONF get nodes -o wide | indent2
+ echo ""
+ if [ -z "$HOST_PATH_BASE_DIR" ]; then
+ HOST_PATH_BASE_DIR="/tmp"
+ echo " Persistent volumes will be mounted to $HOST_PATH_BASE_DIR on applicable node"
+ echo " No guarantee that persistent volume data is available on all nodes in the cluster"
+ else
+ echo "Persistent volumes will be mounted to base dir: $HOST_PATH_BASE_DIR"
+ echo "Assuming this dir is mounted from each node to a dir on the localhost or other"
+ echo "file system available to all nodes"
+ fi
fi
fi
if [ "$5" == "REMOTE_RELEASE_ORAN" ]; then
image=$NEXUS_RELEASE_REPO_ORAN$image
fi
- #No nexus repo added for local images, tag: LOCAL
+ #No nexus repo added for local images, tag: LOCAL and other tags
tmp=$tmp$image"\t"
fi
if [ -z $tag ]; then
((CTR++))
fi
done
+ for im in $USE_EXTERNAL_IMAGES; do
+ if [ "$1" == "$im" ]; then
+ suffix="EXTERNAL"
+ ((CTR++))
+ fi
+ done
echo $suffix
if [ $CTR -gt 1 ]; then
exit 1
elif [ $1 == "remote" ] || [ $1 == "remote-remove" ]; then
if [ $1 == "remote-remove" ]; then
if [ $RUNMODE == "DOCKER" ]; then
+
echo -ne " Attempt to stop and remove container(s), if running - ${SAMELINE}"
tmp=$(docker ps -aq --filter name=${3} --filter network=${DOCKER_SIM_NWNAME})
if [ $? -eq 0 ] && [ ! -z "$tmp" ]; then
- docker stop $tmp &> ./tmp/.dockererr
+ docker stop -t 0 $tmp &> ./tmp/.dockererr
if [ $? -ne 0 ]; then
((IMAGE_ERR++))
echo ""
echo -ne " Attempt to stop and remove container(s), if running - "$GREEN"stopped"$EGREEN"${SAMELINE}"
tmp=$(docker ps -aq --filter name=${3} --filter network=${DOCKER_SIM_NWNAME}) &> /dev/null
if [ $? -eq 0 ] && [ ! -z "$tmp" ]; then
- docker rm $tmp &> ./tmp/.dockererr
+ docker rm -f $tmp &> ./tmp/.dockererr
if [ $? -ne 0 ]; then
((IMAGE_ERR++))
echo ""
if [ "$DELETE_KUBE_NAMESPACES" -eq 1 ]; then
- test_env_namespaces=$(kubectl get ns --no-headers -o custom-columns=":metadata.name" -l autotest=engine) #Get list of ns created by the test env
+ test_env_namespaces=$(kubectl $KUBECONF get ns --no-headers -o custom-columns=":metadata.name" -l autotest=engine) #Get list of ns created by the test env
if [ $? -ne 0 ]; then
echo " Cannot get list of namespaces...ignoring delete"
else
done
fi
else
- echo " Namespace delete option not set"
+ echo " Namespace delete option not set or ignored"
fi
echo ""
echo " Removing stopped containers..."
docker rm $(docker ps -qa --filter "label=nrttest_app") 2> /dev/null
else
- echo " Contatiner delete option not set"
+ echo " Contatiner delete option not set or ignored"
fi
echo ""
# The following sequence pull the configured images
echo -e $BOLD"Pulling configured images, if needed"$EBOLD
+ __exclude_check=0
if [ ! -z "$IMAGE_REPO_ADR" ] && [ $IMAGE_REPO_POLICY == "local" ]; then
- echo -e $YELLOW" Excluding all remote image check/pull when running with image repo: $IMAGE_REPO_ADR and image policy $IMAGE_REPO_POLICY"$EYELLOW
- else
- for imagename in $APP_SHORT_NAMES; do
- __check_included_image $imagename
- incl=$?
- __check_project_image $imagename
- proj=$?
- if [ $incl -eq 0 ]; then
- if [ $proj -eq 0 ]; then
- START_ARG_MOD=$START_ARG
- __check_image_local_override $imagename
- if [ $? -eq 1 ]; then
- START_ARG_MOD="local"
- fi
- else
- START_ARG_MOD=$START_ARG
+ echo -e $YELLOW" Excluding all remote image check/pull (unless local override) when running with image repo: $IMAGE_REPO_ADR and image policy: $IMAGE_REPO_POLICY"$EYELLOW
+ __exclude_check=1
+ fi
+ for imagename in $APP_SHORT_NAMES; do
+ __check_included_image $imagename
+ incl=$?
+ __check_project_image $imagename
+ proj=$?
+ if [ $incl -eq 0 ]; then
+ if [ $proj -eq 0 ]; then
+ START_ARG_MOD=$START_ARG
+ __check_image_local_override $imagename
+ if [ $? -eq 1 ]; then
+ START_ARG_MOD="local"
fi
+ else
+ START_ARG_MOD=$START_ARG
+ fi
+ __exclude_image_check=0
+ if [ $__exclude_check == 1 ] && [ "$START_ARG_MOD" != "local" ]; then
+ # For to handle locally built images, overriding remote images
+ __exclude_image_check=1
+ fi
+ if [ $__exclude_image_check == 0 ]; then
__check_image_local_build $imagename
#No pull of images built locally
if [ $? -ne 0 ]; then
function_pointer="__"$imagename"_imagepull"
$function_pointer $START_ARG_MOD $START_ARG
fi
- else
- echo -e $YELLOW" Excluding $imagename image from image check/pull"$EYELLOW
fi
- done
- fi
+ else
+ echo -e $YELLOW" Excluding $imagename image from image check/pull"$EYELLOW
+ fi
+ done
+
#Errors in image setting - exit
if [ $IMAGE_ERR -ne 0 ]; then
# Create a table of the images used in the script - from remote repo
echo -e $BOLD"Remote repo images used in this test script"$EBOLD
- echo -e $YELLOW"-- Note: These image will be pulled when the container starts. Images not managed by the test engine --"$EYELLOW
-
+ echo -e $YELLOW"-- Note: These image will be pulled when the container starts. Images not managed by the test engine "$EYELLOW
+ echo -e $YELLOW"-- Note: Images with local override will however be re-tagged and managed by the test engine "$EYELLOW
docker_tmp_file=./tmp/.docker-images-table
format_string="{{.Repository}}\\t{{.Tag}}"
echo -e "Application\tRepository\tTag" > $docker_tmp_file
echo -e " Pulling remote snapshot or staging images my in some case result in pulling newer image versions outside the control of the test engine"
export KUBE_IMAGE_PULL_POLICY="Always"
fi
- CLUSTER_IP=$(kubectl config view -o jsonpath={.clusters[0].cluster.server} | awk -F[/:] '{print $4}')
- echo -e $YELLOW" The cluster hostname/ip is: $CLUSTER_IP"$EYELLOW
+ #CLUSTER_IP=$(kubectl $KUBECONF config view -o jsonpath={.clusters[0].cluster.server} | awk -F[/:] '{print $4}')
+ #echo -e $YELLOW" The cluster hostname/ip is: $CLUSTER_IP"$EYELLOW
echo "================================================================================="
echo "================================================================================="
echo "Runtime statistics collected in file: "$TESTLOGS/$ATC/stat_data.csv
echo ""
fi
-
+ TMP_FLAG_FAIL_PASS=0
total=$((RES_PASS+RES_FAIL))
if [ $RES_TEST -eq 0 ]; then
+ TMP_FLAG_FAIL_PASS=1
echo -e "\033[1mNo tests seem to have been executed. Check the script....\033[0m"
echo -e "\033[31m\033[1m ___ ___ ___ ___ ___ _____ ___ _ ___ _ _ _ ___ ___ \033[0m"
echo -e "\033[31m\033[1m/ __|/ __| _ \_ _| _ \_ _| | __/_\ |_ _| | | | | | _ \ __|\033[0m"
echo -e "\033[31m\033[1m\__ \ (__| /| || _/ | | | _/ _ \ | || |_| |_| | / _| \033[0m"
echo -e "\033[31m\033[1m|___/\___|_|_\___|_| |_| |_/_/ \_\___|____\___/|_|_\___|\033[0m"
elif [ $total != $RES_TEST ]; then
+ TMP_FLAG_FAIL_PASS=1
echo -e "\033[1mTotal number of tests does not match the sum of passed and failed tests. Check the script....\033[0m"
echo -e "\033[31m\033[1m ___ ___ ___ ___ ___ _____ ___ _ ___ _ _ _ ___ ___ \033[0m"
echo -e "\033[31m\033[1m/ __|/ __| _ \_ _| _ \_ _| | __/_\ |_ _| | | | | | _ \ __|\033[0m"
echo -e "\033[31m\033[1m\__ \ (__| /| || _/ | | | _/ _ \ | || |_| |_| | / _| \033[0m"
echo -e "\033[31m\033[1m|___/\___|_|_\___|_| |_| |_/_/ \_\___|____\___/|_|_\___|\033[0m"
elif [ $RES_CONF_FAIL -ne 0 ]; then
+ TMP_FLAG_FAIL_PASS=1
echo -e "\033[1mOne or more configurations has failed. Check the script log....\033[0m"
echo -e "\033[31m\033[1m ___ ___ ___ ___ ___ _____ ___ _ ___ _ _ _ ___ ___ \033[0m"
echo -e "\033[31m\033[1m/ __|/ __| _ \_ _| _ \_ _| | __/_\ |_ _| | | | | | _ \ __|\033[0m"
echo -e "\033[31m\033[1m\__ \ (__| /| || _/ | | | _/ _ \ | || |_| |_| | / _| \033[0m"
echo -e "\033[31m\033[1m|___/\___|_|_\___|_| |_| |_/_/ \_\___|____\___/|_|_\___|\033[0m"
elif [ $RES_PASS = $RES_TEST ]; then
+ TMP_FLAG_FAIL_PASS=0
echo -e "All tests \033[32m\033[1mPASS\033[0m"
echo -e "\033[32m\033[1m ___ _ ___ ___ \033[0m"
echo -e "\033[32m\033[1m | _ \/_\ / __/ __| \033[0m"
#Create file with OK exit code
echo "0" > "$AUTOTEST_HOME/.result$ATC.txt"
echo "0" > "$TESTLOGS/$ATC/.result$ATC.txt"
+ echo $(date) > $TESTLOGS/$ATC/endpoint_tc_end.log
else
+ TMP_FLAG_FAIL_PASS=1
echo -e "One or more tests with status \033[31m\033[1mFAIL\033[0m "
echo -e "\033[31m\033[1m ___ _ ___ _ \033[0m"
echo -e "\033[31m\033[1m | __/_\ |_ _| | \033[0m"
echo -e "\033[31m\033[1m | _/ _ \ | || |__ \033[0m"
echo -e "\033[31m\033[1m |_/_/ \_\___|____|\033[0m"
echo ""
+ fi
+
+ if [ $TMP_FLAG_FAIL_PASS -ne 0 ]; then
# Update test suite counter
if [ -f .tmp_tcsuite_fail_ctr ]; then
tmpval=$(< .tmp_tcsuite_fail_ctr)
# Get resource type for scaling
# args: <resource-name> <namespace>
__kube_get_resource_type() {
- kubectl get deployment $1 -n $2 1> /dev/null 2> ./tmp/kubeerr
+ kubectl $KUBECONF get deployment $1 -n $2 1> /dev/null 2> ./tmp/kubeerr
if [ $? -eq 0 ]; then
echo "deployment"
return 0
fi
- kubectl get sts $1 -n $2 1> /dev/null 2> ./tmp/kubeerr
+ kubectl $KUBECONF get sts $1 -n $2 1> /dev/null 2> ./tmp/kubeerr
if [ $? -eq 0 ]; then
echo "sts"
return 0
# (Not for test scripts)
__kube_scale() {
echo -ne " Setting $1 $2 replicas=$4 in namespace $3"$SAMELINE
- kubectl scale $1 $2 -n $3 --replicas=$4 1> /dev/null 2> ./tmp/kubeerr
+ kubectl $KUBECONF scale $1 $2 -n $3 --replicas=$4 1> /dev/null 2> ./tmp/kubeerr
if [ $? -ne 0 ]; then
echo -e " Setting $1 $2 replicas=$4 in namespace $3 $RED Failed $ERED"
((RES_CONF_FAIL++))
TSTART=$SECONDS
for i in {1..500}; do
- count=$(kubectl get $1/$2 -n $3 -o jsonpath='{.status.replicas}' 2> /dev/null)
+ count=$(kubectl $KUBECONF get $1/$2 -n $3 -o jsonpath='{.status.replicas}' 2> /dev/null)
retcode=$?
if [ -z "$count" ]; then
#No value is sometimes returned for some reason, in case the resource has replica 0
labelid=$3
resources="deployment replicaset statefulset"
for restype in $resources; do
- result=$(kubectl get $restype -n $namespace -o jsonpath='{.items[?(@.metadata.labels.'$labelname'=="'$labelid'")].metadata.name}')
+ result=$(kubectl $KUBECONF get $restype -n $namespace -o jsonpath='{.items[?(@.metadata.labels.'$labelname'=="'$labelid'")].metadata.name}')
if [ $? -eq 0 ] && [ ! -z "$result" ]; then
for resid in $result; do
- echo -ne " Ordered caling $restype $resid in namespace $namespace with label $labelname=$labelid to 0"$SAMELINE
- kubectl scale $restype $resid -n $namespace --replicas=0 1> /dev/null 2> ./tmp/kubeerr
+ echo -ne " Ordered scaling $restype $resid in namespace $namespace with label $labelname=$labelid to 0"$SAMELINE
+ kubectl $KUBECONF scale $restype $resid -n $namespace --replicas=0 1> /dev/null 2> ./tmp/kubeerr
echo -e " Ordered scaling $restype $resid in namespace $namespace with label $labelname=$labelid to 0 $GREEN OK $EGREEN"
done
fi
scaled_all=0
for restype in $resources; do
if [ -z "$3" ]; then
- result=$(kubectl get $restype -n $namespace -o jsonpath='{.items[?(@.metadata.labels.'$labelname')].metadata.name}')
+ result=$(kubectl $KUBECONF get $restype -n $namespace -o jsonpath='{.items[?(@.metadata.labels.'$labelname')].metadata.name}')
else
- result=$(kubectl get $restype -n $namespace -o jsonpath='{.items[?(@.metadata.labels.'$labelname'=="'$labelid'")].metadata.name}')
+ result=$(kubectl $KUBECONF get $restype -n $namespace -o jsonpath='{.items[?(@.metadata.labels.'$labelname'=="'$labelid'")].metadata.name}')
fi
if [ $? -eq 0 ] && [ ! -z "$result" ]; then
for resid in $result; do
echo -e " Ordered scaling $restype $resid in namespace $namespace with label $labelname=$labelid to 0"
- kubectl scale $restype $resid -n $namespace --replicas=0 1> /dev/null 2> ./tmp/kubeerr
+ kubectl $KUBECONF scale $restype $resid -n $namespace --replicas=0 1> /dev/null 2> ./tmp/kubeerr
count=1
T_START=$SECONDS
while [ $count -ne 0 ]; do
- count=$(kubectl get $restype $resid -n $namespace -o jsonpath='{.status.replicas}' 2> /dev/null)
+ count=$(kubectl $KUBECONF get $restype $resid -n $namespace -o jsonpath='{.status.replicas}' 2> /dev/null)
echo -ne " Scaling $restype $resid in namespace $namespace with label $labelname=$labelid to 0, current count=$count"$SAMELINE
if [ $? -eq 0 ] && [ ! -z "$count" ]; then
sleep 0.5
namespace=$1
labelname=$2
labelid=$3
- resources="deployments replicaset statefulset services pods configmaps persistentvolumeclaims persistentvolumes serviceaccounts clusterrolebindings"
+ resources="deployments replicaset statefulset services pods configmaps persistentvolumeclaims persistentvolumes serviceaccounts clusterrolebindings secrets authorizationpolicies requestauthentications"
deleted_resourcetypes=""
for restype in $resources; do
ns_flag="-n $namespace"
ns_flag=""
ns_text=""
fi
- result=$(kubectl get $restype $ns_flag -o jsonpath='{.items[?(@.metadata.labels.'$labelname'=="'$labelid'")].metadata.name}')
+ result=$(kubectl $KUBECONF get $restype $ns_flag -o jsonpath='{.items[?(@.metadata.labels.'$labelname'=="'$labelid'")].metadata.name}' 2> /dev/null)
if [ $? -eq 0 ] && [ ! -z "$result" ]; then
deleted_resourcetypes=$deleted_resourcetypes" "$restype
for resid in $result; do
if [ $restype == "replicaset" ] || [ $restype == "statefulset" ]; then
count=1
while [ $count -ne 0 ]; do
- count=$(kubectl get $restype $resid $ns_flag -o jsonpath='{.status.replicas}' 2> /dev/null)
+ count=$(kubectl $KUBECONF get $restype $resid $ns_flag -o jsonpath='{.status.replicas}' 2> /dev/null)
echo -ne " Scaling $restype $resid $ns_text with label $labelname=$labelid to 0, current count=$count"$SAMELINE
if [ $? -eq 0 ] && [ ! -z "$count" ]; then
sleep 0.5
echo -e " Scaled $restype $resid $ns_text with label $labelname=$labelid to 0, current count=$count $GREEN OK $EGREEN"
fi
echo -ne " Deleting $restype $resid $ns_text with label $labelname=$labelid "$SAMELINE
- kubectl delete --grace-period=1 $restype $resid $ns_flag 1> /dev/null 2> ./tmp/kubeerr
+ kubectl $KUBECONF delete --grace-period=1 $restype $resid $ns_flag 1> /dev/null 2> ./tmp/kubeerr
if [ $? -eq 0 ]; then
echo -e " Deleted $restype $resid $ns_text with label $labelname=$labelid $GREEN OK $EGREEN"
else
result="dummy"
while [ ! -z "$result" ]; do
sleep 0.5
- result=$(kubectl get $restype $ns_flag -o jsonpath='{.items[?(@.metadata.labels.'$labelname'=="'$labelid'")].metadata.name}')
+ result=$(kubectl $KUBECONF get $restype $ns_flag -o jsonpath='{.items[?(@.metadata.labels.'$labelname'=="'$labelid'")].metadata.name}')
echo -ne " Waiting for $restype $ns_text with label $labelname=$labelid to be deleted...$(($SECONDS-$T_START)) seconds "$SAMELINE
if [ -z "$result" ]; then
echo -e " Waiting for $restype $ns_text with label $labelname=$labelid to be deleted...$(($SECONDS-$T_START)) seconds $GREEN OK $EGREEN"
__kube_create_namespace() {
#Check if test namespace exists, if not create it
- kubectl get namespace $1 1> /dev/null 2> ./tmp/kubeerr
+ kubectl $KUBECONF get namespace $1 1> /dev/null 2> ./tmp/kubeerr
if [ $? -ne 0 ]; then
echo -ne " Creating namespace "$1 $SAMELINE
- kubectl create namespace $1 1> /dev/null 2> ./tmp/kubeerr
+ kubectl $KUBECONF create namespace $1 1> /dev/null 2> ./tmp/kubeerr
if [ $? -ne 0 ]; then
echo -e " Creating namespace $1 $RED$BOLD FAILED $EBOLD$ERED"
((RES_CONF_FAIL++))
echo " Message: $(<./tmp/kubeerr)"
return 1
else
- kubectl label ns $1 autotest=engine
+ kubectl $KUBECONF label ns $1 autotest=engine > /dev/null
echo -e " Creating namespace $1 $GREEN$BOLD OK $EBOLD$EGREEN"
fi
else
__kube_delete_namespace() {
#Check if test namespace exists, if so remove it
- kubectl get namespace $1 1> /dev/null 2> ./tmp/kubeerr
+ kubectl $KUBECONF get namespace $1 1> /dev/null 2> ./tmp/kubeerr
if [ $? -eq 0 ]; then
echo -ne " Removing namespace "$1 $SAMELINE
- kubectl delete namespace $1 1> /dev/null 2> ./tmp/kubeerr
+ kubectl $KUBECONF delete namespace $1 1> /dev/null 2> ./tmp/kubeerr
if [ $? -ne 0 ]; then
echo -e " Removing namespace $1 $RED$BOLD FAILED $EBOLD$ERED"
((RES_CONF_FAIL++))
return 0
}
-# Removes a namespace
+# Removes and re-create a namespace
# args: <namespace>
# (Not for test scripts)
clean_and_create_namespace() {
if [ $? -ne 0 ]; then
return 1
fi
+}
+# Add/remove label on non-namespaced kube object
+# args: <api> <instance> <label>
+# (Not for test scripts)
+__kube_label_non_ns_instance() {
+ kubectl $KUBECONF label $1 $2 "$3" 1> /dev/null 2> ./tmp/kubeerr
+ return $?
+}
+
+# Add/remove label on namespaced kube object
+# args: <api> <instance> <namespace> <label>
+# (Not for test scripts)
+__kube_label_ns_instance() {
+ kubectl $KUBECONF label $1 $2 -n $3 "$4" 1> /dev/null 2> ./tmp/kubeerr
+ return $?
}
# Find the host ip of an app (using the service resource)
exit 1
fi
for timeout in {1..60}; do
- host=$(kubectl get svc $1 -n $2 -o jsonpath='{.spec.clusterIP}')
+ host=$(kubectl $KUBECONF get svc $1 -n $2 -o jsonpath='{.spec.clusterIP}')
if [ $? -eq 0 ]; then
if [ ! -z "$host" ]; then
echo $host
fi
for timeout in {1..60}; do
- port=$(kubectl get svc $1 -n $2 -o jsonpath='{...ports[?(@.name=="'$3'")].port}')
+ port=$(kubectl $KUBECONF get svc $1 -n $2 -o jsonpath='{...ports[?(@.name=="'$3'")].port}')
if [ $? -eq 0 ]; then
if [ ! -z "$port" ]; then
echo $port
fi
for timeout in {1..60}; do
- port=$(kubectl get svc $1 -n $2 -o jsonpath='{...ports[?(@.name=="'$3'")].nodePort}')
+ port=$(kubectl $KUBECONF get svc $1 -n $2 -o jsonpath='{...ports[?(@.name=="'$3'")].nodePort}')
if [ $? -eq 0 ]; then
if [ ! -z "$port" ]; then
echo $port
__kube_create_instance() {
echo -ne " Creating $1 $2"$SAMELINE
envsubst < $3 > $4
- kubectl apply -f $4 1> /dev/null 2> ./tmp/kubeerr
+ kubectl $KUBECONF apply -f $4 1> /dev/null 2> ./tmp/kubeerr
if [ $? -ne 0 ]; then
((RES_CONF_FAIL++))
echo -e " Creating $1 $2 $RED Failed $ERED"
echo -ne " Creating configmap $1 "$SAMELINE
envsubst < $5 > $5"_tmp"
cp $5"_tmp" $5 #Need to copy back to orig file name since create configmap neeed the original file name
- kubectl create configmap $1 -n $2 --from-file=$5 --dry-run=client -o yaml > $6
+ kubectl $KUBECONF create configmap $1 -n $2 --from-file=$5 --dry-run=client -o yaml > $6
if [ $? -ne 0 ]; then
echo -e " Creating configmap $1 $RED Failed $ERED"
((RES_CONF_FAIL++))
return 1
fi
- kubectl apply -f $6 1> /dev/null 2> ./tmp/kubeerr
+ kubectl $KUBECONF apply -f $6 1> /dev/null 2> ./tmp/kubeerr
if [ $? -ne 0 ]; then
echo -e " Creating configmap $1 $RED Apply failed $ERED"
echo " Message: $(<./tmp/kubeerr)"
((RES_CONF_FAIL++))
return 1
fi
- kubectl label configmap $1 -n $2 $3"="$4 --overwrite 1> /dev/null 2> ./tmp/kubeerr
+ kubectl $KUBECONF label configmap $1 -n $2 $3"="$4 --overwrite 1> /dev/null 2> ./tmp/kubeerr
if [ $? -ne 0 ]; then
echo -e " Creating configmap $1 $RED Labeling failed $ERED"
echo " Message: $(<./tmp/kubeerr)"
return 1
fi
# Log the resulting map
- kubectl get configmap $1 -n $2 -o yaml > $6
+ kubectl $KUBECONF get configmap $1 -n $2 -o yaml > $6
echo -e " Creating configmap $1 $GREEN OK $EGREEN"
return 0
envsubst < $input_yaml > $output_yaml
- kubectl delete -f $output_yaml 1> /dev/null 2> /dev/null # Delete the previous terminated pod - if existing
+ kubectl $KUBECONF delete -f $output_yaml 1> /dev/null 2> /dev/null # Delete the previous terminated pod - if existing
__kube_create_instance pod $PVC_CLEANER_APP_NAME $input_yaml $output_yaml
if [ $? -ne 0 ]; then
term_ts=$(($SECONDS+30))
while [ $term_ts -gt $SECONDS ]; do
- pod_status=$(kubectl get pod pvc-cleaner -n $PVC_CLEANER_NAMESPACE --no-headers -o custom-columns=":status.phase")
+ pod_status=$(kubectl $KUBECONF get pod pvc-cleaner -n $PVC_CLEANER_NAMESPACE --no-headers -o custom-columns=":status.phase")
if [ "$pod_status" == "Succeeded" ]; then
return 0
fi
fi
done
+ # Remove istio label on namespaces
+ test_env_namespaces=$(kubectl $KUBECONF get ns --no-headers -o custom-columns=":metadata.name" -l autotest=engine -l istio-injection=enabled) #Get list of ns created by the test env
+ if [ $? -ne 0 ]; then
+ echo " Cannot get list of namespaces...continues.."
+ else
+ for test_env_ns in $test_env_namespaces; do
+ echo " Removing istio label on ns: "$test_env_ns
+ __kube_label_non_ns_instance ns $test_env_ns "istio-injection-"
+ done
+ fi
+
echo ""
}
__clean_kube
if [ $PRE_CLEAN -eq 1 ]; then
echo " Cleaning docker resouces to free up resources, may take time..."
- ../common/clean_docker.sh 2>&1 /dev/null
+ ../common/clean_docker.sh 2>&1 > /dev/null
echo ""
fi
else
__clean_containers
if [ $PRE_CLEAN -eq 1 ]; then
echo " Cleaning kubernetes resouces to free up resources, may take time..."
- ../common/clean_kube.sh 2>&1 /dev/null
+ ../common/clean_kube.sh $KUBECONF 2>&1 > /dev/null
echo ""
fi
fi
envsubst < $compose_file > "gen_"$compose_file
compose_file="gen_"$compose_file
- if [ $DOCKER_COMPOSE_VERION == "V1" ]; then
+ if [ $DOCKER_COMPOSE_VERSION == "V1" ]; then
docker_compose_cmd="docker-compose"
else
docker_compose_cmd="docker compose"
echo -ne " Container $BOLD${appname}$EBOLD starting${SAMELINE}"
- pa_st=false
+ a1pmsst=false
echo -ne " Waiting for ${ENTITY} ${appname} service status...${SAMELINE}"
TSTART=$SECONDS
loop_ctr=0
fi
echo -ne " Waiting for {ENTITY} $BOLD${appname}$EBOLD service status on ${3}, result: $result${SAMELINE}"
echo -ne " The ${ENTITY} $BOLD${appname}$EBOLD$GREEN is alive$EGREEN, responds to service status:$GREEN $result $EGREEN on ${url} after $(($SECONDS-$TSTART)) seconds"
- pa_st=true
+ a1pmsst=true
break
else
TS_TMP=$SECONDS
let loop_ctr=loop_ctr+1
done
- if [ "$pa_st" = "false" ]; then
+ if [ "$a1pmsst" = "false" ]; then
((RES_CONF_FAIL++))
echo -e $RED" The ${ENTITY} ${appname} did not respond to service status on ${url} in $(($SECONDS-$TSTART)) seconds"$ERED
return 1
#tmp=$(docker ps | grep $appname)
tmp=$(docker ps -q --filter name=$appname) #get the container id
- if [ -z "$tmp" ]; then #Only check logs for running Policy Agent apps
+ if [ -z "$tmp" ]; then #Only check logs for running A1PMS apps
echo " "$dispname" is not running, no check made"
return
fi
done
fi
if [ $RUNMODE == "KUBE" ]; then
- namespaces=$(kubectl get namespaces -o jsonpath='{.items[?(@.metadata.name)].metadata.name}')
+ namespaces=$(kubectl $KUBECONF get namespaces -o jsonpath='{.items[?(@.metadata.name)].metadata.name}')
for nsid in $namespaces; do
- pods=$(kubectl get pods -n $nsid -o jsonpath='{.items[?(@.metadata.labels.autotest)].metadata.name}')
+ pods=$(kubectl $KUBECONF get pods -n $nsid -o jsonpath='{.items[?(@.metadata.labels.autotest)].metadata.name}')
for podid in $pods; do
- kubectl logs -n $nsid $podid > $TESTLOGS/$ATC/$1_${podid}.log
+ kubectl $KUBECONF logs -n $nsid $podid > $TESTLOGS/$ATC/$1_${podid}.log
done
done
fi
## Generic curl
###############
# Generic curl function, assumes all 200-codes are ok
+# Used proxy, set
# args: <valid-curl-args-including full url>
# returns: <returned response (without respose code)> or "<no-response-from-server>" or "<not found, <http-code>>""
# returns: The return code is 0 for ok and 1 for not ok
proxyflag=" --proxy-insecure --proxy $KUBE_PROXY_PATH"
fi
fi
- curlString="curl -skw %{http_code} $proxyflag $@"
+
+ if [ ! -z "$KUBE_PROXY_CURL_JWT" ]; then
+ jwt="-H "\""Authorization: Bearer $KUBE_PROXY_CURL_JWT"\"
+ curlString="curl -skw %{http_code} $proxyflag $@"
+ echo " CMD: $curlString $jwt" >> $HTTPLOG
+ res=$($curlString -H "Authorization: Bearer $KUBE_PROXY_CURL_JWT")
+ retcode=$?
+ else
+ curlString="curl -skw %{http_code} $proxyflag $@"
+ echo " CMD: $curlString" >> $HTTPLOG
+ res=$($curlString)
+ retcode=$?
+ fi
+ echo " RESP: $res" >> $HTTPLOG
+ echo " RETCODE: $retcode" >> $HTTPLOG
+ if [ $retcode -ne 0 ]; then
+ echo "<no-response-from-server>"
+ return 1
+ fi
+ http_code="${res:${#res}-3}"
+ if [ ${#res} -eq 3 ]; then
+ if [ $http_code -lt 200 ] || [ $http_code -gt 299 ]; then
+ echo "<no-response-from-server>"
+ return 1
+ else
+ return 0
+ fi
+ else
+ if [ $http_code -lt 200 ] || [ $http_code -gt 299 ]; then
+ echo "<not found, resp:${http_code}>"
+ return 1
+ fi
+ if [ $# -eq 2 ]; then
+ echo "${res:0:${#res}-3}" | xargs
+ else
+ echo "${res:0:${#res}-3}"
+ fi
+
+ return 0
+ fi
+
+}
+
+# Generic curl function, assumes all 200-codes are ok
+# Uses no proxy, even if it is set
+# args: <valid-curl-args-including full url>
+# returns: <returned response (without respose code)> or "<no-response-from-server>" or "<not found, <http-code>>""
+# returns: The return code is 0 for ok and 1 for not ok
+__do_curl_no_proxy() {
+ echo ${FUNCNAME[1]} "line: "${BASH_LINENO[1]} >> $HTTPLOG
+ curlString="curl -skw %{http_code} $@"
echo " CMD: $curlString" >> $HTTPLOG
res=$($curlString)
retcode=$?
return 1
fi
if [ $# -eq 2 ]; then
- echo "${res:0:${#res}-3}" | xargs
+ echo "${res:0:${#res}-3}" | xargs
else
- echo "${res:0:${#res}-3}"
+ echo "${res:0:${#res}-3}"
fi
return 0
checkjsonarraycount=1
fi
- echo -e $BOLD"TEST $TEST_SEQUENCE_NR (${BASH_LINENO[1]}): ${1}, ${3} ${4} ${5} within ${6} seconds"$EBOLD
- echo "TEST $TEST_SEQUENCE_NR - ${TIMESTAMP}: (${BASH_LINENO[1]}): ${1}, ${3} ${4} ${5} within ${6} seconds" >> $HTTPLOG
+ echo -e $BOLD"TEST $TEST_SEQUENCE_NR - (${BASH_LINENO[1]}) - ${TIMESTAMP}: ${1}, ${3} ${4} ${5} within ${6} seconds"$EBOLD
+ echo "TEST $TEST_SEQUENCE_NR - (${BASH_LINENO[1]}) - ${TIMESTAMP}: ${1}, ${3} ${4} ${5} within ${6} seconds" >> $HTTPLOG
((RES_TEST++))
((TEST_SEQUENCE_NR++))
checkjsonarraycount=1
fi
- echo -e $BOLD"TEST $TEST_SEQUENCE_NR (${BASH_LINENO[1]}): ${1}, ${3} ${4} ${5}"$EBOLD
- echo "TEST $TEST_SEQUENCE_NR - ${TIMESTAMP}: (${BASH_LINENO[1]}): ${1}, ${3} ${4} ${5}" >> $HTTPLOG
+ echo -e $BOLD"TEST $TEST_SEQUENCE_NR - (${BASH_LINENO[1]}) - ${TIMESTAMP}: ${1}, ${3} ${4} ${5}"$EBOLD
+ echo "TEST $TEST_SEQUENCE_NR - (${BASH_LINENO[1]}) - ${TIMESTAMP}: ${1}, ${3} ${4} ${5}" >> $HTTPLOG
((RES_TEST++))
((TEST_SEQUENCE_NR++))
if [ $checkjsonarraycount -eq 0 ]; then