diff options
author | Ashish Singh <1998ashishsingh@gmail.com> | 2020-07-08 13:34:12 +0530 |
---|---|---|
committer | Ashish-uzumaki <1998ashishsingh@gmail.com> | 2020-07-09 12:02:38 +0530 |
commit | ac02a0e376d3ba32a12041824dc919f21401413d (patch) | |
tree | 6b52d1e15c65e46b8e598b31bee8c408d32830eb /scripts/sdc-workflow-d | |
parent | da61def8bfefd1000a882e7108f3f1943392f914 (diff) |
Adding basic E2E test for workflow designer
Change-Id: I45c3fb5a8b26fdd75084c07a9f934b8800259c35
Signed-off-by: Ashish-uzumaki <1998ashishsingh@gmail.com>
Issue-ID: SDC-3193
Signed-off-by: Ashish-uzumaki <1998ashishsingh@gmail.com>
Diffstat (limited to 'scripts/sdc-workflow-d')
-rw-r--r-- | scripts/sdc-workflow-d/cleanup_sdc_workflow.sh | 44 | ||||
-rwxr-xr-x | scripts/sdc-workflow-d/docker_run.sh | 606 | ||||
-rw-r--r-- | scripts/sdc-workflow-d/sdc_workflow_d.sh | 117 |
3 files changed, 767 insertions, 0 deletions
diff --git a/scripts/sdc-workflow-d/cleanup_sdc_workflow.sh b/scripts/sdc-workflow-d/cleanup_sdc_workflow.sh new file mode 100644 index 00000000..3ce28a12 --- /dev/null +++ b/scripts/sdc-workflow-d/cleanup_sdc_workflow.sh @@ -0,0 +1,44 @@ +#!/bin/bash +# +# Copyright 2019 © Samsung Electronics Co., Ltd. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# + +# +# Note! This is only temporary solution for killing SDC DCAE plugin's +# docker containers that must be currently used whenever docker_run.sh is used +# with -dcae option - See SDC-2338 for related image naming issue +# +# DCAE plugin-related parts will also have to be refactored under dedicated +# directories in the future +# + +echo "This is ${WORKSPACE}/scripts/sdc-dcae-d/cleanup_sdc_workflow.sh" + +cp -rf ${WORKSPACE}/data/logs/ ${WORKSPACE}/archives/ + +ls -Rt ${WORKSPACE}/archives/ + +#kill and remove all sdc dockers +docker stop $(docker ps -a -q --filter="name=sdc") +docker rm $(docker ps -a -q --filter="name=sdc") +# kill and remove all sdc dcae dockers +docker stop workflow-frontend +docker stop workflow-backend +docker rm workflow-frontend +docker rm workflow-backend + +#delete data folder + +sudo rm -rf ${WORKSPACE}/data/*
\ No newline at end of file diff --git a/scripts/sdc-workflow-d/docker_run.sh b/scripts/sdc-workflow-d/docker_run.sh new file mode 100755 index 00000000..e641594f --- /dev/null +++ b/scripts/sdc-workflow-d/docker_run.sh @@ -0,0 +1,606 @@ +#!/bin/bash + +# +# Constants: +# + +WORKSPACE="${WORKSPACE:-}" +SUCCESS=0 +FAILURE=1 + +CS_PASSWORD="onap123#@!" +SDC_USER="asdc_user" +SDC_PASSWORD="Aa1234%^!" + +JETTY_BASE="/var/lib/jetty" +SDC_CERT_DIR="onap/cert" + +RELEASE=latest + +LOCAL=false +BE_DEBUG_PORT="--publish 4000:4000" +FE_DEBUG_PORT="--publish 6000:6000" +ONBOARD_DEBUG_PORT="--publish 4001:4001" +CS_PORT=${CS_PORT:-9042} + + +# Java Options: +BE_JAVA_OPTIONS="-Xdebug -agentlib:jdwp=transport=dt_socket,address=4000,server=y,suspend=n -Xmx1536m -Xms1536m" +FE_JAVA_OPTIONS="-Xdebug -agentlib:jdwp=transport=dt_socket,address=6000,server=y,suspend=n -Xmx256m -Xms256m" +ONBOARD_BE_JAVA_OPTIONS="-Xdebug -agentlib:jdwp=transport=dt_socket,address=4001,server=y,suspend=n -Xmx1g -Xms1g" +SIM_JAVA_OPTIONS=" -Xmx128m -Xms128m -Xss1m -Dlog4j.configuration=file:///${JETTY_BASE}/config/sdc-simulator/log4j.properties" +API_TESTS_JAVA_OPTIONS="-Xmx512m -Xms512m" +UI_TESTS_JAVA_OPTIONS="-Xmx1024m -Xms1024m" +#Define this as variable, so it can be excluded in run commands on Docker for OSX, as /etc/localtime cant be mounted there. +LOCAL_TIME_MOUNT_CMD="--volume /etc/localtime:/etc/localtime:ro" +# If os is OSX, unset this, so /etc/localtime is not mounted, otherwise leave it be +if [[ "$OSTYPE" == "darwin"* ]]; then + LOCAL_TIME_MOUNT_CMD="" +fi + + +# +# Functions: +# + + +function usage { + echo "usage: docker_run.sh [ -r|--release <RELEASE-NAME> ] [ -e|--environment <ENV-NAME> ] [ -p|--port <Docker-hub-port>] [ -l|--local <Run-without-pull>] [ -sim|--simulator <Run-with-simulator>] [ -ta <run api tests with the supplied test suit>] [ -tu <run ui tests with the supplied test suit>] [ -ta <run api tests with the supplied test suit>] [ -tu <run ui tests with the supplied test suit>] [ -tad <run api tests with the default test suit>] [ -tu <run ui tests with the default test suit>] [ -h|--help ]" + echo "start dockers built locally example: docker_run.sh -l" + echo "start dockers built locally and simulator example: docker_run.sh -l -sim" + echo "start dockers, pull from onap nexus according to release and simulator example: docker_run.sh -r 1.5-STAGING-latest -sim" + echo "start dockers built locally and run api tests docker example: docker_run.sh -l -tad" + echo "start dockers built locally and run only the catalog be example: docker_run.sh -l -d sdc-BE " +} +# + + +function cleanup { + echo "Performing old dockers cleanup" + + if [ "$1" == "all" ] ; then + docker_ids=`docker ps -a | egrep "ecomp-nexus:${PORT}/sdc|sdc|Exit}|dcae" | awk '{print $1}'` + for X in ${docker_ids} + do + docker rm -f ${X} + done + else + echo "performing $1 docker cleanup" + tmp=`docker ps -a -q --filter="name=$1"` + if [[ ! -z "$tmp" ]]; then + docker rm -f ${tmp} + fi + fi +} +# + + +function dir_perms { + mkdir -p ${WORKSPACE}/data/logs/BE/SDC/SDC-BE + mkdir -p ${WORKSPACE}/data/logs/FE/SDC/SDC-FE + mkdir -p ${WORKSPACE}/data/logs/sdc-api-tests/ExtentReport + mkdir -p ${WORKSPACE}/data/logs/ONBOARD/SDC/ONBOARD-BE + mkdir -p ${WORKSPACE}/data/logs/sdc-api-tests/target + mkdir -p ${WORKSPACE}/data/logs/sdc-ui-tests/ExtentReport + mkdir -p ${WORKSPACE}/data/logs/sdc-ui-tests/target + mkdir -p ${WORKSPACE}/data/logs/docker_logs + mkdir -p ${WORKSPACE}/data/logs/WS + echo "Creating dir '${WORKSPACE}/data/${SDC_CERT_DIR}'" + mkdir -p ${WORKSPACE}/data/${SDC_CERT_DIR} + chmod -R 777 ${WORKSPACE}/data/logs +} +# + + +function docker_logs { + docker logs $1 > ${WORKSPACE}/data/logs/docker_logs/$1_docker.log +} +# + + +# +# Readiness Prob +# + +function ready_probe { + docker exec $1 /var/lib/ready-probe.sh > /dev/null 2>&1 + rc=$? + if [[ ${rc} == 0 ]]; then + echo DOCKER $1 start finished in $2 seconds + return ${SUCCESS} + fi + return ${FAILURE} +} +# + +function ready_probe_jetty { + docker exec $1 /var/lib/jetty/ready-probe.sh > /dev/null 2>&1 + rc=$? + if [[ ${rc} == 0 ]]; then + echo DOCKER $1 start finished in $2 seconds + return ${SUCCESS} + fi + return ${FAILURE} +} +# + +function probe_docker { + MATCH=`docker logs --tail 30 $1 | grep "DOCKER STARTED"` + echo MATCH is -- ${MATCH} + + if [ -n "$MATCH" ] ; then + echo DOCKER start finished in $2 seconds + return ${SUCCESS} + fi + return ${FAILURE} +} +# + +function probe_test_docker { + # This expected logging should be output by startup.sh of the + # respective test docker container + MATCH=`docker logs --tail 30 $1 | grep "Startup completed successfully"` + echo MATCH is -- ${MATCH} + + if [ -n "$MATCH" ] ; then + echo TEST DOCKER start finished in $2 seconds + return ${SUCCESS} + fi + return ${FAILURE} +} +# + + +function probe_sim { + if lsof -Pi :8285 -sTCP:LISTEN -t >/dev/null ; then + echo "Already running" + return ${SUCCESS} + else + echo "Not running" + return ${FAILURE} + fi +} +# + + +function monitor_docker { + DOCKER_NAME=$1 + echo "Monitor ${DOCKER_NAME} Docker" + sleep 5 + TIME_OUT=900 + INTERVAL=20 + TIME=0 + + while [ "$TIME" -lt "$TIME_OUT" ]; do + + case ${DOCKER_NAME} in + + sdc-cs) + ready_probe ${DOCKER_NAME} ${TIME} ; + status=$? ; + ;; + sdc-BE) + ready_probe_jetty ${DOCKER_NAME} ${TIME} ; + status=$? ; + ;; + sdc-FE) + ready_probe_jetty ${DOCKER_NAME} ${TIME} ; + status=$? ; + ;; + sdc-onboard-BE) + ready_probe_jetty ${DOCKER_NAME} ${TIME} ; + status=$? ; + ;; + sdc-api-tests) + probe_test_docker ${DOCKER_NAME} ${TIME}; + status=$? ; + ;; + sdc-ui-tests) + probe_test_docker ${DOCKER_NAME} ${TIME}; + status=$? ; + ;; + *) + probe_docker ${DOCKER_NAME} ${TIME}; + status=$? ; + ;; + + esac + + if [[ ${status} == ${SUCCESS} ]] ; then + break; + fi + + echo "Sleep: ${INTERVAL} seconds before testing if ${DOCKER_NAME} DOCKER is up. Total wait time up now is: ${TIME} seconds. Timeout is: ${TIME_OUT} seconds" + sleep ${INTERVAL} + TIME=$(($TIME+$INTERVAL)) + done + + docker_logs ${DOCKER_NAME} + + if [ "$TIME" -ge "$TIME_OUT" ]; then + echo -e "\e[1;31mTIME OUT: DOCKER was NOT fully started in $TIME_OUT seconds... Could cause problems ...\e[0m" + fi +} +# + +# healthCheck script used the secure connection to send request (https is always turn on) +function healthCheck { + + echo "BE Health Check:" + curl -k --noproxy "*" https://${IP}:8443/sdc2/rest/healthCheck + + echo "" + echo "" + echo "FE Health Check:" + curl -k --noproxy "*" https://${IP}:9443/sdc1/rest/healthCheck + + + echo "" + echo "" + healthCheck_http_code=$(curl -k --noproxy "*" -o /dev/null -w '%{http_code}' -H "Accept: application/json" -H "Content-Type: application/json" -H "USER_ID: jh0003" https://${IP}:8443/sdc2/rest/v1/user/demo;) + if [[ ${healthCheck_http_code} != 200 ]]; then + echo "Error [${healthCheck_http_code}] while checking existence of user" + return ${healthCheck_http_code} + fi + echo "check user existence: OK" + return ${healthCheck_http_code} +} +# + + +function command_exit_status { + status=$1 + docker=$2 + if [ "${status}" != "0" ] ; then + echo "[ ERROR ] Docker ${docker} run command exit with status [${status}]" + fi +} +# + + +# +# Run Containers +# + + +#Cassandra +function sdc-cs { + DOCKER_NAME="sdc-cs" + echo "docker run sdc-cassandra..." + if [ ${LOCAL} = false ]; then + docker pull ${PREFIX}/sdc-cassandra:${RELEASE} + fi + docker run -dit --name ${DOCKER_NAME} --env RELEASE="${RELEASE}" --env CS_PASSWORD="${CS_PASSWORD}" --env ENVNAME="${DEP_ENV}" --env HOST_IP=${IP} --env MAX_HEAP_SIZE="1536M" --env HEAP_NEWSIZE="512M" --log-driver=json-file --log-opt max-size=100m --log-opt max-file=10 --ulimit memlock=-1:-1 --ulimit nofile=4096:100000 ${LOCAL_TIME_MOUNT_CMD} --volume ${WORKSPACE}/data/CS:/var/lib/cassandra --volume ${WORKSPACE}/data/environments:/root/chef-solo/environments --publish ${CS_PORT}:${CS_PORT} ${PREFIX}/sdc-cassandra:${RELEASE} /bin/sh + command_exit_status $? ${DOCKER_NAME} + echo "please wait while CS is starting..." + monitor_docker ${DOCKER_NAME} +} +# + + +#Cassandra-init +function sdc-cs-init { + DOCKER_NAME="sdc-cs-init" + echo "docker run sdc-cassandra-init..." + if [ ${LOCAL} = false ]; then + docker pull ${PREFIX}/sdc-cassandra-init:${RELEASE} + fi + docker run --name ${DOCKER_NAME} --env RELEASE="${RELEASE}" --env SDC_USER="${SDC_USER}" --env SDC_PASSWORD="${SDC_PASSWORD}" --env CS_PASSWORD="${CS_PASSWORD}" --env ENVNAME="${DEP_ENV}" --env HOST_IP=${IP} --log-driver=json-file --log-opt max-size=100m --log-opt max-file=10 --ulimit memlock=-1:-1 --ulimit nofile=4096:100000 ${LOCAL_TIME_MOUNT_CMD} --volume ${WORKSPACE}/data/CS:/var/lib/cassandra --volume ${WORKSPACE}/data/environments:/home/sdc/chef-solo/environments --volume ${WORKSPACE}/data/CS-Init:/root/chef-solo/cache ${PREFIX}/sdc-cassandra-init:${RELEASE} > /dev/null 2>&1 + rc=$? + docker_logs ${DOCKER_NAME} + if [[ ${rc} != 0 ]]; then exit ${rc}; fi +} +# + + +#Onboard Cassandra-init +function sdc-cs-onboard-init { + DOCKER_NAME="sdc-cs-onboard-init" + echo "docker run sdc-cs-onboard-init..." + if [ ${LOCAL} = false ]; then + docker pull ${PREFIX}/sdc-onboard-cassandra-init:${RELEASE} + fi + docker run --name ${DOCKER_NAME} --env RELEASE="${RELEASE}" --env CS_HOST_IP=${IP} --env CS_HOST_PORT=${CS_PORT} --env SDC_USER="${SDC_USER}" --env SDC_PASSWORD="${SDC_PASSWORD}" --env CS_PASSWORD="${CS_PASSWORD}" --env ENVNAME="${DEP_ENV}" --log-driver=json-file --log-opt max-size=100m --log-opt max-file=10 --ulimit memlock=-1:-1 --ulimit nofile=4096:100000 ${LOCAL_TIME_MOUNT_CMD} --volume ${WORKSPACE}/data/CS:/var/lib/cassandra --volume ${WORKSPACE}/data/environments:/home/sdc/chef-solo/environments --volume ${WORKSPACE}/data/CS-Init:/home/sdc/chef-solo/cache ${PREFIX}/sdc-onboard-cassandra-init:${RELEASE} + rc=$? + docker_logs ${DOCKER_NAME} + if [[ ${rc} != 0 ]]; then exit ${rc}; fi +} +# + + +#Back-End +function sdc-BE { + DOCKER_NAME="sdc-BE" + echo "docker run sdc-backend..." + if [ ${LOCAL} = false ]; then + docker pull ${PREFIX}/sdc-backend:${RELEASE} + else + ADDITIONAL_ARGUMENTS=${BE_DEBUG_PORT} + fi + docker run --detach --name ${DOCKER_NAME} --env HOST_IP=${IP} --env ENVNAME="${DEP_ENV}" --env cassandra_ssl_enabled="false" --env JAVA_OPTIONS="${BE_JAVA_OPTIONS}" --log-driver=json-file --log-opt max-size=100m --log-opt max-file=10 --ulimit memlock=-1:-1 --ulimit nofile=4096:100000 ${LOCAL_TIME_MOUNT_CMD} --volume ${WORKSPACE}/data/logs/BE/:${JETTY_BASE}/logs --volume ${WORKSPACE}/data/environments:${JETTY_BASE}/chef-solo/environments --publish 8443:8443 --publish 8080:8080 ${ADDITIONAL_ARGUMENTS} ${PREFIX}/sdc-backend:${RELEASE} + command_exit_status $? ${DOCKER_NAME} + echo "please wait while BE is starting..." + monitor_docker ${DOCKER_NAME} +} +# + + +# Back-End-Init +function sdc-BE-init { + DOCKER_NAME="sdc-BE-init" + echo "docker run sdc-backend-init..." + if [ ${LOCAL} = false ]; then + docker pull ${PREFIX}/sdc-backend-init:${RELEASE} + fi + docker run --name ${DOCKER_NAME} --env HOST_IP=${IP} --env ENVNAME="${DEP_ENV}" --log-driver=json-file --log-opt max-size=100m --log-opt max-file=10 --ulimit memlock=-1:-1 --ulimit nofile=4096:100000 ${LOCAL_TIME_MOUNT_CMD} --volume ${WORKSPACE}/data/logs/BE/:${JETTY_BASE}/logs --volume ${WORKSPACE}/data/environments:/home/sdc/chef-solo/environments ${PREFIX}/sdc-backend-init:${RELEASE} > /dev/null 2>&1 + rc=$? + docker_logs ${DOCKER_NAME} + if [[ ${rc} != 0 ]]; then exit ${rc}; fi +} +# + + +# Onboard Back-End +function sdc-onboard-BE { + DOCKER_NAME="sdc-onboard-BE" + echo "docker run sdc-onboard-BE ..." +# TODO Check the dir_perms action . do we need it here ?? +# dir_perms + if [ ${LOCAL} = false ]; then + docker pull ${PREFIX}/sdc-onboard-backend:${RELEASE} + else + ADDITIONAL_ARGUMENTS=${ONBOARD_DEBUG_PORT} + fi + docker run --detach --name ${DOCKER_NAME} --env HOST_IP=${IP} --env ENVNAME="${DEP_ENV}" --env cassandra_ssl_enabled="false" --env SDC_CLUSTER_NAME="SDC-CS-${DEP_ENV}" --env SDC_USER="${SDC_USER}" --env SDC_PASSWORD="${SDC_PASSWORD}" --env SDC_CERT_DIR="${SDC_CERT_DIR}" --env JAVA_OPTIONS="${ONBOARD_BE_JAVA_OPTIONS}" --log-driver=json-file --log-opt max-size=100m --log-opt max-file=10 --ulimit memlock=-1:-1 --ulimit nofile=4096:100000 ${LOCAL_TIME_MOUNT_CMD} --volume ${WORKSPACE}/data/${SDC_CERT_DIR}:${JETTY_BASE}/onap/cert --volume ${WORKSPACE}/data/logs/ONBOARD:${JETTY_BASE}/logs --volume ${WORKSPACE}/data/environments:/${JETTY_BASE}/chef-solo/environments --publish 8445:8445 --publish 8081:8081 ${ADDITIONAL_ARGUMENTS} ${PREFIX}/sdc-onboard-backend:${RELEASE} + command_exit_status $? ${DOCKER_NAME} + echo "please wait while sdc-onboard-BE is starting..." + monitor_docker ${DOCKER_NAME} +} +# + + +# Front-End +function sdc-FE { + DOCKER_NAME="sdc-FE" + IMAGE_NAME="${PREFIX}/sdc-frontend:${RELEASE}" + echo "Running container '${DOCKER_NAME}' based on '${IMAGE_NAME}' image..." + if [ ${LOCAL} = false ]; then + docker pull ${PREFIX}/sdc-frontend:${RELEASE} + else + ADDITIONAL_ARGUMENTS=${FE_DEBUG_PORT} + fi + + PLUGIN_CONFIG_FILE="${WORKSPACE}/data/environments/plugins-configuration.yaml" + if [[ -f ${WORKSPACE}/data/environments/plugins-configuration.yaml ]]; then + PLUGINS_CONF_VOLUME_MOUNT="--volume ${PLUGIN_CONFIG_FILE}:${JETTY_BASE}/config/catalog-fe/plugins-configuration.yaml" + else + echo "INFO: '${PLUGIN_CONFIG_FILE}' not provided, ignoring..." + fi + + docker run \ + --detach \ + --name ${DOCKER_NAME} \ + --env HOST_IP=${IP} \ + --env ENVNAME="${DEP_ENV}" \ + --env JAVA_OPTIONS="${FE_JAVA_OPTIONS}" \ + --log-driver=json-file \ + --log-opt max-size=100m \ + --log-opt max-file=10 \ + --ulimit memlock=-1:-1 \ + --ulimit nofile=4096:100000 \ + --volume ${WORKSPACE}/data/logs/FE/:${JETTY_BASE}/logs \ + --volume ${WORKSPACE}/data/environments:/${JETTY_BASE}/chef-solo/environments \ + ${LOCAL_TIME_MOUNT_CMD} \ + ${PLUGINS_CONF_VOLUME_MOUNT} \ + --publish 9443:9443 \ + --publish 8181:8181 \ + ${ADDITIONAL_ARGUMENTS} \ + ${IMAGE_NAME} + + command_exit_status $? ${DOCKER_NAME} + echo "Please wait while '${DOCKER_NAME}' container is starting..." + monitor_docker ${DOCKER_NAME} +} +# + + +# apis-sanity +function sdc-api-tests { + if [[ ${RUN_API_TESTS} = true ]] ; then + healthCheck + healthCheck_http_code=$? + if [[ ${healthCheck_http_code} == 200 ]] ; then + echo "docker run sdc-api-tests..." + echo "Trigger sdc-api-tests docker, please wait..." + + if [ ${LOCAL} = false ]; then + docker pull ${PREFIX}/sdc-api-tests:${RELEASE} + fi + + docker run --detach --name sdc-api-tests --env HOST_IP=${IP} --env ENVNAME="${DEP_ENV}" --env JAVA_OPTIONS="${API_TESTS_JAVA_OPTIONS}" --env SUITE_NAME=${API_SUITE} --log-driver=json-file --log-opt max-size=100m --log-opt max-file=10 --ulimit memlock=-1:-1 --ulimit nofile=4096:100000 ${LOCAL_TIME_MOUNT_CMD} --volume ${WORKSPACE}/data/logs/sdc-api-tests/target:/var/lib/tests/target --volume ${WORKSPACE}/data/logs/sdc-api-tests/ExtentReport:/var/lib/tests/ExtentReport --volume ${WORKSPACE}/data/logs/sdc-api-tests/outputCsar:/var/lib/tests/outputCsar --volume ${WORKSPACE}/data/environments:/root/chef-solo/environments --volume ${WORKSPACE}/data/${SDC_CERT_DIR}:/var/lib/tests/cert --publish 9560:9560 ${PREFIX}/sdc-api-tests:${RELEASE} echo "please wait while SDC-API-TESTS is starting....." + monitor_docker sdc-api-tests + fi + fi +} +# + + +# ui-sanity +function sdc-ui-tests { + + if [[ ${RUN_UI_TESTS} = true ]] ; then + healthCheck + healthCheck_http_code=$? + if [[ ${healthCheck_http_code} == 200 ]]; then + echo "docker run sdc-ui-tets..." + echo "Trigger sdc-ui-tests docker, please wait..." + + if [ ${LOCAL} = false ]; then + docker pull ${PREFIX}/sdc-ui-tests:${RELEASE} + fi + RUN_SIMULATOR=true; + sdc-sim + docker run --detach --name sdc-ui-tests --env HOST_IP=${IP} --env ENVNAME="${DEP_ENV}" --env JAVA_OPTIONS="${UI_TESTS_JAVA_OPTIONS}" --env SUITE_NAME=${UI_SUITE} --log-driver=json-file --log-opt max-size=100m --log-opt max-file=10 --ulimit memlock=-1:-1 --ulimit nofile=4096:100000 ${LOCAL_TIME_MOUNT_CMD} --volume ${WORKSPACE}/data/logs/sdc-ui-tests/target:/var/lib/tests/target --volume ${WORKSPACE}/data/logs/sdc-ui-tests/ExtentReport:/var/lib/tests/ExtentReport --volume ${WORKSPACE}/data/environments:/root/chef-solo/environments --publish 5901:5901 --publish 6901:6901 ${PREFIX}/sdc-ui-tests:${RELEASE} + echo "please wait while SDC-UI-TESTS is starting....." + monitor_docker sdc-ui-tests + fi + fi +} +# + + +# SDC-Simulator +function sdc-sim { + if [ "${RUN_SIMULATOR}" == true ]; then + echo "docker run sdc-webSimulator..." + if [ ${LOCAL} = false ]; then + docker pull ${PREFIX}/sdc-simulator:${RELEASE} + fi + + probe_sim + sim_stat=$? + if [ ${sim_stat} == 1 ]; then + docker run \ + --detach \ + --name sdc-sim \ + --env FE_URL="${FE_URL}" \ + --env JAVA_OPTIONS="${SIM_JAVA_OPTIONS}" \ + --env ENVNAME="${DEP_ENV}" \ + ${LOCAL_TIME_MOUNT_CMD} \ + --volume ${WORKSPACE}/data/logs/WS/:${JETTY_BASE}/logs \ + --volume ${WORKSPACE}/data/environments:/root/chef-solo/environments \ + --publish 8285:8080 \ + --publish 8286:8443 ${PREFIX}/sdc-simulator:${RELEASE} + echo "please wait while SDC-WEB-SIMULATOR is starting....." + monitor_docker sdc-sim + fi + fi +} +# + + +# +# Main +# + +# Handle command line arguments +while [ $# -gt 0 ]; do + case $1 in + + # -r | --release - The specific docker version to pull and deploy + -r | --release ) + shift 1 ; + RELEASE=$1; + shift 1;; + + # -e | --environment - The environment name you want to deploy + -e | --environment ) + shift 1; + DEP_ENV=$1; + shift 1 ;; + + # -p | --port - The port from which to connect to the docker nexus + -p | --port ) + shift 1 ; + PORT=$1; + shift 1 ;; + + # -l | --local - Use this for deploying your local dockers without pulling them first + -l | --local ) + LOCAL=true; + shift 1;; + + # -ta - Use this for running the APIs sanity docker after all other dockers have been deployed + -ta ) + shift 1 ; + API_SUITE=$1; + RUN_API_TESTS=true; + shift 1 ;; + + # -tu - Use this for running the UI sanity docker after all other dockers have been deployed + -tu ) + shift 1 ; + UI_SUITE=$1; + RUN_UI_TESTS=true; + shift 1 ;; + + # -tad - Use this for running the DEFAULT suite of tests in APIs sanity docker after all other dockers have been deployed + -tad | -t ) + API_SUITE="onapApiSanity"; + RUN_API_TESTS=true; + shift 1 ;; + + # -tud - Use this for running the DEFAULT suite of tests in UI sanity docker after all other dockers have been deployed + -tud ) + UI_SUITE="onapUiSanity"; + RUN_UI_TESTS=true; + shift 1 ;; + + # -d | --docker - The init specified docker + -d | --docker ) + shift 1 ; + DOCKER=$1; + shift 1 ;; + # -sim | --simulator run the simulator + -sim | --simulator ) + RUN_SIMULATOR=true; + shift 1 ;; + # -sim | --simulator run the simulator + -u | --fe_url ) + shift 1 ; + FE_URL=$1; + shift 1 ;; + + # -h | --help - Display the help message with all the available run options + -h | --help ) + usage; + exit ${SUCCESS};; + + * ) + usage; + exit ${FAILURE};; + esac +done + + +#Prefix those with WORKSPACE so it can be set to something other than /opt +[ -f ${WORKSPACE}/opt/config/env_name.txt ] && DEP_ENV=$(cat ${WORKSPACE}/opt/config/env_name.txt) || echo ${DEP_ENV} +[ -f ${WORKSPACE}/opt/config/nexus_username.txt ] && NEXUS_USERNAME=$(cat ${WORKSPACE}/opt/config/nexus_username.txt) || NEXUS_USERNAME=release +[ -f ${WORKSPACE}/opt/config/nexus_password.txt ] && NEXUS_PASSWD=$(cat ${WORKSPACE}/opt/config/nexus_password.txt) || NEXUS_PASSWD=sfWU3DFVdBr7GVxB85mTYgAW +[ -f ${WORKSPACE}/opt/config/nexus_docker_repo.txt ] && NEXUS_DOCKER_REPO=$(cat ${WORKSPACE}/opt/config/nexus_docker_repo.txt) || NEXUS_DOCKER_REPO=nexus3.onap.org:${PORT} +[ -f ${WORKSPACE}/opt/config/nexus_username.txt ] && docker login -u $NEXUS_USERNAME -p $NEXUS_PASSWD $NEXUS_DOCKER_REPO + + +export IP=`ip route get 8.8.8.8 | awk '/src/{ print $7 }'` +#If OSX, then use this to get IP +if [[ "$OSTYPE" == "darwin"* ]]; then + export IP=$(ipconfig getifaddr en0) +fi +export PREFIX=${NEXUS_DOCKER_REPO}'/onap' + +if [ ${LOCAL} = true ]; then + PREFIX='onap' +fi + +echo "" + +if [ -z "${DOCKER}" ]; then + cleanup all + dir_perms + sdc-cs + sdc-cs-init + sdc-cs-onboard-init + sdc-onboard-BE + sdc-BE + sdc-BE-init + sdc-FE + healthCheck + sdc-sim + sdc-api-tests + sdc-ui-tests +else + cleanup ${DOCKER} + dir_perms + ${DOCKER} + healthCheck +fi diff --git a/scripts/sdc-workflow-d/sdc_workflow_d.sh b/scripts/sdc-workflow-d/sdc_workflow_d.sh new file mode 100644 index 00000000..6432090f --- /dev/null +++ b/scripts/sdc-workflow-d/sdc_workflow_d.sh @@ -0,0 +1,117 @@ +#!/bin/bash +# +# Copyright 2019 © Samsung Electronics Co., Ltd. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# +echo "[INFO] This is sdc_workflow_d.sh" +# run sdc deployment +source "${WORKSPACE}/scripts/sdc/setup_sdc_for_sanity.sh" +export ROBOT_VARIABLES + +# fail quick if error +set -exo pipefail + +export ENV_NAME='CSIT' + +function iss_true { + _value=$(eval echo "\$${1}" | tr '[:upper:]' '[:lower:]') + + case "$_value" in + 1|yes|true|Y) + return 0 + ;; + esac + + return 1 +} + +# returns 0: if SDC_LOCAL_IMAGES is set to true value +# returns 1: otherwise +function using_local_workflow_images { + iss_true WORKFLOW_LOCAL_IMAGES +} + +# cloning workflow directory +mkdir -p "${WORKSPACE}/data/clone/" +cd "${WORKSPACE}/data/clone" +if using_local_workflow_images && [ -n "$WORKFLOW_LOCAL_GITREPO" ] ; then + WORKFLOW_LOCAL_GITREPO=$(realpath "$WORKFLOW_LOCAL_GITREPO") + if [ -d "$WORKFLOW_LOCAL_GITREPO" ] ; then + rm -rf ./workflow + cp -a "$WORKFLOW_LOCAL_GITREPO" ./workflow + else + echo "[ERROR]: Local git repo for workflow does not exist: ${WORKFLOW_LOCAL_GITREPO}" + exit 1 + fi +else + git clone --depth 1 "https://github.com/onap/sdc-sdc-workflow-designer.git" -b ${GERRIT_BRANCH} +fi +# set enviroment variables +source ${WORKSPACE}/data/clone/workflow/version.properties +export WORKFLOW_RELEASE=$major.$minor-STAGING-latest + +SDC_CS=$(docker inspect -f '{{range .NetworkSettings.Networks}}{{.IPAddress}}{{end}}' sdc-cs) +SDC_BE=$(docker inspect -f '{{range .NetworkSettings.Networks}}{{.IPAddress}}{{end}}' sdc-BE) +echo "[INFO] Initialization of workflow init" +echo ${SDC_CS} +echo ${SDC_BE} +docker run -ti \ + -e "CS_HOST=${SDC_CS}" \ + -e "CS_PORT=9042" \ + -e "CS_AUTHENTICATE=true"\ + -e "CS_USER=asdc_user" \ + -e "CS_PASSWORD=Aa1234%^!" nexus3.onap.org:10001/onap/sdc-workflow-init:latest + +echo "[INFO] Initialization of workflow Backend init" +docker run -d --name "workflow-backend" -e "SDC_PROTOCOL=http" \ + -e "SDC_ENDPOINT=${SDC_BE}:8080" \ + -e "SDC_USER=workflow" \ + -e "SDC_PASSWORD=Kp8bJ4SXszM0WXlhak3eHlcse2gAw84vaoGGmJvUy2U" \ + -e "CS_HOSTS=${SDC_CS}" \ + -e "CS_PORT=9042" \ + -e "CS_AUTHENTICATE=true"\ + -e "CS_USER=asdc_user" \ + -e "CS_PASSWORD=Aa1234%^!" \ + -e "CS_SSL_ENABLED=false"\ + -e "SERVER_SSL_ENABLED=false" \ + --env JAVA_OPTIONS="${BE_JAVA_OPTIONS}" --publish 8384:8080 --publish 10443:8443 --publish 8000:8000 nexus3.onap.org:10001/onap/sdc-workflow-backend:latest + +WORKFLOW_BE=$(docker inspect -f '{{range .NetworkSettings.Networks}}{{.IPAddress}}{{end}}' workflow-backend) +echo "[INFO] starting workflow designer fronend" +docker run -d --name "workflow-frontend" \ + -e BACKEND="http://${WORKFLOW_BE}:8080"\ + --publish 8484:8080 --publish 11443:8443 nexus3.onap.org:10001/onap/sdc-workflow-frontend:latest + +cp "${WORKSPACE}/data/clone/sdc/sdc-os-chef/environments/plugins-configuration.yaml" \ + "${WORKSPACE}/data/environments/plugins-configuration.yaml" + +WF_IP=$(docker inspect -f '{{range .NetworkSettings.Networks}}{{.IPAddress}}{{end}}' workflow-frontend) +WFADDR="http:\/\/${WF_IP}:8080\/workflows" +echo ${WFADDR} +sed -i \ + -e "s/<%= @workflow_discovery_url %>/${WFADDR}/g" \ + -e "s/<%= @workflow_source_url %>/${WFADDR}/g" \ + "${WORKSPACE}/data/environments/plugins-configuration.yaml" + +cp "${WORKSPACE}/data/clone/sdc/sdc-os-chef/scripts/docker_run.sh" "${WORKSPACE}/scripts/sdc-workflow-d/" + +echo "[INFO] restarting sdc-FE with updated plugin configuration file with Worflow host ip" +docker stop sdc-FE +"${WORKSPACE}/scripts/sdc-workflow-d/docker_run.sh" \ + --local \ + -e "${ENV_NAME}" \ + -p 10001 -d sdc-FE +# This file is sourced in another script which is out of our control... +set +e +set +o pipefail |