diff options
-rw-r--r-- | plans/sdc-workflow-d/setup.sh | 50 | ||||
-rw-r--r-- | plans/sdc-workflow-d/teardown.sh | 18 | ||||
-rw-r--r-- | plans/sdc-workflow-d/testplan.txt | 3 | ||||
-rw-r--r-- | scripts/sdc-workflow-d/cleanup_sdc_workflow.sh | 44 | ||||
-rwxr-xr-x | scripts/sdc-workflow-d/docker_run.sh | 606 | ||||
-rw-r--r-- | scripts/sdc-workflow-d/sdc_workflow_d.sh | 117 | ||||
-rwxr-xr-x | scripts/sdc/docker_run.sh | 606 | ||||
-rw-r--r-- | scripts/sdc/setup_sdc_for_sanity.sh | 14 | ||||
-rw-r--r-- | tests/sdc-workflow-d/__init__.robot | 2 | ||||
-rw-r--r-- | tests/sdc-workflow-d/global_properties.robot | 43 | ||||
-rw-r--r-- | tests/sdc-workflow-d/test1.robot | 112 |
11 files changed, 1609 insertions, 6 deletions
diff --git a/plans/sdc-workflow-d/setup.sh b/plans/sdc-workflow-d/setup.sh new file mode 100644 index 00000000..30f1c0f2 --- /dev/null +++ b/plans/sdc-workflow-d/setup.sh @@ -0,0 +1,50 @@ +#!/bin/bash +# +# Copyright 2019 © Samsung Electronics Co., Ltd. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# + +# It can enable HTTPS for SDC component +export SDC_TEST_HTTPS="${SDC_TEST_HTTPS:-false}" + +# By default all images are from remote upstream registry, this option +# provides the chance to test locally built images +export SDC_LOCAL_IMAGES="${SDC_LOCAL_IMAGES:-false}" + +export WORKFLOW_LOCAL_IMAGES="${WORKFLOW_LOCAL_IMAGES:-false}" + +# For this to take effect SDC_LOCAL_IMAGES must be enabled... +# +# The path to the local sdc git repo from which the local images have +# been built - it also affects the tag used - if left empty *AND* +# local images are used *AND* SDC_LOCAL_TAG is unset then the tag +# will be set to: 'latest' +# +# BEWARE: Using local images with an incorrect git repo could lead to +# problems...set SDC_LOCAL_GITREPO or GERRIT_BRANCH properly... +export SDC_LOCAL_GITREPO="${SDC_LOCAL_GITREPO}" + +# For this to take effect SDC_LOCAL_IMAGES must be enabled... +# +# This will set the tag for local images - leaving this empty *AND* +# with unset SDC_LOCAL_GITREPO the local images will fallback to the +# tag: 'latest' +export SDC_LOCAL_TAG="${SDC_LOCAL_TAG}" + + +export WORKFLOW_LOCAL_GITREPO="${WORKFLOW_LOCAL_GITREPO}" + + + +source ${WORKSPACE}/scripts/sdc-workflow-d/sdc_workflow_d.sh diff --git a/plans/sdc-workflow-d/teardown.sh b/plans/sdc-workflow-d/teardown.sh new file mode 100644 index 00000000..bcc43fe4 --- /dev/null +++ b/plans/sdc-workflow-d/teardown.sh @@ -0,0 +1,18 @@ +#!/bin/bash +# +# Copyright 2019 © Samsung Electronics Co., Ltd. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# + +source ${WORKSPACE}/scripts/sdc-workflow-d/cleanup_sdc_workflow.sh diff --git a/plans/sdc-workflow-d/testplan.txt b/plans/sdc-workflow-d/testplan.txt new file mode 100644 index 00000000..6d61eb26 --- /dev/null +++ b/plans/sdc-workflow-d/testplan.txt @@ -0,0 +1,3 @@ +# Test suites are relative paths under [integration/csit.git]/tests/. +# Place the suites in run order. +sdc-workflow-d/test1.robot diff --git a/scripts/sdc-workflow-d/cleanup_sdc_workflow.sh b/scripts/sdc-workflow-d/cleanup_sdc_workflow.sh new file mode 100644 index 00000000..3ce28a12 --- /dev/null +++ b/scripts/sdc-workflow-d/cleanup_sdc_workflow.sh @@ -0,0 +1,44 @@ +#!/bin/bash +# +# Copyright 2019 © Samsung Electronics Co., Ltd. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# + +# +# Note! This is only temporary solution for killing SDC DCAE plugin's +# docker containers that must be currently used whenever docker_run.sh is used +# with -dcae option - See SDC-2338 for related image naming issue +# +# DCAE plugin-related parts will also have to be refactored under dedicated +# directories in the future +# + +echo "This is ${WORKSPACE}/scripts/sdc-dcae-d/cleanup_sdc_workflow.sh" + +cp -rf ${WORKSPACE}/data/logs/ ${WORKSPACE}/archives/ + +ls -Rt ${WORKSPACE}/archives/ + +#kill and remove all sdc dockers +docker stop $(docker ps -a -q --filter="name=sdc") +docker rm $(docker ps -a -q --filter="name=sdc") +# kill and remove all sdc dcae dockers +docker stop workflow-frontend +docker stop workflow-backend +docker rm workflow-frontend +docker rm workflow-backend + +#delete data folder + +sudo rm -rf ${WORKSPACE}/data/*
\ No newline at end of file diff --git a/scripts/sdc-workflow-d/docker_run.sh b/scripts/sdc-workflow-d/docker_run.sh new file mode 100755 index 00000000..e641594f --- /dev/null +++ b/scripts/sdc-workflow-d/docker_run.sh @@ -0,0 +1,606 @@ +#!/bin/bash + +# +# Constants: +# + +WORKSPACE="${WORKSPACE:-}" +SUCCESS=0 +FAILURE=1 + +CS_PASSWORD="onap123#@!" +SDC_USER="asdc_user" +SDC_PASSWORD="Aa1234%^!" + +JETTY_BASE="/var/lib/jetty" +SDC_CERT_DIR="onap/cert" + +RELEASE=latest + +LOCAL=false +BE_DEBUG_PORT="--publish 4000:4000" +FE_DEBUG_PORT="--publish 6000:6000" +ONBOARD_DEBUG_PORT="--publish 4001:4001" +CS_PORT=${CS_PORT:-9042} + + +# Java Options: +BE_JAVA_OPTIONS="-Xdebug -agentlib:jdwp=transport=dt_socket,address=4000,server=y,suspend=n -Xmx1536m -Xms1536m" +FE_JAVA_OPTIONS="-Xdebug -agentlib:jdwp=transport=dt_socket,address=6000,server=y,suspend=n -Xmx256m -Xms256m" +ONBOARD_BE_JAVA_OPTIONS="-Xdebug -agentlib:jdwp=transport=dt_socket,address=4001,server=y,suspend=n -Xmx1g -Xms1g" +SIM_JAVA_OPTIONS=" -Xmx128m -Xms128m -Xss1m -Dlog4j.configuration=file:///${JETTY_BASE}/config/sdc-simulator/log4j.properties" +API_TESTS_JAVA_OPTIONS="-Xmx512m -Xms512m" +UI_TESTS_JAVA_OPTIONS="-Xmx1024m -Xms1024m" +#Define this as variable, so it can be excluded in run commands on Docker for OSX, as /etc/localtime cant be mounted there. +LOCAL_TIME_MOUNT_CMD="--volume /etc/localtime:/etc/localtime:ro" +# If os is OSX, unset this, so /etc/localtime is not mounted, otherwise leave it be +if [[ "$OSTYPE" == "darwin"* ]]; then + LOCAL_TIME_MOUNT_CMD="" +fi + + +# +# Functions: +# + + +function usage { + echo "usage: docker_run.sh [ -r|--release <RELEASE-NAME> ] [ -e|--environment <ENV-NAME> ] [ -p|--port <Docker-hub-port>] [ -l|--local <Run-without-pull>] [ -sim|--simulator <Run-with-simulator>] [ -ta <run api tests with the supplied test suit>] [ -tu <run ui tests with the supplied test suit>] [ -ta <run api tests with the supplied test suit>] [ -tu <run ui tests with the supplied test suit>] [ -tad <run api tests with the default test suit>] [ -tu <run ui tests with the default test suit>] [ -h|--help ]" + echo "start dockers built locally example: docker_run.sh -l" + echo "start dockers built locally and simulator example: docker_run.sh -l -sim" + echo "start dockers, pull from onap nexus according to release and simulator example: docker_run.sh -r 1.5-STAGING-latest -sim" + echo "start dockers built locally and run api tests docker example: docker_run.sh -l -tad" + echo "start dockers built locally and run only the catalog be example: docker_run.sh -l -d sdc-BE " +} +# + + +function cleanup { + echo "Performing old dockers cleanup" + + if [ "$1" == "all" ] ; then + docker_ids=`docker ps -a | egrep "ecomp-nexus:${PORT}/sdc|sdc|Exit}|dcae" | awk '{print $1}'` + for X in ${docker_ids} + do + docker rm -f ${X} + done + else + echo "performing $1 docker cleanup" + tmp=`docker ps -a -q --filter="name=$1"` + if [[ ! -z "$tmp" ]]; then + docker rm -f ${tmp} + fi + fi +} +# + + +function dir_perms { + mkdir -p ${WORKSPACE}/data/logs/BE/SDC/SDC-BE + mkdir -p ${WORKSPACE}/data/logs/FE/SDC/SDC-FE + mkdir -p ${WORKSPACE}/data/logs/sdc-api-tests/ExtentReport + mkdir -p ${WORKSPACE}/data/logs/ONBOARD/SDC/ONBOARD-BE + mkdir -p ${WORKSPACE}/data/logs/sdc-api-tests/target + mkdir -p ${WORKSPACE}/data/logs/sdc-ui-tests/ExtentReport + mkdir -p ${WORKSPACE}/data/logs/sdc-ui-tests/target + mkdir -p ${WORKSPACE}/data/logs/docker_logs + mkdir -p ${WORKSPACE}/data/logs/WS + echo "Creating dir '${WORKSPACE}/data/${SDC_CERT_DIR}'" + mkdir -p ${WORKSPACE}/data/${SDC_CERT_DIR} + chmod -R 777 ${WORKSPACE}/data/logs +} +# + + +function docker_logs { + docker logs $1 > ${WORKSPACE}/data/logs/docker_logs/$1_docker.log +} +# + + +# +# Readiness Prob +# + +function ready_probe { + docker exec $1 /var/lib/ready-probe.sh > /dev/null 2>&1 + rc=$? + if [[ ${rc} == 0 ]]; then + echo DOCKER $1 start finished in $2 seconds + return ${SUCCESS} + fi + return ${FAILURE} +} +# + +function ready_probe_jetty { + docker exec $1 /var/lib/jetty/ready-probe.sh > /dev/null 2>&1 + rc=$? + if [[ ${rc} == 0 ]]; then + echo DOCKER $1 start finished in $2 seconds + return ${SUCCESS} + fi + return ${FAILURE} +} +# + +function probe_docker { + MATCH=`docker logs --tail 30 $1 | grep "DOCKER STARTED"` + echo MATCH is -- ${MATCH} + + if [ -n "$MATCH" ] ; then + echo DOCKER start finished in $2 seconds + return ${SUCCESS} + fi + return ${FAILURE} +} +# + +function probe_test_docker { + # This expected logging should be output by startup.sh of the + # respective test docker container + MATCH=`docker logs --tail 30 $1 | grep "Startup completed successfully"` + echo MATCH is -- ${MATCH} + + if [ -n "$MATCH" ] ; then + echo TEST DOCKER start finished in $2 seconds + return ${SUCCESS} + fi + return ${FAILURE} +} +# + + +function probe_sim { + if lsof -Pi :8285 -sTCP:LISTEN -t >/dev/null ; then + echo "Already running" + return ${SUCCESS} + else + echo "Not running" + return ${FAILURE} + fi +} +# + + +function monitor_docker { + DOCKER_NAME=$1 + echo "Monitor ${DOCKER_NAME} Docker" + sleep 5 + TIME_OUT=900 + INTERVAL=20 + TIME=0 + + while [ "$TIME" -lt "$TIME_OUT" ]; do + + case ${DOCKER_NAME} in + + sdc-cs) + ready_probe ${DOCKER_NAME} ${TIME} ; + status=$? ; + ;; + sdc-BE) + ready_probe_jetty ${DOCKER_NAME} ${TIME} ; + status=$? ; + ;; + sdc-FE) + ready_probe_jetty ${DOCKER_NAME} ${TIME} ; + status=$? ; + ;; + sdc-onboard-BE) + ready_probe_jetty ${DOCKER_NAME} ${TIME} ; + status=$? ; + ;; + sdc-api-tests) + probe_test_docker ${DOCKER_NAME} ${TIME}; + status=$? ; + ;; + sdc-ui-tests) + probe_test_docker ${DOCKER_NAME} ${TIME}; + status=$? ; + ;; + *) + probe_docker ${DOCKER_NAME} ${TIME}; + status=$? ; + ;; + + esac + + if [[ ${status} == ${SUCCESS} ]] ; then + break; + fi + + echo "Sleep: ${INTERVAL} seconds before testing if ${DOCKER_NAME} DOCKER is up. Total wait time up now is: ${TIME} seconds. Timeout is: ${TIME_OUT} seconds" + sleep ${INTERVAL} + TIME=$(($TIME+$INTERVAL)) + done + + docker_logs ${DOCKER_NAME} + + if [ "$TIME" -ge "$TIME_OUT" ]; then + echo -e "\e[1;31mTIME OUT: DOCKER was NOT fully started in $TIME_OUT seconds... Could cause problems ...\e[0m" + fi +} +# + +# healthCheck script used the secure connection to send request (https is always turn on) +function healthCheck { + + echo "BE Health Check:" + curl -k --noproxy "*" https://${IP}:8443/sdc2/rest/healthCheck + + echo "" + echo "" + echo "FE Health Check:" + curl -k --noproxy "*" https://${IP}:9443/sdc1/rest/healthCheck + + + echo "" + echo "" + healthCheck_http_code=$(curl -k --noproxy "*" -o /dev/null -w '%{http_code}' -H "Accept: application/json" -H "Content-Type: application/json" -H "USER_ID: jh0003" https://${IP}:8443/sdc2/rest/v1/user/demo;) + if [[ ${healthCheck_http_code} != 200 ]]; then + echo "Error [${healthCheck_http_code}] while checking existence of user" + return ${healthCheck_http_code} + fi + echo "check user existence: OK" + return ${healthCheck_http_code} +} +# + + +function command_exit_status { + status=$1 + docker=$2 + if [ "${status}" != "0" ] ; then + echo "[ ERROR ] Docker ${docker} run command exit with status [${status}]" + fi +} +# + + +# +# Run Containers +# + + +#Cassandra +function sdc-cs { + DOCKER_NAME="sdc-cs" + echo "docker run sdc-cassandra..." + if [ ${LOCAL} = false ]; then + docker pull ${PREFIX}/sdc-cassandra:${RELEASE} + fi + docker run -dit --name ${DOCKER_NAME} --env RELEASE="${RELEASE}" --env CS_PASSWORD="${CS_PASSWORD}" --env ENVNAME="${DEP_ENV}" --env HOST_IP=${IP} --env MAX_HEAP_SIZE="1536M" --env HEAP_NEWSIZE="512M" --log-driver=json-file --log-opt max-size=100m --log-opt max-file=10 --ulimit memlock=-1:-1 --ulimit nofile=4096:100000 ${LOCAL_TIME_MOUNT_CMD} --volume ${WORKSPACE}/data/CS:/var/lib/cassandra --volume ${WORKSPACE}/data/environments:/root/chef-solo/environments --publish ${CS_PORT}:${CS_PORT} ${PREFIX}/sdc-cassandra:${RELEASE} /bin/sh + command_exit_status $? ${DOCKER_NAME} + echo "please wait while CS is starting..." + monitor_docker ${DOCKER_NAME} +} +# + + +#Cassandra-init +function sdc-cs-init { + DOCKER_NAME="sdc-cs-init" + echo "docker run sdc-cassandra-init..." + if [ ${LOCAL} = false ]; then + docker pull ${PREFIX}/sdc-cassandra-init:${RELEASE} + fi + docker run --name ${DOCKER_NAME} --env RELEASE="${RELEASE}" --env SDC_USER="${SDC_USER}" --env SDC_PASSWORD="${SDC_PASSWORD}" --env CS_PASSWORD="${CS_PASSWORD}" --env ENVNAME="${DEP_ENV}" --env HOST_IP=${IP} --log-driver=json-file --log-opt max-size=100m --log-opt max-file=10 --ulimit memlock=-1:-1 --ulimit nofile=4096:100000 ${LOCAL_TIME_MOUNT_CMD} --volume ${WORKSPACE}/data/CS:/var/lib/cassandra --volume ${WORKSPACE}/data/environments:/home/sdc/chef-solo/environments --volume ${WORKSPACE}/data/CS-Init:/root/chef-solo/cache ${PREFIX}/sdc-cassandra-init:${RELEASE} > /dev/null 2>&1 + rc=$? + docker_logs ${DOCKER_NAME} + if [[ ${rc} != 0 ]]; then exit ${rc}; fi +} +# + + +#Onboard Cassandra-init +function sdc-cs-onboard-init { + DOCKER_NAME="sdc-cs-onboard-init" + echo "docker run sdc-cs-onboard-init..." + if [ ${LOCAL} = false ]; then + docker pull ${PREFIX}/sdc-onboard-cassandra-init:${RELEASE} + fi + docker run --name ${DOCKER_NAME} --env RELEASE="${RELEASE}" --env CS_HOST_IP=${IP} --env CS_HOST_PORT=${CS_PORT} --env SDC_USER="${SDC_USER}" --env SDC_PASSWORD="${SDC_PASSWORD}" --env CS_PASSWORD="${CS_PASSWORD}" --env ENVNAME="${DEP_ENV}" --log-driver=json-file --log-opt max-size=100m --log-opt max-file=10 --ulimit memlock=-1:-1 --ulimit nofile=4096:100000 ${LOCAL_TIME_MOUNT_CMD} --volume ${WORKSPACE}/data/CS:/var/lib/cassandra --volume ${WORKSPACE}/data/environments:/home/sdc/chef-solo/environments --volume ${WORKSPACE}/data/CS-Init:/home/sdc/chef-solo/cache ${PREFIX}/sdc-onboard-cassandra-init:${RELEASE} + rc=$? + docker_logs ${DOCKER_NAME} + if [[ ${rc} != 0 ]]; then exit ${rc}; fi +} +# + + +#Back-End +function sdc-BE { + DOCKER_NAME="sdc-BE" + echo "docker run sdc-backend..." + if [ ${LOCAL} = false ]; then + docker pull ${PREFIX}/sdc-backend:${RELEASE} + else + ADDITIONAL_ARGUMENTS=${BE_DEBUG_PORT} + fi + docker run --detach --name ${DOCKER_NAME} --env HOST_IP=${IP} --env ENVNAME="${DEP_ENV}" --env cassandra_ssl_enabled="false" --env JAVA_OPTIONS="${BE_JAVA_OPTIONS}" --log-driver=json-file --log-opt max-size=100m --log-opt max-file=10 --ulimit memlock=-1:-1 --ulimit nofile=4096:100000 ${LOCAL_TIME_MOUNT_CMD} --volume ${WORKSPACE}/data/logs/BE/:${JETTY_BASE}/logs --volume ${WORKSPACE}/data/environments:${JETTY_BASE}/chef-solo/environments --publish 8443:8443 --publish 8080:8080 ${ADDITIONAL_ARGUMENTS} ${PREFIX}/sdc-backend:${RELEASE} + command_exit_status $? ${DOCKER_NAME} + echo "please wait while BE is starting..." + monitor_docker ${DOCKER_NAME} +} +# + + +# Back-End-Init +function sdc-BE-init { + DOCKER_NAME="sdc-BE-init" + echo "docker run sdc-backend-init..." + if [ ${LOCAL} = false ]; then + docker pull ${PREFIX}/sdc-backend-init:${RELEASE} + fi + docker run --name ${DOCKER_NAME} --env HOST_IP=${IP} --env ENVNAME="${DEP_ENV}" --log-driver=json-file --log-opt max-size=100m --log-opt max-file=10 --ulimit memlock=-1:-1 --ulimit nofile=4096:100000 ${LOCAL_TIME_MOUNT_CMD} --volume ${WORKSPACE}/data/logs/BE/:${JETTY_BASE}/logs --volume ${WORKSPACE}/data/environments:/home/sdc/chef-solo/environments ${PREFIX}/sdc-backend-init:${RELEASE} > /dev/null 2>&1 + rc=$? + docker_logs ${DOCKER_NAME} + if [[ ${rc} != 0 ]]; then exit ${rc}; fi +} +# + + +# Onboard Back-End +function sdc-onboard-BE { + DOCKER_NAME="sdc-onboard-BE" + echo "docker run sdc-onboard-BE ..." +# TODO Check the dir_perms action . do we need it here ?? +# dir_perms + if [ ${LOCAL} = false ]; then + docker pull ${PREFIX}/sdc-onboard-backend:${RELEASE} + else + ADDITIONAL_ARGUMENTS=${ONBOARD_DEBUG_PORT} + fi + docker run --detach --name ${DOCKER_NAME} --env HOST_IP=${IP} --env ENVNAME="${DEP_ENV}" --env cassandra_ssl_enabled="false" --env SDC_CLUSTER_NAME="SDC-CS-${DEP_ENV}" --env SDC_USER="${SDC_USER}" --env SDC_PASSWORD="${SDC_PASSWORD}" --env SDC_CERT_DIR="${SDC_CERT_DIR}" --env JAVA_OPTIONS="${ONBOARD_BE_JAVA_OPTIONS}" --log-driver=json-file --log-opt max-size=100m --log-opt max-file=10 --ulimit memlock=-1:-1 --ulimit nofile=4096:100000 ${LOCAL_TIME_MOUNT_CMD} --volume ${WORKSPACE}/data/${SDC_CERT_DIR}:${JETTY_BASE}/onap/cert --volume ${WORKSPACE}/data/logs/ONBOARD:${JETTY_BASE}/logs --volume ${WORKSPACE}/data/environments:/${JETTY_BASE}/chef-solo/environments --publish 8445:8445 --publish 8081:8081 ${ADDITIONAL_ARGUMENTS} ${PREFIX}/sdc-onboard-backend:${RELEASE} + command_exit_status $? ${DOCKER_NAME} + echo "please wait while sdc-onboard-BE is starting..." + monitor_docker ${DOCKER_NAME} +} +# + + +# Front-End +function sdc-FE { + DOCKER_NAME="sdc-FE" + IMAGE_NAME="${PREFIX}/sdc-frontend:${RELEASE}" + echo "Running container '${DOCKER_NAME}' based on '${IMAGE_NAME}' image..." + if [ ${LOCAL} = false ]; then + docker pull ${PREFIX}/sdc-frontend:${RELEASE} + else + ADDITIONAL_ARGUMENTS=${FE_DEBUG_PORT} + fi + + PLUGIN_CONFIG_FILE="${WORKSPACE}/data/environments/plugins-configuration.yaml" + if [[ -f ${WORKSPACE}/data/environments/plugins-configuration.yaml ]]; then + PLUGINS_CONF_VOLUME_MOUNT="--volume ${PLUGIN_CONFIG_FILE}:${JETTY_BASE}/config/catalog-fe/plugins-configuration.yaml" + else + echo "INFO: '${PLUGIN_CONFIG_FILE}' not provided, ignoring..." + fi + + docker run \ + --detach \ + --name ${DOCKER_NAME} \ + --env HOST_IP=${IP} \ + --env ENVNAME="${DEP_ENV}" \ + --env JAVA_OPTIONS="${FE_JAVA_OPTIONS}" \ + --log-driver=json-file \ + --log-opt max-size=100m \ + --log-opt max-file=10 \ + --ulimit memlock=-1:-1 \ + --ulimit nofile=4096:100000 \ + --volume ${WORKSPACE}/data/logs/FE/:${JETTY_BASE}/logs \ + --volume ${WORKSPACE}/data/environments:/${JETTY_BASE}/chef-solo/environments \ + ${LOCAL_TIME_MOUNT_CMD} \ + ${PLUGINS_CONF_VOLUME_MOUNT} \ + --publish 9443:9443 \ + --publish 8181:8181 \ + ${ADDITIONAL_ARGUMENTS} \ + ${IMAGE_NAME} + + command_exit_status $? ${DOCKER_NAME} + echo "Please wait while '${DOCKER_NAME}' container is starting..." + monitor_docker ${DOCKER_NAME} +} +# + + +# apis-sanity +function sdc-api-tests { + if [[ ${RUN_API_TESTS} = true ]] ; then + healthCheck + healthCheck_http_code=$? + if [[ ${healthCheck_http_code} == 200 ]] ; then + echo "docker run sdc-api-tests..." + echo "Trigger sdc-api-tests docker, please wait..." + + if [ ${LOCAL} = false ]; then + docker pull ${PREFIX}/sdc-api-tests:${RELEASE} + fi + + docker run --detach --name sdc-api-tests --env HOST_IP=${IP} --env ENVNAME="${DEP_ENV}" --env JAVA_OPTIONS="${API_TESTS_JAVA_OPTIONS}" --env SUITE_NAME=${API_SUITE} --log-driver=json-file --log-opt max-size=100m --log-opt max-file=10 --ulimit memlock=-1:-1 --ulimit nofile=4096:100000 ${LOCAL_TIME_MOUNT_CMD} --volume ${WORKSPACE}/data/logs/sdc-api-tests/target:/var/lib/tests/target --volume ${WORKSPACE}/data/logs/sdc-api-tests/ExtentReport:/var/lib/tests/ExtentReport --volume ${WORKSPACE}/data/logs/sdc-api-tests/outputCsar:/var/lib/tests/outputCsar --volume ${WORKSPACE}/data/environments:/root/chef-solo/environments --volume ${WORKSPACE}/data/${SDC_CERT_DIR}:/var/lib/tests/cert --publish 9560:9560 ${PREFIX}/sdc-api-tests:${RELEASE} echo "please wait while SDC-API-TESTS is starting....." + monitor_docker sdc-api-tests + fi + fi +} +# + + +# ui-sanity +function sdc-ui-tests { + + if [[ ${RUN_UI_TESTS} = true ]] ; then + healthCheck + healthCheck_http_code=$? + if [[ ${healthCheck_http_code} == 200 ]]; then + echo "docker run sdc-ui-tets..." + echo "Trigger sdc-ui-tests docker, please wait..." + + if [ ${LOCAL} = false ]; then + docker pull ${PREFIX}/sdc-ui-tests:${RELEASE} + fi + RUN_SIMULATOR=true; + sdc-sim + docker run --detach --name sdc-ui-tests --env HOST_IP=${IP} --env ENVNAME="${DEP_ENV}" --env JAVA_OPTIONS="${UI_TESTS_JAVA_OPTIONS}" --env SUITE_NAME=${UI_SUITE} --log-driver=json-file --log-opt max-size=100m --log-opt max-file=10 --ulimit memlock=-1:-1 --ulimit nofile=4096:100000 ${LOCAL_TIME_MOUNT_CMD} --volume ${WORKSPACE}/data/logs/sdc-ui-tests/target:/var/lib/tests/target --volume ${WORKSPACE}/data/logs/sdc-ui-tests/ExtentReport:/var/lib/tests/ExtentReport --volume ${WORKSPACE}/data/environments:/root/chef-solo/environments --publish 5901:5901 --publish 6901:6901 ${PREFIX}/sdc-ui-tests:${RELEASE} + echo "please wait while SDC-UI-TESTS is starting....." + monitor_docker sdc-ui-tests + fi + fi +} +# + + +# SDC-Simulator +function sdc-sim { + if [ "${RUN_SIMULATOR}" == true ]; then + echo "docker run sdc-webSimulator..." + if [ ${LOCAL} = false ]; then + docker pull ${PREFIX}/sdc-simulator:${RELEASE} + fi + + probe_sim + sim_stat=$? + if [ ${sim_stat} == 1 ]; then + docker run \ + --detach \ + --name sdc-sim \ + --env FE_URL="${FE_URL}" \ + --env JAVA_OPTIONS="${SIM_JAVA_OPTIONS}" \ + --env ENVNAME="${DEP_ENV}" \ + ${LOCAL_TIME_MOUNT_CMD} \ + --volume ${WORKSPACE}/data/logs/WS/:${JETTY_BASE}/logs \ + --volume ${WORKSPACE}/data/environments:/root/chef-solo/environments \ + --publish 8285:8080 \ + --publish 8286:8443 ${PREFIX}/sdc-simulator:${RELEASE} + echo "please wait while SDC-WEB-SIMULATOR is starting....." + monitor_docker sdc-sim + fi + fi +} +# + + +# +# Main +# + +# Handle command line arguments +while [ $# -gt 0 ]; do + case $1 in + + # -r | --release - The specific docker version to pull and deploy + -r | --release ) + shift 1 ; + RELEASE=$1; + shift 1;; + + # -e | --environment - The environment name you want to deploy + -e | --environment ) + shift 1; + DEP_ENV=$1; + shift 1 ;; + + # -p | --port - The port from which to connect to the docker nexus + -p | --port ) + shift 1 ; + PORT=$1; + shift 1 ;; + + # -l | --local - Use this for deploying your local dockers without pulling them first + -l | --local ) + LOCAL=true; + shift 1;; + + # -ta - Use this for running the APIs sanity docker after all other dockers have been deployed + -ta ) + shift 1 ; + API_SUITE=$1; + RUN_API_TESTS=true; + shift 1 ;; + + # -tu - Use this for running the UI sanity docker after all other dockers have been deployed + -tu ) + shift 1 ; + UI_SUITE=$1; + RUN_UI_TESTS=true; + shift 1 ;; + + # -tad - Use this for running the DEFAULT suite of tests in APIs sanity docker after all other dockers have been deployed + -tad | -t ) + API_SUITE="onapApiSanity"; + RUN_API_TESTS=true; + shift 1 ;; + + # -tud - Use this for running the DEFAULT suite of tests in UI sanity docker after all other dockers have been deployed + -tud ) + UI_SUITE="onapUiSanity"; + RUN_UI_TESTS=true; + shift 1 ;; + + # -d | --docker - The init specified docker + -d | --docker ) + shift 1 ; + DOCKER=$1; + shift 1 ;; + # -sim | --simulator run the simulator + -sim | --simulator ) + RUN_SIMULATOR=true; + shift 1 ;; + # -sim | --simulator run the simulator + -u | --fe_url ) + shift 1 ; + FE_URL=$1; + shift 1 ;; + + # -h | --help - Display the help message with all the available run options + -h | --help ) + usage; + exit ${SUCCESS};; + + * ) + usage; + exit ${FAILURE};; + esac +done + + +#Prefix those with WORKSPACE so it can be set to something other than /opt +[ -f ${WORKSPACE}/opt/config/env_name.txt ] && DEP_ENV=$(cat ${WORKSPACE}/opt/config/env_name.txt) || echo ${DEP_ENV} +[ -f ${WORKSPACE}/opt/config/nexus_username.txt ] && NEXUS_USERNAME=$(cat ${WORKSPACE}/opt/config/nexus_username.txt) || NEXUS_USERNAME=release +[ -f ${WORKSPACE}/opt/config/nexus_password.txt ] && NEXUS_PASSWD=$(cat ${WORKSPACE}/opt/config/nexus_password.txt) || NEXUS_PASSWD=sfWU3DFVdBr7GVxB85mTYgAW +[ -f ${WORKSPACE}/opt/config/nexus_docker_repo.txt ] && NEXUS_DOCKER_REPO=$(cat ${WORKSPACE}/opt/config/nexus_docker_repo.txt) || NEXUS_DOCKER_REPO=nexus3.onap.org:${PORT} +[ -f ${WORKSPACE}/opt/config/nexus_username.txt ] && docker login -u $NEXUS_USERNAME -p $NEXUS_PASSWD $NEXUS_DOCKER_REPO + + +export IP=`ip route get 8.8.8.8 | awk '/src/{ print $7 }'` +#If OSX, then use this to get IP +if [[ "$OSTYPE" == "darwin"* ]]; then + export IP=$(ipconfig getifaddr en0) +fi +export PREFIX=${NEXUS_DOCKER_REPO}'/onap' + +if [ ${LOCAL} = true ]; then + PREFIX='onap' +fi + +echo "" + +if [ -z "${DOCKER}" ]; then + cleanup all + dir_perms + sdc-cs + sdc-cs-init + sdc-cs-onboard-init + sdc-onboard-BE + sdc-BE + sdc-BE-init + sdc-FE + healthCheck + sdc-sim + sdc-api-tests + sdc-ui-tests +else + cleanup ${DOCKER} + dir_perms + ${DOCKER} + healthCheck +fi diff --git a/scripts/sdc-workflow-d/sdc_workflow_d.sh b/scripts/sdc-workflow-d/sdc_workflow_d.sh new file mode 100644 index 00000000..6432090f --- /dev/null +++ b/scripts/sdc-workflow-d/sdc_workflow_d.sh @@ -0,0 +1,117 @@ +#!/bin/bash +# +# Copyright 2019 © Samsung Electronics Co., Ltd. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# +echo "[INFO] This is sdc_workflow_d.sh" +# run sdc deployment +source "${WORKSPACE}/scripts/sdc/setup_sdc_for_sanity.sh" +export ROBOT_VARIABLES + +# fail quick if error +set -exo pipefail + +export ENV_NAME='CSIT' + +function iss_true { + _value=$(eval echo "\$${1}" | tr '[:upper:]' '[:lower:]') + + case "$_value" in + 1|yes|true|Y) + return 0 + ;; + esac + + return 1 +} + +# returns 0: if SDC_LOCAL_IMAGES is set to true value +# returns 1: otherwise +function using_local_workflow_images { + iss_true WORKFLOW_LOCAL_IMAGES +} + +# cloning workflow directory +mkdir -p "${WORKSPACE}/data/clone/" +cd "${WORKSPACE}/data/clone" +if using_local_workflow_images && [ -n "$WORKFLOW_LOCAL_GITREPO" ] ; then + WORKFLOW_LOCAL_GITREPO=$(realpath "$WORKFLOW_LOCAL_GITREPO") + if [ -d "$WORKFLOW_LOCAL_GITREPO" ] ; then + rm -rf ./workflow + cp -a "$WORKFLOW_LOCAL_GITREPO" ./workflow + else + echo "[ERROR]: Local git repo for workflow does not exist: ${WORKFLOW_LOCAL_GITREPO}" + exit 1 + fi +else + git clone --depth 1 "https://github.com/onap/sdc-sdc-workflow-designer.git" -b ${GERRIT_BRANCH} +fi +# set enviroment variables +source ${WORKSPACE}/data/clone/workflow/version.properties +export WORKFLOW_RELEASE=$major.$minor-STAGING-latest + +SDC_CS=$(docker inspect -f '{{range .NetworkSettings.Networks}}{{.IPAddress}}{{end}}' sdc-cs) +SDC_BE=$(docker inspect -f '{{range .NetworkSettings.Networks}}{{.IPAddress}}{{end}}' sdc-BE) +echo "[INFO] Initialization of workflow init" +echo ${SDC_CS} +echo ${SDC_BE} +docker run -ti \ + -e "CS_HOST=${SDC_CS}" \ + -e "CS_PORT=9042" \ + -e "CS_AUTHENTICATE=true"\ + -e "CS_USER=asdc_user" \ + -e "CS_PASSWORD=Aa1234%^!" nexus3.onap.org:10001/onap/sdc-workflow-init:latest + +echo "[INFO] Initialization of workflow Backend init" +docker run -d --name "workflow-backend" -e "SDC_PROTOCOL=http" \ + -e "SDC_ENDPOINT=${SDC_BE}:8080" \ + -e "SDC_USER=workflow" \ + -e "SDC_PASSWORD=Kp8bJ4SXszM0WXlhak3eHlcse2gAw84vaoGGmJvUy2U" \ + -e "CS_HOSTS=${SDC_CS}" \ + -e "CS_PORT=9042" \ + -e "CS_AUTHENTICATE=true"\ + -e "CS_USER=asdc_user" \ + -e "CS_PASSWORD=Aa1234%^!" \ + -e "CS_SSL_ENABLED=false"\ + -e "SERVER_SSL_ENABLED=false" \ + --env JAVA_OPTIONS="${BE_JAVA_OPTIONS}" --publish 8384:8080 --publish 10443:8443 --publish 8000:8000 nexus3.onap.org:10001/onap/sdc-workflow-backend:latest + +WORKFLOW_BE=$(docker inspect -f '{{range .NetworkSettings.Networks}}{{.IPAddress}}{{end}}' workflow-backend) +echo "[INFO] starting workflow designer fronend" +docker run -d --name "workflow-frontend" \ + -e BACKEND="http://${WORKFLOW_BE}:8080"\ + --publish 8484:8080 --publish 11443:8443 nexus3.onap.org:10001/onap/sdc-workflow-frontend:latest + +cp "${WORKSPACE}/data/clone/sdc/sdc-os-chef/environments/plugins-configuration.yaml" \ + "${WORKSPACE}/data/environments/plugins-configuration.yaml" + +WF_IP=$(docker inspect -f '{{range .NetworkSettings.Networks}}{{.IPAddress}}{{end}}' workflow-frontend) +WFADDR="http:\/\/${WF_IP}:8080\/workflows" +echo ${WFADDR} +sed -i \ + -e "s/<%= @workflow_discovery_url %>/${WFADDR}/g" \ + -e "s/<%= @workflow_source_url %>/${WFADDR}/g" \ + "${WORKSPACE}/data/environments/plugins-configuration.yaml" + +cp "${WORKSPACE}/data/clone/sdc/sdc-os-chef/scripts/docker_run.sh" "${WORKSPACE}/scripts/sdc-workflow-d/" + +echo "[INFO] restarting sdc-FE with updated plugin configuration file with Worflow host ip" +docker stop sdc-FE +"${WORKSPACE}/scripts/sdc-workflow-d/docker_run.sh" \ + --local \ + -e "${ENV_NAME}" \ + -p 10001 -d sdc-FE +# This file is sourced in another script which is out of our control... +set +e +set +o pipefail diff --git a/scripts/sdc/docker_run.sh b/scripts/sdc/docker_run.sh new file mode 100755 index 00000000..e641594f --- /dev/null +++ b/scripts/sdc/docker_run.sh @@ -0,0 +1,606 @@ +#!/bin/bash + +# +# Constants: +# + +WORKSPACE="${WORKSPACE:-}" +SUCCESS=0 +FAILURE=1 + +CS_PASSWORD="onap123#@!" +SDC_USER="asdc_user" +SDC_PASSWORD="Aa1234%^!" + +JETTY_BASE="/var/lib/jetty" +SDC_CERT_DIR="onap/cert" + +RELEASE=latest + +LOCAL=false +BE_DEBUG_PORT="--publish 4000:4000" +FE_DEBUG_PORT="--publish 6000:6000" +ONBOARD_DEBUG_PORT="--publish 4001:4001" +CS_PORT=${CS_PORT:-9042} + + +# Java Options: +BE_JAVA_OPTIONS="-Xdebug -agentlib:jdwp=transport=dt_socket,address=4000,server=y,suspend=n -Xmx1536m -Xms1536m" +FE_JAVA_OPTIONS="-Xdebug -agentlib:jdwp=transport=dt_socket,address=6000,server=y,suspend=n -Xmx256m -Xms256m" +ONBOARD_BE_JAVA_OPTIONS="-Xdebug -agentlib:jdwp=transport=dt_socket,address=4001,server=y,suspend=n -Xmx1g -Xms1g" +SIM_JAVA_OPTIONS=" -Xmx128m -Xms128m -Xss1m -Dlog4j.configuration=file:///${JETTY_BASE}/config/sdc-simulator/log4j.properties" +API_TESTS_JAVA_OPTIONS="-Xmx512m -Xms512m" +UI_TESTS_JAVA_OPTIONS="-Xmx1024m -Xms1024m" +#Define this as variable, so it can be excluded in run commands on Docker for OSX, as /etc/localtime cant be mounted there. +LOCAL_TIME_MOUNT_CMD="--volume /etc/localtime:/etc/localtime:ro" +# If os is OSX, unset this, so /etc/localtime is not mounted, otherwise leave it be +if [[ "$OSTYPE" == "darwin"* ]]; then + LOCAL_TIME_MOUNT_CMD="" +fi + + +# +# Functions: +# + + +function usage { + echo "usage: docker_run.sh [ -r|--release <RELEASE-NAME> ] [ -e|--environment <ENV-NAME> ] [ -p|--port <Docker-hub-port>] [ -l|--local <Run-without-pull>] [ -sim|--simulator <Run-with-simulator>] [ -ta <run api tests with the supplied test suit>] [ -tu <run ui tests with the supplied test suit>] [ -ta <run api tests with the supplied test suit>] [ -tu <run ui tests with the supplied test suit>] [ -tad <run api tests with the default test suit>] [ -tu <run ui tests with the default test suit>] [ -h|--help ]" + echo "start dockers built locally example: docker_run.sh -l" + echo "start dockers built locally and simulator example: docker_run.sh -l -sim" + echo "start dockers, pull from onap nexus according to release and simulator example: docker_run.sh -r 1.5-STAGING-latest -sim" + echo "start dockers built locally and run api tests docker example: docker_run.sh -l -tad" + echo "start dockers built locally and run only the catalog be example: docker_run.sh -l -d sdc-BE " +} +# + + +function cleanup { + echo "Performing old dockers cleanup" + + if [ "$1" == "all" ] ; then + docker_ids=`docker ps -a | egrep "ecomp-nexus:${PORT}/sdc|sdc|Exit}|dcae" | awk '{print $1}'` + for X in ${docker_ids} + do + docker rm -f ${X} + done + else + echo "performing $1 docker cleanup" + tmp=`docker ps -a -q --filter="name=$1"` + if [[ ! -z "$tmp" ]]; then + docker rm -f ${tmp} + fi + fi +} +# + + +function dir_perms { + mkdir -p ${WORKSPACE}/data/logs/BE/SDC/SDC-BE + mkdir -p ${WORKSPACE}/data/logs/FE/SDC/SDC-FE + mkdir -p ${WORKSPACE}/data/logs/sdc-api-tests/ExtentReport + mkdir -p ${WORKSPACE}/data/logs/ONBOARD/SDC/ONBOARD-BE + mkdir -p ${WORKSPACE}/data/logs/sdc-api-tests/target + mkdir -p ${WORKSPACE}/data/logs/sdc-ui-tests/ExtentReport + mkdir -p ${WORKSPACE}/data/logs/sdc-ui-tests/target + mkdir -p ${WORKSPACE}/data/logs/docker_logs + mkdir -p ${WORKSPACE}/data/logs/WS + echo "Creating dir '${WORKSPACE}/data/${SDC_CERT_DIR}'" + mkdir -p ${WORKSPACE}/data/${SDC_CERT_DIR} + chmod -R 777 ${WORKSPACE}/data/logs +} +# + + +function docker_logs { + docker logs $1 > ${WORKSPACE}/data/logs/docker_logs/$1_docker.log +} +# + + +# +# Readiness Prob +# + +function ready_probe { + docker exec $1 /var/lib/ready-probe.sh > /dev/null 2>&1 + rc=$? + if [[ ${rc} == 0 ]]; then + echo DOCKER $1 start finished in $2 seconds + return ${SUCCESS} + fi + return ${FAILURE} +} +# + +function ready_probe_jetty { + docker exec $1 /var/lib/jetty/ready-probe.sh > /dev/null 2>&1 + rc=$? + if [[ ${rc} == 0 ]]; then + echo DOCKER $1 start finished in $2 seconds + return ${SUCCESS} + fi + return ${FAILURE} +} +# + +function probe_docker { + MATCH=`docker logs --tail 30 $1 | grep "DOCKER STARTED"` + echo MATCH is -- ${MATCH} + + if [ -n "$MATCH" ] ; then + echo DOCKER start finished in $2 seconds + return ${SUCCESS} + fi + return ${FAILURE} +} +# + +function probe_test_docker { + # This expected logging should be output by startup.sh of the + # respective test docker container + MATCH=`docker logs --tail 30 $1 | grep "Startup completed successfully"` + echo MATCH is -- ${MATCH} + + if [ -n "$MATCH" ] ; then + echo TEST DOCKER start finished in $2 seconds + return ${SUCCESS} + fi + return ${FAILURE} +} +# + + +function probe_sim { + if lsof -Pi :8285 -sTCP:LISTEN -t >/dev/null ; then + echo "Already running" + return ${SUCCESS} + else + echo "Not running" + return ${FAILURE} + fi +} +# + + +function monitor_docker { + DOCKER_NAME=$1 + echo "Monitor ${DOCKER_NAME} Docker" + sleep 5 + TIME_OUT=900 + INTERVAL=20 + TIME=0 + + while [ "$TIME" -lt "$TIME_OUT" ]; do + + case ${DOCKER_NAME} in + + sdc-cs) + ready_probe ${DOCKER_NAME} ${TIME} ; + status=$? ; + ;; + sdc-BE) + ready_probe_jetty ${DOCKER_NAME} ${TIME} ; + status=$? ; + ;; + sdc-FE) + ready_probe_jetty ${DOCKER_NAME} ${TIME} ; + status=$? ; + ;; + sdc-onboard-BE) + ready_probe_jetty ${DOCKER_NAME} ${TIME} ; + status=$? ; + ;; + sdc-api-tests) + probe_test_docker ${DOCKER_NAME} ${TIME}; + status=$? ; + ;; + sdc-ui-tests) + probe_test_docker ${DOCKER_NAME} ${TIME}; + status=$? ; + ;; + *) + probe_docker ${DOCKER_NAME} ${TIME}; + status=$? ; + ;; + + esac + + if [[ ${status} == ${SUCCESS} ]] ; then + break; + fi + + echo "Sleep: ${INTERVAL} seconds before testing if ${DOCKER_NAME} DOCKER is up. Total wait time up now is: ${TIME} seconds. Timeout is: ${TIME_OUT} seconds" + sleep ${INTERVAL} + TIME=$(($TIME+$INTERVAL)) + done + + docker_logs ${DOCKER_NAME} + + if [ "$TIME" -ge "$TIME_OUT" ]; then + echo -e "\e[1;31mTIME OUT: DOCKER was NOT fully started in $TIME_OUT seconds... Could cause problems ...\e[0m" + fi +} +# + +# healthCheck script used the secure connection to send request (https is always turn on) +function healthCheck { + + echo "BE Health Check:" + curl -k --noproxy "*" https://${IP}:8443/sdc2/rest/healthCheck + + echo "" + echo "" + echo "FE Health Check:" + curl -k --noproxy "*" https://${IP}:9443/sdc1/rest/healthCheck + + + echo "" + echo "" + healthCheck_http_code=$(curl -k --noproxy "*" -o /dev/null -w '%{http_code}' -H "Accept: application/json" -H "Content-Type: application/json" -H "USER_ID: jh0003" https://${IP}:8443/sdc2/rest/v1/user/demo;) + if [[ ${healthCheck_http_code} != 200 ]]; then + echo "Error [${healthCheck_http_code}] while checking existence of user" + return ${healthCheck_http_code} + fi + echo "check user existence: OK" + return ${healthCheck_http_code} +} +# + + +function command_exit_status { + status=$1 + docker=$2 + if [ "${status}" != "0" ] ; then + echo "[ ERROR ] Docker ${docker} run command exit with status [${status}]" + fi +} +# + + +# +# Run Containers +# + + +#Cassandra +function sdc-cs { + DOCKER_NAME="sdc-cs" + echo "docker run sdc-cassandra..." + if [ ${LOCAL} = false ]; then + docker pull ${PREFIX}/sdc-cassandra:${RELEASE} + fi + docker run -dit --name ${DOCKER_NAME} --env RELEASE="${RELEASE}" --env CS_PASSWORD="${CS_PASSWORD}" --env ENVNAME="${DEP_ENV}" --env HOST_IP=${IP} --env MAX_HEAP_SIZE="1536M" --env HEAP_NEWSIZE="512M" --log-driver=json-file --log-opt max-size=100m --log-opt max-file=10 --ulimit memlock=-1:-1 --ulimit nofile=4096:100000 ${LOCAL_TIME_MOUNT_CMD} --volume ${WORKSPACE}/data/CS:/var/lib/cassandra --volume ${WORKSPACE}/data/environments:/root/chef-solo/environments --publish ${CS_PORT}:${CS_PORT} ${PREFIX}/sdc-cassandra:${RELEASE} /bin/sh + command_exit_status $? ${DOCKER_NAME} + echo "please wait while CS is starting..." + monitor_docker ${DOCKER_NAME} +} +# + + +#Cassandra-init +function sdc-cs-init { + DOCKER_NAME="sdc-cs-init" + echo "docker run sdc-cassandra-init..." + if [ ${LOCAL} = false ]; then + docker pull ${PREFIX}/sdc-cassandra-init:${RELEASE} + fi + docker run --name ${DOCKER_NAME} --env RELEASE="${RELEASE}" --env SDC_USER="${SDC_USER}" --env SDC_PASSWORD="${SDC_PASSWORD}" --env CS_PASSWORD="${CS_PASSWORD}" --env ENVNAME="${DEP_ENV}" --env HOST_IP=${IP} --log-driver=json-file --log-opt max-size=100m --log-opt max-file=10 --ulimit memlock=-1:-1 --ulimit nofile=4096:100000 ${LOCAL_TIME_MOUNT_CMD} --volume ${WORKSPACE}/data/CS:/var/lib/cassandra --volume ${WORKSPACE}/data/environments:/home/sdc/chef-solo/environments --volume ${WORKSPACE}/data/CS-Init:/root/chef-solo/cache ${PREFIX}/sdc-cassandra-init:${RELEASE} > /dev/null 2>&1 + rc=$? + docker_logs ${DOCKER_NAME} + if [[ ${rc} != 0 ]]; then exit ${rc}; fi +} +# + + +#Onboard Cassandra-init +function sdc-cs-onboard-init { + DOCKER_NAME="sdc-cs-onboard-init" + echo "docker run sdc-cs-onboard-init..." + if [ ${LOCAL} = false ]; then + docker pull ${PREFIX}/sdc-onboard-cassandra-init:${RELEASE} + fi + docker run --name ${DOCKER_NAME} --env RELEASE="${RELEASE}" --env CS_HOST_IP=${IP} --env CS_HOST_PORT=${CS_PORT} --env SDC_USER="${SDC_USER}" --env SDC_PASSWORD="${SDC_PASSWORD}" --env CS_PASSWORD="${CS_PASSWORD}" --env ENVNAME="${DEP_ENV}" --log-driver=json-file --log-opt max-size=100m --log-opt max-file=10 --ulimit memlock=-1:-1 --ulimit nofile=4096:100000 ${LOCAL_TIME_MOUNT_CMD} --volume ${WORKSPACE}/data/CS:/var/lib/cassandra --volume ${WORKSPACE}/data/environments:/home/sdc/chef-solo/environments --volume ${WORKSPACE}/data/CS-Init:/home/sdc/chef-solo/cache ${PREFIX}/sdc-onboard-cassandra-init:${RELEASE} + rc=$? + docker_logs ${DOCKER_NAME} + if [[ ${rc} != 0 ]]; then exit ${rc}; fi +} +# + + +#Back-End +function sdc-BE { + DOCKER_NAME="sdc-BE" + echo "docker run sdc-backend..." + if [ ${LOCAL} = false ]; then + docker pull ${PREFIX}/sdc-backend:${RELEASE} + else + ADDITIONAL_ARGUMENTS=${BE_DEBUG_PORT} + fi + docker run --detach --name ${DOCKER_NAME} --env HOST_IP=${IP} --env ENVNAME="${DEP_ENV}" --env cassandra_ssl_enabled="false" --env JAVA_OPTIONS="${BE_JAVA_OPTIONS}" --log-driver=json-file --log-opt max-size=100m --log-opt max-file=10 --ulimit memlock=-1:-1 --ulimit nofile=4096:100000 ${LOCAL_TIME_MOUNT_CMD} --volume ${WORKSPACE}/data/logs/BE/:${JETTY_BASE}/logs --volume ${WORKSPACE}/data/environments:${JETTY_BASE}/chef-solo/environments --publish 8443:8443 --publish 8080:8080 ${ADDITIONAL_ARGUMENTS} ${PREFIX}/sdc-backend:${RELEASE} + command_exit_status $? ${DOCKER_NAME} + echo "please wait while BE is starting..." + monitor_docker ${DOCKER_NAME} +} +# + + +# Back-End-Init +function sdc-BE-init { + DOCKER_NAME="sdc-BE-init" + echo "docker run sdc-backend-init..." + if [ ${LOCAL} = false ]; then + docker pull ${PREFIX}/sdc-backend-init:${RELEASE} + fi + docker run --name ${DOCKER_NAME} --env HOST_IP=${IP} --env ENVNAME="${DEP_ENV}" --log-driver=json-file --log-opt max-size=100m --log-opt max-file=10 --ulimit memlock=-1:-1 --ulimit nofile=4096:100000 ${LOCAL_TIME_MOUNT_CMD} --volume ${WORKSPACE}/data/logs/BE/:${JETTY_BASE}/logs --volume ${WORKSPACE}/data/environments:/home/sdc/chef-solo/environments ${PREFIX}/sdc-backend-init:${RELEASE} > /dev/null 2>&1 + rc=$? + docker_logs ${DOCKER_NAME} + if [[ ${rc} != 0 ]]; then exit ${rc}; fi +} +# + + +# Onboard Back-End +function sdc-onboard-BE { + DOCKER_NAME="sdc-onboard-BE" + echo "docker run sdc-onboard-BE ..." +# TODO Check the dir_perms action . do we need it here ?? +# dir_perms + if [ ${LOCAL} = false ]; then + docker pull ${PREFIX}/sdc-onboard-backend:${RELEASE} + else + ADDITIONAL_ARGUMENTS=${ONBOARD_DEBUG_PORT} + fi + docker run --detach --name ${DOCKER_NAME} --env HOST_IP=${IP} --env ENVNAME="${DEP_ENV}" --env cassandra_ssl_enabled="false" --env SDC_CLUSTER_NAME="SDC-CS-${DEP_ENV}" --env SDC_USER="${SDC_USER}" --env SDC_PASSWORD="${SDC_PASSWORD}" --env SDC_CERT_DIR="${SDC_CERT_DIR}" --env JAVA_OPTIONS="${ONBOARD_BE_JAVA_OPTIONS}" --log-driver=json-file --log-opt max-size=100m --log-opt max-file=10 --ulimit memlock=-1:-1 --ulimit nofile=4096:100000 ${LOCAL_TIME_MOUNT_CMD} --volume ${WORKSPACE}/data/${SDC_CERT_DIR}:${JETTY_BASE}/onap/cert --volume ${WORKSPACE}/data/logs/ONBOARD:${JETTY_BASE}/logs --volume ${WORKSPACE}/data/environments:/${JETTY_BASE}/chef-solo/environments --publish 8445:8445 --publish 8081:8081 ${ADDITIONAL_ARGUMENTS} ${PREFIX}/sdc-onboard-backend:${RELEASE} + command_exit_status $? ${DOCKER_NAME} + echo "please wait while sdc-onboard-BE is starting..." + monitor_docker ${DOCKER_NAME} +} +# + + +# Front-End +function sdc-FE { + DOCKER_NAME="sdc-FE" + IMAGE_NAME="${PREFIX}/sdc-frontend:${RELEASE}" + echo "Running container '${DOCKER_NAME}' based on '${IMAGE_NAME}' image..." + if [ ${LOCAL} = false ]; then + docker pull ${PREFIX}/sdc-frontend:${RELEASE} + else + ADDITIONAL_ARGUMENTS=${FE_DEBUG_PORT} + fi + + PLUGIN_CONFIG_FILE="${WORKSPACE}/data/environments/plugins-configuration.yaml" + if [[ -f ${WORKSPACE}/data/environments/plugins-configuration.yaml ]]; then + PLUGINS_CONF_VOLUME_MOUNT="--volume ${PLUGIN_CONFIG_FILE}:${JETTY_BASE}/config/catalog-fe/plugins-configuration.yaml" + else + echo "INFO: '${PLUGIN_CONFIG_FILE}' not provided, ignoring..." + fi + + docker run \ + --detach \ + --name ${DOCKER_NAME} \ + --env HOST_IP=${IP} \ + --env ENVNAME="${DEP_ENV}" \ + --env JAVA_OPTIONS="${FE_JAVA_OPTIONS}" \ + --log-driver=json-file \ + --log-opt max-size=100m \ + --log-opt max-file=10 \ + --ulimit memlock=-1:-1 \ + --ulimit nofile=4096:100000 \ + --volume ${WORKSPACE}/data/logs/FE/:${JETTY_BASE}/logs \ + --volume ${WORKSPACE}/data/environments:/${JETTY_BASE}/chef-solo/environments \ + ${LOCAL_TIME_MOUNT_CMD} \ + ${PLUGINS_CONF_VOLUME_MOUNT} \ + --publish 9443:9443 \ + --publish 8181:8181 \ + ${ADDITIONAL_ARGUMENTS} \ + ${IMAGE_NAME} + + command_exit_status $? ${DOCKER_NAME} + echo "Please wait while '${DOCKER_NAME}' container is starting..." + monitor_docker ${DOCKER_NAME} +} +# + + +# apis-sanity +function sdc-api-tests { + if [[ ${RUN_API_TESTS} = true ]] ; then + healthCheck + healthCheck_http_code=$? + if [[ ${healthCheck_http_code} == 200 ]] ; then + echo "docker run sdc-api-tests..." + echo "Trigger sdc-api-tests docker, please wait..." + + if [ ${LOCAL} = false ]; then + docker pull ${PREFIX}/sdc-api-tests:${RELEASE} + fi + + docker run --detach --name sdc-api-tests --env HOST_IP=${IP} --env ENVNAME="${DEP_ENV}" --env JAVA_OPTIONS="${API_TESTS_JAVA_OPTIONS}" --env SUITE_NAME=${API_SUITE} --log-driver=json-file --log-opt max-size=100m --log-opt max-file=10 --ulimit memlock=-1:-1 --ulimit nofile=4096:100000 ${LOCAL_TIME_MOUNT_CMD} --volume ${WORKSPACE}/data/logs/sdc-api-tests/target:/var/lib/tests/target --volume ${WORKSPACE}/data/logs/sdc-api-tests/ExtentReport:/var/lib/tests/ExtentReport --volume ${WORKSPACE}/data/logs/sdc-api-tests/outputCsar:/var/lib/tests/outputCsar --volume ${WORKSPACE}/data/environments:/root/chef-solo/environments --volume ${WORKSPACE}/data/${SDC_CERT_DIR}:/var/lib/tests/cert --publish 9560:9560 ${PREFIX}/sdc-api-tests:${RELEASE} echo "please wait while SDC-API-TESTS is starting....." + monitor_docker sdc-api-tests + fi + fi +} +# + + +# ui-sanity +function sdc-ui-tests { + + if [[ ${RUN_UI_TESTS} = true ]] ; then + healthCheck + healthCheck_http_code=$? + if [[ ${healthCheck_http_code} == 200 ]]; then + echo "docker run sdc-ui-tets..." + echo "Trigger sdc-ui-tests docker, please wait..." + + if [ ${LOCAL} = false ]; then + docker pull ${PREFIX}/sdc-ui-tests:${RELEASE} + fi + RUN_SIMULATOR=true; + sdc-sim + docker run --detach --name sdc-ui-tests --env HOST_IP=${IP} --env ENVNAME="${DEP_ENV}" --env JAVA_OPTIONS="${UI_TESTS_JAVA_OPTIONS}" --env SUITE_NAME=${UI_SUITE} --log-driver=json-file --log-opt max-size=100m --log-opt max-file=10 --ulimit memlock=-1:-1 --ulimit nofile=4096:100000 ${LOCAL_TIME_MOUNT_CMD} --volume ${WORKSPACE}/data/logs/sdc-ui-tests/target:/var/lib/tests/target --volume ${WORKSPACE}/data/logs/sdc-ui-tests/ExtentReport:/var/lib/tests/ExtentReport --volume ${WORKSPACE}/data/environments:/root/chef-solo/environments --publish 5901:5901 --publish 6901:6901 ${PREFIX}/sdc-ui-tests:${RELEASE} + echo "please wait while SDC-UI-TESTS is starting....." + monitor_docker sdc-ui-tests + fi + fi +} +# + + +# SDC-Simulator +function sdc-sim { + if [ "${RUN_SIMULATOR}" == true ]; then + echo "docker run sdc-webSimulator..." + if [ ${LOCAL} = false ]; then + docker pull ${PREFIX}/sdc-simulator:${RELEASE} + fi + + probe_sim + sim_stat=$? + if [ ${sim_stat} == 1 ]; then + docker run \ + --detach \ + --name sdc-sim \ + --env FE_URL="${FE_URL}" \ + --env JAVA_OPTIONS="${SIM_JAVA_OPTIONS}" \ + --env ENVNAME="${DEP_ENV}" \ + ${LOCAL_TIME_MOUNT_CMD} \ + --volume ${WORKSPACE}/data/logs/WS/:${JETTY_BASE}/logs \ + --volume ${WORKSPACE}/data/environments:/root/chef-solo/environments \ + --publish 8285:8080 \ + --publish 8286:8443 ${PREFIX}/sdc-simulator:${RELEASE} + echo "please wait while SDC-WEB-SIMULATOR is starting....." + monitor_docker sdc-sim + fi + fi +} +# + + +# +# Main +# + +# Handle command line arguments +while [ $# -gt 0 ]; do + case $1 in + + # -r | --release - The specific docker version to pull and deploy + -r | --release ) + shift 1 ; + RELEASE=$1; + shift 1;; + + # -e | --environment - The environment name you want to deploy + -e | --environment ) + shift 1; + DEP_ENV=$1; + shift 1 ;; + + # -p | --port - The port from which to connect to the docker nexus + -p | --port ) + shift 1 ; + PORT=$1; + shift 1 ;; + + # -l | --local - Use this for deploying your local dockers without pulling them first + -l | --local ) + LOCAL=true; + shift 1;; + + # -ta - Use this for running the APIs sanity docker after all other dockers have been deployed + -ta ) + shift 1 ; + API_SUITE=$1; + RUN_API_TESTS=true; + shift 1 ;; + + # -tu - Use this for running the UI sanity docker after all other dockers have been deployed + -tu ) + shift 1 ; + UI_SUITE=$1; + RUN_UI_TESTS=true; + shift 1 ;; + + # -tad - Use this for running the DEFAULT suite of tests in APIs sanity docker after all other dockers have been deployed + -tad | -t ) + API_SUITE="onapApiSanity"; + RUN_API_TESTS=true; + shift 1 ;; + + # -tud - Use this for running the DEFAULT suite of tests in UI sanity docker after all other dockers have been deployed + -tud ) + UI_SUITE="onapUiSanity"; + RUN_UI_TESTS=true; + shift 1 ;; + + # -d | --docker - The init specified docker + -d | --docker ) + shift 1 ; + DOCKER=$1; + shift 1 ;; + # -sim | --simulator run the simulator + -sim | --simulator ) + RUN_SIMULATOR=true; + shift 1 ;; + # -sim | --simulator run the simulator + -u | --fe_url ) + shift 1 ; + FE_URL=$1; + shift 1 ;; + + # -h | --help - Display the help message with all the available run options + -h | --help ) + usage; + exit ${SUCCESS};; + + * ) + usage; + exit ${FAILURE};; + esac +done + + +#Prefix those with WORKSPACE so it can be set to something other than /opt +[ -f ${WORKSPACE}/opt/config/env_name.txt ] && DEP_ENV=$(cat ${WORKSPACE}/opt/config/env_name.txt) || echo ${DEP_ENV} +[ -f ${WORKSPACE}/opt/config/nexus_username.txt ] && NEXUS_USERNAME=$(cat ${WORKSPACE}/opt/config/nexus_username.txt) || NEXUS_USERNAME=release +[ -f ${WORKSPACE}/opt/config/nexus_password.txt ] && NEXUS_PASSWD=$(cat ${WORKSPACE}/opt/config/nexus_password.txt) || NEXUS_PASSWD=sfWU3DFVdBr7GVxB85mTYgAW +[ -f ${WORKSPACE}/opt/config/nexus_docker_repo.txt ] && NEXUS_DOCKER_REPO=$(cat ${WORKSPACE}/opt/config/nexus_docker_repo.txt) || NEXUS_DOCKER_REPO=nexus3.onap.org:${PORT} +[ -f ${WORKSPACE}/opt/config/nexus_username.txt ] && docker login -u $NEXUS_USERNAME -p $NEXUS_PASSWD $NEXUS_DOCKER_REPO + + +export IP=`ip route get 8.8.8.8 | awk '/src/{ print $7 }'` +#If OSX, then use this to get IP +if [[ "$OSTYPE" == "darwin"* ]]; then + export IP=$(ipconfig getifaddr en0) +fi +export PREFIX=${NEXUS_DOCKER_REPO}'/onap' + +if [ ${LOCAL} = true ]; then + PREFIX='onap' +fi + +echo "" + +if [ -z "${DOCKER}" ]; then + cleanup all + dir_perms + sdc-cs + sdc-cs-init + sdc-cs-onboard-init + sdc-onboard-BE + sdc-BE + sdc-BE-init + sdc-FE + healthCheck + sdc-sim + sdc-api-tests + sdc-ui-tests +else + cleanup ${DOCKER} + dir_perms + ${DOCKER} + healthCheck +fi diff --git a/scripts/sdc/setup_sdc_for_sanity.sh b/scripts/sdc/setup_sdc_for_sanity.sh index 90ffa230..1a323975 100644 --- a/scripts/sdc/setup_sdc_for_sanity.sh +++ b/scripts/sdc/setup_sdc_for_sanity.sh @@ -48,7 +48,7 @@ function using_https { set -exo pipefail echo "This is ${WORKSPACE}/scripts/sdc/setup_sdc_for_sanity.sh" - +echo "lets check what is ${1} ${2}" ENABLE_SIMULATOR= case "$1" in tad|tud) @@ -58,16 +58,17 @@ case "$1" in '') # we will just setup sdc - no tests export TEST_SUITE="" - - # this is mandatory ENABLE_SIMULATOR="--simulator" + # this is mandatory ;; *) - usage - exit 1 + export TEST_SUITE="" + ENABLE_SIMULATOR="--simulator" + # # usage + # exit 1 ;; esac - +echo "Lets check is simulator is enabled or not ${ENABLE_SIMULATOR}" # Clone sdc enviroment template mkdir -p "${WORKSPACE}/data/environments/" mkdir -p "${WORKSPACE}/data/clone/" @@ -78,6 +79,7 @@ if using_local_images && [ -n "$SDC_LOCAL_GITREPO" ] ; then if [ -d "$SDC_LOCAL_GITREPO" ] ; then rm -rf ./sdc cp -a "$SDC_LOCAL_GITREPO" ./sdc + # echo "[skipping copying git repo of sdc]" else echo "[ERROR]: Local git repo for sdc does not exist: ${SDC_LOCAL_GITREPO}" exit 1 diff --git a/tests/sdc-workflow-d/__init__.robot b/tests/sdc-workflow-d/__init__.robot new file mode 100644 index 00000000..0ee6767d --- /dev/null +++ b/tests/sdc-workflow-d/__init__.robot @@ -0,0 +1,2 @@ +*** Settings *** +Documentation sdc-workflow-D diff --git a/tests/sdc-workflow-d/global_properties.robot b/tests/sdc-workflow-d/global_properties.robot new file mode 100644 index 00000000..03de4c41 --- /dev/null +++ b/tests/sdc-workflow-d/global_properties.robot @@ -0,0 +1,43 @@ +*** Settings *** +Documentation store all properties that can change or are used in multiple places here +... format is all caps with underscores between words and prepended with GLOBAL +... make sure you prepend them with GLOBAL so that other files can easily see it is from this file. + + +*** Variables *** +${GLOBAL_APPLICATION_ID} robot-ete +${GLOBAL_SO_STATUS_PATH} /onap/so/infra/orchestrationRequests/v6/ +${GLOBAL_SELENIUM_BROWSER} chrome +${GLOBAL_SELENIUM_BROWSER_CAPABILITIES} Create Dictionary +${GLOBAL_SELENIUM_DELAY} 0 +${GLOBAL_SELENIUM_BROWSER_IMPLICIT_WAIT} 5 +${GLOBAL_SELENIUM_BROWSER_WAIT_TIMEOUT} 15 +${GLOBAL_OPENSTACK_HEAT_SERVICE_TYPE} orchestration +${GLOBAL_OPENSTACK_CINDER_SERVICE_TYPE} volume +${GLOBAL_OPENSTACK_NOVA_SERVICE_TYPE} compute +${GLOBAL_OPENSTACK_NEUTRON_SERVICE_TYPE} network +${GLOBAL_OPENSTACK_GLANCE_SERVICE_TYPE} image +${GLOBAL_OPENSTACK_KEYSTONE_SERVICE_TYPE} identity +${GLOBAL_OPENSTACK_STACK_DEPLOYMENT_TIMEOUT} 600s +${GLOBAL_AAI_CLOUD_OWNER} CloudOwner +${GLOBAL_AAI_CLOUD_OWNER_DEFINED_TYPE} OwnerType +${GLOBAL_AAI_COMPLEX_NAME} clli1 +${GLOBAL_AAI_PHYSICAL_LOCATION_ID} clli1 +${GLOBAL_AAI_AVAILABILITY_ZONE_NAME} nova +${GLOBAL_BUILD_NUMBER} 0 +${GLOBAL_OWNING_ENTITY_NAME} OE-Demonstration +${GLOBAL_VID_UI_TIMEOUT_SHORT} 20s +${GLOBAL_VID_UI_TIMEOUT_MEDIUM} 60s +${GLOBAL_VID_UI_TIMEOUT_LONG} 120s +${GLOBAL_AAI_INDEX_PATH} /aai/v14 +${GLOBAL_AAI_ZONE_ID} nova1 +${GLOBAL_AAI_ZONE_NAME} nova +${GLOBAL_AAI_DESIGN_TYPE} integration +${GLOBAL_AAI_ZONE_CONTEXT} labs +${GLOBAL_TEMPLATE_FOLDER} robot/assets/templates +${GLOBAL_ASSETS_FOLDER} robot/assets +${GLOBAL_SERVICE_MAPPING_DIRECTORY} ./demo/service_mapping +${GLOBAL_SO_HEALTH_CHECK_PATH} /manage/health +${GLOBAL_SO_CLOUD_CONFIG_PATH} /cloudSite +${GLOBAL_SO_CLOUD_CONFIG_TEMPLATE} so/create_cloud_config.jinja +${GLOBAL_SO_ORCHESTRATION_REQUESTS_PATH} /onap/so/infra/orchestrationRequests/v7 diff --git a/tests/sdc-workflow-d/test1.robot b/tests/sdc-workflow-d/test1.robot new file mode 100644 index 00000000..6217f2a9 --- /dev/null +++ b/tests/sdc-workflow-d/test1.robot @@ -0,0 +1,112 @@ +*** Settings *** +Documentation This is the basic test for workflow designer +Library RequestsLibrary +Library Collections +Library SeleniumLibrary +Resource global_properties.robot + +*** Variables *** +${HOMEPAGE} http://localhost:8285 +${HEADLESS} True + +***Keywords*** + +Open SDC GUI + [Documentation] Logs in to SDC GUI + [Arguments] ${PATH} + ## Setup Browever now being managed by the test case + ##Setup Browser + Go To ${HOMEPAGE}${PATH} + Maximize Browser Window + + # Set Browser Implicit Wait ${GLOBAL_SELENIUM_BROWSER_IMPLICIT_WAIT} + # Log Logging in to ${SDC_FE_ENDPOINT}${PATH} + Wait Until Page Contains Jimmy + # Log Logged in to ${SDC_FE_ENDPOINT}${PATH} + +Setup Browser + [Documentation] Sets up browser based upon the value of ${GLOBAL_SELENIUM_BROWSER} + Run Keyword If '${GLOBAL_SELENIUM_BROWSER}' == 'firefox' Setup Browser Firefox + Run Keyword If '${GLOBAL_SELENIUM_BROWSER}' == 'chrome' Setup Browser Chrome + Log Running with ${GLOBAL_SELENIUM_BROWSER} + +Setup Browser Firefox + ${caps}= Evaluate sys.modules['selenium.webdriver'].common.desired_capabilities.DesiredCapabilities.FIREFOX sys + Set To Dictionary ${caps} marionette= + Set To Dictionary ${caps} elementScrollBehavior 1 + ${wd}= Create WebDriver Firefox capabilities=${caps} + Set Global Variable ${GLOBAL_SELENIUM_BROWSER_CAPABILITIES} ${caps} + + +Setup Browser Chrome + ${chrome options}= Evaluate sys.modules['selenium.webdriver'].ChromeOptions() sys + Call Method ${chrome options} add_argument no-sandbox + Call Method ${chrome options} add_argument ignore-certificate-errors + Run Keyword If ${HEADLESS}==True Call Method ${chrome options} add_argument headless + ${dc} Evaluate sys.modules['selenium.webdriver'].DesiredCapabilities.CHROME sys, selenium.webdriver + Set To Dictionary ${dc} elementScrollBehavior 1 + Set To Dictionary ${dc} ACCEPT_SSL_CERTS True + Create Webdriver Chrome chrome_options=${chrome_options} desired_capabilities=${dc} + Set Global Variable ${GLOBAL_SELENIUM_BROWSER_CAPABILITIES} ${dc} + +Input Username + [Arguments] ${username} + Input Text name=userId ${username} + +Input Password + [Arguments] ${password} + Input Text name=password ${password} + +Input Name + [Arguments] ${workflowName} + Input Text id=workflowName ${workflowName} + +Input Description + [Arguments] ${description} + Input Text xpath=/html/body/div[2]/div/div[2]/div/form/div/div[1]/div[2]/div/textarea ${description} + +Input WFdescription + [Arguments] ${description} + Input Text xpath=//*[@id="root"]/div[1]/div/div[2]/div[2]/div/div[1]/div/textarea + +Submit Login Button + Click Element xpath=/html/body/form/input[3] + +Submit WorkFlow Button + Click Element xpath=/html/body/div/home-page/div/top-nav/nav/ul/li[5]/a + +Add WorkFlow + Click Element xpath=//*[@id="root"]/div[1]/div/div[2]/div/div[2]/div[1] + # Click Element xpath=//*[@id="root"]/div[1]/div/div[2]/div/div[2]/div[1]/div[1]/div/svg + +Create Workflow + Click Element xpath=/html/body/div[2]/div/div[2]/div/form/div/div[2]/button[1] + +Goto Frame + Select Frame xpath=/html/body/div/plugin-tab-view/div/plugin-frame/div/div/iframe + +Save WorkFlow + Click Element xpath=//*[@id="root"]/div[1]/div/div[1]/div[2]/div[2]/div/div/div[2]/div/div/span + +*** Test Cases *** +Workflow Designer Testing + [Documentation] User can homepage and see the tag line + Setup Browser + Open SDC GUI /login + Input Username cs0008 + Input Password 123123a + Submit Login Button + Wait Until Page Contains WORKFLOW + Submit WorkFlow Button + BuiltIn.Sleep 5s + Goto Frame + Add WorkFlow + BuiltIn.Sleep 5s + Input Name testing7 + Input Description first test through selenium + Create Workflow + # Wait Until Page Contains General + # Input Description2 write some dummy description + # Save WorkFlow + # BuiltIn.Sleep 5s + Close Browser
\ No newline at end of file |