diff options
Diffstat (limited to 'boot')
-rw-r--r-- | boot/aaf_install.sh | 426 | ||||
-rw-r--r-- | boot/aaf_vm_init.sh | 18 | ||||
-rwxr-xr-x | boot/dcae2_install.sh | 59 | ||||
-rwxr-xr-x | boot/dcae2_vm_init.sh | 236 | ||||
-rw-r--r-- | boot/music_vm_init.sh | 7 | ||||
-rw-r--r-- | boot/nbi_vm_init.sh | 9 | ||||
-rwxr-xr-x | boot/vfc_vm_init.sh | 26 |
7 files changed, 655 insertions, 126 deletions
diff --git a/boot/aaf_install.sh b/boot/aaf_install.sh index 14363269..0378a4e7 100644 --- a/boot/aaf_install.sh +++ b/boot/aaf_install.sh @@ -5,7 +5,9 @@ NEXUS_REPO=$(cat /opt/config/nexus_repo.txt) ARTIFACTS_VERSION=$(cat /opt/config/artifacts_version.txt) DNS_IP_ADDR=$(cat /opt/config/dns_ip_addr.txt) CLOUD_ENV=$(cat /opt/config/cloud_env.txt) +GERRIT_BRANCH=$(cat /opt/config/gerrit_branch.txt) MTU=$(/sbin/ifconfig | grep MTU | sed 's/.*MTU://' | sed 's/ .*//' | sort -n | head -1) +CODE_REPO=$(cat /opt/config/remote_repo.txt) # Add host name to /etc/host to avoid warnings in openstack images if [[ $CLOUD_ENV != "rackspace" ]] @@ -49,7 +51,7 @@ fi echo "deb http://ppa.launchpad.net/openjdk-r/ppa/ubuntu $(lsb_release -c -s) main" >> /etc/apt/sources.list.d/java.list echo "deb-src http://ppa.launchpad.net/openjdk-r/ppa/ubuntu $(lsb_release -c -s) main" >> /etc/apt/sources.list.d/java.list apt-get update -apt-get install --allow-unauthenticated -y apt-transport-https ca-certificates wget openjdk-8-jdk git ntp ntpdate make +apt-get install --allow-unauthenticated -y apt-transport-https ca-certificates wget openjdk-8-jdk git ntp ntpdate make maven # Download scripts from Nexus curl -k $NEXUS_REPO/org.onap.demo/boot/$ARTIFACTS_VERSION/aaf_vm_init.sh -o /opt/aaf_vm_init.sh @@ -60,7 +62,7 @@ mv /opt/aaf_serv.sh /etc/init.d update-rc.d aaf_serv.sh defaults # Download and install docker-engine and docker-compose -echo "deb https://apt.dockerproject.org/repo ubuntu-trusty main" | tee /etc/apt/sources.list.d/docker.list +echo "deb https://apt.dockerproject.org/repo ubuntu-xenial main" | tee /etc/apt/sources.list.d/docker.list apt-get update apt-get install -y linux-image-extra-$(uname -r) linux-image-extra-virtual apt-get install -y --allow-unauthenticated docker-engine @@ -91,4 +93,424 @@ resolvconf -u # Clone Gerrit repository and run docker containers cd /opt +git clone -b $GERRIT_BRANCH --single-branch $CODE_REPO +chmod +x /opt/authz/auth/auth-cass/docker/dinstall.sh +chmod +x /opt/authz/auth/docker/drun.sh +chmod +x /opt/authz/auth/docker/d.props + +#Update maven settings +cat > /usr/share/maven/conf/settings.xml << EOF +<?xml version="1.0" encoding="UTF-8"?> + +<!-- +Licensed to the Apache Software Foundation (ASF) under one +or more contributor license agreements. See the NOTICE file +distributed with this work for additional information +regarding copyright ownership. The ASF licenses this file +to you under the Apache License, Version 2.0 (the +"License"); you may not use this file except in compliance +with the License. You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, +software distributed under the License is distributed on an +"AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +KIND, either express or implied. See the License for the +specific language governing permissions and limitations +under the License. +--> + +<!-- + | This is the configuration file for Maven. It can be specified at two levels: + | + | 1. User Level. This settings.xml file provides configuration for a single user, +| and is normally provided in \${user.home}/.m2/settings.xml. + | + | NOTE: This location can be overridden with the CLI option: + | + | -s /path/to/user/settings.xml + | + | 2. Global Level. This settings.xml file provides configuration for all Maven + | users on a machine (assuming they're all using the same Maven + | installation). It's normally provided in +| \${maven.home}/conf/settings.xml. + | + | NOTE: This location can be overridden with the CLI option: + | + | -gs /path/to/global/settings.xml + | + | The sections in this sample file are intended to give you a running start at + | getting the most out of your Maven installation. Where appropriate, the default + | values (values used when the setting is not specified) are provided. + | + |--> +<settings xmlns="http://maven.apache.org/SETTINGS/1.0.0" + xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" + xsi:schemaLocation="http://maven.apache.org/SETTINGS/1.0.0 http://maven.apache.org/xsd/settings-1.0.0.xsd"> + <!-- localRepository + | The path to the local repository maven will use to store artifacts. + | +| Default: \${user.home}/.m2/repository + <localRepository>/path/to/local/repo</localRepository> + --> + + <!-- interactiveMode + | This will determine whether maven prompts you when it needs input. If set to false, + | maven will use a sensible default value, perhaps based on some other setting, for + | the parameter in question. + | + | Default: true + <interactiveMode>true</interactiveMode> + --> + + <!-- offline + | Determines whether maven should attempt to connect to the network when executing a build. + | This will have an effect on artifact downloads, artifact deployment, and others. + | + | Default: false + <offline>false</offline> + --> + + <!-- pluginGroups + | This is a list of additional group identifiers that will be searched when resolving plugins by their prefix, i.e. + | when invoking a command line like "mvn prefix:goal". Maven will automatically add the group identifiers + | "org.apache.maven.plugins" and "org.codehaus.mojo" if these are not already contained in the list. + |--> + <pluginGroups> + <!-- pluginGroup + | Specifies a further group identifier to use for plugin lookup. + <pluginGroup>com.your.plugins</pluginGroup> + --> + </pluginGroups> + + <!-- proxies + | This is a list of proxies which can be used on this machine to connect to the network. + | Unless otherwise specified (by system property or command-line switch), the first proxy + | specification in this list marked as active will be used. + |--> + <proxies> + <!-- proxy + | Specification for one proxy, to be used in connecting to the network. + | + <proxy> + <id>optional</id> + <active>true</active> + <protocol>http</protocol> + <username>proxyuser</username> + <password>proxypass</password> + <host>proxy.host.net</host> + <port>80</port> + <nonProxyHosts>local.net|some.host.com</nonProxyHosts> + </proxy> + --> + </proxies> + + <!-- servers + | This is a list of authentication profiles, keyed by the server-id used within the system. + | Authentication profiles can be used whenever maven must make a connection to a remote server. + |--> + <servers> + <!-- server + | Specifies the authentication information to use when connecting to a particular server, identified by + | a unique name within the system (referred to by the 'id' attribute below). + | + | NOTE: You should either specify username/password OR privateKey/passphrase, since these pairings are + | used together. + | + <server> + <id>deploymentRepo</id> + <username>repouser</username> + <password>repopwd</password> + </server> + --> + + <!-- Another sample, using keys to authenticate. + <server> + <id>siteServer</id> + <privateKey>/path/to/private/key</privateKey> + <passphrase>optional; leave empty if not used.</passphrase> + </server> + --> + </servers> + + <!-- mirrors + | This is a list of mirrors to be used in downloading artifacts from remote repositories. + | + | It works like this: a POM may declare a repository to use in resolving certain artifacts. + | However, this repository may have problems with heavy traffic at times, so people have mirrored + | it to several places. + | + | That repository definition will have a unique id, so we can create a mirror reference for that + | repository, to be used as an alternate download site. The mirror site will be the preferred + | server for that repository. + |--> + + <!-- profiles + | This is a list of profiles which can be activated in a variety of ways, and which can modify + | the build process. Profiles provided in the settings.xml are intended to provide local machine- + | specific paths and repository locations which allow the build to work in the local environment. + | + | For example, if you have an integration testing plugin - like cactus - that needs to know where + | your Tomcat instance is installed, you can provide a variable here such that the variable is + | dereferenced during the build process to configure the cactus plugin. + | + | As noted above, profiles can be activated in a variety of ways. One way - the activeProfiles + | section of this document (settings.xml) - will be discussed later. Another way essentially + | relies on the detection of a system property, either matching a particular value for the property, + | or merely testing its existence. Profiles can also be activated by JDK version prefix, where a + | value of '1.4' might activate a profile when the build is executed on a JDK version of '1.4.2_07'. + | Finally, the list of active profiles can be specified directly from the command line. + | + | NOTE: For profiles defined in the settings.xml, you are restricted to specifying only artifact + | repositories, plugin repositories, and free-form properties to be used as configuration + | variables for plugins in the POM. + | + |--> + + + + + + <profiles> + <profile> + + <id>10_nexus</id> + <!--Enable snapshots for the built in central repo to direct --> + <!--all requests to nexus via the mirror --> + <repositories> + <repository> + <id>10_nexus</id> + <url>http://repo.maven.apache.org/maven2/</url> + <releases><enabled>true</enabled></releases> + <snapshots><enabled>true</enabled></snapshots> + </repository> + </repositories> + + <pluginRepositories> + <pluginRepository> + <id>10_nexus</id> + <url>http://repo.maven.apache.org/maven2/</url> + <releases><enabled>true</enabled></releases> + <snapshots><enabled>true</enabled></snapshots> + </pluginRepository> + </pluginRepositories> + + </profile> + <profile> + <id>20_openecomp-public</id> + <repositories> + <repository> + <id>20_openecomp-public</id> + <name>20_openecomp-public</name> + <url>https://nexus.onap.org/content/repositories/public/</url> + <releases> + <enabled>true</enabled> + <updatePolicy>daily</updatePolicy> + </releases> + <snapshots> + <enabled>false</enabled> + </snapshots> + </repository> + </repositories> + <pluginRepositories> + <pluginRepository> + <id>20_openecomp-public</id> + <name>20_openecomp-public</name> + <url>https://nexus.onap.org/content/repositories/public/</url> + <releases> + <enabled>true</enabled> + <updatePolicy>daily</updatePolicy> + </releases> + <snapshots> + <enabled>false</enabled> + </snapshots> + </pluginRepository> + </pluginRepositories> + </profile> + <profile> + <id>30_openecomp-staging</id> + <repositories> + <repository> + <id>30_openecomp-staging</id> + <name>30_openecomp-staging</name> + <url>https://nexus.onap.org/content/repositories/staging/</url> + <releases> + <enabled>true</enabled> + <updatePolicy>daily</updatePolicy> + </releases> + <snapshots> + <enabled>false</enabled> + </snapshots> + </repository> + </repositories> + <pluginRepositories> + <pluginRepository> + <id>30_openecomp-staging</id> + <name>30_openecomp-staging</name> + <url>https://nexus.onap.org/content/repositories/staging/</url> + <releases> + <enabled>true</enabled> + <updatePolicy>daily</updatePolicy> + </releases> + <snapshots> + <enabled>false</enabled> + </snapshots> + </pluginRepository> + </pluginRepositories> + </profile> + <profile> + <id>40_openecomp-release</id> + <repositories> + <repository> + <id>40_openecomp-release</id> + <name>40_openecomp-release</name> + <url>https://nexus.onap.org/content/repositories/releases/</url> + <releases> + <enabled>true</enabled> + <updatePolicy>daily</updatePolicy> + </releases> + <snapshots> + <enabled>false</enabled> + </snapshots> + </repository> + </repositories> + <pluginRepositories> + <pluginRepository> + <id>40_openecomp-release</id> + <name>40_openecomp-release</name> + <url>https://nexus.onap.org/content/repositories/releases/</url> + <releases> + <enabled>true</enabled> + <updatePolicy>daily</updatePolicy> + </releases> + <snapshots> + <enabled>false</enabled> + </snapshots> + </pluginRepository> + </pluginRepositories> + </profile> + + <profile> + <id>50_openecomp-snapshots</id> + <repositories> + <repository> + <id>50_openecomp-snapshot</id> + <name>50_openecomp-snapshot</name> + <url>https://nexus.onap.org/content/repositories/snapshots/</url> + <releases> + <enabled>false</enabled> + </releases> + <snapshots> + <enabled>true</enabled> + </snapshots> + </repository> + </repositories> + <pluginRepositories> + <pluginRepository> + <id>50_openecomp-snapshot</id> + <name>50_openecomp-snapshot</name> + <url>https://nexus.onap.org/content/repositories/snapshots/</url> + <releases> + <enabled>false</enabled> + </releases> + <snapshots> + <enabled>true</enabled> + </snapshots> + </pluginRepository> + </pluginRepositories> + </profile> + <profile> + <id>60_opendaylight-release</id> + <repositories> + <repository> + <id>60_opendaylight-mirror</id> + <name>60_opendaylight-mirror</name> + <url>https://nexus.opendaylight.org/content/repositories/public/</url> + <releases> + <enabled>true</enabled> + <updatePolicy>daily</updatePolicy> + </releases> + <snapshots> + <enabled>false</enabled> + </snapshots> + </repository> + </repositories> + <pluginRepositories> + <pluginRepository> + <id>60_opendaylight-mirror</id> + <name>60_opendaylight-mirror</name> + <url>https://nexus.opendaylight.org/content/repositories/public/</url> + <releases> + <enabled>true</enabled> + <updatePolicy>daily</updatePolicy> + </releases> + <snapshots> + <enabled>false</enabled> + </snapshots> + </pluginRepository> + </pluginRepositories> + </profile> + + <profile> + <id>70_opendaylight-snapshots</id> + <repositories> + <repository> + <id>70_opendaylight-snapshot</id> + <name>70_opendaylight-snapshot</name> + <url>https://nexus.opendaylight.org/content/repositories/opendaylight.snapshot/</url> + <releases> + <enabled>false</enabled> + </releases> + <snapshots> + <enabled>true</enabled> + </snapshots> + </repository> + </repositories> + <pluginRepositories> + <pluginRepository> + <id>70_opendaylight-snapshot</id> + <name>70_opendaylight-snapshot</name> + <url>https://nexus.opendaylight.org/content/repositories/opendaylight.snapshot/</url> + <releases> + <enabled>false</enabled> + </releases> + <snapshots> + <enabled>true</enabled> + </snapshots> + </pluginRepository> + </pluginRepositories> + </profile> + </profiles> + + <activeProfiles> + <activeProfile>10_nexus</activeProfile> + <activeProfile>20_openecomp-public</activeProfile> + <activeProfile>30_openecomp-staging</activeProfile> + <activeProfile>40_openecomp-release</activeProfile> + <activeProfile>50_openecomp-snapshots</activeProfile> + <activeProfile>60_opendaylight-release</activeProfile> + <activeProfile>70_opendaylight-snapshots</activeProfile> + + </activeProfiles> + +</settings> +EOF + +cd /opt/authz +mvn install -Dmaven.test.skip=true + +# Rename network interface in openstack Ubuntu 16.04 images. Then, reboot the VM to pick up changes +if [[ $CLOUD_ENV != "rackspace" ]] +then + sed -i "s/GRUB_CMDLINE_LINUX=.*/GRUB_CMDLINE_LINUX=\"net.ifnames=0 biosdevname=0\"/g" /etc/default/grub + grub-mkconfig -o /boot/grub/grub.cfg + sed -i "s/ens[0-9]*/eth0/g" /etc/network/interfaces.d/*.cfg + sed -i "s/ens[0-9]*/eth0/g" /etc/udev/rules.d/70-persistent-net.rules + echo 'network: {config: disabled}' >> /etc/cloud/cloud.cfg.d/99-disable-network-config.cfg + echo "APT::Periodic::Unattended-Upgrade \"0\";" >> /etc/apt/apt.conf.d/10periodic + reboot +fi + +cd /opt ./aaf_vm_init.sh
\ No newline at end of file diff --git a/boot/aaf_vm_init.sh b/boot/aaf_vm_init.sh index d0f23953..7bc868a8 100644 --- a/boot/aaf_vm_init.sh +++ b/boot/aaf_vm_init.sh @@ -3,4 +3,20 @@ NEXUS_USERNAME=$(cat /opt/config/nexus_username.txt) NEXUS_PASSWD=$(cat /opt/config/nexus_password.txt) NEXUS_DOCKER_REPO=$(cat /opt/config/nexus_docker_repo.txt) -DOCKER_IMAGE_VERSION=$(cat /opt/config/docker_version.txt)
\ No newline at end of file +DOCKER_IMAGE_VERSION=$(cat /opt/config/docker_version.txt) + +docker login -u $NEXUS_USERNAME -p $NEXUS_PASSWD $NEXUS_DOCKER_REPO + +cd /opt/authz +git pull + +sed -i "s/DOCKER_REPOSITORY=.*/DOCKER_REPOSITORY="$NEXUS_DOCKER_REPO"/g" /opt/authz/auth/docker/d.props +sed -i "s/VERSION=.*/VERSION="$DOCKER_IMAGE_VERSION"/g" /opt/authz/auth/docker/d.props + +cd /opt/authz/auth/auth-cass/docker +./dinstall.sh + +sleep 2 + +cd /opt/authz/auth/docker +./drun.sh
\ No newline at end of file diff --git a/boot/dcae2_install.sh b/boot/dcae2_install.sh index 3446b6c2..ac3c3b56 100755 --- a/boot/dcae2_install.sh +++ b/boot/dcae2_install.sh @@ -46,7 +46,9 @@ apt-get update apt-get install --allow-unauthenticated -y apt-transport-https ca-certificates wget git ntp ntpdate python python-pip # Download scripts from Nexus +if [ ! -e /opt/dcae2_vm_init.sh ]; then curl -k "$NEXUS_REPO/org.onap.demo/boot/$ARTIFACTS_VERSION/dcae2_vm_init.sh" -o /opt/dcae2_vm_init.sh +fi curl -k "$NEXUS_REPO/org.onap.demo/boot/$ARTIFACTS_VERSION/dcae2_serv.sh" -o /opt/dcae2_serv.sh chmod +x /opt/dcae2_vm_init.sh chmod +x /opt/dcae2_serv.sh @@ -75,13 +77,17 @@ if [ -s /opt/config/external_dns.txt ] then DNS_FLAG=$DNS_FLAG"--dns $(cat /opt/config/external_dns.txt) " fi -echo "DOCKER_OPTS=\"$DNS_FLAG--mtu=$MTU\"" >> /etc/default/docker +echo "DOCKER_OPTS=\"$DNS_FLAG--mtu=$MTU --raw-logs -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock\"" >> /etc/default/docker cp /lib/systemd/system/docker.service /etc/systemd/system sed -i "/ExecStart/s/$/ --mtu=$MTU/g" /etc/systemd/system/docker.service +sed -i "/ExecStart/s/$/ -H tcp:\/\/0.0.0.0:2376 --raw-logs/g" /etc/systemd/system/docker.service systemctl daemon-reload service docker restart +# add hostname aliases +echo "$(cat /opt/config/dcae_ip_addr.txt) consul" >>/etc/hosts +echo "$(cat /opt/config/dcae_ip_addr.txt) dockerhost" >>/etc/hosts # DNS IP address configuration echo "nameserver $DNS_IP_ADDR" >> /etc/resolvconf/resolv.conf.d/head @@ -101,52 +107,5 @@ chmod 777 /opt/app/config/key #cp /opt/config/keystone_url.txt /opt/app/config/keystone_url.txt -URL_ROOT='nexus.onap.org/service/local/repositories/raw/content' -REPO_BLUEPRINTS='org.onap.dcaegen2.platform.blueprints' -REPO_DEPLOYMENTS='org.onap.dcaegen2.deployments' - -if [ -e /opt/config/dcae_deployment_profile.txt ]; then - DEPLOYMENT_PROFILE=$(cat /opt/config/dcae_deployment_profile.txt) -fi -DEPLOYMENT_PROFILE=${DEPLOYMENT_PROFILE:-R1} -if [ "$DEPLOYMENT_PROFILE" == "R1" ]; then - RELEASE_TAG='releases' - # download blueprint input template files - rm -rf /opt/app/inputs-templates - mkdir -p /opt/app/inputs-templates - wget -P /opt/app/inputs-templates https://${URL_ROOT}/${REPO_BLUEPRINTS}/${RELEASE_TAG}/input-templates/inputs.yaml - wget -P /opt/app/inputs-templates https://${URL_ROOT}/${REPO_BLUEPRINTS}/${RELEASE_TAG}/input-templates/cdapinputs.yaml - wget -P /opt/app/inputs-templates https://${URL_ROOT}/${REPO_BLUEPRINTS}/${RELEASE_TAG}/input-templates/phinputs.yaml - wget -P /opt/app/inputs-templates https://${URL_ROOT}/${REPO_BLUEPRINTS}/${RELEASE_TAG}/input-templates/dhinputs.yaml - wget -P /opt/app/inputs-templates https://${URL_ROOT}/${REPO_BLUEPRINTS}/${RELEASE_TAG}/input-templates/invinputs.yaml - wget -P /opt/app/inputs-templates https://${URL_ROOT}/${REPO_BLUEPRINTS}/${RELEASE_TAG}/input-templates/vesinput.yaml - wget -P /opt/app/inputs-templates https://${URL_ROOT}/${REPO_BLUEPRINTS}/${RELEASE_TAG}/input-templates/tcainputs.yaml - wget -P /opt/app/inputs-templates https://${URL_ROOT}/${REPO_BLUEPRINTS}/${RELEASE_TAG}/input-templates/he-ip.yaml - wget -P /opt/app/inputs-templates https://${URL_ROOT}/${REPO_BLUEPRINTS}/${RELEASE_TAG}/input-templates/hr-ip.yaml - - - # generate blueprint input files - pip install --upgrade jinja2 - wget https://${URL_ROOT}/${REPO_DEPLOYMENTS}/${RELEASE_TAG}/scripts/detemplate-bpinputs.py && \ - (python detemplate-bpinputs.py /opt/config /opt/app/inputs-templates /opt/app/config; rm detemplate-bpinputs.py) - - # Run docker containers - cd /opt - ./dcae2_vm_init.sh -fi - -if [ "$DEPLOYMENT_PROFILE" == "R2MVP" ]; then - RELEASE_TAG='R2' - rm -rf /opt/app/inputs-templates - mkdir -p /opt/app/inputs-templates - wget -P /opt/app/inputs-templates https://${URL_ROOT}/${REPO_DEPLOYMENTS}/${RELEASE_TAG}/heat/${DEPLOYMENT_PROFILE}/docker-compose-1.yaml - wget -P /opt/app/inputs-templates https://${URL_ROOT}/${REPO_DEPLOYMENTS}/${RELEASE_TAG}/heat/${DEPLOYMENT_PROFILE}/docker-compose-2.yaml - - pip install --upgrade jinja2 - wget https://${URL_ROOT}/${REPO_DEPLOYMENTS}/${RELEASE_TAG}/scripts/detemplate-bpinputs.py && \ - (python detemplate-bpinputs.py /opt/config /opt/app/inputs-templates /opt/app/config; rm detemplate-bpinputs.py) - - cd /opt - ./dcae2_vm_init.sh -fi - +cd /opt +#./dcae2_vm_init.sh diff --git a/boot/dcae2_vm_init.sh b/boot/dcae2_vm_init.sh index ac6713d0..d26bb47a 100755 --- a/boot/dcae2_vm_init.sh +++ b/boot/dcae2_vm_init.sh @@ -595,6 +595,19 @@ list_dns_zone() curl -v -s -H "Content-Type: application/json" -H "X-Auth-Token: $TOKEN" -X GET "${MULTICLOUD_PLUGIN_ENDPOINT}/dns-delegate/v2/zones/${ZONEID}/recordsets" } +################################## start of vm_init ##################################### + +# prepare the configurations needed by DCAEGEN2 installer +rm -rf /opt/app/config +mkdir -p /opt/app/config + + +# private key +sed -e 's/\\n/\n/g' /opt/config/priv_key | sed -e 's/^[ \t]*//g; s/[ \t]*$//g' > /opt/app/config/key +chmod 777 /opt/app/config/key + +# move keystone url file +#cp /opt/config/keystone_url.txt /opt/app/config/keystone_url.txt URL_ROOT='nexus.onap.org/service/local/repositories/raw/content' @@ -614,13 +627,39 @@ ZONE=$(cat /opt/config/rand_str.txt) MYFLOATIP=$(cat /opt/config/dcae_float_ip.txt) MYLOCALIP=$(cat /opt/config/dcae_ip_addr.txt) + # start docker image pulling while we are waiting for A&AI to come online docker login -u "$NEXUS_USER" -p "$NEXUS_PASSWORD" "$NEXUS_DOCKER_REPO" - if [ "$DEPLOYMENT_PROFILE" == "R1" ]; then - docker pull "$NEXUS_DOCKER_REPO/onap/org.onap.dcaegen2.deployments.bootstrap:$DOCKER_VERSION" && docker pull nginx & + RELEASE_TAG='releases' + # download blueprint input template files + rm -rf /opt/app/inputs-templates + mkdir -p /opt/app/inputs-templates + wget -P /opt/app/inputs-templates https://${URL_ROOT}/${REPO_BLUEPRINTS}/${RELEASE_TAG}/input-templates/inputs.yaml + wget -P /opt/app/inputs-templates https://${URL_ROOT}/${REPO_BLUEPRINTS}/${RELEASE_TAG}/input-templates/cdapinputs.yaml + wget -P /opt/app/inputs-templates https://${URL_ROOT}/${REPO_BLUEPRINTS}/${RELEASE_TAG}/input-templates/phinputs.yaml + wget -P /opt/app/inputs-templates https://${URL_ROOT}/${REPO_BLUEPRINTS}/${RELEASE_TAG}/input-templates/dhinputs.yaml + wget -P /opt/app/inputs-templates https://${URL_ROOT}/${REPO_BLUEPRINTS}/${RELEASE_TAG}/input-templates/invinputs.yaml + wget -P /opt/app/inputs-templates https://${URL_ROOT}/${REPO_BLUEPRINTS}/${RELEASE_TAG}/input-templates/vesinput.yaml + wget -P /opt/app/inputs-templates https://${URL_ROOT}/${REPO_BLUEPRINTS}/${RELEASE_TAG}/input-templates/tcainputs.yaml + wget -P /opt/app/inputs-templates https://${URL_ROOT}/${REPO_BLUEPRINTS}/${RELEASE_TAG}/input-templates/he-ip.yaml + wget -P /opt/app/inputs-templates https://${URL_ROOT}/${REPO_BLUEPRINTS}/${RELEASE_TAG}/input-templates/hr-ip.yaml + + # generate blueprint input files + pip install --upgrade jinja2 + wget https://${URL_ROOT}/${REPO_DEPLOYMENTS}/${RELEASE_TAG}/scripts/detemplate-bpinputs.py \ + && \ + (python detemplate-bpinputs.py /opt/config /opt/app/inputs-templates /opt/app/config; \ + rm detemplate-bpinputs.py) + + # Run docker containers + cd /opt + + + docker pull "$NEXUS_DOCKER_REPO/onap/org.onap.dcaegen2.deployments.bootstrap:$DOCKER_VERSION" \ + && docker pull nginx & ######################################### # Wait for then register with A&AI @@ -666,25 +705,40 @@ if [ "$DEPLOYMENT_PROFILE" == "R1" ]; then # start proxy for consul's health check CONSULIP=$(head -1 /opt/app/config/runtime.ip.consul | sed 's/[[:space:]]//g') echo "Consul is available at $CONSULIP" - - cat >./nginx.conf <<EOL -server { - listen 80; - server_name dcae.simpledemo.onap.org; - location /healthcheck { - proxy_pass http://${CONSULIP}:8500/v1/health/state/passing; - } -} -EOL - docker run --name dcae-proxy -p 8080:80 -v "$(pwd)/nginx.conf:/etc/nginx/conf.d/default.conf" -d nginx - echo "Healthcheck API available at http://${MYFLOATIP}:8080/healthcheck" - echo " or http://${MYLOCALIP}:8080/healthcheck" - fi +if [[ $DEPLOYMENT_PROFILE == R2* ]]; then + RELEASE_TAG='R2' + set +e + rm -rf /opt/app/inputs-templates + mkdir -p /opt/app/inputs-templates + wget -P /opt/app/inputs-templates https://${URL_ROOT}/${REPO_DEPLOYMENTS}/${RELEASE_TAG}/heat/docker-compose-1.yaml + wget -P /opt/app/inputs-templates https://${URL_ROOT}/${REPO_DEPLOYMENTS}/${RELEASE_TAG}/heat/docker-compose-2.yaml + wget -P /opt/app/inputs-templates https://${URL_ROOT}/${REPO_DEPLOYMENTS}/${RELEASE_TAG}/heat/docker-compose-3.yaml + wget -P /opt/app/inputs-templates https://${URL_ROOT}/${REPO_DEPLOYMENTS}/${RELEASE_TAG}/heat/docker-compose-4.yaml + wget -P /opt/app/inputs-templates https://${URL_ROOT}/${REPO_DEPLOYMENTS}/${RELEASE_TAG}/heat/register.sh + wget -P /opt/app/inputs-templates https://${URL_ROOT}/${REPO_DEPLOYMENTS}/${RELEASE_TAG}/heat/setup.sh + wget -P /opt/app/inputs-templates https://${URL_ROOT}/${REPO_DEPLOYMENTS}/${RELEASE_TAG}/heat/build-plugins.sh + + pip install --upgrade jinja2 + wget https://${URL_ROOT}/${REPO_DEPLOYMENTS}/${RELEASE_TAG}/scripts/detemplate-bpinputs.py \ + && \ + (python detemplate-bpinputs.py /opt/config /opt/app/inputs-templates /opt/app/config; \ + rm detemplate-bpinputs.py) + + if [ -e /opt/app/config/register.sh ]; then + chmod +x /opt/app/config/register.sh + fi + if [ -e /opt/app/config/setup.sh ]; then + chmod +x /opt/app/config/setup.sh + fi + if [ -e /opt/app/config/build-plugins.sh ]; then + chmod +x /opt/app/config/build-plugins.sh + fi + set -e -if [ "$DEPLOYMENT_PROFILE" == "R2MVP" ]; then cd /opt/app/config + # deploy essentials /opt/docker/docker-compose -f docker-compose-1.yaml up -d echo "Waiting for Consul to come up ready" while ! nc -z localhost 8500; do sleep 1; done @@ -695,48 +749,116 @@ if [ "$DEPLOYMENT_PROFILE" == "R2MVP" ]; then echo "All dependencies are up, proceed to the next phase" sleep 5 - NAME='config_binding_service' - PORT='10000' - ID=$(sudo docker ps |grep "$NAME" |cut -b1-12) - while [ -z "$ID" ]; do echo "Waiting for $NAME container to be deployed"; sleep 1; ID=$(sudo docker ps |grep "$NAME" |cut -b1-12); done - REG='{"ID": "'"$NAME"'0", "Name": "'"$NAME"'", "Address": "'"$NAME"'", "Port": '"$PORT"'}' - curl -v -X PUT -H "Content-Type: application/json" --data "${REG}" http://localhost:8500/v1/agent/service/register + echo "Setup CloudifyManager and Registrator" + ./setup.sh + ./register.sh - sleep 5 - echo "Now bring up DCAE service components" + echo "Bring up DCAE MIN service components for R2 use cases" /opt/docker/docker-compose -f docker-compose-2.yaml up -d - - NAME='ves' - PORT='8080' - echo "Registering for $NAME:$PORT" - ID=$(sudo docker ps |grep "$NAME" |cut -b1-12) - while [ -z "$ID" ]; do echo "Waiting for $NAME container to be deployed"; sleep 1; ID=$(sudo docker ps |grep "$NAME" |cut -b1-12); done - REG='{"ID": "'"$NAME"'", "Name": "'"$NAME"'", "Address": "'"$NAME"'", "Port": '"$PORT"'}' - curl -v -X PUT -H "Content-Type: application/json" --data "${REG}" http://localhost:8500/v1/agent/service/register - - NAME='tca' - PORT='11011' - echo "Registering for $NAME:$PORT" - ID=$(sudo docker ps |grep "$NAME" |cut -b1-12) - while [ -z "$ID" ]; do echo "Waiting for $NAME container to be deployed"; sleep 1; ID=$(sudo docker ps |grep "$NAME" |cut -b1-12); done - REG='{"ID": "'"$NAME"'", "Name": "'"$NAME"'", "Address": "'"$NAME"'", "Port": '"$PORT"'}' - curl -v -X PUT -H "Content-Type: application/json" --data "${REG}" http://localhost:8500/v1/agent/service/register - - NAME='hr' - PORT='9101' - echo "Registering for $NAME:$PORT" - ID=$(sudo docker ps |grep "$NAME" |cut -b1-12) - while [ -z "$ID" ]; do echo "Waiting for $NAME container to be deployed"; sleep 1; ID=$(sudo docker ps |grep "$NAME" |cut -b1-12); done - REG='{"ID": "'"$NAME"'", "Name": "'"$NAME"'", "Address": "'"$NAME"'", "Port": '"$PORT"'}' - curl -v -X PUT -H "Content-Type: application/json" --data "${REG}" http://localhost:8500/v1/agent/service/register - - NAME='he' - PORT='9102' - echo "Registering for $NAME:$PORT" - ID=$(sudo docker ps |grep "$NAME" |cut -b1-12) - while [ -z "$ID" ]; do echo "Waiting for $NAME container to be deployed"; sleep 1; ID=$(sudo docker ps |grep "$NAME" |cut -b1-12); done - REG='{"ID": "'"$NAME"'", "Name": "'"$NAME"'", "Address": "'"$NAME"'", "Port": '"$PORT"'}' - curl -v -X PUT -H "Content-Type: application/json" --data "${REG}" http://localhost:8500/v1/agent/service/register + if [[ "$DEPLOYMENT_PROFILE" == "R2" || "$DEPLOYMENT_PROFILE" == "R2PLUS" ]]; then + echo "Bring up DCAE platform components" + /opt/docker/docker-compose -f docker-compose-3.yaml up -d + + if [ "$DEPLOYMENT_PROFILE" == "R2PLUS" ]; then + echo "Bring up additional (plus) DCAE service components" + /opt/docker/docker-compose -f docker-compose-4.yaml up -d + fi + fi + + # start proxy for consul's health check + CONSULIP=$(cat /opt/config/dcae_ip_addr.txt) + echo "Consul is available at $CONSULIP" fi +cat >./nginx.conf <<EOL +server { + listen 80; + server_name dcae.simpledemo.onap.org; + root /www/healthcheck; + + location /healthcheck { + try_files /services.yaml =404; + } + location /R1 { + proxy_pass http://${CONSULIP}:8500/v1/health/state/passing; + } + location /R2MIN{ + try_files /r2mvp_healthy.yaml =404; + } + location /R2 { + try_files /r2_healthy.yaml =404; + } + location /R2PLUS { + try_files /r2plus_healthy.yaml =404; + } +} +EOL + +HEALTHPORT=8000 +docker run -d \ +--name dcae-health \ +-p ${HEALTHPORT}:80 \ +-v "$(pwd)/nginx.conf:/etc/nginx/conf.d/default.conf" \ +-v "/tmp/healthcheck:/www/healthcheck" \ + nginx + +echo "Healthcheck API available at http://${MYFLOATIP}:${HEALTHPORT}/healthcheck" +echo " http://${MYFLOATIP}:${HEALTHPORT}/R1" +echo " http://${MYFLOATIP}:${HEALTHPORT}/R2MIN" +echo " http://${MYFLOATIP}:${HEALTHPORT}/R2PLUS" + +# run forever for updating health status based on consul +while : +do + rm -rf /tmp/healthcheck/* + # all registered services + SERVICES=$(curl -s http://consul:8500/v1/agent/services |jq '. | to_entries[] | .value.Service') + # passing services + SERVICES=$(curl -s http://consul:8500/v1/health/state/passing | jq '.[] | .ServiceName') + + # remove empty lines/entries + SERVICES=$(echo "$SERVICES" | sed '/^\s*\"\"\s*$/d' |sed '/^\s*$/d') + + SERVICES_JSON=$(echo "$SERVICES" | sed 's/\"$/\",/g' | sed '$ s/.$//') + + + PLT_CONSUL=$(echo "$SERVICES" |grep consul) + PLT_CBS=$(echo "$SERVICES" |grep "config_binding_service") + MVP_PG_HOLMES=$(echo "$SERVICES" |grep "pgHolmes") + MVP_VES=$(echo "$SERVICES" |grep "mvp.*ves") + MVP_TCA=$(echo "$SERVICES" |grep "mvp.*tca") + MVP_HR=$(echo "$SERVICES" |grep "mvp.*holmes-rule") + MVP_HE=$(echo "$SERVICES" |grep "mvp.*holmes-engine") + + PLT_CM=$(echo "$SERVICES" |grep "cloudify.*manager") + PLT_DH=$(echo "$SERVICES" |grep "deployment.*handler") + PLT_PH=$(echo "$SERVICES" |grep "policy.*handler") + PLT_SCH=$(echo "$SERVICES" |grep "service.*change.*handler") + PLT_INV=$(echo "$SERVICES" |grep "inventory") + PLT_PG_INVENTORY=$(echo "$SERVICES" |grep "pgInventory") + + PLUS_MHB=$(echo "$SERVICES" |grep "heartbeat") + PLUS_PRH=$(echo "$SERVICES" |grep "prh") + PLUS_MPR=$(echo "$SERVICES" |grep "mapper") + PLUS_TRAP=$(echo "$SERVICES" |grep "snmptrap") + + DATA="{\"healthy\" : \"$(date)\", \"healthy_services\": [${SERVICES_JSON}]}" + if [[ -n $PLT_CONSUL && -n $PLT_CBS && -n $MVP_PG_HOLMES && -n $MVP_VES && \ + -n $MVP_TCA && -n $MVP_HR && -n $MVP_HE ]]; then + echo "${DATA}" > /tmp/healthcheck/r2mvp_healthy.yaml + echo "${DATA}" > /tmp/healthcheck/services.yaml + fi + + if [[ -n $CONSUL && -n $CBS && -n $PLT_CM && -n $PLT_DH && \ + -n $PLT_PH && -n $PLT_SCH && -n $PLT_INV && -n $PLT_PG_INVENTORY ]]; then + echo "${DATA}" > /tmp/healthcheck/r2_healthy.yaml + + if [[ -n $PLUS_MHB && -n $PLUS_PRH && -n $PLUS_MPR && -n $PLUS_TRAP ]]; then + echo "${DATA}" > /tmp/healthcheck/r2plus_healthy.yaml + fi + fi + + sleep 60 +done + diff --git a/boot/music_vm_init.sh b/boot/music_vm_init.sh index 372cd33e..2241ed13 100644 --- a/boot/music_vm_init.sh +++ b/boot/music_vm_init.sh @@ -14,4 +14,11 @@ docker pull $NEXUS_DOCKER_REPO/onap/music/music:$DOCKER_IMAGE_VERSION cd /opt/music/distribution/dockermusic git pull + +#Stop existing docker containers (if any) +if [ $(docker ps | wc -l) > 1 ]; then + ./music.sh stop + sleep 2 +fi + ./music.sh start
\ No newline at end of file diff --git a/boot/nbi_vm_init.sh b/boot/nbi_vm_init.sh index 576ec1da..9fac717b 100644 --- a/boot/nbi_vm_init.sh +++ b/boot/nbi_vm_init.sh @@ -5,9 +5,12 @@ NEXUS_PASSWD=$(cat /opt/config/nexus_password.txt) NEXUS_DOCKER_REPO=$(cat /opt/config/nexus_docker_repo.txt) DOCKER_IMAGE_VERSION=$(cat /opt/config/docker_version.txt) -# Fetch the latest docker-compose.yml +# Create nbi directory +mkdir /opt/nbi cd /opt/nbi -wget https://git.onap.org/externalapi/nbi/tree/docker-compose.yml?h=master + +# Fetch the latest docker-compose.yml +wget -o docker-compose.yml https://git.onap.org/externalapi/nbi/plain/docker-compose.yml?h=master # Pull the nbi docker image from nexus # MariaDB and mongoDB will be pulled automatically from docker.io during docker-compose @@ -16,4 +19,4 @@ docker login -u $NEXUS_USERNAME -p $NEXUS_PASSWD $NEXUS_DOCKER_REPO docker pull $NEXUS_DOCKER_REPO/onap/externalapi/nbi:$DOCKER_IMAGE_VERSION # Start nbi, MariaDB and MongoDB containers with docker compose and nbi/docker-compose.yml -/opt/docker/docker-compose up -d
\ No newline at end of file +/opt/docker/docker-compose up -d diff --git a/boot/vfc_vm_init.sh b/boot/vfc_vm_init.sh index 63e8b54e..288404b7 100755 --- a/boot/vfc_vm_init.sh +++ b/boot/vfc_vm_init.sh @@ -50,19 +50,19 @@ docker rm -f vfc_svnfm_nokiav2 docker run -i -t -d --name vfc_wfengine_activiti -p 8804:8080 -e SERVICE_IP=$OPENO_IP -e SERVICE_PORT=8804 -e OPENPALETTE_MSB_IP=$OPENO_IP -e OPENPALETTE_MSB_PORT=80 $NEXUS_DOCKER_REPO/onap/vfc/wfengine-activiti:$ACTIVITI_DOCKER_VER docker run -i -t -d --name vfc_wfengine_mgrservice -p 8805:10550 -e SERVICE_IP=$OPENO_IP -e SERVICE_PORT=8805 -e OPENPALETTE_MSB_IP=$OPENO_IP -e OPENPALETTE_MSB_PORT=80 $NEXUS_DOCKER_REPO/onap/vfc/wfengine-mgrservice:$MGRSERVICE_DOCKER_VER docker run -i -t -d --name vfc_catalog -p 8806:8806 -e MSB_ADDR=$OPENO_IP:80 -e SERVICE_IP=$OPENO_IP $NEXUS_DOCKER_REPO/onap/vfc/catalog:$CATALOG_DOCKER_VER -docker run -i -t -d --name vfc_emsdriver -p 8206:8206 -e MSB_ADDR=$OPENO_IP:80 -e VES_ADDR=$DCAE_COLL_IP:8080 -e VES_AUTHINFO="":"" $NEXUS_DOCKER_REPO/onap/vfc/emsdriver:$EMSDRIVER_DOCKER_VER -docker run -i -t -d --name vfc_gvnfmdriver -p 8484:8484 -e MSB_ADDR=$OPENO_IP:80 $NEXUS_DOCKER_REPO/onap/vfc/gvnfmdriver:$GVNFMDRIVER_DOCKER_VER -docker run -i -t -d --name vfc_jujudriver -p 8483:8483 -e MSB_ADDR=$OPENO_IP:80 $NEXUS_DOCKER_REPO/onap/vfc/jujudriver:$JUJUDRIVER_DOCKER_VER -docker run -i -t -d --name vfc_svnfm_huawei -p 8482:8482 -p 8443:8443 -e MSB_ADDR=$OPENO_IP:80 $NEXUS_DOCKER_REPO/onap/vfc/nfvo/svnfm/huawei:$HUAWEI_DOCKER_VER -docker run -i -t -d --name vfc_nslcm -p 8403:8403 -e MSB_ADDR=$OPENO_IP:80 $NEXUS_DOCKER_REPO/onap/vfc/nslcm:$NSLCM_DOCKER_VER -docker run -i -t -d --name vfc_resmanagement -p 8480:8480 -e MSB_ADDR=$OPENO_IP:80 $NEXUS_DOCKER_REPO/onap/vfc/resmanagement:$RESMANAGEMENT_DOCKER_VER -docker run -i -t -d --name vfc_vnflcm -p 8801:8801 -e MSB_ADDR=$OPENO_IP:80 $NEXUS_DOCKER_REPO/onap/vfc/vnflcm:$VNFLCM_DOCKER_VER -docker run -i -t -d --name vfc_vnfmgr -p 8803:8803 -e MSB_ADDR=$OPENO_IP:80 $NEXUS_DOCKER_REPO/onap/vfc/vnfmgr:$VNFMGR_DOCKER_VER -docker run -i -t -d --name vfc_vnfres -p 8802:8802 -e MSB_ADDR=$OPENO_IP:80 $NEXUS_DOCKER_REPO/onap/vfc/vnfres:$VNFRES_DOCKER_VER -docker run -i -t -d --name vfc_ztesdncdriver -p 8411:8411 -e MSB_ADDR=$OPENO_IP:80 $NEXUS_DOCKER_REPO/onap/vfc/ztesdncdriver:$ZTESDNCDRIVER_DOCKER_VER -docker run -i -t -d --name vfc_ztevnfmdriver -p 8410:8410 -e MSB_ADDR=$OPENO_IP:80 $NEXUS_DOCKER_REPO/onap/vfc/ztevnfmdriver:$ZTEVNFMDRIVER_DOCKER_VER -docker run -i -t -d --name vfc_svnfm_nokia -p 8486:8486 -e MSB_ADDR=$OPENO_IP:80 $NEXUS_DOCKER_REPO/onap/vfc/nfvo/svnfm/nokia:$NOKIA_DOCKER_VER -docker run -i -t -d --name vfc_multivimproxy -p 8481:8481 -e MSB_ADDR=$MSB_IP:80 $NEXUS_DOCKER_REPO/onap/vfc/multivimproxy:$MULTIVIMPROXY_DOCKER_VER +docker run -i -t -d --name vfc_emsdriver -p 8206:8206 -e MSB_ADDR=$OPENO_IP:80 -e SERVICE_IP=$OPENO_IP -e VES_ADDR=$DCAE_COLL_IP:8080 -e VES_AUTHINFO="":"" $NEXUS_DOCKER_REPO/onap/vfc/emsdriver:$EMSDRIVER_DOCKER_VER +docker run -i -t -d --name vfc_gvnfmdriver -p 8484:8484 -e MSB_ADDR=$OPENO_IP:80 -e SERVICE_IP=$OPENO_IP $NEXUS_DOCKER_REPO/onap/vfc/gvnfmdriver:$GVNFMDRIVER_DOCKER_VER +docker run -i -t -d --name vfc_jujudriver -p 8483:8483 -e MSB_ADDR=$OPENO_IP:80 -e SERVICE_IP=$OPENO_IP $NEXUS_DOCKER_REPO/onap/vfc/jujudriver:$JUJUDRIVER_DOCKER_VER +docker run -i -t -d --name vfc_svnfm_huawei -p 8482:8482 -p 8443:8443 -e MSB_ADDR=$OPENO_IP:80 -e SERVICE_IP=$OPENO_IP $NEXUS_DOCKER_REPO/onap/vfc/nfvo/svnfm/huawei:$HUAWEI_DOCKER_VER +docker run -i -t -d --name vfc_nslcm -p 8403:8403 -e MSB_ADDR=$OPENO_IP:80 -e SERVICE_IP=$OPENO_IP $NEXUS_DOCKER_REPO/onap/vfc/nslcm:$NSLCM_DOCKER_VER +docker run -i -t -d --name vfc_resmanagement -p 8480:8480 -e MSB_ADDR=$OPENO_IP:80 -e SERVICE_IP=$OPENO_IP $NEXUS_DOCKER_REPO/onap/vfc/resmanagement:$RESMANAGEMENT_DOCKER_VER +docker run -i -t -d --name vfc_vnflcm -p 8801:8801 -e MSB_ADDR=$OPENO_IP:80 -e SERVICE_IP=$OPENO_IP $NEXUS_DOCKER_REPO/onap/vfc/vnflcm:$VNFLCM_DOCKER_VER +docker run -i -t -d --name vfc_vnfmgr -p 8803:8803 -e MSB_ADDR=$OPENO_IP:80 -e SERVICE_IP=$OPENO_IP $NEXUS_DOCKER_REPO/onap/vfc/vnfmgr:$VNFMGR_DOCKER_VER +docker run -i -t -d --name vfc_vnfres -p 8802:8802 -e MSB_ADDR=$OPENO_IP:80 -e SERVICE_IP=$OPENO_IP $NEXUS_DOCKER_REPO/onap/vfc/vnfres:$VNFRES_DOCKER_VER +docker run -i -t -d --name vfc_ztesdncdriver -p 8411:8411 -e MSB_ADDR=$OPENO_IP:80 -e SERVICE_IP=$OPENO_IP $NEXUS_DOCKER_REPO/onap/vfc/ztesdncdriver:$ZTESDNCDRIVER_DOCKER_VER +docker run -i -t -d --name vfc_ztevnfmdriver -p 8410:8410 -e MSB_ADDR=$OPENO_IP:80 -e SERVICE_IP=$OPENO_IP $NEXUS_DOCKER_REPO/onap/vfc/ztevnfmdriver:$ZTEVNFMDRIVER_DOCKER_VER +docker run -i -t -d --name vfc_svnfm_nokia -p 8486:8486 -e MSB_ADDR=$OPENO_IP:80 -e SERVICE_IP=$OPENO_IP $NEXUS_DOCKER_REPO/onap/vfc/nfvo/svnfm/nokia:$NOKIA_DOCKER_VER +docker run -i -t -d --name vfc_multivimproxy -p 8481:8481 -e MSB_ADDR=$MSB_IP:80 -e SERVICE_IP=$OPENO_IP $NEXUS_DOCKER_REPO/onap/vfc/multivimproxy:$MULTIVIMPROXY_DOCKER_VER CBAM_IP=$(cat /opt/config/cbam_ip.txt) VNFM_ID=vnfmUuid |