diff options
Diffstat (limited to 'deployment')
19 files changed, 120 insertions, 217 deletions
diff --git a/deployment/heat/onap-oom/env/gwu/onap.env b/deployment/heat/onap-oom/env/gwu/onap.env index 3ca447d70..1a9aa9a82 100644 --- a/deployment/heat/onap-oom/env/gwu/onap.env +++ b/deployment/heat/onap-oom/env/gwu/onap.env @@ -13,7 +13,7 @@ parameters: apt_proxy: 192.168.1.51:3142 docker_proxy: 192.168.1.51:5000 - rancher_vm_flavor: m1.small + rancher_vm_flavor: m1.large k8s_vm_flavor: m2.xxlarge public_net_id: 024582bd-ef9b-48b9-9e70-e6732559d9df diff --git a/deployment/heat/onap-oom/env/huawei/onap-beijing-oom.env b/deployment/heat/onap-oom/env/huawei/onap-beijing-oom.env index 0365c751c..88d6b7492 100644 --- a/deployment/heat/onap-oom/env/huawei/onap-beijing-oom.env +++ b/deployment/heat/onap-oom/env/huawei/onap-beijing-oom.env @@ -13,7 +13,7 @@ parameters: apt_proxy: 10.145.122.118:3142 docker_proxy: 10.145.122.118:5000 - rancher_vm_flavor: m1.small + rancher_vm_flavor: m1.large k8s_vm_flavor: m2.xxlarge public_net_id: 3a6247f1-fac6-4167-a49f-33cc8415ccf4 diff --git a/deployment/heat/onap-oom/env/tlab/ETE-HEAT-Test.env b/deployment/heat/onap-oom/env/tlab/ETE-HEAT-Test.env index 442d9fd76..405008380 100644 --- a/deployment/heat/onap-oom/env/tlab/ETE-HEAT-Test.env +++ b/deployment/heat/onap-oom/env/tlab/ETE-HEAT-Test.env @@ -13,7 +13,7 @@ parameters: apt_proxy: 192.168.31.204:3142 docker_proxy: 192.168.31.204:5000 - rancher_vm_flavor: m1.small + rancher_vm_flavor: m1.large k8s_vm_flavor: m2.xxlarge public_net_id: fbe8fd92-6636-4e63-ab28-bb6a5b0888a9 diff --git a/deployment/heat/onap-oom/env/tlab/ETE-OOM-Test.env b/deployment/heat/onap-oom/env/tlab/ETE-OOM-Test.env index 3cd226113..63d2f671a 100644 --- a/deployment/heat/onap-oom/env/tlab/ETE-OOM-Test.env +++ b/deployment/heat/onap-oom/env/tlab/ETE-OOM-Test.env @@ -13,7 +13,7 @@ parameters: apt_proxy: 192.168.31.204:3142 docker_proxy: 192.168.31.204:5000 - rancher_vm_flavor: m1.small + rancher_vm_flavor: m1.large k8s_vm_flavor: m2.xxlarge public_net_id: fbe8fd92-6636-4e63-ab28-bb6a5b0888a9 diff --git a/deployment/heat/onap-oom/env/tlab/EXTONAP_DEV.env b/deployment/heat/onap-oom/env/tlab/EXTONAP_DEV.env index 2d360eaca..cb9244299 100644 --- a/deployment/heat/onap-oom/env/tlab/EXTONAP_DEV.env +++ b/deployment/heat/onap-oom/env/tlab/EXTONAP_DEV.env @@ -13,7 +13,7 @@ parameters: apt_proxy: 192.168.31.204:3142 docker_proxy: 192.168.31.204:5000 - rancher_vm_flavor: m1.small + rancher_vm_flavor: m1.large k8s_vm_flavor: m2.xxlarge public_net_id: fbe8fd92-6636-4e63-ab28-bb6a5b0888a9 diff --git a/deployment/heat/onap-oom/env/windriver/Integration-Jenkins.env b/deployment/heat/onap-oom/env/windriver/Integration-Jenkins.env index 846d77b6e..22a5bdeb2 100644 --- a/deployment/heat/onap-oom/env/windriver/Integration-Jenkins.env +++ b/deployment/heat/onap-oom/env/windriver/Integration-Jenkins.env @@ -13,7 +13,7 @@ parameters: apt_proxy: 10.12.5.2:3142 docker_proxy: 10.12.5.2:5000 - rancher_vm_flavor: m1.small + rancher_vm_flavor: m1.large k8s_vm_flavor: m1.xxlarge public_net_id: 971040b2-7059-49dc-b220-4fab50cb2ad4 diff --git a/deployment/heat/onap-oom/env/windriver/Integration-SB-03.env b/deployment/heat/onap-oom/env/windriver/Integration-SB-03.env index 98c6c5f25..350bd475a 100644 --- a/deployment/heat/onap-oom/env/windriver/Integration-SB-03.env +++ b/deployment/heat/onap-oom/env/windriver/Integration-SB-03.env @@ -13,7 +13,7 @@ parameters: apt_proxy: 10.12.5.2:3142 docker_proxy: 10.12.5.2:5000 - rancher_vm_flavor: m1.small + rancher_vm_flavor: m1.large k8s_vm_flavor: m1.xxlarge public_net_id: 971040b2-7059-49dc-b220-4fab50cb2ad4 diff --git a/deployment/heat/onap-oom/env/windriver/Integration-SB-04.env b/deployment/heat/onap-oom/env/windriver/Integration-SB-04.env index ea7aa7ddb..4e6e1cdfd 100644 --- a/deployment/heat/onap-oom/env/windriver/Integration-SB-04.env +++ b/deployment/heat/onap-oom/env/windriver/Integration-SB-04.env @@ -13,7 +13,7 @@ parameters: apt_proxy: 10.12.5.2:3142 docker_proxy: 10.12.5.2:5000 - rancher_vm_flavor: m1.small + rancher_vm_flavor: m1.large k8s_vm_flavor: m1.xxlarge public_net_id: 971040b2-7059-49dc-b220-4fab50cb2ad4 diff --git a/deployment/heat/onap-oom/env/windriver/Integration-SB-05.env b/deployment/heat/onap-oom/env/windriver/Integration-SB-05.env index 5369182b4..47147ed3a 100644 --- a/deployment/heat/onap-oom/env/windriver/Integration-SB-05.env +++ b/deployment/heat/onap-oom/env/windriver/Integration-SB-05.env @@ -13,7 +13,7 @@ parameters: apt_proxy: 10.12.5.2:3142 docker_proxy: 10.12.5.2:5000 - rancher_vm_flavor: m1.small + rancher_vm_flavor: m1.large k8s_vm_flavor: m1.xxlarge public_net_id: 971040b2-7059-49dc-b220-4fab50cb2ad4 diff --git a/deployment/heat/onap-oom/env/windriver/Integration-SB-06-DNS-openrc b/deployment/heat/onap-oom/env/windriver/Integration-SB-06-DNS-openrc new file mode 120000 index 000000000..850c25544 --- /dev/null +++ b/deployment/heat/onap-oom/env/windriver/Integration-SB-06-DNS-openrc @@ -0,0 +1 @@ +../../../../../test/ete/labs/windriver/Integration-SB-06-DNS-openrc
\ No newline at end of file diff --git a/deployment/heat/onap-oom/env/windriver/Integration-SB-06-openrc b/deployment/heat/onap-oom/env/windriver/Integration-SB-06-openrc new file mode 120000 index 000000000..98ea5829a --- /dev/null +++ b/deployment/heat/onap-oom/env/windriver/Integration-SB-06-openrc @@ -0,0 +1 @@ +../../../../../test/ete/labs/windriver/Integration-SB-06-openrc
\ No newline at end of file diff --git a/deployment/heat/onap-oom/env/windriver/Integration-SB-06.env b/deployment/heat/onap-oom/env/windriver/Integration-SB-06.env new file mode 100644 index 000000000..6cf405386 --- /dev/null +++ b/deployment/heat/onap-oom/env/windriver/Integration-SB-06.env @@ -0,0 +1,33 @@ +parameters: + + keystone_url: http://10.12.25.2:5000 + openstack_tenant_id: 7fad299815104c0a8f90a8df80343f03 + openstack_tenant_name: Integration-SB-06 + openstack_username: demo + openstack_api_key: onapdemo + + ubuntu_1404_image: ubuntu-14-04-cloud-amd64 + ubuntu_1604_image: ubuntu-16-04-cloud-amd64 + centos_7_image: CentOS-7 + + apt_proxy: 10.12.5.2:3142 + docker_proxy: 10.12.5.2:5000 + + rancher_vm_flavor: m1.large + k8s_vm_flavor: m1.xxlarge + + public_net_id: 971040b2-7059-49dc-b220-4fab50cb2ad4 + public_net_name: external + + oam_network_cidr: 10.0.0.0/16 + dns_forwarder: 10.12.25.5 + external_dns: 8.8.8.8 + + dnsaas_proxy_enable: "true" + dnsaas_proxied_keystone_url_path: "api/multicloud-titanium_cloud/v0/pod25_RegionOne/identity/v2.0" + dnsaas_keystone_url: http://10.12.25.5:5000 + dnsaas_region: RegionOne + dnsaas_tenant_id: 2d7c6ec2cf9147fa974fa059b3673d7a + dnsaas_tenant_name: Integration-SB-06 + dnsaas_username: demo + dnsaas_password: onapdemo diff --git a/deployment/heat/onap-oom/env/windriver/Integration-SB-07-DNS-openrc b/deployment/heat/onap-oom/env/windriver/Integration-SB-07-DNS-openrc new file mode 120000 index 000000000..7e4a3af20 --- /dev/null +++ b/deployment/heat/onap-oom/env/windriver/Integration-SB-07-DNS-openrc @@ -0,0 +1 @@ +../../../../../test/ete/labs/windriver/Integration-SB-07-DNS-openrc
\ No newline at end of file diff --git a/deployment/heat/onap-oom/env/windriver/Integration-SB-07-openrc b/deployment/heat/onap-oom/env/windriver/Integration-SB-07-openrc new file mode 120000 index 000000000..87292901c --- /dev/null +++ b/deployment/heat/onap-oom/env/windriver/Integration-SB-07-openrc @@ -0,0 +1 @@ +../../../../../test/ete/labs/windriver/Integration-SB-07-openrc
\ No newline at end of file diff --git a/deployment/heat/onap-oom/env/windriver/Integration-SB-07.env b/deployment/heat/onap-oom/env/windriver/Integration-SB-07.env new file mode 100644 index 000000000..10b691e89 --- /dev/null +++ b/deployment/heat/onap-oom/env/windriver/Integration-SB-07.env @@ -0,0 +1,33 @@ +parameters: + + keystone_url: http://10.12.25.2:5000 + openstack_tenant_id: 1e097c6713e74fd7ac8e4295e605ee1e + openstack_tenant_name: Integration-SB-07 + openstack_username: demo + openstack_api_key: onapdemo + + ubuntu_1404_image: ubuntu-14-04-cloud-amd64 + ubuntu_1604_image: ubuntu-16-04-cloud-amd64 + centos_7_image: CentOS-7 + + apt_proxy: 10.12.5.2:3142 + docker_proxy: 10.12.5.2:5000 + + rancher_vm_flavor: m1.large + k8s_vm_flavor: m1.xxlarge + + public_net_id: 971040b2-7059-49dc-b220-4fab50cb2ad4 + public_net_name: external + + oam_network_cidr: 10.0.0.0/16 + dns_forwarder: 10.12.25.5 + external_dns: 8.8.8.8 + + dnsaas_proxy_enable: "true" + dnsaas_proxied_keystone_url_path: "api/multicloud-titanium_cloud/v0/pod25_RegionOne/identity/v2.0" + dnsaas_keystone_url: http://10.12.25.5:5000 + dnsaas_region: RegionOne + dnsaas_tenant_id: 885777d5e9074f31ad7801f9e6876b84 + dnsaas_tenant_name: Integration-SB-07 + dnsaas_username: demo + dnsaas_password: onapdemo diff --git a/deployment/heat/onap-oom/k8s_vm_entrypoint.sh b/deployment/heat/onap-oom/k8s_vm_entrypoint.sh index 9118c8e24..d483e73df 100644 --- a/deployment/heat/onap-oom/k8s_vm_entrypoint.sh +++ b/deployment/heat/onap-oom/k8s_vm_entrypoint.sh @@ -19,12 +19,12 @@ Acquire::https::Proxy "DIRECT"; EOF fi apt-get -y update -apt-get -y install linux-image-extra-$(uname -r) jq +apt-get -y install linux-image-extra-$(uname -r) jq make cd ~ -# install docker 1.12 -curl -s https://releases.rancher.com/install-docker/1.12.sh | sh +# install docker 17.03 +curl -s https://releases.rancher.com/install-docker/17.03.sh | sh usermod -aG docker ubuntu # install kubernetes 1.8.6 @@ -33,9 +33,9 @@ chmod +x ./kubectl sudo mv ./kubectl /usr/local/bin/kubectl mkdir ~/.kube -# install helm 2.3 -wget -q http://storage.googleapis.com/kubernetes-helm/helm-v2.3.0-linux-amd64.tar.gz -tar -zxvf helm-v2.3.0-linux-amd64.tar.gz +# install helm +wget -q http://storage.googleapis.com/kubernetes-helm/helm-v2.7.2-linux-amd64.tar.gz +tar -zxvf helm-v2.7.2-linux-amd64.tar.gz sudo mv linux-amd64/helm /usr/local/bin/helm # Fix virtual memory allocation for onap-log:elasticsearch: @@ -84,18 +84,6 @@ done RANCHER_AGENT_CMD=$(jq -r .command token.json) eval $RANCHER_AGENT_CMD -# download rancher CLI -wget -q https://github.com/rancher/cli/releases/download/v0.6.7/rancher-linux-amd64-v0.6.7.tar.xz -unxz rancher-linux-amd64-v0.6.7.tar.xz -tar xvf rancher-linux-amd64-v0.6.7.tar - -# Clone OOM: -cd ~ -git clone -b amsterdam http://gerrit.onap.org/r/oom - -# Update values.yaml to point to docker-proxy instead of nexus3: -cd ~/oom/kubernetes -perl -p -i -e 's/nexus3.onap.org:10001/__docker_proxy__/g' `find ./ -name values.yaml` oneclick/setenv.bash KUBETOKEN=$(echo -n 'Basic '$(echo -n "$RANCHER_ACCESS_KEY:$RANCHER_SECRET_KEY" | base64 -w 0) | base64 -w 0) @@ -124,197 +112,32 @@ EOF export KUBECONFIG=/root/.kube/config kubectl config view -# Update ~/oom/kubernetes/kube2msb/values.yaml kubeMasterAuthToken to use the token from ~/.kube/config -sed -i "s/kubeMasterAuthToken:.*/kubeMasterAuthToken: $KUBETOKEN/" ~/oom/kubernetes/kube2msb/values.yaml - -# Put your onap_key ssh private key in ~/.ssh/onap_key - -# Create or edit ~/oom/kubernetes/config/onap-parameters.yaml -cat > ~/oom/kubernetes/config/onap-parameters.yaml <<EOF -# For information regarding those parameters, please visit http://onap.readthedocs.io/en/latest/submodules/dcaegen2.git/docs/sections/installation_heat.html - -################# -# COMMON CONFIG # -################# - -# NEXUS -NEXUS_HTTP_REPO: https://nexus.onap.org/content/sites/raw -NEXUS_DOCKER_REPO: nexus3.onap.org:10001 -NEXUS_USERNAME: docker -NEXUS_PASSWORD: docker - -# ONAP config -# Do not change unless you know what you're doing -DMAAP_TOPIC: "AUTO" -DEMO_ARTIFACTS_VERSION: "1.1.1" - -# ------------------------------------------------# -# OpenStack Config on which VNFs will be deployed # -# ------------------------------------------------# - -# The four below parameters are only used by Robot. -# As Robot is able to perform some automated actions, -# e.g. onboard/distribute/instantiate, it has to be -# configured with four below parameters (in addition -# to the OPENSTACK ones). -# If you don't intend to use Robot for those actions, -# you can put dummy values, but you will have to provide -# those values when deploying VNF anyway. -# -------------------------------------------------- -# This is the OAM Network ID used for internal network by VNFs. -# You could create 10.10.10.0/24 (256 IPs should be enough) in your cloud instance. -OPENSTACK_OAM_NETWORK_ID: "__oam_network_id__" -# This is the public Network ID. Public = external network in OpenStack. -# Floating IPs will be created and assigned to VNFs from this network, -# to provide external reachability. -OPENSTACK_PUBLIC_NETWORK_ID: "__public_net_id__" -# VM Flavor to be used by VNF. -OPENSTACK_FLAVOR: "m1.medium" -# VM image to be used by VNF. Here ubuntu 14.04 is provided. -OPENSTACK_IMAGE: "__ubuntu_1604_image__" - -OPENSTACK_USERNAME: "__openstack_username__" -OPENSTACK_PASSWORD: "__openstack_api_key__" -OPENSTACK_TENANT_NAME: "__openstack_tenant_name__" -OPENSTACK_TENANT_ID: "__openstack_tenant_id__" -OPENSTACK_REGION: "RegionOne" -# Either v2.0 or v3 -OPENSTACK_API_VERSION: "v2.0" -OPENSTACK_KEYSTONE_URL: "__keystone_url__" -# Don't change this if you don't know what it is -OPENSTACK_SERVICE_TENANT_NAME: "service" - -######## -# DCAE # -######## - -# Whether or not to deploy DCAE -# If set to false, all the parameters below can be left empty or removed -# If set to false, update ../dcaegen2/values.yaml disableDcae value to true, -# this is to avoid deploying the DCAE deployments and services. -DEPLOY_DCAE: "true" - -# DCAE Config -DCAE_DOCKER_VERSION: v1.1.1 -DCAE_VM_BASE_NAME: "dcae" - -# ------------------------------------------------# -# OpenStack Config on which DCAE will be deployed # -# ------------------------------------------------# - -# Whether to have DCAE deployed on the same OpenStack instance on which VNF will be deployed. -# (e.g. re-use the same config as defined above) -# If set to true, discard the next config block, else provide the values. -IS_SAME_OPENSTACK_AS_VNF: "true" - -# Fill in the values in below block only if IS_SAME_OPENSTACK_AS_VNF set to "false" -# --- -# Either v2.0 or v3 -DCAE_OS_API_VERSION: "v2.0" -DCAE_OS_KEYSTONE_URL: "__keystone_url__" -DCAE_OS_USERNAME: "" -DCAE_OS_PASSWORD: "" -DCAE_OS_TENANT_NAME: "" -DCAE_OS_TENANT_ID: "" -DCAE_OS_REGION: "" -# --- - -# We need to provide the config of the public network here, because the DCAE VMs will be -# assigned a floating IP on this network so one can access them, to debug for instance. -# The ID of the public network. -DCAE_OS_PUBLIC_NET_ID: "__public_net_id__" -# The name of the public network. -DCAE_OS_PUBLIC_NET_NAME: "__public_net_name__" -# This is the private network that will be used by DCAE VMs. The network will be created during the DCAE boostrap process, -# and will the subnet created will use this CIDR. (/28 provides 16 IPs, DCAE requires 15.) -DCAE_OS_OAM_NETWORK_CIDR: "10.99.0.0/16" -# This will be the private ip of the DCAE boostrap VM. This VM is responsible for spinning up the whole DCAE stack (14 VMs total) -DCAE_IP_ADDR: "10.99.4.1" - -# The flavors' name to be used by DCAE VMs -DCAE_OS_FLAVOR_SMALL: "m1.small" -DCAE_OS_FLAVOR_MEDIUM: "m1.medium" -DCAE_OS_FLAVOR_LARGE: "m1.large" -# The images' name to be used by DCAE VMs -DCAE_OS_UBUNTU_14_IMAGE: "__ubuntu_1404_image__" -DCAE_OS_UBUNTU_16_IMAGE: "__ubuntu_1604_image__" -DCAE_OS_CENTOS_7_IMAGE: "__centos_7_image__" - -# This is the keypair that will be created in OpenStack, and that one can use to access DCAE VMs using ssh. -# The private key needs to be in a specific format so at the end of the process, it's formatted properly -# when ending up in the DCAE HEAT stack. The best way is to do the following: -# - copy paste your key -# - surround it with quote -# - add \n at the end of each line -# - escape the result using https://www.freeformatter.com/java-dotnet-escape.html#ad-output -DCAE_OS_KEY_NAME: "onap_key" -DCAE_OS_PUB_KEY: "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDKXDgoo3+WOqcUG8/5uUbk81+yczgwC4Y8ywTmuQqbNxlY1oQ0YxdMUqUnhitSXs5S/yRuAVOYHwGg2mCs20oAINrP+mxBI544AMIb9itPjCtgqtE2EWo6MmnFGbHB4Sx3XioE7F4VPsh7japsIwzOjbrQe+Mua1TGQ5d4nfEOQaaglXLLPFfuc7WbhbJbK6Q7rHqZfRcOwAMXgDoBqlyqKeiKwnumddo2RyNT8ljYmvB6buz7KnMinzo7qB0uktVT05FH9Rg0CTWH5norlG5qXgP2aukL0gk1ph8iAt7uYLf1ktp+LJI2gaF6L0/qli9EmVCSLr1uJ38Q8CBflhkh" -DCAE_OS_PRIVATE_KEY: \"-----BEGIN RSA PRIVATE KEY-----\\n\r\nMIIEpQIBAAKCAQEAylw4KKN/ljqnFBvP+blG5PNfsnM4MAuGPMsE5rkKmzcZWNaE\\n\r\nNGMXTFKlJ4YrUl7OUv8kbgFTmB8BoNpgrNtKACDaz/psQSOeOADCG/YrT4wrYKrR\\n\r\nNhFqOjJpxRmxweEsd14qBOxeFT7Ie42qbCMMzo260HvjLmtUxkOXeJ3xDkGmoJVy\\n\r\nyzxX7nO1m4WyWyukO6x6mX0XDsADF4A6AapcqinoisJ7pnXaNkcjU/JY2Jrwem7s\\n\r\n+ypzIp86O6gdLpLVU9ORR/UYNAk1h+Z6K5Rual4D9mrpC9IJNaYfIgLe7mC39ZLa\\n\r\nfiySNoGhei9P6pYvRJlQki69bid/EPAgX5YZIQIDAQABAoIBAQClDekkhI9ZqseC\\n\r\nqFjPuKaxsizZMg+faJb6WSHLSxzyk1OSWY6F6FklgLeC8HW/fuLNYZyGOYDEsG20\\n\r\nlMqL02Wdiy7OutS3oOS5iyzIf9a90HfFJi706el6RIpvINETcaXCS0T8tQrcS1Rd\\n\r\nKqTaBRC6HXJGAPbBcvw3pwQSdskatU6a/Kt2a3x6DsqqinQcgEB/SbrDaJCUX9sb\\n\r\nF2HVUwdq7aZK1Lk0ozr1FID9mrhjwWuQ6XC+vjG0FqtyXeMpR5iaQ73hex3FXQ8z\\n\r\nOjkFbMwuHWSh1DSx70r5yFrrBqwQKnMsBqx4QDRf3fIENUnWviaL+n+gwcXA07af\\n\r\n4kaNUFUtAoGBAPuNNRAGhZnyZ9zguns9PM56nmeMUikV5dPN2DTbQb79cpfV+7pC\\n\r\n6PeSH/dTKFLz62d6qAM2EsNXQvewf8fipBVBRPsRqKOv+uepd01dHNy62I5B+zRm\\n\r\nbe9Kbe+EN60qdzvyPM+2hV6CnvGv1dirimS9pu6RrxD2Rmz1ectnJE+rAoGBAM3w\\n\r\nUbSEemyZ6EKjck2RfdipzY0MNBnIZ2cUqHh8mmPXjdTLzpXb9vmPbHb01Qwo8MP+\\n\r\ngMnTbTBOzyNAaHdIrCO9FHW6C85j3ot5Yzcr+EcBVcua+7KHU0Sgn44JNH8DisJ7\\n\r\nY63UP/1Xb4d1/QvHfxYy3WOvvRdVZ7pPo8JNX95jAoGAIe5CIg8/JizUZa7KeKUh\\n\r\n9pgDleQPkQsrHQ6/AyIwFBsLwf9THSS5V+uV9D57SfUs46Bf2U8J6N90YQSlt8iS\\n\r\naWuManFPVgT+yxDIzt6obf2mCEpOIBtQ6N4ZRh2HhQwdWTCrkzkDdGQaHG+jYL6C\\n\r\nxGPwiG2ON7OAfGIAM7eN5lECgYEAhoRLWlaOgRGnHKAWsYQvZ67CjTdDcPPuVu6v\\n\r\nfMQnNMA/7JeTwV+E205L0wfpgZ/cZKmBBlQMJlnUA3q2wfO+PTnse1mjDJU/cGtB\\n\r\n22/lJLxChlQdxGeQhGtGzUhF+hEeOhrO6WSSx7CtMRZoy6Dr6lwfMFZCdVNcBd6v\\n\r\nYOOZk3ECgYEAseUKGb6E80XTVVNziyuiVbQCsI0ZJuRfqMZ2IIDQJU9u6AnGAway\\n\r\nitqHbkGsmDT+4HUz01+1JKnnw42RdSrHdU/LaOonD+RIGqe2x800QXzqASKLdCXr\\n\r\ny7RoiFqJtkdFQykzJemA+xOXvHLgKi/MXFsU90PCD0VJKLj8vwpX78Y=\\n\r\n-----END RSA PRIVATE KEY-----\\n\r\n\" - -# This below settings allows one to configure the /etc/resolv.conf nameserver resolution for all the DCAE VMs. -# - -# In the HEAT setup, it's meant to be a DNS list, as the HEAT setup deploys a DNS Server VM in addition to DNS Designate -# and this DNS Server is setup to forward request to the DNS Designate backend when it cannot resolve, hence the -# DNS_FORWARDER config here. The DCAE Boostrap requires both inputs, even though they are now similar, we have to pass -# them. -# - -# ATTENTION: Assumption is made the DNS Designate backend is configure to forward request to a public DNS (e.g. 8.8.8.8) -# - -# Put the IP of the DNS Designate backend (e.g. the OpenStack IP supporting DNS Designate) -DNS_IP: "__dns_forwarder__" -DNS_FORWARDER: "__dns_forwarder__" - -# Public DNS - not used but required by the DCAE boostrap container -EXTERNAL_DNS: "__external_dns__" - -# DNS domain for the DCAE VMs -DCAE_DOMAIN: "dcaeg2.onap.org" - -# Proxy DNS Designate. This means DCAE will run in an instance not support Designate, and Designate will be provided by another instance. -# Set to true if you wish to use it -DNSAAS_PROXY_ENABLE: "__dnsaas_proxy_enable__" -# Provide this only if DNSAAS_PROXY_ENABLE set to true. The IP has to be the IP of one of the K8S hosts. -# e.g. http://10.195.197.164/api/multicloud-titanium_cloud/v0/pod25_RegionOne/identity/v2.0 -DCAE_PROXIED_KEYSTONE_URL: "http://__k8s_ip_addr__/__dnsaas_proxied_keystone_url_path__" - -# -----------------------------------------------------# -# OpenStack Config on which DNS Designate is supported # -# -----------------------------------------------------# - -# If this is the same OpenStack used for the VNF or DCAE, please re-enter the values here. - -DNSAAS_API_VERSION: "v3" -DNSAAS_REGION: "RegionOne" -DNSAAS_KEYSTONE_URL: "__dnsaas_keystone_url__" -DNSAAS_TENANT_ID: "__dnsaas_tenant_id__" -DNSAAS_TENANT_NAME: "__dnsaas_tenant_name__" -DNSAAS_USERNAME: "__dnsaas_username__" -DNSAAS_PASSWORD: "__dnsaas_password__" -EOF -cat ~/oom/kubernetes/config/onap-parameters.yaml - - # wait for kubernetes to initialze sleep 100 until [ $(kubectl get pods --namespace kube-system | tail -n +2 | grep -c Running) -ge 6 ]; do sleep 10 done -# Source the environment file: -cd ~/oom/kubernetes/oneclick/ -source setenv.bash -# run the config pod creation -cd ~/oom/kubernetes/config -./createConfig.sh -n onap +# Install using OOM +export HOME=/root + +# Clone OOM: +cd ~ +git clone -b master http://gerrit.onap.org/r/oom +git log -1 + +# Update values.yaml to point to docker-proxy instead of nexus3: +cd ~/oom/kubernetes +#perl -p -i -e 's/nexus3.onap.org:10001/__docker_proxy__/g' `find ./ -name values.yaml` +sed -i 's/nexus3.onap.org:10001/__docker_proxy__/g' onap/values.yaml +sed -i 's/#repository:/repository:/g' onap/values.yaml +sed -i 's/#repositorySecret:/repositorySecret:/g' onap/values.yaml +git diff -# Wait until the config container completes. -sleep 20 -until [ $(kubectl get pods --namespace onap -a | tail -n +2 | grep -c Completed) -eq 1 ]; do - sleep 10 -done -# version control the config to see what's happening +# version control the persistence volume to see what's happening +mkdir -p /dockerdata-nfs/ cd /dockerdata-nfs/ git init git config user.email "root@k8s" @@ -322,11 +145,19 @@ git config user.name "root" git add -A git commit -m "initial commit" -cat /dockerdata-nfs/onap/dcaegen2/heat/onap_dcae.env - # Run ONAP: -cd ~/oom/kubernetes/oneclick/ -./createAll.bash -n onap +cd ~/oom/kubernetes/ +# verify version +helm version +helm init --client-only +helm init --upgrade +helm serve & +sleep 3 +helm repo add local http://127.0.0.1:8879 +helm repo list +make all +helm search -l | grep local +helm install local/onap -n dev --namespace onap # Check ONAP status: sleep 3 diff --git a/deployment/heat/onap-oom/onap-oom.yaml b/deployment/heat/onap-oom/onap-oom.yaml index 760a9575d..0a40ee224 100644 --- a/deployment/heat/onap-oom/onap-oom.yaml +++ b/deployment/heat/onap-oom/onap-oom.yaml @@ -238,6 +238,7 @@ resources: __public_net_id__: { get_param: public_net_id } __public_net_name__: { get_param: public_net_name } __oam_network_id__: { get_resource: oam_network } + __oam_subnet_id__: { get_resource: oam_subnet } __oam_network_cidr__: { get_param: oam_network_cidr } __ubuntu_1404_image__: { get_param: ubuntu_1404_image } __ubuntu_1604_image__: { get_param: ubuntu_1604_image } diff --git a/deployment/heat/onap-oom/rancher_vm_entrypoint.sh b/deployment/heat/onap-oom/rancher_vm_entrypoint.sh index ac704f106..12ed51f13 100644 --- a/deployment/heat/onap-oom/rancher_vm_entrypoint.sh +++ b/deployment/heat/onap-oom/rancher_vm_entrypoint.sh @@ -19,4 +19,4 @@ fi apt-get -y update apt-get -y install docker.io usermod -aG docker ubuntu -docker run --restart unless-stopped -d -p 8080:8080 rancher/server:v1.6.10 +docker run --restart unless-stopped -d -p 8080:8080 rancher/server:v1.6.14 diff --git a/deployment/heat/onap-oom/scripts/deploy.sh b/deployment/heat/onap-oom/scripts/deploy.sh index de33b08e0..c81e65c8b 100755 --- a/deployment/heat/onap-oom/scripts/deploy.sh +++ b/deployment/heat/onap-oom/scripts/deploy.sh @@ -38,13 +38,14 @@ fi ssh-keygen -R $K8S_IP for n in $(seq 1 10); do - timeout 2m ssh -o StrictHostKeychecking=no -i ~/.ssh/onap_key ubuntu@$K8S_IP 'sudo su -l root -c "/root/oom/kubernetes/robot/ete-k8s.sh health"' + timeout 2m ssh -o StrictHostKeychecking=no -i ~/.ssh/onap_key ubuntu@$K8S_IP 'sudo su -l root -c "/root/oom/kubernetes/robot/ete-k8s.sh onap health"' RESULT=$? if [ $RESULT -eq 0 ]; then break fi sleep 15m done -LOG_DIR=$(ssh -o StrictHostKeychecking=no -i ~/.ssh/onap_key ubuntu@$K8S_IP "ls -1t /dockerdata-nfs/onap/robot/eteshare/logs | head -1") -rsync -e "ssh -i ~/.ssh/onap_key" -avPz ubuntu@$K8S_IP:/dockerdata-nfs/onap/robot/eteshare/logs/${LOG_DIR}/ $WORKSPACE/archives/ +ROBOT_POD=$(ssh -o StrictHostKeychecking=no -i ~/.ssh/onap_key ubuntu@$K8S_IP 'sudo su -c "kubectl --namespace onap get pods"' | grep robot | sed 's/ .*//') +LOG_DIR=$(ssh -o StrictHostKeychecking=no -i ~/.ssh/onap_key ubuntu@$K8S_IP "sudo su -c \"kubectl exec $ROBOT_POD --namespace onap -- ls -1t /share/logs | head -1\"") +wget --user=robot --password=robot -r -np -nH --cut-dirs=2 -R "index.html*" -P $WORKSPACE/archives/ http://$K8S_IP:30209/logs/$LOG_DIR/ exit 0 |