diff options
67 files changed, 1034 insertions, 338 deletions
diff --git a/docs/Access_Rancher_server_via_web_browser.jpeg b/docs/Access_Rancher_server_via_web_browser.jpeg Binary files differnew file mode 100644 index 0000000000..103fc2731f --- /dev/null +++ b/docs/Access_Rancher_server_via_web_browser.jpeg diff --git a/docs/Add_Kubernetes_Environment_to_Rancher.png b/docs/Add_Kubernetes_Environment_to_Rancher.png Binary files differnew file mode 100644 index 0000000000..0d60f156fb --- /dev/null +++ b/docs/Add_Kubernetes_Environment_to_Rancher.png diff --git a/docs/Apply_customization_script_for_the_Rancher_VM.jpeg b/docs/Apply_customization_script_for_the_Rancher_VM.jpeg Binary files differnew file mode 100644 index 0000000000..832dda3404 --- /dev/null +++ b/docs/Apply_customization_script_for_the_Rancher_VM.jpeg diff --git a/docs/Click_create.jpeg b/docs/Click_create.jpeg Binary files differnew file mode 100644 index 0000000000..fad185674c --- /dev/null +++ b/docs/Click_create.jpeg diff --git a/docs/Click_on_CLI_and_then_click_on_Generate_Config.jpeg b/docs/Click_on_CLI_and_then_click_on_Generate_Config.jpeg Binary files differnew file mode 100644 index 0000000000..fec6a35433 --- /dev/null +++ b/docs/Click_on_CLI_and_then_click_on_Generate_Config.jpeg diff --git a/docs/Click_on_Close_button.jpeg b/docs/Click_on_Close_button.jpeg Binary files differnew file mode 100644 index 0000000000..cb02480923 --- /dev/null +++ b/docs/Click_on_Close_button.jpeg diff --git a/docs/Click_on_Copy_to_Clipboard-wait_until_you_see_a_token-do_not_copy_user+password-the_server_is_not_ready_at_that_point.jpeg b/docs/Click_on_Copy_to_Clipboard-wait_until_you_see_a_token-do_not_copy_user+password-the_server_is_not_ready_at_that_point.jpeg Binary files differnew file mode 100644 index 0000000000..bb0f60c4eb --- /dev/null +++ b/docs/Click_on_Copy_to_Clipboard-wait_until_you_see_a_token-do_not_copy_user+password-the_server_is_not_ready_at_that_point.jpeg diff --git a/docs/K8s-Assign_Floating_IP_for_external_access.jpeg b/docs/K8s-Assign_Floating_IP_for_external_access.jpeg Binary files differnew file mode 100644 index 0000000000..fe84d11707 --- /dev/null +++ b/docs/K8s-Assign_Floating_IP_for_external_access.jpeg diff --git a/docs/K8s-Key_Pair.jpeg b/docs/K8s-Key_Pair.jpeg Binary files differnew file mode 100644 index 0000000000..1e335d36a6 --- /dev/null +++ b/docs/K8s-Key_Pair.jpeg diff --git a/docs/K8s-Launch_Instance.jpeg b/docs/K8s-Launch_Instance.jpeg Binary files differnew file mode 100644 index 0000000000..7e2122c429 --- /dev/null +++ b/docs/K8s-Launch_Instance.jpeg diff --git a/docs/K8s-Launch_new_VM_instance_to_create_a_Kubernetes_single_host_or_cluster.jpeg b/docs/K8s-Launch_new_VM_instance_to_create_a_Kubernetes_single_host_or_cluster.jpeg Binary files differnew file mode 100644 index 0000000000..8d154d4a2d --- /dev/null +++ b/docs/K8s-Launch_new_VM_instance_to_create_a_Kubernetes_single_host_or_cluster.jpeg diff --git a/docs/K8s-Manage_Floating_IP_Associations.jpeg b/docs/K8s-Manage_Floating_IP_Associations.jpeg Binary files differnew file mode 100644 index 0000000000..f15e5e34c7 --- /dev/null +++ b/docs/K8s-Manage_Floating_IP_Associations.jpeg diff --git a/docs/K8s-Networking.jpeg b/docs/K8s-Networking.jpeg Binary files differnew file mode 100644 index 0000000000..ee4cafca54 --- /dev/null +++ b/docs/K8s-Networking.jpeg diff --git a/docs/K8s-Security_Group.jpeg b/docs/K8s-Security_Group.jpeg Binary files differnew file mode 100644 index 0000000000..5e852185b2 --- /dev/null +++ b/docs/K8s-Security_Group.jpeg diff --git a/docs/K8s-Select_Flavor.jpeg b/docs/K8s-Select_Flavor.jpeg Binary files differnew file mode 100644 index 0000000000..c3cf4b2c83 --- /dev/null +++ b/docs/K8s-Select_Flavor.jpeg diff --git a/docs/K8s-Select_Ubuntu_16.04_as_base_image.jpeg b/docs/K8s-Select_Ubuntu_16.04_as_base_image.jpeg Binary files differnew file mode 100644 index 0000000000..e9a03e782c --- /dev/null +++ b/docs/K8s-Select_Ubuntu_16.04_as_base_image.jpeg diff --git a/docs/Rancher-Allocate_Floating_IP.jpeg b/docs/Rancher-Allocate_Floating_IP.jpeg Binary files differnew file mode 100644 index 0000000000..9d6ff3ff95 --- /dev/null +++ b/docs/Rancher-Allocate_Floating_IP.jpeg diff --git a/docs/Rancher-Key_Pair.jpeg b/docs/Rancher-Key_Pair.jpeg Binary files differnew file mode 100644 index 0000000000..1bcb6b6eb7 --- /dev/null +++ b/docs/Rancher-Key_Pair.jpeg diff --git a/docs/Rancher-Launch_Instance.jpeg b/docs/Rancher-Launch_Instance.jpeg Binary files differnew file mode 100644 index 0000000000..38943699e6 --- /dev/null +++ b/docs/Rancher-Launch_Instance.jpeg diff --git a/docs/Rancher-Launch_new_VM_instance_to_host_the_Rancher_Server.jpeg b/docs/Rancher-Launch_new_VM_instance_to_host_the_Rancher_Server.jpeg Binary files differnew file mode 100644 index 0000000000..2b3fde3fb7 --- /dev/null +++ b/docs/Rancher-Launch_new_VM_instance_to_host_the_Rancher_Server.jpeg diff --git a/docs/Rancher-Manage_Floating_IP_Associations.jpeg b/docs/Rancher-Manage_Floating_IP_Associations.jpeg Binary files differnew file mode 100644 index 0000000000..ae062c1d39 --- /dev/null +++ b/docs/Rancher-Manage_Floating_IP_Associations.jpeg diff --git a/docs/Rancher-Networking.jpeg b/docs/Rancher-Networking.jpeg Binary files differnew file mode 100644 index 0000000000..16ef380ef5 --- /dev/null +++ b/docs/Rancher-Networking.jpeg diff --git a/docs/Rancher-Security_Groups.jpeg b/docs/Rancher-Security_Groups.jpeg Binary files differnew file mode 100644 index 0000000000..c8ff2e8d76 --- /dev/null +++ b/docs/Rancher-Security_Groups.jpeg diff --git a/docs/Rancher-Select_Flavor.jpeg b/docs/Rancher-Select_Flavor.jpeg Binary files differnew file mode 100644 index 0000000000..a3f35b53ce --- /dev/null +++ b/docs/Rancher-Select_Flavor.jpeg diff --git a/docs/Rancher-Select_Ubuntu_16.04_as_base_image.jpeg b/docs/Rancher-Select_Ubuntu_16.04_as_base_image.jpeg Binary files differnew file mode 100644 index 0000000000..b80fae4a5c --- /dev/null +++ b/docs/Rancher-Select_Ubuntu_16.04_as_base_image.jpeg diff --git a/docs/Return_to_Rancher_environment_eg_SB4_and_wait_for_services_to_complete_10-15_mins.jpeg b/docs/Return_to_Rancher_environment_eg_SB4_and_wait_for_services_to_complete_10-15_mins.jpeg Binary files differnew file mode 100644 index 0000000000..6d62b99f55 --- /dev/null +++ b/docs/Return_to_Rancher_environment_eg_SB4_and_wait_for_services_to_complete_10-15_mins.jpeg diff --git a/docs/Select_Add_Environment.png b/docs/Select_Add_Environment.png Binary files differnew file mode 100644 index 0000000000..48fffa41bd --- /dev/null +++ b/docs/Select_Add_Environment.png diff --git a/docs/and_click_on_Save_accept_defaults.jpeg b/docs/and_click_on_Save_accept_defaults.jpeg Binary files differnew file mode 100644 index 0000000000..9b57927aac --- /dev/null +++ b/docs/and_click_on_Save_accept_defaults.jpeg diff --git a/docs/index.rst b/docs/index.rst index 712802d619..340b43be67 100644 --- a/docs/index.rst +++ b/docs/index.rst @@ -14,3 +14,4 @@ OOM Documentation Repository oom_developer_guide.rst oom_cloud_setup_guide.rst release-notes.rst + oom_setup_kubernetes_rancher.rst diff --git a/docs/k8s-topology.jpg b/docs/k8s-topology.jpg Binary files differnew file mode 100644 index 0000000000..eba8e72b68 --- /dev/null +++ b/docs/k8s-topology.jpg diff --git a/docs/master_nfs_node.sh b/docs/master_nfs_node.sh new file mode 100644 index 0000000000..4a7a8dbc12 --- /dev/null +++ b/docs/master_nfs_node.sh @@ -0,0 +1,32 @@ +#!/bin/bash + +usage () { + echo "Usage:" + echo " ./$(basename $0) node1_ip node2_ip ... nodeN_ip" + exit 1 +} + +if [ "$#" -lt 1 ]; then + echo "Missing NFS slave nodes" + usage +fi + +#Install NFS kernel +sudo apt-get update +sudo apt-get install -y nfs-kernel-server + +#Create /dockerdata-nfs and set permissions +sudo mkdir -p /dockerdata-nfs +sudo chmod 777 -R /dockerdata-nfs +sudo chown nobody:nogroup /dockerdata-nfs/ + +#Update the /etc/exports +NFS_EXP="" +for i in $@; do + NFS_EXP+="$i(rw,sync,no_root_squash,no_subtree_check) " +done +echo "/dockerdata-nfs "$NFS_EXP | sudo tee -a /etc/exports + +#Restart the NFS service +sudo exportfs -a +sudo systemctl restart nfs-kernel-server diff --git a/docs/oom_cloud_setup_guide.rst b/docs/oom_cloud_setup_guide.rst index 7582a4a03d..73509669e8 100644 --- a/docs/oom_cloud_setup_guide.rst +++ b/docs/oom_cloud_setup_guide.rst @@ -12,6 +12,7 @@ .. _Setting Up Kubernetes with Rancher: https://wiki.onap.org/display/DW/ONAP+on+Kubernetes+on+Rancher .. _Setting Up Kubernetes with Kubeadm: https://wiki.onap.org/display/DW/Deploying+Kubernetes+Cluster+with+kubeadm .. _Setting Up Kubernetes with Cloudify: https://wiki.onap.org/display/DW/ONAP+on+Kubernetes+on+Cloudify +.. _ONAP on Kubernetes Wiki: https://wiki.onap.org/display/DW/ONAP+on+Kubernetes .. figure:: oomLogoV2-medium.png :align: right @@ -63,26 +64,38 @@ least three if not six nodes to ensure there is no single point of failure. ===== ===== ====== ==================== RAM HD vCores Ports ===== ===== ====== ==================== - 128GB 160GB 16 0.0.0.0/0 (all open) + 128GB 160GB 32 0.0.0.0/0 (all open) ===== ===== ====== ==================== +.. note:: + Kubernetes supports a maximum of 110 pods per node which forces one to use at + least two nodes to deploy all of ONAP although at least three are recommended + (for example 4x32GB - 8 vCores each). Subsets of ONAP may still be deployed + on a single node. + Cloud Installation ================== -#. OOM supports deployment on major public clouds. The following guides - provide instructions on how to deploy ONAP on these clouds: - - - `Microsoft Azure`_, - - `Amazon AWS`_, - - `Google GCE`_, - - `VMware VIO`_, - - IBM, and - - `Openstack`_. - -#. Alternatively, OOM can be deployed on a private set of physical hosts or VMs - (or even a combination of the two). The following guides describe how to - create a Kubernetes cluster with popular tools: - - - `Setting up Kubernetes with Rancher`_ (recommended) - - `Setting up Kubernetes with Kubeadm`_ - - `Setting up Kubernetes with Cloudify`_ +.. #. OOM supports deployment on major public clouds. The following guides +.. provide instructions on how to deploy ONAP on these clouds: +.. +.. - `Microsoft Azure`_, +.. - `Amazon AWS`_, +.. - `Google GCE`_, +.. - `VMware VIO`_, +.. - IBM, and +.. - `Openstack`_. +.. +.. #. Alternatively, OOM can be deployed on a private set of physical hosts or VMs +.. (or even a combination of the two). The following guides describe how to +.. create a Kubernetes cluster with popular tools: +.. +.. - `Setting up Kubernetes with Rancher`_ (recommended) +.. - `Setting up Kubernetes with Kubeadm`_ +.. - `Setting up Kubernetes with Cloudify`_ + +OOM can be deployed on a private set of physical hosts or VMs (or even a +combination of the two). The following guide describe the recommended method to +setup a Kubernetes cluster: :ref:`onap-on-kubernetes-with-rancher`. + +There are alternative deployment methods described on the `ONAP on Kubernetes Wiki`_ diff --git a/docs/oom_setup_kubernetes_rancher.rst b/docs/oom_setup_kubernetes_rancher.rst new file mode 100644 index 0000000000..8e55bf1b66 --- /dev/null +++ b/docs/oom_setup_kubernetes_rancher.rst @@ -0,0 +1,478 @@ +.. This work is licensed under a Creative Commons Attribution 4.0 International License. +.. http://creativecommons.org/licenses/by/4.0 +.. Copyright 2018 Amdocs, Bell Canada + +.. Links +.. _HELM Best Practices Guide: https://docs.helm.sh/chart_best_practices/#requirements +.. _kubectl Cheat Sheet: https://kubernetes.io/docs/reference/kubectl/cheatsheet/ +.. _Kubernetes documentation for emptyDir: https://kubernetes.io/docs/concepts/storage/volumes/#emptydir +.. _Docker DevOps: https://wiki.onap.org/display/DW/Docker+DevOps#DockerDevOps-DockerBuild +.. _http://cd.onap.info:30223/mso/logging/debug: http://cd.onap.info:30223/mso/logging/debug +.. _Onboarding and Distributing a Vendor Software Product: https://wiki.onap.org/pages/viewpage.action?pageId=1018474 +.. _README.md: https://gerrit.onap.org/r/gitweb?p=oom.git;a=blob;f=kubernetes/README.md + +.. figure:: oomLogoV2-medium.png + :align: right + +.. _onap-on-kubernetes-with-rancher: + +ONAP on Kubernetes with Rancher +############################### + +The following instructions will step you through the installation of Kubernetes +on an OpenStack environment with Rancher. The development lab used for this +installation is the ONAP Windriver lab. + +This guide does not cover all of the steps required to setup your OpenStack +environment: e.g. OAM networks and security groups but there is a wealth of +OpenStack information on the web. + +Rancher Installation +==================== + +The following instructions describe how to create an Openstack VM running +Rancher. This node will not be used to host ONAP itself, it will be used +exclusively by Rancher. + +Launch new VM instance to host the Rancher Server +------------------------------------------------- + +.. image:: Rancher-Launch_new_VM_instance_to_host_the_Rancher_Server.jpeg + +Select Ubuntu 16.04 as base image +--------------------------------- +Select "No" on "Create New Volume" + +.. image:: Rancher-Select_Ubuntu_16.04_as_base_image.jpeg + +Select Flavor +------------- +Known issues exist if flavor is too small for Rancher. Please select a flavor +with at least 4 vCPU and 8GB ram. + +.. image:: Rancher-Select_Flavor.jpeg + +Networking +---------- + +.. image:: Rancher-Networking.jpeg + +Security Groups +--------------- + +.. image:: Rancher-Security_Groups.jpeg + +Key Pair +-------- +Use an existing key pair (e.g. onap_key), import an existing one or create a +new one to assign. + +.. image:: Rancher-Key_Pair.jpeg + +Apply customization script for the Rancher VM +--------------------------------------------- + +Click :download:`openstack-rancher.sh <openstack-rancher.sh>` to download the script. + +.. literalinclude:: openstack-rancher.sh + :language: bash + +This customization script will: + +* setup root access to the VM (comment out if you wish to disable this + capability and restrict access to ssh access only) +* install docker +* install rancher +* install kubectl +* install helm +* install nfs server + +.. note:: + The Beijing release of OOM only supports Helm 2.8.2 not the 2.7.2 shown in + the screen capture below. The supported versions of all the software components + are listed in the :ref:`cloud-setup-guide-label`. + +.. image:: Apply_customization_script_for_the_Rancher_VM.jpeg + +Launch Instance +--------------- + +.. image:: Rancher-Launch_Instance.jpeg + +Assign Floating IP for external access +-------------------------------------- + +.. image:: Rancher-Allocate_Floating_IP.jpeg + +.. image:: Rancher-Manage_Floating_IP_Associations.jpeg + +.. image:: Rancher-Launch_Instance.jpeg + +Kubernetes Installation +======================= + +Launch new VM instance(s) to create a Kubernetes single host or cluster +----------------------------------------------------------------------- + +To create a cluster: + +.. note:: + #. do not append a '-1' suffix (e.g. sb4-k8s) + #. increase count to the # of of kubernetes worker nodes you want (eg. 3) + +.. image:: K8s-Launch_new_VM_instance_to_create_a_Kubernetes_single_host_or_cluster.jpeg + +Select Ubuntu 16.04 as base image +--------------------------------- +Select "No" on "Create New Volume" + +.. image:: K8s-Select_Ubuntu_16.04_as_base_image.jpeg + +Select Flavor +------------- +The size of a Kubernetes host depends on the size of the ONAP deployment that +will be installed. + +As of the Beijing release a minimum of 3 x 32GB hosts will be needed to run a +full ONAP deployment (all components). + +If a small subset of ONAP components are being deployed for testing purposes, +then a single 16GB or 32GB host should suffice. + +.. image:: K8s-Select_Flavor.jpeg + +Networking +----------- + +.. image:: K8s-Networking.jpeg + +Security Group +--------------- + +.. image:: K8s-Security_Group.jpeg + +Key Pair +-------- +Use an existing key pair (e.g. onap_key), import an existing one or create a +new one to assign. + +.. image:: K8s-Key_Pair.jpeg + +Apply customization script for Kubernetes VM(s) +----------------------------------------------- + +Click :download:`openstack-k8s-node.sh <openstack-k8s-node.sh>` to +download the script. + +.. literalinclude:: openstack-k8s-node.sh + :language: bash + +This customization script will: + +* setup root access to the VM (comment out if you wish to disable this + capability and restrict access to ssh access only) +* install docker +* install kubectl +* install helm +* install nfs common (see configuration step here) + +.. note:: + Ensure you are using the correct versions as described in the + :ref:`cloud-setup-guide-label` + +Launch Instance +--------------- + +.. image:: K8s-Launch_Instance.jpeg + +Assign Floating IP for external access +-------------------------------------- + +.. image:: K8s-Assign_Floating_IP_for_external_access.jpeg + +.. image:: K8s-Manage_Floating_IP_Associations.jpeg + +.. image:: K8s-Launch_Instance.jpeg + +Setting up an NFS share for Multinode Kubernetes Clusters +========================================================= +The figure below illustrates a possible topology of a multinode Kubernetes +cluster. + +.. image:: k8s-topology.jpg + +One node, the Master Node, runs Rancher and Helm clients and connects to all +the Kubernetes nodes in the cluster. Kubernetes nodes, in turn, run Rancher, +Kubernetes and Tiller (Helm) agents, which receive, execute, and respond to +commands issued by the Master Node (e.g. kubectl or helm operations). Note that +the Master Node can be either a remote machine that the user can log in to or a +local machine (e.g. laptop, desktop) that has access to the Kubernetes cluster. + +Deploying applications to a Kubernetes cluster requires Kubernetes nodes to +share a common, distributed filesystem. One node in the cluster plays the role +of NFS Master (not to confuse with the Master Node that runs Rancher and Helm +clients, which is located outside the cluster), while all the other cluster +nodes play the role of NFS slaves. In the figure above, the left-most cluster +node plays the role of NFS Master (indicated by the crown symbol). To properly +set up an NFS share on Master and Slave nodes, the user can run the scripts +below. + +Click :download:`master_nfs_node.sh <master_nfs_node.sh>` to download the script. + +.. literalinclude:: master_nfs_node.sh + :language: bash + +Click :download:`slave_nfs_node.sh <slave_nfs_node.sh>` to download the script. + +.. literalinclude:: slave_nfs_node.sh + :language: bash + +The master_nfs_node.sh script runs in the NFS Master node and needs the list of +NFS Slave nodes as input, e.g.:: + + > sudo ./master_nfs_node.sh node1_ip node2_ip ... nodeN_ip + +The slave_nfs_node.sh script runs in each NFS Slave node and needs the IP of +the NFS Master node as input, e.g.:: + + > sudo ./slave_nfs_node.sh master_node_ip + +Configuration (Rancher and Kubernetes) +====================================== + +Access Rancher server via web browser +------------------------------------- +(e.g. http://10.12.6.16:8080/env/1a5/apps/stacks) + +.. image:: Access_Rancher_server_via_web_browser.jpeg + +Add Kubernetes Environment to Rancher +------------------------------------- + +1. Select “Manage Environments” + +.. image:: Add_Kubernetes_Environment_to_Rancher.png + +2. Select “Add Environment” + +.. image:: Select_Add_Environment.png + +3. Add unique name for your new Rancher environment + +4. Select the Kubernetes template + +5. Click "create" + +.. image:: Click_create.jpeg + +6. Select the new named environment (ie. SB4) from the dropdown list (top left). + +Rancher is now waiting for a Kubernetes Host to be added. + +.. image:: K8s-Assign_Floating_IP_for_external_access.jpeg + +Add Kubernetes Host +------------------- + +1. If this is the first (or only) host being added - click on the "Add a host" link + +.. image:: K8s-Assign_Floating_IP_for_external_access.jpeg + +and click on "Save" (accept defaults). + +.. image:: and_click_on_Save_accept_defaults.jpeg + +otherwise select INFRASTRUCTURE→ Hosts and click on "Add Host" + +.. image:: otherwise_select_INFRASTRUCTURE_Hosts_and_click_on_Add_Host.jpg + +2. Enter the management IP for the k8s VM (e.g. 10.0.0.4) that was just created. + +3. Click on “Copy to Clipboard” button + +4. Click on “Close” button + +.. image:: Click_on_Close_button.jpeg + +Without the 10.0.0.4 IP - the CATTLE_AGENT will be derived on the host - but it +may not be a routable IP. + +Configure Kubernetes Host +------------------------- + +1. Login to the new Kubernetes Host:: + + > ssh -i ~/oom-key.pem ubuntu@10.12.5.1 + The authenticity of host '10.12.5.172 (10.12.5.172)' can't be established. + ECDSA key fingerprint is SHA256:tqxayN58nCJKOJcWrEZzImkc0qKQHDDfUTHqk4WMcEI. + Are you sure you want to continue connecting (yes/no)? yes + Warning: Permanently added '10.12.5.172' (ECDSA) to the list of known hosts. + Welcome to Ubuntu 16.04.2 LTS (GNU/Linux 4.4.0-64-generic x86_64) + + * Documentation: https://help.ubuntu.com + * Management: https://landscape.canonical.com + * Support: https://ubuntu.com/advantage + + Get cloud support with Ubuntu Advantage Cloud Guest: + http://www.ubuntu.com/business/services/cloud + + 180 packages can be updated. + 100 updates are security updates. + + The programs included with the Ubuntu system are free software; + the exact distribution terms for each program are described in the + individual files in /usr/share/doc/*/copyright. + + Ubuntu comes with ABSOLUTELY NO WARRANTY, to the extent permitted by + applicable law. + + To run a command as administrator (user "root"), use "sudo <command>". + See "man sudo_root" for details. + + ubuntu@sb4-k8s-1:~$ + + +2. Paste Clipboard content and hit enter to install Rancher Agent:: + + ubuntu@sb4-k8s-1:~$ sudo docker run -e CATTLE_AGENT_IP="10.0.0.4“ --rm --privileged -v /var/run/docker.sock:/var/run/docker.sock -v /var/lib/rancher:/var/lib/rancher rancher/agent:v1.2.9 http://10.12.6.16:8080/v1/scripts/5D757C68BD0A2125602A:1514678400000:yKW9xHGJDLvq6drz2eDzR2mjato + Unable to find image 'rancher/agent:v1.2.9' locally + v1.2.9: Pulling From rancher/agent + b3e1c725a85f: Pull complete + 6071086409fc: Pull complete + d0ac3b234321: Pull complete + 87f567b5cf58: Pull complete + a63e24b217c4: Pull complete + d0a3f58caef0: Pull complete + 16914729cfd3: Pull complete + dc5c21984c5b: Pull complete + d7e8f9784b20: Pull complete + Digest: sha256:c21255ac4d94ffbc7b523F870F20ea5189b68Fa3d642800adb4774aab4748e66 + Status: Downloaded newer image for rancher/agent:v1.2.9 + + INFO: Running Agent Registration Process, CATTLE_URL=http://10.12.6.16:8080/v1 + INFO: Attempting to connect to: http://10.12.6.16:8080/v1 + INFO: http://10.12.6.16:8080/v1 is accessible + INFO: Inspecting host capabilities + INFO: Boot2Docker: false + INFO: Host writable: true + INFO: Token: xxxxxxxx + INFO: Running registration + INFO: Printing Environment + INFO: ENV: CATTLE_ACCESS_KEY=98B35AC484FBF820E0AD + INFO: ENV: CATTLE_AGENT_IP=10.0.9.4 + INFO: ENV: CATTLE_HOME=/var/lib/cattle + INFO: ENV: CATTLE_REGISTRATION_ACCESS_KEY=registrationToken + INFO: ENV: CATTLE_REGISTRATION_SECRET_KEY=xxxxxxx + INFO: ENV: CATTLE_SECRET_KEY=xxxxxxx + INFO: ENV: CATTLE_URL=http://10.12.6.16:8080/v1 + INFO: ENV: DETECTED_CATTLE_AGENT_IP=10.12.5.172 + INFO: ENV: RANCHER_AGENT_IMAGE=rancher/agent:v1.2.9 + INFO: Launched Rancher Agent: c27ee0f3dc4c783b0db647ea1f73c35b3843a4b8d60b96375b1a05aa77d83136 + ubuntu@sb4-k8s-1:~$ + +3. Return to Rancher environment (e.g. SB4) and wait for services to complete + (~ 10-15 mins) + +.. image:: Return_to_Rancher_environment_eg_SB4_and_wait_for_services_to_complete_10-15_mins.jpeg + +Configure kubectl and helm +========================== +In this example we are configuring kubectl and helm that have been installed +(as a convenience) onto the rancher and kubernetes hosts. Typically you would +install them both on your PC and remotely connect to the cluster. The following +procedure would remain the same. + +1. Click on CLI and then click on “Generate Config” + +.. image:: Click_on_CLI_and_then_click_on_Generate_Config.jpeg + +2. Click on “Copy to Clipboard” - wait until you see a "token" - do not copy + user+password - the server is not ready at that point + +.. image:: Click_on_Copy_to_Clipboard-wait_until_you_see_a_token-do_not_copy_user+password-the_server_is_not_ready_at_that_point.jpeg + +3. Create a .kube directory in user directory (if one does not exist):: + + ubuntu@sb4-kSs-1:~$ mkdir .kube + ubuntu@sb4-kSs-1:~$ vi .kube/config + +4. Paste contents of Clipboard into a file called “config” and save the file:: + + apiVersion: v1 + kind : Config + clusters: + - cluster: + api-version: v1 + insecure-skip-tls-verify: true + server: "https://10.12.6.16:8080/r/projects/1a7/kubernetes:6443" + name: "SB4" + contexts: + - context: + cluster: "SB4" + user: "SB4" + name: "SB4" + current-context: "SB4" + users: + - name: "SB4" + user: + token: "QmFzaWMgTlRBd01qZzBOemc)TkRrMk1UWkNOMFpDTlVFNlExcHdSa1JhVZreE5XSm1TRGhWU2t0Vk1sQjVhalZaY0dWaFVtZGFVMHQzWW1WWVJtVmpSQT09" + ~ + ~ + ~ + - INSERT -- + +5. Validate that kubectl is able to connect to the kubernetes cluster:: + + ubuntu@sb4-k8s-1:~$ kubectl config get-contexts + CURRENT NAME CLUSTER AUTHINFO NAMESPACE + * SB4 SB4 SB4 + ubuntu@sb4-kSs-1:~$ + +and show running pods:: + + ubuntu@sb4-k8s-1:~$ kubectl get pods --all-namespaces -o=wide + NAMESPACE NAME READY STATUS RESTARTS AGE IP NODE + kube-system heapster—7Gb8cd7b5 -q7p42 1/1 Running 0 13m 10.42.213.49 sb4-k8s-1 + kube-system kube-dns-5d7bM87c9-c6f67 3/3 Running 0 13m 10.42.181.110 sb4-k8s-1 + kube-system kubernetes-dashboard-f9577fffd-kswjg 1/1 Running 0 13m 10.42.105.113 sb4-k8s-1 + kube-system monitoring-grafana-997796fcf-vg9h9 1/1 Running 0 13m 10.42,141.58 sb4-k8s-1 + kube-system monitoring-influxdb-56chd96b-hk66b 1/1 Running 0 13m 10.4Z.246.90 sb4-k8s-1 + kube-system tiller-deploy-cc96d4f6b-v29k9 1/1 Running 0 13m 10.42.147.248 sb4-k8s-1 + ubuntu@sb4-k8s-1:~$ + +6. Validate helm is running at the right version. If not, an error like this + will be displayed:: + + ubuntu@sb4-k8s-1:~$ helm list + Error: incompatible versions c1ient[v2.8.2] server[v2.6.1] + ubuntu@sb4-k8s-1:~$ + +7. Upgrade the server-side component of helm (tiller) via `helm init --upgrade`:: + + ubuntu@sb4-k8s-1:~$ helm init --upgrade + Creating /home/ubuntu/.helm + Creating /home/ubuntu/.helm/repository + Creating /home/ubuntu/.helm/repository/cache + Creating /home/ubuntu/.helm/repository/local + Creating /home/ubuntu/.helm/plugins + Creating /home/ubuntu/.helm/starters + Creating /home/ubuntu/.helm/cache/archive + Creating /home/ubuntu/.helm/repository/repositories.yaml + Adding stable repo with URL: https://kubernetes-charts.storage.googleapis.com + Adding local repo with URL: http://127.0.0.1:8879/charts + $HELM_HOME has been configured at /home/ubuntu/.helm. + + Tiller (the Helm server-side component) has been upgraded to the current version. + Happy Helming! + ubuntu@sb4-k8s-1:~$ + +ONAP Deployment via OOM +======================= +Now that kubernetes and Helm are installed and configured you can prepare to +deploy ONAP. Follow the instructions in the README.md_ or look at the official +documentation to get started: + +- :ref:`quick-start-label` - deploy ONAP on an existing cloud +- :ref:`user-guide-label` - a guide for operators of an ONAP instance + + diff --git a/docs/oom_user_guide.rst b/docs/oom_user_guide.rst index fbd3d9ea97..e63120f1c6 100644 --- a/docs/oom_user_guide.rst +++ b/docs/oom_user_guide.rst @@ -75,8 +75,8 @@ Install Helm ~~~~~~~~~~~~ Helm is used by OOM for package and configuration management. To install Helm, enter the following:: - > wget http://storage.googleapis.com/kubernetes-helm/helm-v2.7.2-linux-amd64.tar.gz - > tar -zxvf helm-v2.7.2-linux-amd64.tar.gz + > wget http://storage.googleapis.com/kubernetes-helm/helm-v2.8.2-linux-amd64.tar.gz + > tar -zxvf helm-v2.8.2-linux-amd64.tar.gz > sudo mv linux-amd64/helm /usr/local/bin/helm Verify the Helm version with:: diff --git a/docs/openstack-k8s-node.sh b/docs/openstack-k8s-node.sh new file mode 100644 index 0000000000..7d49ad5033 --- /dev/null +++ b/docs/openstack-k8s-node.sh @@ -0,0 +1,46 @@ +#!/bin/bash + +DOCKER_VERSION=17.03 +KUBECTL_VERSION=1.8.10 +HELM_VERSION=2.8.2 + +# setup root access - default login: oom/oom - comment out to restrict access too ssh key only +sed -i 's/PermitRootLogin.*/PermitRootLogin yes/' /etc/ssh/sshd_config +sed -i 's/PasswordAuthentication.*/PasswordAuthentication yes/' /etc/ssh/sshd_config +service sshd restart +echo -e "oom\noom" | passwd root + +apt-get update +curl https://releases.rancher.com/install-docker/$DOCKER_VERSION.sh | sh +mkdir -p /etc/systemd/system/docker.service.d/ +cat > /etc/systemd/system/docker.service.d/docker.conf << EOF +[Service] +ExecStart= +ExecStart=/usr/bin/dockerd -H fd:// --insecure-registry=nexus3.onap.org:10001 +EOF +systemctl daemon-reload +systemctl restart docker +apt-mark hold docker-ce + +IP_ADDY=`ip address |grep ens|grep inet|awk '{print $2}'| awk -F / '{print $1}'` +HOSTNAME=`hostname` + +echo "$IP_ADDY $HOSTNAME" >> /etc/hosts + +docker login -u docker -p docker nexus3.onap.org:10001 + +sudo apt-get install make -y + +sudo curl -LO https://storage.googleapis.com/kubernetes-release/release/v$KUBECTL_VERSION/bin/linux/amd64/kubectl +sudo chmod +x ./kubectl +sudo mv ./kubectl /usr/local/bin/kubectl +sudo mkdir ~/.kube +wget http://storage.googleapis.com/kubernetes-helm/helm-v${HELM_VERSION}-linux-amd64.tar.gz +sudo tar -zxvf helm-v${HELM_VERSION}-linux-amd64.tar.gz +sudo mv linux-amd64/helm /usr/local/bin/helm + +# install nfs +sudo apt-get install nfs-common -y + + +exit 0 diff --git a/docs/openstack-rancher.sh b/docs/openstack-rancher.sh new file mode 100644 index 0000000000..d51abe4d68 --- /dev/null +++ b/docs/openstack-rancher.sh @@ -0,0 +1,51 @@ +#!/bin/bash + +DOCKER_VERSION=17.03 +RANCHER_VERSION=1.6.14 +KUBECTL_VERSION=1.8.10 +HELM_VERSION=2.8.2 + +# setup root access - default login: oom/oom - comment out to restrict access too ssh key only +sed -i 's/PermitRootLogin.*/PermitRootLogin yes/' /etc/ssh/sshd_config +sed -i 's/PasswordAuthentication.*/PasswordAuthentication yes/' /etc/ssh/sshd_config +service sshd restart +echo -e "oom\noom" | passwd root + +apt-get update +curl https://releases.rancher.com/install-docker/$DOCKER_VERSION.sh | sh +mkdir -p /etc/systemd/system/docker.service.d/ +cat > /etc/systemd/system/docker.service.d/docker.conf << EOF +[Service] +ExecStart= +ExecStart=/usr/bin/dockerd -H fd:// --insecure-registry=nexus3.onap.org:10001 +EOF +systemctl daemon-reload +systemctl restart docker +apt-mark hold docker-ce + +IP_ADDY=`ip address |grep ens|grep inet|awk '{print $2}'| awk -F / '{print $1}'` +HOSTNAME=`hostname` + +echo "$IP_ADDY $HOSTNAME" >> /etc/hosts + +docker login -u docker -p docker nexus3.onap.org:10001 + +sudo apt-get install make -y + +sudo docker run -d --restart=unless-stopped -p 8080:8080 --name rancher_server rancher/server:v$RANCHER_VERSION +sudo curl -LO https://storage.googleapis.com/kubernetes-release/release/v$KUBECTL_VERSION/bin/linux/amd64/kubectl +sudo chmod +x ./kubectl +sudo mv ./kubectl /usr/local/bin/kubectl +sudo mkdir ~/.kube +wget http://storage.googleapis.com/kubernetes-helm/helm-v${HELM_VERSION}-linux-amd64.tar.gz +sudo tar -zxvf helm-v${HELM_VERSION}-linux-amd64.tar.gz +sudo mv linux-amd64/helm /usr/local/bin/helm + +# nfs server +sudo apt-get install nfs-kernel-server -y + +sudo mkdir -p /nfs_share +sudo chown nobody:nogroup /nfs_share/ + + +exit 0 diff --git a/docs/otherwise_select_INFRASTRUCTURE_Hosts_and_click_on_Add_Host.jpg b/docs/otherwise_select_INFRASTRUCTURE_Hosts_and_click_on_Add_Host.jpg Binary files differnew file mode 100644 index 0000000000..18af758a0d --- /dev/null +++ b/docs/otherwise_select_INFRASTRUCTURE_Hosts_and_click_on_Add_Host.jpg diff --git a/docs/release-notes.rst b/docs/release-notes.rst index 2573c2ca02..57a1fab997 100644 --- a/docs/release-notes.rst +++ b/docs/release-notes.rst @@ -8,66 +8,10 @@ Release Notes ============= -Version: 1.1.0 --------------- - -:Release Date: 2017-11-16 - -**New Features** - -The Amsterdam release is the first release of the ONAP Operations Manager (OOM). - -The main goal of the Amsterdam release was to: - - - Support Flexible Platform Deployment via Kubernetes of fully containerized OMAP components - on any type of environment. - - Support State Management of ONAP platform components. - - Support full production ONAP deployment and any variation of component level deployment for development. - - Platform Operations Orchestration / Control Loop Actions. - - Platform centralized logging with ELK stack. - -**Bug Fixes** - - The full list of implemented user stories and epics is available on `JIRA <https://jira.onap.org/secure/RapidBoard.jspa?rapidView=41&view=planning.nodetail&epics=visible>`_ - This is the first release of OOM, the defects fixed in this release were raised during the course of the release. - Anything not closed is captured below under Known Issues. If you want to review the defects fixed in the Amsterdam release, refer to Jira link above. - -**Known Issues** - - `OOM-6 <https://jira.onap.org/browse/OOM-6>`_ Automated platform deployment on Docker/Kubernetes - - VFC, AAF, MSB minor issues. - - Workaround: Manual configuration changes - however the reference vFirewall use case does not currently require these components. - - - `OOM-10 <https://jira.onap.org/browse/OOM-10>`_ Platform configuration management. - - OOM ONAP Configuration Management - Handling of Secrets. - - Workaround: Automated workaround to be able to pull from protected docker repositories. - - -**Security Issues** - N/A - - -**Upgrade Notes** - - N/A - -**Deprecation Notes** - - N/A - -**Other** - - - -=========== - - Version 2.0.0 Beijing Release ----------------------------- -:Release Date: 2018-05-24 +:Release Date: 2018-06-07 Epic **** @@ -444,5 +388,66 @@ Sub-task * [`OOM-1036 <https://jira.onap.org/browse/OOM-1036>`_] - update helm from 2.7.2 to 2.8.2 wiki/rtd * [`OOM-1063 <https://jira.onap.org/browse/OOM-1063>`_] - Document Portal LoadBalancer Ingress IP Settings +**Security Notes** + +OOM code has been formally scanned during build time using NexusIQ and no Critical vulnerability was found. + +Quick Links: + - `OOM project page <https://wiki.onap.org/display/DW/ONAP+Operations+Manager+Project>`_ + + - `Passing Badge information for OOM <https://bestpractices.coreinfrastructure.org/en/projects/1631>`_ + +Version: 1.1.0 +-------------- + +:Release Date: 2017-11-16 + +**New Features** + +The Amsterdam release is the first release of the ONAP Operations Manager (OOM). + +The main goal of the Amsterdam release was to: + + - Support Flexible Platform Deployment via Kubernetes of fully containerized OMAP components - on any type of environment. + - Support State Management of ONAP platform components. + - Support full production ONAP deployment and any variation of component level deployment for development. + - Platform Operations Orchestration / Control Loop Actions. + - Platform centralized logging with ELK stack. + +**Bug Fixes** + + The full list of implemented user stories and epics is available on `JIRA <https://jira.onap.org/secure/RapidBoard.jspa?rapidView=41&view=planning.nodetail&epics=visible>`_ + This is the first release of OOM, the defects fixed in this release were raised during the course of the release. + Anything not closed is captured below under Known Issues. If you want to review the defects fixed in the Amsterdam release, refer to Jira link above. + +**Known Issues** + - `OOM-6 <https://jira.onap.org/browse/OOM-6>`_ Automated platform deployment on Docker/Kubernetes + + VFC, AAF, MSB minor issues. + + Workaround: Manual configuration changes - however the reference vFirewall use case does not currently require these components. + + - `OOM-10 <https://jira.onap.org/browse/OOM-10>`_ Platform configuration management. + + OOM ONAP Configuration Management - Handling of Secrets. + + Workaround: Automated workaround to be able to pull from protected docker repositories. + + +**Security Issues** + N/A + + +**Upgrade Notes** + + N/A + +**Deprecation Notes** + + N/A + +**Other** + + N/A End of Release Notes diff --git a/docs/slave_nfs_node.sh b/docs/slave_nfs_node.sh new file mode 100644 index 0000000000..5cb164ccac --- /dev/null +++ b/docs/slave_nfs_node.sh @@ -0,0 +1,25 @@ +#!/bin/bash + +usage () { + echo "Usage:" + echo " ./$(basename $0) nfs_master_ip" + exit 1 +} + +if [ "$#" -ne 1 ]; then + echo "Missing NFS mater node" + usage +fi + +MASTER_IP=$1 + +#Install NFS common +sudo apt-get update +sudo apt-get install -y nfs-common + +#Create NFS directory +sudo mkdir -p /dockerdata-nfs + +#Mount the remote NFS directory to the local one +sudo mount $MASTER_IP:/dockerdata-nfs /dockerdata-nfs/ +echo "$MASTER_IP:/dockerdata-nfs /dockerdata-nfs nfs auto,nofail,noatime,nolock,intr,tcp,actimeo=1800 0 0" | sudo tee -a /etc/fstab diff --git a/kubernetes/aai/charts/aai-babel/values.yaml b/kubernetes/aai/charts/aai-babel/values.yaml index 7118d1669b..2239c9a11d 100644 --- a/kubernetes/aai/charts/aai-babel/values.yaml +++ b/kubernetes/aai/charts/aai-babel/values.yaml @@ -24,7 +24,7 @@ global: ################################################################# # application image -image: onap/babel:1.2-STAGING-latest +image: onap/babel:1.2.0 # application configuration config: diff --git a/kubernetes/aai/charts/aai-champ/values.yaml b/kubernetes/aai/charts/aai-champ/values.yaml index 13cc9e77c0..78afd419ca 100644 --- a/kubernetes/aai/charts/aai-champ/values.yaml +++ b/kubernetes/aai/charts/aai-champ/values.yaml @@ -24,7 +24,7 @@ global: ################################################################# # application image -image: onap/champ:1.2-STAGING-latest +image: onap/champ:1.2.3 # application configuration config: diff --git a/kubernetes/aai/charts/aai-data-router/values.yaml b/kubernetes/aai/charts/aai-data-router/values.yaml index 5ce7689822..51fc65a408 100644 --- a/kubernetes/aai/charts/aai-data-router/values.yaml +++ b/kubernetes/aai/charts/aai-data-router/values.yaml @@ -7,7 +7,7 @@ global: # global defaults # application image repository: nexus3.onap.org:10001 -image: onap/data-router:1.2-STAGING-latest +image: onap/data-router:1.2.2 pullPolicy: Always restartPolicy: Always diff --git a/kubernetes/aai/charts/aai-gizmo/resources/config/log/logback.xml b/kubernetes/aai/charts/aai-gizmo/resources/config/log/logback.xml index f63afd3c6b..a04d44cefa 100644 --- a/kubernetes/aai/charts/aai-gizmo/resources/config/log/logback.xml +++ b/kubernetes/aai/charts/aai-gizmo/resources/config/log/logback.xml @@ -1,213 +1,179 @@ -<?xml version="1.0" encoding="UTF-8"?> -<configuration scan="true" scanPeriod="3 seconds" debug="true"> - <!--<jmxConfigurator /> --> - <!-- directory path for all other type logs --> - <property name="logDir" value="/var/log/onap" /> - <property name="componentName" value="aai" /> - <property name="subComponentName" value="aai-CRUD" /> - <property name="logDirectory" value="${logDir}/${componentName}/${subComponentName}" /> - <!-- default eelf log file names --> - <property name="generalLogName" value="application" /> - <property name="errorLogName" value="error" /> - <property name="metricsLogName" value="metrics" /> - <property name="auditLogName" value="audit" /> - <property name="debugLogName" value="debug" /> - <property name="queueSize" value="256" /> - <property name="maxFileSize" value="50MB" /> - <property name="maxHistory" value="30" /> - <property name="totalSizeCap" value="10GB" /> - <property name="pattern" value="%d{"yyyy-MM-dd'T'HH:mm:ss.SSSXXX", UTC}\t[%thread]\t%-5level\t%logger\t%replace(%replace(%replace(%mdc){'\t','\\\\t'}){', ','\t'}){'\n', '\\\\n'}\t%replace(%replace(%msg){'\n', '\\\\n'}){'\t','\\\\t'}%n" /> - <!-- Example evaluator filter applied against console appender --> - <appender name="STDOUT" class="ch.qos.logback.core.ConsoleAppender"> - <encoder> - <pattern>${pattern}</pattern> - </encoder> - </appender> - <!-- ============================================================================ --> - <!-- EELF Appenders --> - <!-- ============================================================================ --> - <!-- The EELFAppender is used to record events to the general application +<configuration scan="true" scanPeriod="3 seconds" debug="false"> + <!--<jmxConfigurator /> --> + <!-- directory path for all other type logs --> + + <property name="logDir" value="/var/log/onap" /> + + + <!-- specify the component name + <ECOMP-component-name>::= "MSO" | "DCAE" | "ASDC " | "AAI" |"Policy" | "SDNC" | "AC" --> + <property name="componentName" value="AAI-CRUD" /> + + <!-- default eelf log file names --> + <property name="generalLogName" value="error" /> + <property name="metricsLogName" value="metrics" /> + <property name="auditLogName" value="audit" /> + <property name="debugLogName" value="debug" /> + + <property name="errorLogPattern" value="%d{yyyy-MM-dd'T'HH:mm:ss.SSSXXX}|%mdc{RequestId}|%thread|GIZMO|%mdc{PartnerName}|%logger||%.-5level|%msg%n" /> + <property name="auditMetricPattern" value="%m%n" /> + + <property name="logDirectory" value="${logDir}/${componentName}" /> + + <!-- Example evaluator filter applied against console appender --> + <appender name="STDOUT" class="ch.qos.logback.core.ConsoleAppender"> + <encoder> + <pattern>${errorLogPattern}</pattern> + </encoder> + </appender> + + <!-- ============================================================================ --> + <!-- EELF Appenders --> + <!-- ============================================================================ --> + + <!-- The EELFAppender is used to record events to the general application log --> - <appender name="EELF" class="ch.qos.logback.core.rolling.RollingFileAppender"> - <file>${logDirectory}/${generalLogName}.log</file> - <rollingPolicy class="ch.qos.logback.core.rolling.TimeBasedRollingPolicy"> - <fileNamePattern>${logDirectory}/${generalLogName}.%d{yyyy-MM-dd}.%i.log</fileNamePattern> - <timeBasedFileNamingAndTriggeringPolicy class="ch.qos.logback.core.rolling.SizeAndTimeBasedFNATP"> - <maxFileSize>${maxFileSize}</maxFileSize> - </timeBasedFileNamingAndTriggeringPolicy> - <maxHistory>${maxHistory}</maxHistory> - <totalSizeCap>${totalSizeCap}</totalSizeCap> - </rollingPolicy> - <encoder> - <pattern>${pattern}</pattern> - </encoder> - </appender> - <appender name="asyncEELF" class="ch.qos.logback.classic.AsyncAppender"> - <!-- deny all events with a level below INFO, that is TRACE and DEBUG --> - <filter class="ch.qos.logback.classic.filter.ThresholdFilter"> - <level>INFO</level> - </filter> - <queueSize>${queueSize}</queueSize> - <appender-ref ref="EELF" /> - </appender> - <!-- EELF Audit Appender. This appender is used to record audit engine - related logging events. The audit logger and appender are specializations - of the EELF application root logger and appender. This can be used to segregate - Policy engine events from other components, or it can be eliminated to record + + <appender name="EELF" + class="ch.qos.logback.core.rolling.RollingFileAppender"> + <file>${logDirectory}/${generalLogName}.log</file> + <rollingPolicy + class="ch.qos.logback.core.rolling.TimeBasedRollingPolicy"> + <fileNamePattern>${logDirectory}/${generalLogName}.%d{yyyy-MM-dd}.log.zip + </fileNamePattern> + <maxHistory>60</maxHistory> + </rollingPolicy> + <encoder> + <pattern>${errorLogPattern}</pattern> + </encoder> + </appender> + <appender name="asyncEELF" class="ch.qos.logback.classic.AsyncAppender"> + <!-- deny all events with a level below INFO, that is TRACE and DEBUG --> + <filter class="ch.qos.logback.classic.filter.ThresholdFilter"> + <level>INFO</level> + </filter> + <queueSize>256</queueSize> + <appender-ref ref="EELF" /> + </appender> + + + <!-- EELF Audit Appender. This appender is used to record audit engine + related logging events. The audit logger and appender are specializations + of the EELF application root logger and appender. This can be used to segregate + Policy engine events from other components, or it can be eliminated to record these events as part of the application root log. --> - <appender name="EELFAudit" class="ch.qos.logback.core.rolling.RollingFileAppender"> - <file>${logDirectory}/${auditLogName}.log</file> - <rollingPolicy class="ch.qos.logback.core.rolling.TimeBasedRollingPolicy"> - <fileNamePattern>${logDirectory}/${auditLogName}.%d{yyyy-MM-dd}.%i.log</fileNamePattern> - <timeBasedFileNamingAndTriggeringPolicy class="ch.qos.logback.core.rolling.SizeAndTimeBasedFNATP"> - <maxFileSize>${maxFileSize}</maxFileSize> - </timeBasedFileNamingAndTriggeringPolicy> - <maxHistory>${maxHistory}</maxHistory> - <totalSizeCap>${totalSizeCap}</totalSizeCap> - </rollingPolicy> - <encoder> - <pattern>${pattern}</pattern> - </encoder> - </appender> - <appender name="asyncEELFAudit" class="ch.qos.logback.classic.AsyncAppender"> - <queueSize>${queueSize}</queueSize> - <appender-ref ref="EELFAudit" /> - </appender> - <appender name="EELFMetrics" class="ch.qos.logback.core.rolling.RollingFileAppender"> - <file>${logDirectory}/${metricsLogName}.log</file> - <rollingPolicy class="ch.qos.logback.core.rolling.TimeBasedRollingPolicy"> - <fileNamePattern>${logDirectory}/${metricsLogName}.%d{yyyy-MM-dd}.%i.log</fileNamePattern> - <timeBasedFileNamingAndTriggeringPolicy class="ch.qos.logback.core.rolling.SizeAndTimeBasedFNATP"> - <maxFileSize>${maxFileSize}</maxFileSize> - </timeBasedFileNamingAndTriggeringPolicy> - <maxHistory>${maxHistory}</maxHistory> - <totalSizeCap>${totalSizeCap}</totalSizeCap> - </rollingPolicy> - <encoder> - <pattern>${pattern}</pattern> - </encoder> - </appender> - <appender name="asyncEELFMetrics" class="ch.qos.logback.classic.AsyncAppender"> - <queueSize>${queueSize}</queueSize> - <appender-ref ref="EELFMetrics" /> - </appender> - <appender name="EELFError" class="ch.qos.logback.core.rolling.RollingFileAppender"> - <file>${logDirectory}/${errorLogName}.log</file> - <rollingPolicy class="ch.qos.logback.core.rolling.TimeBasedRollingPolicy"> - <fileNamePattern>${logDirectory}/${errorLogName}.%d{yyyy-MM-dd}.%i.log</fileNamePattern> - <timeBasedFileNamingAndTriggeringPolicy class="ch.qos.logback.core.rolling.SizeAndTimeBasedFNATP"> - <maxFileSize>${maxFileSize}</maxFileSize> - </timeBasedFileNamingAndTriggeringPolicy> - <maxHistory>${maxHistory}</maxHistory> - <totalSizeCap>${totalSizeCap}</totalSizeCap> - </rollingPolicy> - <encoder> - <pattern>${pattern}</pattern> - </encoder> - <filter class="ch.qos.logback.classic.filter.ThresholdFilter"> - <level>INFO</level> - </filter> - </appender> - <appender name="asyncEELFError" class="ch.qos.logback.classic.AsyncAppender"> - <queueSize>${queueSize}</queueSize> - <appender-ref ref="EELFError" /> - </appender> - <appender name="EELFDebug" class="ch.qos.logback.core.rolling.RollingFileAppender"> - <file>${logDirectory}/${debugLogName}.log</file> - <rollingPolicy class="ch.qos.logback.core.rolling.TimeBasedRollingPolicy"> - <fileNamePattern>${logDirectory}/${debugLogName}.%d{yyyy-MM-dd}.%i.log</fileNamePattern> - <timeBasedFileNamingAndTriggeringPolicy class="ch.qos.logback.core.rolling.SizeAndTimeBasedFNATP"> - <maxFileSize>${maxFileSize}</maxFileSize> - </timeBasedFileNamingAndTriggeringPolicy> - <maxHistory>${maxHistory}</maxHistory> - <totalSizeCap>${totalSizeCap}</totalSizeCap> - </rollingPolicy> - <encoder> - <pattern>${pattern}</pattern> - </encoder> - </appender> - <appender name="asyncEELFDebug" class="ch.qos.logback.classic.AsyncAppender"> - <queueSize>${queueSize}</queueSize> - <appender-ref ref="EELFDebug" /> - <includeCallerData>true</includeCallerData> - </appender> - <!-- ============================================================================ --> - <!-- EELF loggers --> - <!-- ============================================================================ --> - <logger name="com.att.eelf" level="info" additivity="false"> - <appender-ref ref="asyncEELF" /> - </logger> - <logger name="com.att.eelf.security" level="info" additivity="false"> - <appender-ref ref="asyncEELFSecurity" /> - </logger> - <logger name="com.att.eelf.perf" level="info" additivity="false"> - <appender-ref ref="asyncEELFPerformance" /> - </logger> - <logger name="com.att.eelf.server" level="info" additivity="false"> - <appender-ref ref="asyncEELFServer" /> - </logger> - <logger name="com.att.eelf.policy" level="info" additivity="false"> - <appender-ref ref="asyncEELFPolicy" /> - </logger> - <logger name="com.att.eelf.debug" level="debug" additivity="false"> - <appender-ref ref="asyncEELFDebug" /> - </logger> - <logger name="com.att.eelf.audit" level="info" additivity="false"> - <appender-ref ref="asyncEELFAudit" /> - </logger> - <logger name="com.att.eelf.metrics" level="info" additivity="false"> - <appender-ref ref="asyncEELFMetrics" /> - </logger> - <logger name="com.att.eelf.error" level="info" additivity="false"> - <appender-ref ref="asyncEELFError" /> - </logger> - <!-- Spring related loggers --> - <logger name="org.springframework" level="WARN" /> - <logger name="org.springframework.beans" level="WARN" /> - <logger name="org.springframework.web" level="WARN" /> - <logger name="com.blog.spring.jms" level="WARN" /> - <!-- AJSC Services (bootstrap services) --> - <logger name="ajsc" level="WARN" /> - <logger name="ajsc.RouteMgmtService" level="WARN" /> - <logger name="ajsc.ComputeService" level="WARN" /> - <logger name="ajsc.VandelayService" level="WARN" /> - <logger name="ajsc.FilePersistenceService" level="WARN" /> - <logger name="ajsc.UserDefinedJarService" level="WARN" /> - <logger name="ajsc.UserDefinedBeansDefService" level="WARN" /> - <logger name="ajsc.LoggingConfigurationService" level="WARN" /> - <!-- AJSC related loggers (DME2 Registration, csi logging, restlet, servlet - logging) --> - <logger name="ajsc.utils" level="WARN" /> - <logger name="ajsc.utils.DME2Helper" level="WARN" /> - <logger name="ajsc.filters" level="WARN" /> - <logger name="ajsc.beans.interceptors" level="WARN" /> - <logger name="ajsc.restlet" level="WARN" /> - <logger name="ajsc.servlet" level="WARN" /> - <logger name="com.att" level="INFO" /> - <logger name="com.att.ajsc.csi.logging" level="WARN" /> - <logger name="com.att.ajsc.filemonitor" level="WARN" /> - <!-- CRUD Service loggers --> - <logger name="org.openecomp.crud" level="INFO" /> - <!-- Other Loggers that may help troubleshoot --> - <logger name="net.sf" level="WARN" /> - <logger name="org.apache" level="WARN" /> - <logger name="org.apache.commons.httpclient" level="WARN" /> - <logger name="org.apache.commons" level="WARN" /> - <logger name="org.apache.coyote" level="WARN" /> - <logger name="org.apache.jasper" level="WARN" /> - <!-- Camel Related Loggers (including restlet/servlet/jaxrs/cxf logging. + + <appender name="EELFAudit" + class="ch.qos.logback.core.rolling.RollingFileAppender"> + <file>${logDirectory}/${auditLogName}.log</file> + <rollingPolicy + class="ch.qos.logback.core.rolling.TimeBasedRollingPolicy"> + <fileNamePattern>${logDirectory}/${auditLogName}.%d{yyyy-MM-dd}.log.zip + </fileNamePattern> + <maxHistory>60</maxHistory> + </rollingPolicy> + <encoder> + <pattern>${auditMetricPattern}</pattern> + </encoder> + </appender> + <appender name="asyncEELFAudit" class="ch.qos.logback.classic.AsyncAppender"> + <queueSize>256</queueSize> + <appender-ref ref="EELFAudit" /> + </appender> + + <appender name="EELFMetrics" + class="ch.qos.logback.core.rolling.RollingFileAppender"> + <file>${logDirectory}/${metricsLogName}.log</file> + <rollingPolicy + class="ch.qos.logback.core.rolling.TimeBasedRollingPolicy"> + <fileNamePattern>${logDirectory}/${metricsLogName}.%d{yyyy-MM-dd}.log.zip + </fileNamePattern> + <maxHistory>60</maxHistory> + </rollingPolicy> + <encoder> + <!-- <pattern>"%d{HH:mm:ss.SSS} [%thread] %-5level %logger{1024} - + %msg%n"</pattern> --> + <pattern>${auditMetricPattern}</pattern> + </encoder> + </appender> + + + <appender name="asyncEELFMetrics" class="ch.qos.logback.classic.AsyncAppender"> + <queueSize>256</queueSize> + <appender-ref ref="EELFMetrics"/> + </appender> + + <appender name="EELFDebug" + class="ch.qos.logback.core.rolling.RollingFileAppender"> + <file>${logDirectory}/${debugLogName}.log</file> + <rollingPolicy + class="ch.qos.logback.core.rolling.TimeBasedRollingPolicy"> + <fileNamePattern>${logDirectory}/${debugLogName}.%d{yyyy-MM-dd}.log.zip + </fileNamePattern> + <maxHistory>60</maxHistory> + </rollingPolicy> + <encoder> + <pattern>${errorLogPattern}</pattern> + </encoder> + </appender> + + <appender name="asyncEELFDebug" class="ch.qos.logback.classic.AsyncAppender"> + <queueSize>256</queueSize> + <appender-ref ref="EELFDebug" /> + <includeCallerData>false</includeCallerData> + </appender> + + + <!-- ============================================================================ --> + <!-- EELF loggers --> + <!-- ============================================================================ --> + <logger name="com.att.eelf" level="info" additivity="false"> + <appender-ref ref="asyncEELF" /> + <appender-ref ref="asyncEELFDebug" /> + </logger> + + <logger name="com.att.eelf.audit" level="info" additivity="false"> + <appender-ref ref="asyncEELFAudit" /> + </logger> + <logger name="com.att.eelf.metrics" level="info" additivity="false"> + <appender-ref ref="asyncEELFMetrics" /> + </logger> + + <!-- Spring related loggers --> + <logger name="org.springframework" level="WARN" /> + <logger name="org.springframework.beans" level="WARN" /> + <logger name="org.springframework.web" level="WARN" /> + <logger name="com.blog.spring.jms" level="WARN" /> + + <!-- ARMAdapter Service loggers --> + <logger name="com.amdocs.aai.armadapter" level="INFO" /> + + <!-- Other Loggers that may help troubleshoot --> + <logger name="net.sf" level="WARN" /> + <logger name="org.apache" level="WARN" /> + <logger name="org.apache.commons.httpclient" level="WARN" /> + <logger name="org.apache.commons" level="WARN" /> + <logger name="org.apache.coyote" level="WARN" /> + <logger name="org.apache.jasper" level="WARN" /> + + <!-- Camel Related Loggers (including restlet/servlet/jaxrs/cxf logging. May aid in troubleshooting) --> - <logger name="org.apache.camel" level="WARN" /> - <logger name="org.apache.cxf" level="WARN" /> - <logger name="org.apache.camel.processor.interceptor" level="WARN" /> - <logger name="org.apache.cxf.jaxrs.interceptor" level="WARN" /> - <logger name="org.apache.cxf.service" level="WARN" /> - <logger name="org.restlet" level="WARN" /> - <logger name="org.apache.camel.component.restlet" level="WARN" /> - <!-- logback internals logging --> - <logger name="ch.qos.logback.classic" level="WARN" /> - <logger name="ch.qos.logback.core" level="WARN" /> - <root level="INFO"> - <appender-ref ref="asyncEELFDebug" /> - </root> + <logger name="org.apache.camel" level="WARN" /> + <logger name="org.apache.cxf" level="WARN" /> + <logger name="org.apache.camel.processor.interceptor" level="WARN" /> + <logger name="org.apache.cxf.jaxrs.interceptor" level="WARN" /> + <logger name="org.apache.cxf.service" level="WARN" /> + <logger name="org.restlet" level="WARN" /> + <logger name="org.apache.camel.component.restlet" level="WARN" /> + + + <!-- logback internals logging --> + <logger name="ch.qos.logback.classic" level="WARN" /> + <logger name="ch.qos.logback.core" level="WARN" /> + + <root> + <appender-ref ref="asyncEELF" /> + <!-- <appender-ref ref="asyncEELFDebug" /> --> + </root> + </configuration> diff --git a/kubernetes/aai/charts/aai-gizmo/values.yaml b/kubernetes/aai/charts/aai-gizmo/values.yaml index 1ea924fc13..acf64172cc 100644 --- a/kubernetes/aai/charts/aai-gizmo/values.yaml +++ b/kubernetes/aai/charts/aai-gizmo/values.yaml @@ -22,7 +22,7 @@ global: ################################################################# # application image -image: onap/gizmo:1.1-STAGING-latest +image: onap/gizmo:1.2.1 # application configuration config: diff --git a/kubernetes/aai/charts/aai-modelloader/values.yaml b/kubernetes/aai/charts/aai-modelloader/values.yaml index 676fc6fdfc..8acb28f9d2 100644 --- a/kubernetes/aai/charts/aai-modelloader/values.yaml +++ b/kubernetes/aai/charts/aai-modelloader/values.yaml @@ -7,7 +7,7 @@ global: # global defaults # application image repository: nexus3.onap.org:10001 -image: onap/model-loader:1.2-STAGING-latest +image: onap/model-loader:1.2.1 pullPolicy: Always restartPolicy: Always diff --git a/kubernetes/aai/charts/aai-resources/values.yaml b/kubernetes/aai/charts/aai-resources/values.yaml index 7b55aa75e2..50719e4bfb 100644 --- a/kubernetes/aai/charts/aai-resources/values.yaml +++ b/kubernetes/aai/charts/aai-resources/values.yaml @@ -9,7 +9,7 @@ global: # global defaults # application image repository: nexus3.onap.org:10001 -image: onap/aai-resources:1.2-STAGING-latest +image: onap/aai-resources:1.2.1 pullPolicy: Always restartPolicy: Always diff --git a/kubernetes/aai/charts/aai-search-data/values.yaml b/kubernetes/aai/charts/aai-search-data/values.yaml index c6f42bd38b..f781662e79 100644 --- a/kubernetes/aai/charts/aai-search-data/values.yaml +++ b/kubernetes/aai/charts/aai-search-data/values.yaml @@ -9,7 +9,7 @@ global: # global defaults # application image repository: nexus3.onap.org:10001 -image: onap/search-data-service:1.2-STAGING-latest +image: onap/search-data-service:1.2.1 pullPolicy: Always restartPolicy: Always diff --git a/kubernetes/aai/charts/aai-sparky-be/values.yaml b/kubernetes/aai/charts/aai-sparky-be/values.yaml index b9ef4ca85d..c87ec5bf02 100644 --- a/kubernetes/aai/charts/aai-sparky-be/values.yaml +++ b/kubernetes/aai/charts/aai-sparky-be/values.yaml @@ -14,7 +14,7 @@ global: # global defaults # application image repository: nexus3.onap.org:10001 -image: onap/sparky-be:1.2-STAGING-latest +image: onap/sparky-be:1.2.1 pullPolicy: Always restartPolicy: Always diff --git a/kubernetes/aai/charts/aai-traversal/values.yaml b/kubernetes/aai/charts/aai-traversal/values.yaml index c22ae75c3f..943d34a3f5 100644 --- a/kubernetes/aai/charts/aai-traversal/values.yaml +++ b/kubernetes/aai/charts/aai-traversal/values.yaml @@ -9,7 +9,7 @@ global: # global defaults # application image repository: nexus3.onap.org:10001 -image: onap/aai-traversal:1.2-STAGING-latest +image: onap/aai-traversal:1.2.1 pullPolicy: Always restartPolicy: Always diff --git a/kubernetes/appc/charts/appc-cdt/templates/deployment.yaml b/kubernetes/appc/charts/appc-cdt/templates/deployment.yaml index 2226b75bbf..a2cb9fc83a 100644 --- a/kubernetes/appc/charts/appc-cdt/templates/deployment.yaml +++ b/kubernetes/appc/charts/appc-cdt/templates/deployment.yaml @@ -21,6 +21,8 @@ spec: - name: {{ include "common.name" . }} image: "{{ include "common.repository" . }}/{{ .Values.image }}" imagePullPolicy: {{ .Values.global.pullPolicy | default .Values.pullPolicy }} + command: + - /opt/startCdt.sh ports: - containerPort: {{ .Values.service.internalPort }} name: {{ .Values.service.name }} @@ -39,6 +41,11 @@ spec: initialDelaySeconds: {{ .Values.readiness.initialDelaySeconds }} periodSeconds: {{ .Values.readiness.periodSeconds }} env: + # This sets the port that CDT will use to connect to the main appc container. + # The 32 is the node port suffix that is used in the main appc oom templates + # for nodePort3. + - name: CDT_PORT + value: "{{ .Values.global.nodePortPrefix | default .Values.nodePortPrefix }}{{ .Values.nodePort3 }}" volumeMounts: - mountPath: /etc/localtime name: localtime diff --git a/kubernetes/appc/resources/config/appc/opt/onap/appc/bin/startODL.sh b/kubernetes/appc/resources/config/appc/opt/onap/appc/bin/startODL.sh index e13193f364..a990739d55 100755 --- a/kubernetes/appc/resources/config/appc/opt/onap/appc/bin/startODL.sh +++ b/kubernetes/appc/resources/config/appc/opt/onap/appc/bin/startODL.sh @@ -173,5 +173,8 @@ fi appcInstallEndTime=$(date +%s) echo "Total Appc install took $(expr $appcInstallEndTime - $appcInstallStartTime) seconds" +echo "Starting cdt-proxy-service jar, logging to ${APPC_HOME}/cdt-proxy-service/jar.log" +java -jar ${APPC_HOME}/cdt-proxy-service/cdt-proxy-service.jar > ${APPC_HOME}/cdt-proxy-service/jar.log & + exec ${ODL_HOME}/bin/karaf diff --git a/kubernetes/appc/templates/service.yaml b/kubernetes/appc/templates/service.yaml index 6c85985854..733c4ca3be 100644 --- a/kubernetes/appc/templates/service.yaml +++ b/kubernetes/appc/templates/service.yaml @@ -31,6 +31,10 @@ spec: - port: {{ .Values.service.externalPort2 }} nodePort: {{ .Values.global.nodePortPrefix | default .Values.nodePortPrefix }}{{ .Values.service.nodePort2 }} name: "{{ .Values.service.portName }}-1830" + - port: {{ .Values.service.externalPort3 }} + targetPort: {{ .Values.service.internalPort3 }} + nodePort: {{ .Values.global.nodePortPrefix | default .Values.nodePortPrefix }}{{ .Values.service.nodePort3 }} + name: "{{ .Values.service.portName }}-9090" {{- else -}} - port: {{ .Values.service.externalPort }} targetPort: {{ .Values.service.internalPort }} @@ -38,6 +42,9 @@ spec: - port: {{ .Values.service.externalPort2 }} targetPort: {{ .Values.service.internalPort2 }} name: {{ .Values.service.portName }}-1830 + - port: {{ .Values.service.externalPort3 }} + targetPort: {{ .Values.service.internalPort3 }} + name: {{ .Values.service.portName }}-9090 {{- end}} selector: app: {{ include "common.name" . }} diff --git a/kubernetes/appc/values.yaml b/kubernetes/appc/values.yaml index 79991f51a4..9b7b54706a 100644 --- a/kubernetes/appc/values.yaml +++ b/kubernetes/appc/values.yaml @@ -69,7 +69,8 @@ dgbuilder: dbServiceName: appc-dbhost service: name: appc-dgbuilder - +appc-cdt: + nodePort3: 32 # default number of instances replicaCount: 1 @@ -103,6 +104,10 @@ service: nodePort2: 31 clusterPort: 2550 + internalPort3: 9191 + externalPort3: 9090 + nodePort3: 11 + ## Persist data to a persitent volume persistence: enabled: true diff --git a/kubernetes/dcaegen2/charts/dcae-redis/resources/redis/scripts/redis-cluster-config.sh b/kubernetes/dcaegen2/charts/dcae-redis/resources/redis/scripts/redis-cluster-config.sh index f4547525c9..a63ab1668d 100755 --- a/kubernetes/dcaegen2/charts/dcae-redis/resources/redis/scripts/redis-cluster-config.sh +++ b/kubernetes/dcaegen2/charts/dcae-redis/resources/redis/scripts/redis-cluster-config.sh @@ -25,9 +25,7 @@ do echo "======> $(echo $NODES |wc -w) / {{.Values.replicaCount}} pods up" sleep 5 - RESP=$(wget -vO- --ca-certificate /var/run/secrets/kubernetes.io/serviceaccount/ca.crt --header "Authorization -: Bearer $(</var/run/secrets/kubernetes.io/serviceaccount/token)" https://$KUBERNETES_SERVICE_HOST:$KUBERNETES_PORT -_443_TCP_PORT/api/v1/namespaces/{{.Release.Namespace}}/pods?labelSelector=app={{.Chart.Name}}) + RESP=$(wget -vO- --ca-certificate /var/run/secrets/kubernetes.io/serviceaccount/ca.crt --header "Authorization: Bearer $(</var/run/secrets/kubernetes.io/serviceaccount/token)" https://$KUBERNETES_SERVICE_HOST:$KUBERNETES_PORT_443_TCP_PORT/api/v1/namespaces/{{.Release.Namespace}}/pods?labelSelector=app={{.Chart.Name}}) IPS=$(echo $RESP | jq -r '.items[].status.podIP') IPS2=$(echo $IPS | sed -e 's/[a-zA-Z]*//g') diff --git a/kubernetes/log/charts/log-logstash/resources/config/logstash.yml b/kubernetes/log/charts/log-logstash/resources/config/logstash.yml index 3ddf63f9cc..7e5cf67da5 100644 --- a/kubernetes/log/charts/log-logstash/resources/config/logstash.yml +++ b/kubernetes/log/charts/log-logstash/resources/config/logstash.yml @@ -1,4 +1,5 @@ http.host: "0.0.0.0" +pipeline.workers: 3 ## Path where pipeline configurations reside path.config: /usr/share/logstash/pipeline diff --git a/kubernetes/oof/charts/oof-has/charts/oof-has-data/templates/deployment.yaml b/kubernetes/oof/charts/oof-has/charts/oof-has-data/templates/deployment.yaml index 4effba825e..fde094fd16 100755 --- a/kubernetes/oof/charts/oof-has/charts/oof-has-data/templates/deployment.yaml +++ b/kubernetes/oof/charts/oof-has/charts/oof-has-data/templates/deployment.yaml @@ -92,9 +92,9 @@ spec: - mountPath: /usr/local/bin/aai_key.key name: {{ .Values.global.commonConfigPrefix }}-config subPath: aai_key.key - - mountPath: /usr/local/bin/bundle.pem + - mountPath: /usr/local/bin/AAF_RootCA.cer name: {{ .Values.global.commonConfigPrefix }}-config - subPath: bundle.pem + subPath: AAF_RootCA.cer resources: {{ toYaml .Values.resources | indent 12 }} {{- if .Values.nodeSelector }} @@ -123,7 +123,7 @@ spec: path: aai_cert.cer - key: aai_key.key path: aai_key.key - - key: bundle.pem - path: bundle.pem + - key: AAF_RootCA.cer + path: AAF_RootCA.cer imagePullSecrets: - name: "{{ include "common.namespace" . }}-docker-registry-key" diff --git a/kubernetes/oof/charts/oof-has/resources/config/AAF_RootCA.cer b/kubernetes/oof/charts/oof-has/resources/config/AAF_RootCA.cer new file mode 100755 index 0000000000..e9a50d7ea0 --- /dev/null +++ b/kubernetes/oof/charts/oof-has/resources/config/AAF_RootCA.cer @@ -0,0 +1,31 @@ +-----BEGIN CERTIFICATE----- +MIIFPjCCAyagAwIBAgIJAJ6u7cCnzrWdMA0GCSqGSIb3DQEBCwUAMCwxDjAMBgNV +BAsMBU9TQUFGMQ0wCwYDVQQKDARPTkFQMQswCQYDVQQGEwJVUzAeFw0xODA0MDUx +NDE1MjhaFw0zODAzMzExNDE1MjhaMCwxDjAMBgNVBAsMBU9TQUFGMQ0wCwYDVQQK +DARPTkFQMQswCQYDVQQGEwJVUzCCAiIwDQYJKoZIhvcNAQEBBQADggIPADCCAgoC +ggIBAMA5pkgRs7NhGG4ew5JouhyYakgYUyFaG121+/h8qbSdt0hVQv56+EA41Yq7 +XGie7RYDQK9NmAFF3gruE+6X7wvJiChp+Cyd7sFMnb65uWhxEdxWTM2BJFrgfzUn +H8ZCxgaCo3XH4PzlKRy2LQQJEJECwl/RZmRCXijMt5e9h8XoZY/fKkKcZZUsWNCM +pTo266wjvA9MXLmdgReRj0+vrCjrNqy+htwJDztoiHWiYPqT6o8EvGcgjNqjlZx7 +NUNf8MfLDByqKF6+wRbHv1GKjn3/Vijd45Fv8riyRYROiFanvbV6jIfBkv8PZbXg +2VDWsYsgp8NAvMxK+iV8cO+Ck3lBI2GOPZbCEqpPVTYbLUz6sczAlCXwQoPzDIZY +wYa3eR/gYLY1gP2iEVHORag3bLPap9ZX5E8DZkzTNTjovvLk8KaCmfcaUMJsBtDd +ApcUitz10cnRyZc1sX3gE1f3DpzQM6t9C5sOVyRhDcSrKqqwb9m0Ss04XAS9FsqM +P3UWYQyqDXSxlUAYaX892u8mV1hxnt2gjb22RloXMM6TovM3sSrJS0wH+l1nznd6 +aFXftS/G4ZVIVZ/LfT1is4StoyPWZCwwwly1z8qJQ/zhip5NgZTxQw4mi7ww35DY +PdAQOCoajfSvFjqslQ/cPRi/MRCu079heVb5fQnnzVtnpFQRAgMBAAGjYzBhMB0G +A1UdDgQWBBRTVTPyS+vQUbHBeJrBKDF77+rtSTAfBgNVHSMEGDAWgBRTVTPyS+vQ +UbHBeJrBKDF77+rtSTAPBgNVHRMBAf8EBTADAQH/MA4GA1UdDwEB/wQEAwIBhjAN +BgkqhkiG9w0BAQsFAAOCAgEAPx/IaK94n02wPxpnYTy+LVLIxwdq/kawNd6IbiMz +L87zmNMDmHcGbfoRCj8OkhuggX9Lx1/CkhpXimuYsZOFQi5blr/u+v4mIbsgbmi9 +7j+cUHDP0zLycvSvxKHty51LwmaX9a4wkJl5zBU4O1sd/H9tWcEmwJ39ltKoBKBx +c94Zc3iMm5ytRWGj+0rKzLDAXEWpoZ5bE5PLJauA6UDCxDLfs3FwhbS7uDggxYvf +jySF5FCNET94oJ+m8s7VeHvoa8iPGKvXrIqdd7XDHnqJJlVKr7m9S0fMbyEB8ci2 +RtOXDt93ifY1uhoEtEykn4dqBSp8ezvNMnwoXdYPDvTd9uCAFeWFLVreBAWxd25h +PsBTkZA5hpa/rA+mKv6Af4VBViYr8cz4dZCsFChuioVebe9ighrfjB//qKepFjPF +CyjzKN1u0JKm/2x/ORqxkTONG8p3uDwoIOyimUcTtTMv42bfYD88RKakqSFXE9G+ +Z0LlaKABqfjK49o/tsAp+c5LoNlYllKhnetO3QAdraHwdmC36BhoghzR1jpX751A +cZn2VH3Q4XKyp01cJNCJIrua+A+bx6zh3RyW6zIIkbRCbET+UD+4mr8WIcSE3mtR +ZVlnhUDO4z9//WKMVzwS9Rh8/kuszrGFI1KQozXCHLrce3YP6RYZfOed79LXaRwX +dYY= +-----END CERTIFICATE----- diff --git a/kubernetes/oof/charts/oof-has/resources/config/conductor.conf b/kubernetes/oof/charts/oof-has/resources/config/conductor.conf index 4d32deea33..588b9360ba 100755 --- a/kubernetes/oof/charts/oof-has/resources/config/conductor.conf +++ b/kubernetes/oof/charts/oof-has/resources/config/conductor.conf @@ -149,22 +149,29 @@ complex_cache_refresh_interval = 60 # Base URL for A&AI, up to and not including the version, and without a # trailing slash. (string value) -server_url = https://oof.api.simpledemo.onap.org:8443/aai +server_url = https://{{.Values.config.aai.serviceName}}.{{ include "common.namespace" . }}:{{.Values.config.aai.port}}/aai # The version of A&AI in v# format. (string value) -server_url_version = v11 +server_url_version = v13 # SSL/TLS certificate file in pem format. This certificate must be registered # with the A&AI endpoint. (string value) -certificate_file = /usr/local/bin/aai_cert.cer +certificate_file = # Private Certificate Key file in pem format. (string value) -certificate_key_file = /usr/local/bin/aai_key.key +certificate_key_file = # Certificate Authority Bundle file in pem format. Must contain the appropriate # trust chain for the Certificate file. (string value) #certificate_authority_bundle_file = /opt/app/conductor/etc/certs/ca_bundle.pem -certificate_authority_bundle_file =/usr/local/bin/bundle.pem +certificate_authority_bundle_file =/usr/local/bin/AAF_RootCA.cer + +# Basic Authentication Username (string value) +username = OOF + +# Basic Authentication Password (string value) +password = OOF + [api] @@ -355,3 +362,31 @@ concurrent = true [reservation] concurrent = true + +[multicloud] + +# +# From conductor +# + +# Base URL for Multicloud without a trailing slash. (string value) +server_url = http://{{.Values.config.msb.serviceName}}:{{.Values.config.msb.port}}/api/multicloud + +# Timeout for Multicloud Rest Call (string value) +multicloud_rest_timeout = 30 + +# Number of retry for Multicloud Rest Call (string value) +multicloud_retries = 3 + +# The version of Multicloud API. (string value) +server_url_version = v0 + + +[vim_controller] +# +# From conductor +# + +# Extensions list to use (list value) +extensions = multicloud + diff --git a/kubernetes/oof/charts/oof-has/values.yaml b/kubernetes/oof/charts/oof-has/values.yaml index 25a076b6cc..00f59a650c 100755 --- a/kubernetes/oof/charts/oof-has/values.yaml +++ b/kubernetes/oof/charts/oof-has/values.yaml @@ -30,3 +30,11 @@ global: pullPolicy: Always nodePortPrefix: 302 dataRootDir: /dockerdata-nfs +config: + aai: + serviceName: aai + port: 8443 + msb: + serviceName: msb-iag + port: 80 + diff --git a/kubernetes/oof/resources/config/osdf_config.yaml b/kubernetes/oof/resources/config/osdf_config.yaml index 7fe63a50c8..f286be4595 100755 --- a/kubernetes/oof/resources/config/osdf_config.yaml +++ b/kubernetes/oof/resources/config/osdf_config.yaml @@ -1,37 +1,18 @@ -# Copyright © 2017 Amdocs, Bell Canada -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - - -# Credentials for the OOF placement service – SO. Newly added. osdfPlacementSOUsername: {{ .Values.config.osdfPlacementSOUsername }} osdfPlacementSOPassword: {{ .Values.config.osdfPlacementSOPassword }} -# Credentials for the OOF CM scheduling service – Generic. Newly added osdfCMSchedulerUsername: {{ .Values.config.osdfCMSchedulerUsername }} osdfCMSchedulerPassword: {{ .Values.config.osdfCMSchedulerPassword }} -# msoUrl: "" # The SO url for call back. This will be part of the request, so no need soUsername: {{ .Values.config.soUsername }} soPassword: {{ .Values.config.soPassword }} -conductorUrl: "{{.Release.Name}}-has-api.{{include "common.namespace" .}}:TODO-PORT-FOR-HAS-API" +conductorUrl: {{ .Values.config.conductorUrl }} conductorUsername: {{ .Values.config.conductorUsername }} conductorPassword: {{ .Values.config.conductorPassword }} conductorPingWaitTime: {{ .Values.config.conductorPingWaitTime }} conductorMaxRetries: {{ .Values.config.conductorMaxRetries }} -# Policy Platform -- requires ClientAuth, Authorization, and Environment policyPlatformUrl: {{ .Values.config.policyPlatformUrl }} policyPlatformEnv: {{ .Values.config.policyPlatformEnv }} policyPlatformUsername: {{ .Values.config.policyPlatformUsername }} diff --git a/kubernetes/oof/values.yaml b/kubernetes/oof/values.yaml index 9f74271418..51114a6279 100644 --- a/kubernetes/oof/values.yaml +++ b/kubernetes/oof/values.yaml @@ -34,7 +34,7 @@ debugEnabled: false # application configuration config: - msbgateway: msb-iag.{{ include "common.namespace" . }} + msbgateway: msb-iag msbPort: 80 # Credentials for the OOF placement service – SO. osdfPlacementSOUsername: so_test @@ -46,7 +46,7 @@ config: soUsername: "" # SO username for call back. soPassword: "" # SO password for call back. # Url and credentials for Conductor. - conductorUrl: https://oof-has-api.{{ include "common.namespace" . }}:8091/v1/plans/ + conductorUrl: http://oof-has-api:8091/v1/plans/ conductorUsername: admin1 conductorPassword: plan.15 conductorPingWaitTime: 60 diff --git a/kubernetes/policy/charts/brmsgw/values.yaml b/kubernetes/policy/charts/brmsgw/values.yaml index efc397654a..d8cbda4aaf 100644 --- a/kubernetes/policy/charts/brmsgw/values.yaml +++ b/kubernetes/policy/charts/brmsgw/values.yaml @@ -25,7 +25,7 @@ global: ################################################################# # application image repository: nexus3.onap.org:10001 -image: onap/policy-pe:1.2.2 +image: onap/policy-pe:1.2.3 pullPolicy: Always # flag to enable debugging - application support required diff --git a/kubernetes/policy/charts/drools/values.yaml b/kubernetes/policy/charts/drools/values.yaml index ca62f2b01d..1bc09a7d0b 100644 --- a/kubernetes/policy/charts/drools/values.yaml +++ b/kubernetes/policy/charts/drools/values.yaml @@ -27,7 +27,7 @@ global: ################################################################# # application image repository: nexus3.onap.org:10001 -image: onap/policy-drools:1.2.2 +image: onap/policy-drools:1.2.3 pullPolicy: Always # flag to enable debugging - application support required diff --git a/kubernetes/policy/charts/pdp/values.yaml b/kubernetes/policy/charts/pdp/values.yaml index 811830ea89..707b975726 100644 --- a/kubernetes/policy/charts/pdp/values.yaml +++ b/kubernetes/policy/charts/pdp/values.yaml @@ -27,7 +27,7 @@ global: ################################################################# # application image repository: nexus3.onap.org:10001 -image: onap/policy-pe:1.2.2 +image: onap/policy-pe:1.2.3 pullPolicy: Always # flag to enable debugging - application support required diff --git a/kubernetes/policy/values.yaml b/kubernetes/policy/values.yaml index 42d58472ed..6aa8c7e61a 100644 --- a/kubernetes/policy/values.yaml +++ b/kubernetes/policy/values.yaml @@ -40,7 +40,7 @@ global: ################################################################# # application image repository: nexus3.onap.org:10001 -image: onap/policy-pe:1.2.2 +image: onap/policy-pe:1.2.3 pullPolicy: Always subChartsOnly: diff --git a/kubernetes/sdnc/resources/config/bin/installSdncDb.sh b/kubernetes/sdnc/resources/config/bin/installSdncDb.sh index e574bd7aec..ab0bc35880 100644 --- a/kubernetes/sdnc/resources/config/bin/installSdncDb.sh +++ b/kubernetes/sdnc/resources/config/bin/installSdncDb.sh @@ -53,3 +53,6 @@ done # Create VNIs 100-199 ${SDNC_HOME}/bin/addVnis.sh 100 199 + +# Drop FK_NETWORK_MODEL foreign key as workaround for SDNC-291. +${SDNC_HOME}/bin/rmForeignKey.sh NETWORK_MODEL FK_NETWORK_MODEL
\ No newline at end of file diff --git a/kubernetes/so/resources/config/mso/mso-docker.json b/kubernetes/so/resources/config/mso/mso-docker.json index 8276d25a7c..167a8edc09 100755 --- a/kubernetes/so/resources/config/mso/mso-docker.json +++ b/kubernetes/so/resources/config/mso/mso-docker.json @@ -191,10 +191,10 @@ "sdncTimeoutFirewall": "20", "callbackRetryAttempts": "30", "callbackRetrySleepTime": "1000", - "appcClientTopicRead": "APPC-LCM-READ", - "appcClientTopicWrite": "APPC-LCM-WRITE", - "appcClientTopicSdncRead": "SDNC-LCM-READ", - "appcClientTopicSdncWrite": "SDNC-LCM-WRITE", + "appcClientTopicRead": "APPC-LCM-WRITE", + "appcClientTopicWrite": "APPC-LCM-READ", + "appcClientTopicSdncRead": "SDNC-LCM-WRITE", + "appcClientTopicSdncWrite": "SDNC-LCM-READ", "appcClientTopicReadTimeout": "360000", "appcClientResponseTime": "360000", "appcClientPoolMembers": "{{.Values.config.messagerouter.serviceName}}:{{.Values.config.messagerouter.port}}", |