summaryrefslogtreecommitdiffstats
path: root/docs
diff options
context:
space:
mode:
Diffstat (limited to 'docs')
-rw-r--r--docs/Access_Rancher_server_via_web_browser.jpegbin0 -> 135026 bytes
-rw-r--r--docs/Add_Kubernetes_Environment_to_Rancher.pngbin0 -> 49242 bytes
-rw-r--r--docs/Apply_customization_script_for_the_Rancher_VM.jpegbin0 -> 153028 bytes
-rw-r--r--docs/Click_create.jpegbin0 -> 131952 bytes
-rw-r--r--docs/Click_on_CLI_and_then_click_on_Generate_Config.jpegbin0 -> 91262 bytes
-rw-r--r--docs/Click_on_Close_button.jpegbin0 -> 158513 bytes
-rw-r--r--docs/Click_on_Copy_to_Clipboard-wait_until_you_see_a_token-do_not_copy_user+password-the_server_is_not_ready_at_that_point.jpegbin0 -> 92239 bytes
-rw-r--r--docs/K8s-Assign_Floating_IP_for_external_access.jpegbin0 -> 64999 bytes
-rw-r--r--docs/K8s-Key_Pair.jpegbin0 -> 87469 bytes
-rw-r--r--docs/K8s-Launch_Instance.jpegbin0 -> 53520 bytes
-rw-r--r--docs/K8s-Launch_new_VM_instance_to_create_a_Kubernetes_single_host_or_cluster.jpegbin0 -> 84468 bytes
-rw-r--r--docs/K8s-Manage_Floating_IP_Associations.jpegbin0 -> 33961 bytes
-rw-r--r--docs/K8s-Networking.jpegbin0 -> 96431 bytes
-rw-r--r--docs/K8s-Security_Group.jpegbin0 -> 73732 bytes
-rw-r--r--docs/K8s-Select_Flavor.jpegbin0 -> 148973 bytes
-rw-r--r--docs/K8s-Select_Ubuntu_16.04_as_base_image.jpegbin0 -> 136070 bytes
-rw-r--r--docs/Rancher-Allocate_Floating_IP.jpegbin0 -> 32404 bytes
-rw-r--r--docs/Rancher-Key_Pair.jpegbin0 -> 90739 bytes
-rw-r--r--docs/Rancher-Launch_Instance.jpegbin0 -> 51541 bytes
-rw-r--r--docs/Rancher-Launch_new_VM_instance_to_host_the_Rancher_Server.jpegbin0 -> 448601 bytes
-rw-r--r--docs/Rancher-Manage_Floating_IP_Associations.jpegbin0 -> 33712 bytes
-rw-r--r--docs/Rancher-Networking.jpegbin0 -> 93256 bytes
-rw-r--r--docs/Rancher-Security_Groups.jpegbin0 -> 77292 bytes
-rw-r--r--docs/Rancher-Select_Flavor.jpegbin0 -> 142727 bytes
-rw-r--r--docs/Rancher-Select_Ubuntu_16.04_as_base_image.jpegbin0 -> 139303 bytes
-rw-r--r--docs/Return_to_Rancher_environment_eg_SB4_and_wait_for_services_to_complete_10-15_mins.jpegbin0 -> 60818 bytes
-rw-r--r--docs/Select_Add_Environment.pngbin0 -> 38073 bytes
-rw-r--r--docs/and_click_on_Save_accept_defaults.jpegbin0 -> 77083 bytes
-rw-r--r--docs/index.rst1
-rw-r--r--docs/k8s-topology.jpgbin0 -> 57806 bytes
-rw-r--r--docs/master_nfs_node.sh32
-rw-r--r--docs/oom_cloud_setup_guide.rst54
-rw-r--r--docs/oom_quickstart_guide.rst2
-rw-r--r--docs/oom_setup_kubernetes_rancher.rst478
-rw-r--r--docs/oom_user_guide.rst12
-rw-r--r--docs/openstack-k8s-node.sh46
-rw-r--r--docs/openstack-rancher.sh51
-rw-r--r--docs/otherwise_select_INFRASTRUCTURE_Hosts_and_click_on_Add_Host.jpgbin0 -> 46511 bytes
-rw-r--r--docs/release-notes.rst154
-rw-r--r--docs/slave_nfs_node.sh25
40 files changed, 762 insertions, 93 deletions
diff --git a/docs/Access_Rancher_server_via_web_browser.jpeg b/docs/Access_Rancher_server_via_web_browser.jpeg
new file mode 100644
index 0000000000..103fc2731f
--- /dev/null
+++ b/docs/Access_Rancher_server_via_web_browser.jpeg
Binary files differ
diff --git a/docs/Add_Kubernetes_Environment_to_Rancher.png b/docs/Add_Kubernetes_Environment_to_Rancher.png
new file mode 100644
index 0000000000..0d60f156fb
--- /dev/null
+++ b/docs/Add_Kubernetes_Environment_to_Rancher.png
Binary files differ
diff --git a/docs/Apply_customization_script_for_the_Rancher_VM.jpeg b/docs/Apply_customization_script_for_the_Rancher_VM.jpeg
new file mode 100644
index 0000000000..832dda3404
--- /dev/null
+++ b/docs/Apply_customization_script_for_the_Rancher_VM.jpeg
Binary files differ
diff --git a/docs/Click_create.jpeg b/docs/Click_create.jpeg
new file mode 100644
index 0000000000..fad185674c
--- /dev/null
+++ b/docs/Click_create.jpeg
Binary files differ
diff --git a/docs/Click_on_CLI_and_then_click_on_Generate_Config.jpeg b/docs/Click_on_CLI_and_then_click_on_Generate_Config.jpeg
new file mode 100644
index 0000000000..fec6a35433
--- /dev/null
+++ b/docs/Click_on_CLI_and_then_click_on_Generate_Config.jpeg
Binary files differ
diff --git a/docs/Click_on_Close_button.jpeg b/docs/Click_on_Close_button.jpeg
new file mode 100644
index 0000000000..cb02480923
--- /dev/null
+++ b/docs/Click_on_Close_button.jpeg
Binary files differ
diff --git a/docs/Click_on_Copy_to_Clipboard-wait_until_you_see_a_token-do_not_copy_user+password-the_server_is_not_ready_at_that_point.jpeg b/docs/Click_on_Copy_to_Clipboard-wait_until_you_see_a_token-do_not_copy_user+password-the_server_is_not_ready_at_that_point.jpeg
new file mode 100644
index 0000000000..bb0f60c4eb
--- /dev/null
+++ b/docs/Click_on_Copy_to_Clipboard-wait_until_you_see_a_token-do_not_copy_user+password-the_server_is_not_ready_at_that_point.jpeg
Binary files differ
diff --git a/docs/K8s-Assign_Floating_IP_for_external_access.jpeg b/docs/K8s-Assign_Floating_IP_for_external_access.jpeg
new file mode 100644
index 0000000000..fe84d11707
--- /dev/null
+++ b/docs/K8s-Assign_Floating_IP_for_external_access.jpeg
Binary files differ
diff --git a/docs/K8s-Key_Pair.jpeg b/docs/K8s-Key_Pair.jpeg
new file mode 100644
index 0000000000..1e335d36a6
--- /dev/null
+++ b/docs/K8s-Key_Pair.jpeg
Binary files differ
diff --git a/docs/K8s-Launch_Instance.jpeg b/docs/K8s-Launch_Instance.jpeg
new file mode 100644
index 0000000000..7e2122c429
--- /dev/null
+++ b/docs/K8s-Launch_Instance.jpeg
Binary files differ
diff --git a/docs/K8s-Launch_new_VM_instance_to_create_a_Kubernetes_single_host_or_cluster.jpeg b/docs/K8s-Launch_new_VM_instance_to_create_a_Kubernetes_single_host_or_cluster.jpeg
new file mode 100644
index 0000000000..8d154d4a2d
--- /dev/null
+++ b/docs/K8s-Launch_new_VM_instance_to_create_a_Kubernetes_single_host_or_cluster.jpeg
Binary files differ
diff --git a/docs/K8s-Manage_Floating_IP_Associations.jpeg b/docs/K8s-Manage_Floating_IP_Associations.jpeg
new file mode 100644
index 0000000000..f15e5e34c7
--- /dev/null
+++ b/docs/K8s-Manage_Floating_IP_Associations.jpeg
Binary files differ
diff --git a/docs/K8s-Networking.jpeg b/docs/K8s-Networking.jpeg
new file mode 100644
index 0000000000..ee4cafca54
--- /dev/null
+++ b/docs/K8s-Networking.jpeg
Binary files differ
diff --git a/docs/K8s-Security_Group.jpeg b/docs/K8s-Security_Group.jpeg
new file mode 100644
index 0000000000..5e852185b2
--- /dev/null
+++ b/docs/K8s-Security_Group.jpeg
Binary files differ
diff --git a/docs/K8s-Select_Flavor.jpeg b/docs/K8s-Select_Flavor.jpeg
new file mode 100644
index 0000000000..c3cf4b2c83
--- /dev/null
+++ b/docs/K8s-Select_Flavor.jpeg
Binary files differ
diff --git a/docs/K8s-Select_Ubuntu_16.04_as_base_image.jpeg b/docs/K8s-Select_Ubuntu_16.04_as_base_image.jpeg
new file mode 100644
index 0000000000..e9a03e782c
--- /dev/null
+++ b/docs/K8s-Select_Ubuntu_16.04_as_base_image.jpeg
Binary files differ
diff --git a/docs/Rancher-Allocate_Floating_IP.jpeg b/docs/Rancher-Allocate_Floating_IP.jpeg
new file mode 100644
index 0000000000..9d6ff3ff95
--- /dev/null
+++ b/docs/Rancher-Allocate_Floating_IP.jpeg
Binary files differ
diff --git a/docs/Rancher-Key_Pair.jpeg b/docs/Rancher-Key_Pair.jpeg
new file mode 100644
index 0000000000..1bcb6b6eb7
--- /dev/null
+++ b/docs/Rancher-Key_Pair.jpeg
Binary files differ
diff --git a/docs/Rancher-Launch_Instance.jpeg b/docs/Rancher-Launch_Instance.jpeg
new file mode 100644
index 0000000000..38943699e6
--- /dev/null
+++ b/docs/Rancher-Launch_Instance.jpeg
Binary files differ
diff --git a/docs/Rancher-Launch_new_VM_instance_to_host_the_Rancher_Server.jpeg b/docs/Rancher-Launch_new_VM_instance_to_host_the_Rancher_Server.jpeg
new file mode 100644
index 0000000000..2b3fde3fb7
--- /dev/null
+++ b/docs/Rancher-Launch_new_VM_instance_to_host_the_Rancher_Server.jpeg
Binary files differ
diff --git a/docs/Rancher-Manage_Floating_IP_Associations.jpeg b/docs/Rancher-Manage_Floating_IP_Associations.jpeg
new file mode 100644
index 0000000000..ae062c1d39
--- /dev/null
+++ b/docs/Rancher-Manage_Floating_IP_Associations.jpeg
Binary files differ
diff --git a/docs/Rancher-Networking.jpeg b/docs/Rancher-Networking.jpeg
new file mode 100644
index 0000000000..16ef380ef5
--- /dev/null
+++ b/docs/Rancher-Networking.jpeg
Binary files differ
diff --git a/docs/Rancher-Security_Groups.jpeg b/docs/Rancher-Security_Groups.jpeg
new file mode 100644
index 0000000000..c8ff2e8d76
--- /dev/null
+++ b/docs/Rancher-Security_Groups.jpeg
Binary files differ
diff --git a/docs/Rancher-Select_Flavor.jpeg b/docs/Rancher-Select_Flavor.jpeg
new file mode 100644
index 0000000000..a3f35b53ce
--- /dev/null
+++ b/docs/Rancher-Select_Flavor.jpeg
Binary files differ
diff --git a/docs/Rancher-Select_Ubuntu_16.04_as_base_image.jpeg b/docs/Rancher-Select_Ubuntu_16.04_as_base_image.jpeg
new file mode 100644
index 0000000000..b80fae4a5c
--- /dev/null
+++ b/docs/Rancher-Select_Ubuntu_16.04_as_base_image.jpeg
Binary files differ
diff --git a/docs/Return_to_Rancher_environment_eg_SB4_and_wait_for_services_to_complete_10-15_mins.jpeg b/docs/Return_to_Rancher_environment_eg_SB4_and_wait_for_services_to_complete_10-15_mins.jpeg
new file mode 100644
index 0000000000..6d62b99f55
--- /dev/null
+++ b/docs/Return_to_Rancher_environment_eg_SB4_and_wait_for_services_to_complete_10-15_mins.jpeg
Binary files differ
diff --git a/docs/Select_Add_Environment.png b/docs/Select_Add_Environment.png
new file mode 100644
index 0000000000..48fffa41bd
--- /dev/null
+++ b/docs/Select_Add_Environment.png
Binary files differ
diff --git a/docs/and_click_on_Save_accept_defaults.jpeg b/docs/and_click_on_Save_accept_defaults.jpeg
new file mode 100644
index 0000000000..9b57927aac
--- /dev/null
+++ b/docs/and_click_on_Save_accept_defaults.jpeg
Binary files differ
diff --git a/docs/index.rst b/docs/index.rst
index 712802d619..340b43be67 100644
--- a/docs/index.rst
+++ b/docs/index.rst
@@ -14,3 +14,4 @@ OOM Documentation Repository
oom_developer_guide.rst
oom_cloud_setup_guide.rst
release-notes.rst
+ oom_setup_kubernetes_rancher.rst
diff --git a/docs/k8s-topology.jpg b/docs/k8s-topology.jpg
new file mode 100644
index 0000000000..eba8e72b68
--- /dev/null
+++ b/docs/k8s-topology.jpg
Binary files differ
diff --git a/docs/master_nfs_node.sh b/docs/master_nfs_node.sh
new file mode 100644
index 0000000000..4a7a8dbc12
--- /dev/null
+++ b/docs/master_nfs_node.sh
@@ -0,0 +1,32 @@
+#!/bin/bash
+
+usage () {
+ echo "Usage:"
+ echo " ./$(basename $0) node1_ip node2_ip ... nodeN_ip"
+ exit 1
+}
+
+if [ "$#" -lt 1 ]; then
+ echo "Missing NFS slave nodes"
+ usage
+fi
+
+#Install NFS kernel
+sudo apt-get update
+sudo apt-get install -y nfs-kernel-server
+
+#Create /dockerdata-nfs and set permissions
+sudo mkdir -p /dockerdata-nfs
+sudo chmod 777 -R /dockerdata-nfs
+sudo chown nobody:nogroup /dockerdata-nfs/
+
+#Update the /etc/exports
+NFS_EXP=""
+for i in $@; do
+ NFS_EXP+="$i(rw,sync,no_root_squash,no_subtree_check) "
+done
+echo "/dockerdata-nfs "$NFS_EXP | sudo tee -a /etc/exports
+
+#Restart the NFS service
+sudo exportfs -a
+sudo systemctl restart nfs-kernel-server
diff --git a/docs/oom_cloud_setup_guide.rst b/docs/oom_cloud_setup_guide.rst
index f967931234..73509669e8 100644
--- a/docs/oom_cloud_setup_guide.rst
+++ b/docs/oom_cloud_setup_guide.rst
@@ -1,4 +1,5 @@
-.. This work is licensed under a Creative Commons Attribution 4.0 International License.
+.. This work is licensed under a Creative Commons Attribution 4.0
+.. International License.
.. http://creativecommons.org/licenses/by/4.0
.. Copyright 2018 Amdocs, Bell Canada
@@ -11,6 +12,7 @@
.. _Setting Up Kubernetes with Rancher: https://wiki.onap.org/display/DW/ONAP+on+Kubernetes+on+Rancher
.. _Setting Up Kubernetes with Kubeadm: https://wiki.onap.org/display/DW/Deploying+Kubernetes+Cluster+with+kubeadm
.. _Setting Up Kubernetes with Cloudify: https://wiki.onap.org/display/DW/ONAP+on+Kubernetes+on+Cloudify
+.. _ONAP on Kubernetes Wiki: https://wiki.onap.org/display/DW/ONAP+on+Kubernetes
.. figure:: oomLogoV2-medium.png
:align: right
@@ -47,7 +49,7 @@ The versions of Kubernetes that are supported by OOM are as follows:
Release Kubernetes Helm kubectl Docker
============== ========== ===== ======= ========
amsterdam 1.7.x 2.3.x 1.7.x 1.12.x
- beijing/master 1.8.10 2.7.x 1.8.10 17.03.x
+ beijing/master 1.8.10 2.8.2 1.8.10 17.03.x
============== ========== ===== ======= ========
Minimum Hardware Configuration
@@ -62,26 +64,38 @@ least three if not six nodes to ensure there is no single point of failure.
===== ===== ====== ====================
RAM HD vCores Ports
===== ===== ====== ====================
- 128GB 160GB 16 0.0.0.0/0 (all open)
+ 128GB 160GB 32 0.0.0.0/0 (all open)
===== ===== ====== ====================
+.. note::
+ Kubernetes supports a maximum of 110 pods per node which forces one to use at
+ least two nodes to deploy all of ONAP although at least three are recommended
+ (for example 4x32GB - 8 vCores each). Subsets of ONAP may still be deployed
+ on a single node.
+
Cloud Installation
==================
-#. OOM supports deployment on major public clouds. The following guides
- provide instructions on how to deploy ONAP on these clouds:
-
- - `Microsoft Azure`_,
- - `Amazon AWS`_,
- - `Google GCE`_,
- - `VMware VIO`_,
- - IBM, and
- - `Openstack`_.
-
-#. Alternatively, OOM can be deployed on a private set of physical hosts or VMs
- (or even a combination of the two). The following guides describe how to
- create a Kubernetes cluster with popular tools:
-
- - `Setting up Kubernetes with Rancher`_ (recommended)
- - `Setting up Kubernetes with Kubeadm`_
- - `Setting up Kubernetes with Cloudify`_
+.. #. OOM supports deployment on major public clouds. The following guides
+.. provide instructions on how to deploy ONAP on these clouds:
+..
+.. - `Microsoft Azure`_,
+.. - `Amazon AWS`_,
+.. - `Google GCE`_,
+.. - `VMware VIO`_,
+.. - IBM, and
+.. - `Openstack`_.
+..
+.. #. Alternatively, OOM can be deployed on a private set of physical hosts or VMs
+.. (or even a combination of the two). The following guides describe how to
+.. create a Kubernetes cluster with popular tools:
+..
+.. - `Setting up Kubernetes with Rancher`_ (recommended)
+.. - `Setting up Kubernetes with Kubeadm`_
+.. - `Setting up Kubernetes with Cloudify`_
+
+OOM can be deployed on a private set of physical hosts or VMs (or even a
+combination of the two). The following guide describe the recommended method to
+setup a Kubernetes cluster: :ref:`onap-on-kubernetes-with-rancher`.
+
+There are alternative deployment methods described on the `ONAP on Kubernetes Wiki`_
diff --git a/docs/oom_quickstart_guide.rst b/docs/oom_quickstart_guide.rst
index eb5d373721..9672a0e55f 100644
--- a/docs/oom_quickstart_guide.rst
+++ b/docs/oom_quickstart_guide.rst
@@ -16,7 +16,7 @@ available), follow the following instructions to deploy ONAP.
**Step 1.** Clone the OOM repository from ONAP gerrit::
- > git clone http://gerrit.onap.org/r/oom
+ > git clone -b beijing http://gerrit.onap.org/r/oom
> cd oom/kubernetes
diff --git a/docs/oom_setup_kubernetes_rancher.rst b/docs/oom_setup_kubernetes_rancher.rst
new file mode 100644
index 0000000000..8e55bf1b66
--- /dev/null
+++ b/docs/oom_setup_kubernetes_rancher.rst
@@ -0,0 +1,478 @@
+.. This work is licensed under a Creative Commons Attribution 4.0 International License.
+.. http://creativecommons.org/licenses/by/4.0
+.. Copyright 2018 Amdocs, Bell Canada
+
+.. Links
+.. _HELM Best Practices Guide: https://docs.helm.sh/chart_best_practices/#requirements
+.. _kubectl Cheat Sheet: https://kubernetes.io/docs/reference/kubectl/cheatsheet/
+.. _Kubernetes documentation for emptyDir: https://kubernetes.io/docs/concepts/storage/volumes/#emptydir
+.. _Docker DevOps: https://wiki.onap.org/display/DW/Docker+DevOps#DockerDevOps-DockerBuild
+.. _http://cd.onap.info:30223/mso/logging/debug: http://cd.onap.info:30223/mso/logging/debug
+.. _Onboarding and Distributing a Vendor Software Product: https://wiki.onap.org/pages/viewpage.action?pageId=1018474
+.. _README.md: https://gerrit.onap.org/r/gitweb?p=oom.git;a=blob;f=kubernetes/README.md
+
+.. figure:: oomLogoV2-medium.png
+ :align: right
+
+.. _onap-on-kubernetes-with-rancher:
+
+ONAP on Kubernetes with Rancher
+###############################
+
+The following instructions will step you through the installation of Kubernetes
+on an OpenStack environment with Rancher. The development lab used for this
+installation is the ONAP Windriver lab.
+
+This guide does not cover all of the steps required to setup your OpenStack
+environment: e.g. OAM networks and security groups but there is a wealth of
+OpenStack information on the web.
+
+Rancher Installation
+====================
+
+The following instructions describe how to create an Openstack VM running
+Rancher. This node will not be used to host ONAP itself, it will be used
+exclusively by Rancher.
+
+Launch new VM instance to host the Rancher Server
+-------------------------------------------------
+
+.. image:: Rancher-Launch_new_VM_instance_to_host_the_Rancher_Server.jpeg
+
+Select Ubuntu 16.04 as base image
+---------------------------------
+Select "No" on "Create New Volume"
+
+.. image:: Rancher-Select_Ubuntu_16.04_as_base_image.jpeg
+
+Select Flavor
+-------------
+Known issues exist if flavor is too small for Rancher. Please select a flavor
+with at least 4 vCPU and 8GB ram.
+
+.. image:: Rancher-Select_Flavor.jpeg
+
+Networking
+----------
+
+.. image:: Rancher-Networking.jpeg
+
+Security Groups
+---------------
+
+.. image:: Rancher-Security_Groups.jpeg
+
+Key Pair
+--------
+Use an existing key pair (e.g. onap_key), import an existing one or create a
+new one to assign.
+
+.. image:: Rancher-Key_Pair.jpeg
+
+Apply customization script for the Rancher VM
+---------------------------------------------
+
+Click :download:`openstack-rancher.sh <openstack-rancher.sh>` to download the script.
+
+.. literalinclude:: openstack-rancher.sh
+ :language: bash
+
+This customization script will:
+
+* setup root access to the VM (comment out if you wish to disable this
+ capability and restrict access to ssh access only)
+* install docker
+* install rancher
+* install kubectl
+* install helm
+* install nfs server
+
+.. note::
+ The Beijing release of OOM only supports Helm 2.8.2 not the 2.7.2 shown in
+ the screen capture below. The supported versions of all the software components
+ are listed in the :ref:`cloud-setup-guide-label`.
+
+.. image:: Apply_customization_script_for_the_Rancher_VM.jpeg
+
+Launch Instance
+---------------
+
+.. image:: Rancher-Launch_Instance.jpeg
+
+Assign Floating IP for external access
+--------------------------------------
+
+.. image:: Rancher-Allocate_Floating_IP.jpeg
+
+.. image:: Rancher-Manage_Floating_IP_Associations.jpeg
+
+.. image:: Rancher-Launch_Instance.jpeg
+
+Kubernetes Installation
+=======================
+
+Launch new VM instance(s) to create a Kubernetes single host or cluster
+-----------------------------------------------------------------------
+
+To create a cluster:
+
+.. note::
+ #. do not append a '-1' suffix (e.g. sb4-k8s)
+ #. increase count to the # of of kubernetes worker nodes you want (eg. 3)
+
+.. image:: K8s-Launch_new_VM_instance_to_create_a_Kubernetes_single_host_or_cluster.jpeg
+
+Select Ubuntu 16.04 as base image
+---------------------------------
+Select "No" on "Create New Volume"
+
+.. image:: K8s-Select_Ubuntu_16.04_as_base_image.jpeg
+
+Select Flavor
+-------------
+The size of a Kubernetes host depends on the size of the ONAP deployment that
+will be installed.
+
+As of the Beijing release a minimum of 3 x 32GB hosts will be needed to run a
+full ONAP deployment (all components).
+
+If a small subset of ONAP components are being deployed for testing purposes,
+then a single 16GB or 32GB host should suffice.
+
+.. image:: K8s-Select_Flavor.jpeg
+
+Networking
+-----------
+
+.. image:: K8s-Networking.jpeg
+
+Security Group
+---------------
+
+.. image:: K8s-Security_Group.jpeg
+
+Key Pair
+--------
+Use an existing key pair (e.g. onap_key), import an existing one or create a
+new one to assign.
+
+.. image:: K8s-Key_Pair.jpeg
+
+Apply customization script for Kubernetes VM(s)
+-----------------------------------------------
+
+Click :download:`openstack-k8s-node.sh <openstack-k8s-node.sh>` to
+download the script.
+
+.. literalinclude:: openstack-k8s-node.sh
+ :language: bash
+
+This customization script will:
+
+* setup root access to the VM (comment out if you wish to disable this
+ capability and restrict access to ssh access only)
+* install docker
+* install kubectl
+* install helm
+* install nfs common (see configuration step here)
+
+.. note::
+ Ensure you are using the correct versions as described in the
+ :ref:`cloud-setup-guide-label`
+
+Launch Instance
+---------------
+
+.. image:: K8s-Launch_Instance.jpeg
+
+Assign Floating IP for external access
+--------------------------------------
+
+.. image:: K8s-Assign_Floating_IP_for_external_access.jpeg
+
+.. image:: K8s-Manage_Floating_IP_Associations.jpeg
+
+.. image:: K8s-Launch_Instance.jpeg
+
+Setting up an NFS share for Multinode Kubernetes Clusters
+=========================================================
+The figure below illustrates a possible topology of a multinode Kubernetes
+cluster.
+
+.. image:: k8s-topology.jpg
+
+One node, the Master Node, runs Rancher and Helm clients and connects to all
+the Kubernetes nodes in the cluster. Kubernetes nodes, in turn, run Rancher,
+Kubernetes and Tiller (Helm) agents, which receive, execute, and respond to
+commands issued by the Master Node (e.g. kubectl or helm operations). Note that
+the Master Node can be either a remote machine that the user can log in to or a
+local machine (e.g. laptop, desktop) that has access to the Kubernetes cluster.
+
+Deploying applications to a Kubernetes cluster requires Kubernetes nodes to
+share a common, distributed filesystem. One node in the cluster plays the role
+of NFS Master (not to confuse with the Master Node that runs Rancher and Helm
+clients, which is located outside the cluster), while all the other cluster
+nodes play the role of NFS slaves. In the figure above, the left-most cluster
+node plays the role of NFS Master (indicated by the crown symbol). To properly
+set up an NFS share on Master and Slave nodes, the user can run the scripts
+below.
+
+Click :download:`master_nfs_node.sh <master_nfs_node.sh>` to download the script.
+
+.. literalinclude:: master_nfs_node.sh
+ :language: bash
+
+Click :download:`slave_nfs_node.sh <slave_nfs_node.sh>` to download the script.
+
+.. literalinclude:: slave_nfs_node.sh
+ :language: bash
+
+The master_nfs_node.sh script runs in the NFS Master node and needs the list of
+NFS Slave nodes as input, e.g.::
+
+ > sudo ./master_nfs_node.sh node1_ip node2_ip ... nodeN_ip
+
+The slave_nfs_node.sh script runs in each NFS Slave node and needs the IP of
+the NFS Master node as input, e.g.::
+
+ > sudo ./slave_nfs_node.sh master_node_ip
+
+Configuration (Rancher and Kubernetes)
+======================================
+
+Access Rancher server via web browser
+-------------------------------------
+(e.g. http://10.12.6.16:8080/env/1a5/apps/stacks)
+
+.. image:: Access_Rancher_server_via_web_browser.jpeg
+
+Add Kubernetes Environment to Rancher
+-------------------------------------
+
+1. Select “Manage Environments”
+
+.. image:: Add_Kubernetes_Environment_to_Rancher.png
+
+2. Select “Add Environment”
+
+.. image:: Select_Add_Environment.png
+
+3. Add unique name for your new Rancher environment
+
+4. Select the Kubernetes template
+
+5. Click "create"
+
+.. image:: Click_create.jpeg
+
+6. Select the new named environment (ie. SB4) from the dropdown list (top left).
+
+Rancher is now waiting for a Kubernetes Host to be added.
+
+.. image:: K8s-Assign_Floating_IP_for_external_access.jpeg
+
+Add Kubernetes Host
+-------------------
+
+1. If this is the first (or only) host being added - click on the "Add a host" link
+
+.. image:: K8s-Assign_Floating_IP_for_external_access.jpeg
+
+and click on "Save" (accept defaults).
+
+.. image:: and_click_on_Save_accept_defaults.jpeg
+
+otherwise select INFRASTRUCTURE→ Hosts and click on "Add Host"
+
+.. image:: otherwise_select_INFRASTRUCTURE_Hosts_and_click_on_Add_Host.jpg
+
+2. Enter the management IP for the k8s VM (e.g. 10.0.0.4) that was just created.
+
+3. Click on “Copy to Clipboard” button
+
+4. Click on “Close” button
+
+.. image:: Click_on_Close_button.jpeg
+
+Without the 10.0.0.4 IP - the CATTLE_AGENT will be derived on the host - but it
+may not be a routable IP.
+
+Configure Kubernetes Host
+-------------------------
+
+1. Login to the new Kubernetes Host::
+
+ > ssh -i ~/oom-key.pem ubuntu@10.12.5.1
+ The authenticity of host '10.12.5.172 (10.12.5.172)' can't be established.
+ ECDSA key fingerprint is SHA256:tqxayN58nCJKOJcWrEZzImkc0qKQHDDfUTHqk4WMcEI.
+ Are you sure you want to continue connecting (yes/no)? yes
+ Warning: Permanently added '10.12.5.172' (ECDSA) to the list of known hosts.
+ Welcome to Ubuntu 16.04.2 LTS (GNU/Linux 4.4.0-64-generic x86_64)
+
+ * Documentation: https://help.ubuntu.com
+ * Management: https://landscape.canonical.com
+ * Support: https://ubuntu.com/advantage
+
+ Get cloud support with Ubuntu Advantage Cloud Guest:
+ http://www.ubuntu.com/business/services/cloud
+
+ 180 packages can be updated.
+ 100 updates are security updates.
+
+ The programs included with the Ubuntu system are free software;
+ the exact distribution terms for each program are described in the
+ individual files in /usr/share/doc/*/copyright.
+
+ Ubuntu comes with ABSOLUTELY NO WARRANTY, to the extent permitted by
+ applicable law.
+
+ To run a command as administrator (user "root"), use "sudo <command>".
+ See "man sudo_root" for details.
+
+ ubuntu@sb4-k8s-1:~$
+
+
+2. Paste Clipboard content and hit enter to install Rancher Agent::
+
+ ubuntu@sb4-k8s-1:~$ sudo docker run -e CATTLE_AGENT_IP="10.0.0.4“ --rm --privileged -v /var/run/docker.sock:/var/run/docker.sock -v /var/lib/rancher:/var/lib/rancher rancher/agent:v1.2.9 http://10.12.6.16:8080/v1/scripts/5D757C68BD0A2125602A:1514678400000:yKW9xHGJDLvq6drz2eDzR2mjato
+ Unable to find image 'rancher/agent:v1.2.9' locally
+ v1.2.9: Pulling From rancher/agent
+ b3e1c725a85f: Pull complete
+ 6071086409fc: Pull complete
+ d0ac3b234321: Pull complete
+ 87f567b5cf58: Pull complete
+ a63e24b217c4: Pull complete
+ d0a3f58caef0: Pull complete
+ 16914729cfd3: Pull complete
+ dc5c21984c5b: Pull complete
+ d7e8f9784b20: Pull complete
+ Digest: sha256:c21255ac4d94ffbc7b523F870F20ea5189b68Fa3d642800adb4774aab4748e66
+ Status: Downloaded newer image for rancher/agent:v1.2.9
+
+ INFO: Running Agent Registration Process, CATTLE_URL=http://10.12.6.16:8080/v1
+ INFO: Attempting to connect to: http://10.12.6.16:8080/v1
+ INFO: http://10.12.6.16:8080/v1 is accessible
+ INFO: Inspecting host capabilities
+ INFO: Boot2Docker: false
+ INFO: Host writable: true
+ INFO: Token: xxxxxxxx
+ INFO: Running registration
+ INFO: Printing Environment
+ INFO: ENV: CATTLE_ACCESS_KEY=98B35AC484FBF820E0AD
+ INFO: ENV: CATTLE_AGENT_IP=10.0.9.4
+ INFO: ENV: CATTLE_HOME=/var/lib/cattle
+ INFO: ENV: CATTLE_REGISTRATION_ACCESS_KEY=registrationToken
+ INFO: ENV: CATTLE_REGISTRATION_SECRET_KEY=xxxxxxx
+ INFO: ENV: CATTLE_SECRET_KEY=xxxxxxx
+ INFO: ENV: CATTLE_URL=http://10.12.6.16:8080/v1
+ INFO: ENV: DETECTED_CATTLE_AGENT_IP=10.12.5.172
+ INFO: ENV: RANCHER_AGENT_IMAGE=rancher/agent:v1.2.9
+ INFO: Launched Rancher Agent: c27ee0f3dc4c783b0db647ea1f73c35b3843a4b8d60b96375b1a05aa77d83136
+ ubuntu@sb4-k8s-1:~$
+
+3. Return to Rancher environment (e.g. SB4) and wait for services to complete
+ (~ 10-15 mins)
+
+.. image:: Return_to_Rancher_environment_eg_SB4_and_wait_for_services_to_complete_10-15_mins.jpeg
+
+Configure kubectl and helm
+==========================
+In this example we are configuring kubectl and helm that have been installed
+(as a convenience) onto the rancher and kubernetes hosts. Typically you would
+install them both on your PC and remotely connect to the cluster. The following
+procedure would remain the same.
+
+1. Click on CLI and then click on “Generate Config”
+
+.. image:: Click_on_CLI_and_then_click_on_Generate_Config.jpeg
+
+2. Click on “Copy to Clipboard” - wait until you see a "token" - do not copy
+ user+password - the server is not ready at that point
+
+.. image:: Click_on_Copy_to_Clipboard-wait_until_you_see_a_token-do_not_copy_user+password-the_server_is_not_ready_at_that_point.jpeg
+
+3. Create a .kube directory in user directory (if one does not exist)::
+
+ ubuntu@sb4-kSs-1:~$ mkdir .kube
+ ubuntu@sb4-kSs-1:~$ vi .kube/config
+
+4. Paste contents of Clipboard into a file called “config” and save the file::
+
+ apiVersion: v1
+ kind : Config
+ clusters:
+ - cluster:
+ api-version: v1
+ insecure-skip-tls-verify: true
+ server: "https://10.12.6.16:8080/r/projects/1a7/kubernetes:6443"
+ name: "SB4"
+ contexts:
+ - context:
+ cluster: "SB4"
+ user: "SB4"
+ name: "SB4"
+ current-context: "SB4"
+ users:
+ - name: "SB4"
+ user:
+ token: "QmFzaWMgTlRBd01qZzBOemc)TkRrMk1UWkNOMFpDTlVFNlExcHdSa1JhVZreE5XSm1TRGhWU2t0Vk1sQjVhalZaY0dWaFVtZGFVMHQzWW1WWVJtVmpSQT09"
+ ~
+ ~
+ ~
+ - INSERT --
+
+5. Validate that kubectl is able to connect to the kubernetes cluster::
+
+ ubuntu@sb4-k8s-1:~$ kubectl config get-contexts
+ CURRENT NAME CLUSTER AUTHINFO NAMESPACE
+ * SB4 SB4 SB4
+ ubuntu@sb4-kSs-1:~$
+
+and show running pods::
+
+ ubuntu@sb4-k8s-1:~$ kubectl get pods --all-namespaces -o=wide
+ NAMESPACE NAME READY STATUS RESTARTS AGE IP NODE
+ kube-system heapster—7Gb8cd7b5 -q7p42 1/1 Running 0 13m 10.42.213.49 sb4-k8s-1
+ kube-system kube-dns-5d7bM87c9-c6f67 3/3 Running 0 13m 10.42.181.110 sb4-k8s-1
+ kube-system kubernetes-dashboard-f9577fffd-kswjg 1/1 Running 0 13m 10.42.105.113 sb4-k8s-1
+ kube-system monitoring-grafana-997796fcf-vg9h9 1/1 Running 0 13m 10.42,141.58 sb4-k8s-1
+ kube-system monitoring-influxdb-56chd96b-hk66b 1/1 Running 0 13m 10.4Z.246.90 sb4-k8s-1
+ kube-system tiller-deploy-cc96d4f6b-v29k9 1/1 Running 0 13m 10.42.147.248 sb4-k8s-1
+ ubuntu@sb4-k8s-1:~$
+
+6. Validate helm is running at the right version. If not, an error like this
+ will be displayed::
+
+ ubuntu@sb4-k8s-1:~$ helm list
+ Error: incompatible versions c1ient[v2.8.2] server[v2.6.1]
+ ubuntu@sb4-k8s-1:~$
+
+7. Upgrade the server-side component of helm (tiller) via `helm init --upgrade`::
+
+ ubuntu@sb4-k8s-1:~$ helm init --upgrade
+ Creating /home/ubuntu/.helm
+ Creating /home/ubuntu/.helm/repository
+ Creating /home/ubuntu/.helm/repository/cache
+ Creating /home/ubuntu/.helm/repository/local
+ Creating /home/ubuntu/.helm/plugins
+ Creating /home/ubuntu/.helm/starters
+ Creating /home/ubuntu/.helm/cache/archive
+ Creating /home/ubuntu/.helm/repository/repositories.yaml
+ Adding stable repo with URL: https://kubernetes-charts.storage.googleapis.com
+ Adding local repo with URL: http://127.0.0.1:8879/charts
+ $HELM_HOME has been configured at /home/ubuntu/.helm.
+
+ Tiller (the Helm server-side component) has been upgraded to the current version.
+ Happy Helming!
+ ubuntu@sb4-k8s-1:~$
+
+ONAP Deployment via OOM
+=======================
+Now that kubernetes and Helm are installed and configured you can prepare to
+deploy ONAP. Follow the instructions in the README.md_ or look at the official
+documentation to get started:
+
+- :ref:`quick-start-label` - deploy ONAP on an existing cloud
+- :ref:`user-guide-label` - a guide for operators of an ONAP instance
+
+
diff --git a/docs/oom_user_guide.rst b/docs/oom_user_guide.rst
index bb8ff0b27c..f79de5950f 100644
--- a/docs/oom_user_guide.rst
+++ b/docs/oom_user_guide.rst
@@ -75,8 +75,8 @@ Install Helm
~~~~~~~~~~~~
Helm is used by OOM for package and configuration management. To install Helm, enter the following::
- > wget http://storage.googleapis.com/kubernetes-helm/helm-v2.7.2-linux-amd64.tar.gz
- > tar -zxvf helm-v2.7.2-linux-amd64.tar.gz
+ > wget http://storage.googleapis.com/kubernetes-helm/helm-v2.8.2-linux-amd64.tar.gz
+ > tar -zxvf helm-v2.8.2-linux-amd64.tar.gz
> sudo mv linux-amd64/helm /usr/local/bin/helm
Verify the Helm version with::
@@ -108,7 +108,7 @@ stable which should be removed to avoid confusion::
To prepare your system for an installation of ONAP, you'll need to::
- > git clone http://gerrit.onap.org/r/oom
+ > git clone -b beijing http://gerrit.onap.org/r/oom
> cd oom/kubernetes
@@ -196,7 +196,7 @@ Many technologies are used across the projects resulting in significant
operational complexity and an inability to apply global parameters across the
entire ONAP deployment. OOM solves this problem by introducing a common
configuration technology, Helm charts, that provide a hierarchical
-configuration configuration with the ability to override values with higher
+configuration with the ability to override values with higher
level charts or command line options.
The structure of the configuration of ONAP is shown in the following diagram.
@@ -420,10 +420,10 @@ to monitor the real-time health of an ONAP deployment:
- a set of liveness probes which feed into the Kubernetes manager which
are described in the Heal section.
-Within ONAP Consul is the monitoring system of choice and deployed by OOM in two parts:
+Within ONAP, Consul is the monitoring system of choice and deployed by OOM in two parts:
- a three-way, centralized Consul server cluster is deployed as a highly
- available monitor of all of the ONAP components,and
+ available monitor of all of the ONAP components, and
- a number of Consul agents.
The Consul server provides a user interface that allows a user to graphically
diff --git a/docs/openstack-k8s-node.sh b/docs/openstack-k8s-node.sh
new file mode 100644
index 0000000000..7d49ad5033
--- /dev/null
+++ b/docs/openstack-k8s-node.sh
@@ -0,0 +1,46 @@
+#!/bin/bash
+
+DOCKER_VERSION=17.03
+KUBECTL_VERSION=1.8.10
+HELM_VERSION=2.8.2
+
+# setup root access - default login: oom/oom - comment out to restrict access too ssh key only
+sed -i 's/PermitRootLogin.*/PermitRootLogin yes/' /etc/ssh/sshd_config
+sed -i 's/PasswordAuthentication.*/PasswordAuthentication yes/' /etc/ssh/sshd_config
+service sshd restart
+echo -e "oom\noom" | passwd root
+
+apt-get update
+curl https://releases.rancher.com/install-docker/$DOCKER_VERSION.sh | sh
+mkdir -p /etc/systemd/system/docker.service.d/
+cat > /etc/systemd/system/docker.service.d/docker.conf << EOF
+[Service]
+ExecStart=
+ExecStart=/usr/bin/dockerd -H fd:// --insecure-registry=nexus3.onap.org:10001
+EOF
+systemctl daemon-reload
+systemctl restart docker
+apt-mark hold docker-ce
+
+IP_ADDY=`ip address |grep ens|grep inet|awk '{print $2}'| awk -F / '{print $1}'`
+HOSTNAME=`hostname`
+
+echo "$IP_ADDY $HOSTNAME" >> /etc/hosts
+
+docker login -u docker -p docker nexus3.onap.org:10001
+
+sudo apt-get install make -y
+
+sudo curl -LO https://storage.googleapis.com/kubernetes-release/release/v$KUBECTL_VERSION/bin/linux/amd64/kubectl
+sudo chmod +x ./kubectl
+sudo mv ./kubectl /usr/local/bin/kubectl
+sudo mkdir ~/.kube
+wget http://storage.googleapis.com/kubernetes-helm/helm-v${HELM_VERSION}-linux-amd64.tar.gz
+sudo tar -zxvf helm-v${HELM_VERSION}-linux-amd64.tar.gz
+sudo mv linux-amd64/helm /usr/local/bin/helm
+
+# install nfs
+sudo apt-get install nfs-common -y
+
+
+exit 0
diff --git a/docs/openstack-rancher.sh b/docs/openstack-rancher.sh
new file mode 100644
index 0000000000..d51abe4d68
--- /dev/null
+++ b/docs/openstack-rancher.sh
@@ -0,0 +1,51 @@
+#!/bin/bash
+
+DOCKER_VERSION=17.03
+RANCHER_VERSION=1.6.14
+KUBECTL_VERSION=1.8.10
+HELM_VERSION=2.8.2
+
+# setup root access - default login: oom/oom - comment out to restrict access too ssh key only
+sed -i 's/PermitRootLogin.*/PermitRootLogin yes/' /etc/ssh/sshd_config
+sed -i 's/PasswordAuthentication.*/PasswordAuthentication yes/' /etc/ssh/sshd_config
+service sshd restart
+echo -e "oom\noom" | passwd root
+
+apt-get update
+curl https://releases.rancher.com/install-docker/$DOCKER_VERSION.sh | sh
+mkdir -p /etc/systemd/system/docker.service.d/
+cat > /etc/systemd/system/docker.service.d/docker.conf << EOF
+[Service]
+ExecStart=
+ExecStart=/usr/bin/dockerd -H fd:// --insecure-registry=nexus3.onap.org:10001
+EOF
+systemctl daemon-reload
+systemctl restart docker
+apt-mark hold docker-ce
+
+IP_ADDY=`ip address |grep ens|grep inet|awk '{print $2}'| awk -F / '{print $1}'`
+HOSTNAME=`hostname`
+
+echo "$IP_ADDY $HOSTNAME" >> /etc/hosts
+
+docker login -u docker -p docker nexus3.onap.org:10001
+
+sudo apt-get install make -y
+
+sudo docker run -d --restart=unless-stopped -p 8080:8080 --name rancher_server rancher/server:v$RANCHER_VERSION
+sudo curl -LO https://storage.googleapis.com/kubernetes-release/release/v$KUBECTL_VERSION/bin/linux/amd64/kubectl
+sudo chmod +x ./kubectl
+sudo mv ./kubectl /usr/local/bin/kubectl
+sudo mkdir ~/.kube
+wget http://storage.googleapis.com/kubernetes-helm/helm-v${HELM_VERSION}-linux-amd64.tar.gz
+sudo tar -zxvf helm-v${HELM_VERSION}-linux-amd64.tar.gz
+sudo mv linux-amd64/helm /usr/local/bin/helm
+
+# nfs server
+sudo apt-get install nfs-kernel-server -y
+
+sudo mkdir -p /nfs_share
+sudo chown nobody:nogroup /nfs_share/
+
+
+exit 0
diff --git a/docs/otherwise_select_INFRASTRUCTURE_Hosts_and_click_on_Add_Host.jpg b/docs/otherwise_select_INFRASTRUCTURE_Hosts_and_click_on_Add_Host.jpg
new file mode 100644
index 0000000000..18af758a0d
--- /dev/null
+++ b/docs/otherwise_select_INFRASTRUCTURE_Hosts_and_click_on_Add_Host.jpg
Binary files differ
diff --git a/docs/release-notes.rst b/docs/release-notes.rst
index 2573c2ca02..3cb547cce3 100644
--- a/docs/release-notes.rst
+++ b/docs/release-notes.rst
@@ -8,66 +8,10 @@
Release Notes
=============
-Version: 1.1.0
---------------
-
-:Release Date: 2017-11-16
-
-**New Features**
-
-The Amsterdam release is the first release of the ONAP Operations Manager (OOM).
-
-The main goal of the Amsterdam release was to:
-
- - Support Flexible Platform Deployment via Kubernetes of fully containerized OMAP components - on any type of environment.
- - Support State Management of ONAP platform components.
- - Support full production ONAP deployment and any variation of component level deployment for development.
- - Platform Operations Orchestration / Control Loop Actions.
- - Platform centralized logging with ELK stack.
-
-**Bug Fixes**
-
- The full list of implemented user stories and epics is available on `JIRA <https://jira.onap.org/secure/RapidBoard.jspa?rapidView=41&view=planning.nodetail&epics=visible>`_
- This is the first release of OOM, the defects fixed in this release were raised during the course of the release.
- Anything not closed is captured below under Known Issues. If you want to review the defects fixed in the Amsterdam release, refer to Jira link above.
-
-**Known Issues**
- - `OOM-6 <https://jira.onap.org/browse/OOM-6>`_ Automated platform deployment on Docker/Kubernetes
-
- VFC, AAF, MSB minor issues.
-
- Workaround: Manual configuration changes - however the reference vFirewall use case does not currently require these components.
-
- - `OOM-10 <https://jira.onap.org/browse/OOM-10>`_ Platform configuration management.
-
- OOM ONAP Configuration Management - Handling of Secrets.
-
- Workaround: Automated workaround to be able to pull from protected docker repositories.
-
-
-**Security Issues**
- N/A
-
-
-**Upgrade Notes**
-
- N/A
-
-**Deprecation Notes**
-
- N/A
-
-**Other**
-
-
-
-===========
-
-
Version 2.0.0 Beijing Release
-----------------------------
-:Release Date: 2018-05-24
+:Release Date: 2018-06-07
Epic
****
@@ -75,7 +19,7 @@ Epic
* [`OOM-6 <https://jira.onap.org/browse/OOM-6>`_] - Automated platform deployment on Docker/Kubernetes
* [`OOM-7 <https://jira.onap.org/browse/OOM-7>`_] - Platform monitoring and auto-healing
* [`OOM-8 <https://jira.onap.org/browse/OOM-8>`_] - Automated platform scalability
-* [`OOM-9 <https://jira.onap.org/browse/OOM-9>`_] - Platform upgradability&rollbacks
+* [`OOM-9 <https://jira.onap.org/browse/OOM-9>`_] - Platform upgradability & rollbacks
* [`OOM-10 <https://jira.onap.org/browse/OOM-10>`_] - Platform configuration management
* [`OOM-46 <https://jira.onap.org/browse/OOM-46>`_] - Platform infrastructure deployment with TOSCA
* [`OOM-109 <https://jira.onap.org/browse/OOM-109>`_] - Platform Centralized Logging
@@ -237,7 +181,6 @@ Task
* [`OOM-624 <https://jira.onap.org/browse/OOM-624>`_] - CII security badging: cleartext password for keystone and docker repo creds
* [`OOM-726 <https://jira.onap.org/browse/OOM-726>`_] - Mirror AAI docker version changes into OOM from AAI-791
* [`OOM-772 <https://jira.onap.org/browse/OOM-772>`_] - Remove old DCAE from Release
-* [`OOM-793 <https://jira.onap.org/browse/OOM-793>`_] - Document OOM-722 health/ete script changes for onap-discuss/wiki/rtd/integration team
* [`OOM-801 <https://jira.onap.org/browse/OOM-801>`_] - Policy docker images rename - key off new name in POLICY-674
* [`OOM-810 <https://jira.onap.org/browse/OOM-810>`_] - Improve emsdriver code
* [`OOM-819 <https://jira.onap.org/browse/OOM-819>`_] - expose log/logstash 5044 as nodeport for external log producers outside of the kubernetes cluster
@@ -247,6 +190,7 @@ Task
* [`OOM-975 <https://jira.onap.org/browse/OOM-975>`_] - Notes are missing in multicloud
* [`OOM-1031 <https://jira.onap.org/browse/OOM-1031>`_] - Config Changes for consul to make vid, so, log health checks pass
* [`OOM-1032 <https://jira.onap.org/browse/OOM-1032>`_] - Making consul Stateful
+* [`OOM-1122 <https://jira.onap.org/browse/OOM-1122>`_] - Update APPC OOM chart to use Beijing release artifacts
Bug
***
@@ -324,7 +268,7 @@ Bug
* [`OOM-874 <https://jira.onap.org/browse/OOM-874>`_] - Inconsistent repository references in ONAP charts
* [`OOM-875 <https://jira.onap.org/browse/OOM-875>`_] - Cannot retrieve robot logs
* [`OOM-876 <https://jira.onap.org/browse/OOM-876>`_] - Some containers ignore the repository setting
-* [`OOM-878 <https://jira.onap.org/browse/OOM-878>`_] - MySQL slave nodes don't deploy when mysql.replicaCount>1
+* [`OOM-878 <https://jira.onap.org/browse/OOM-878>`_] - MySQL slave nodes don't deploy when mysql.replicaCount > 1
* [`OOM-881 <https://jira.onap.org/browse/OOM-881>`_] - SDN-C Portal pod fails to come up
* [`OOM-882 <https://jira.onap.org/browse/OOM-882>`_] - Some SDNC service names should be prefixed with the helm release name
* [`OOM-884 <https://jira.onap.org/browse/OOM-884>`_] - VID-VID mariadb pv is pointing to a wrong location
@@ -369,7 +313,6 @@ Bug
* [`OOM-964 <https://jira.onap.org/browse/OOM-964>`_] - SDC Healthcheck failure on sdc-be and sdc-kb containers down
* [`OOM-968 <https://jira.onap.org/browse/OOM-968>`_] - warning on default deployment values.yaml
* [`OOM-969 <https://jira.onap.org/browse/OOM-969>`_] - oomk8s images have no Dockerfile's
-* [`OOM-970 <https://jira.onap.org/browse/OOM-970>`_] - Can't configure mysql password for sdnctl user
* [`OOM-971 <https://jira.onap.org/browse/OOM-971>`_] - Common service name template should allow for chart name override
* [`OOM-974 <https://jira.onap.org/browse/OOM-974>`_] - Cassandra bootstrap is done incorrectly
* [`OOM-977 <https://jira.onap.org/browse/OOM-977>`_] - The esr-gui annotations should include a "path" param when register to MSB
@@ -403,7 +346,6 @@ Bug
* [`OOM-1039 <https://jira.onap.org/browse/OOM-1039>`_] - Service distribution to SO fails
* [`OOM-1041 <https://jira.onap.org/browse/OOM-1041>`_] - aai-service was renamed, but old references remain
* [`OOM-1042 <https://jira.onap.org/browse/OOM-1042>`_] - portalapps service was renamed, but old references remain
-* [`OOM-1044 <https://jira.onap.org/browse/OOM-1044>`_] - Fix image/table warning during deploy - since helm install switch a month ago - non-affecting - but check the yaml
* [`OOM-1045 <https://jira.onap.org/browse/OOM-1045>`_] - top level values.yaml missing entry for dmaap chart
* [`OOM-1049 <https://jira.onap.org/browse/OOM-1049>`_] - SDNC_UEB_LISTENER db
* [`OOM-1050 <https://jira.onap.org/browse/OOM-1050>`_] - Impossible to deploy consul using cache docker registry
@@ -415,12 +357,33 @@ Bug
* [`OOM-1064 <https://jira.onap.org/browse/OOM-1064>`_] - Improve docker registry secret management
* [`OOM-1066 <https://jira.onap.org/browse/OOM-1066>`_] - Updating TOSCA blueprint to sync up with helm configuration changes (add dmaap and oof/delete message-router)
* [`OOM-1068 <https://jira.onap.org/browse/OOM-1068>`_] - Update SO with new AAI cert
-* [`OOM-1070 <https://jira.onap.org/browse/OOM-1070>`_] - SO logs partially going to /var/log/ecomp/MSO
* [`OOM-1076 <https://jira.onap.org/browse/OOM-1076>`_] - some charts still using readiness check image from amsterdam 1.x
* [`OOM-1077 <https://jira.onap.org/browse/OOM-1077>`_] - AAI resources and traversal deployment failure on non-rancher envs
* [`OOM-1079 <https://jira.onap.org/browse/OOM-1079>`_] - Robot charts dont allow over ride of pub_key, dcae_collector_ip and dcae_collector_port
-* [`OOM-1081 <https://jira.onap.org/browse/OOM-1081>`_] - Remove component'mock'from TOSCA deployment
+* [`OOM-1081 <https://jira.onap.org/browse/OOM-1081>`_] - Remove component 'mock' from TOSCA deployment
* [`OOM-1082 <https://jira.onap.org/browse/OOM-1082>`_] - Wrong pv location of dcae postgres
+* [`OOM-1085 <https://jira.onap.org/browse/OOM-1085>`_] - appc hostname is incorrect in url
+* [`OOM-1086 <https://jira.onap.org/browse/OOM-1086>`_] - clamp deployment changes /dockerdata-nfs/ReleaseName dir permissions
+* [`OOM-1088 <https://jira.onap.org/browse/OOM-1088>`_] - APPC returns error for vCPE restart message from Policy
+* [`OOM-1089 <https://jira.onap.org/browse/OOM-1089>`_] - DCAE pods are not getting purged
+* [`OOM-1093 <https://jira.onap.org/browse/OOM-1093>`_] - Line wrapping issue in redis-cluster-config.sh script
+* [`OOM-1094 <https://jira.onap.org/browse/OOM-1094>`_] - Fix postgres startup
+* [`OOM-1095 <https://jira.onap.org/browse/OOM-1095>`_] - common makefile builds out of order
+* [`OOM-1096 <https://jira.onap.org/browse/OOM-1096>`_] - node port conflict SDNC (Geo enabled) & other charts
+* [`OOM-1097 <https://jira.onap.org/browse/OOM-1097>`_] - Nbi needs dep-nbi - crash on make all
+* [`OOM-1099 <https://jira.onap.org/browse/OOM-1099>`_] - Add External Interface NBI project into OOM TOSCA
+* [`OOM-1102 <https://jira.onap.org/browse/OOM-1102>`_] - Incorrect AAI services
+* [`OOM-1103 <https://jira.onap.org/browse/OOM-1103>`_] - Cannot disable NBI
+* [`OOM-1104 <https://jira.onap.org/browse/OOM-1104>`_] - Policy DROOLS configuration across container restarts
+* [`OOM-1110 <https://jira.onap.org/browse/OOM-1110>`_] - Clamp issue when connecting Policy
+* [`OOM-1111 <https://jira.onap.org/browse/OOM-1111>`_] - Please revert to using VNFSDK Postgres container
+* [`OOM-1114 <https://jira.onap.org/browse/OOM-1114>`_] - APPC is broken in latest helm chart
+* [`OOM-1115 <https://jira.onap.org/browse/OOM-1115>`_] - SDNC DGBuilder cant operate on DGs in database - need NodePort
+* [`OOM-1116 <https://jira.onap.org/browse/OOM-1116>`_] - Correct values needed by NBI chart
+* [`OOM-1124 <https://jira.onap.org/browse/OOM-1124>`_] - Update OOM APPC chart to enhance AAF support
+* [`OOM-1126 <https://jira.onap.org/browse/OOM-1126>`_] - Incorrect Port mapping between CDT Application and APPC main application
+* [`OOM-1127 <https://jira.onap.org/browse/OOM-1127>`_] - SO fails healthcheck
+* [`OOM-1128 <https://jira.onap.org/browse/OOM-1128>`_] - AAF CS fails to start in OpenLab
Sub-task
********
@@ -433,16 +396,75 @@ Sub-task
* [`OOM-655 <https://jira.onap.org/browse/OOM-655>`_] - Create alternate prepull script which provides more user feedback and logging
* [`OOM-753 <https://jira.onap.org/browse/OOM-753>`_] - Create Helm Sub-Chart for SO's embedded mariadb
* [`OOM-754 <https://jira.onap.org/browse/OOM-754>`_] - Create Helm Chart for SO
-* [`OOM-763 <https://jira.onap.org/browse/OOM-763>`_] - Work with Robot team to minimize/optimize configuration requirements
* [`OOM-774 <https://jira.onap.org/browse/OOM-774>`_] - Create Helm Sub-Chart for APPC's embedded mySQL database
* [`OOM-775 <https://jira.onap.org/browse/OOM-775>`_] - Create Helm Chart for APPC
* [`OOM-778 <https://jira.onap.org/browse/OOM-778>`_] - Replace NFS Provisioner with configurable PV storage solution
* [`OOM-825 <https://jira.onap.org/browse/OOM-825>`_] - Apache 2 License updation for All sqls and .js file
-* [`OOM-833 <https://jira.onap.org/browse/OOM-833>`_] - Apache 2 license addition for all configuration
* [`OOM-849 <https://jira.onap.org/browse/OOM-849>`_] - Policy Nexus component needs persistent volume for /sonatype-work
* [`OOM-991 <https://jira.onap.org/browse/OOM-991>`_] - Adjust SDC-BE init job timing from 10 to 30s to avoid restarts on single node systems
* [`OOM-1036 <https://jira.onap.org/browse/OOM-1036>`_] - update helm from 2.7.2 to 2.8.2 wiki/rtd
* [`OOM-1063 <https://jira.onap.org/browse/OOM-1063>`_] - Document Portal LoadBalancer Ingress IP Settings
+**Security Notes**
+
+OOM code has been formally scanned during build time using NexusIQ and no Critical vulnerability was found.
+
+Quick Links:
+ - `OOM project page <https://wiki.onap.org/display/DW/ONAP+Operations+Manager+Project>`_
+
+ - `Passing Badge information for OOM <https://bestpractices.coreinfrastructure.org/en/projects/1631>`_
+
+Version: 1.1.0
+--------------
+
+:Release Date: 2017-11-16
+
+**New Features**
+
+The Amsterdam release is the first release of the ONAP Operations Manager (OOM).
+
+The main goal of the Amsterdam release was to:
+
+ - Support Flexible Platform Deployment via Kubernetes of fully containerized OMAP components - on any type of environment.
+ - Support State Management of ONAP platform components.
+ - Support full production ONAP deployment and any variation of component level deployment for development.
+ - Platform Operations Orchestration / Control Loop Actions.
+ - Platform centralized logging with ELK stack.
+
+**Bug Fixes**
+
+ The full list of implemented user stories and epics is available on `JIRA <https://jira.onap.org/secure/RapidBoard.jspa?rapidView=41&view=planning.nodetail&epics=visible>`_
+ This is the first release of OOM, the defects fixed in this release were raised during the course of the release.
+ Anything not closed is captured below under Known Issues. If you want to review the defects fixed in the Amsterdam release, refer to Jira link above.
+
+**Known Issues**
+ - `OOM-6 <https://jira.onap.org/browse/OOM-6>`_ Automated platform deployment on Docker/Kubernetes
+
+ VFC, AAF, MSB minor issues.
+
+ Workaround: Manual configuration changes - however the reference vFirewall use case does not currently require these components.
+
+ - `OOM-10 <https://jira.onap.org/browse/OOM-10>`_ Platform configuration management.
+
+ OOM ONAP Configuration Management - Handling of Secrets.
+
+ Workaround: Automated workaround to be able to pull from protected docker repositories.
+
+
+**Security Issues**
+ N/A
+
+
+**Upgrade Notes**
+
+ N/A
+
+**Deprecation Notes**
+
+ N/A
+
+**Other**
+
+ N/A
End of Release Notes
diff --git a/docs/slave_nfs_node.sh b/docs/slave_nfs_node.sh
new file mode 100644
index 0000000000..5cb164ccac
--- /dev/null
+++ b/docs/slave_nfs_node.sh
@@ -0,0 +1,25 @@
+#!/bin/bash
+
+usage () {
+ echo "Usage:"
+ echo " ./$(basename $0) nfs_master_ip"
+ exit 1
+}
+
+if [ "$#" -ne 1 ]; then
+ echo "Missing NFS mater node"
+ usage
+fi
+
+MASTER_IP=$1
+
+#Install NFS common
+sudo apt-get update
+sudo apt-get install -y nfs-common
+
+#Create NFS directory
+sudo mkdir -p /dockerdata-nfs
+
+#Mount the remote NFS directory to the local one
+sudo mount $MASTER_IP:/dockerdata-nfs /dockerdata-nfs/
+echo "$MASTER_IP:/dockerdata-nfs /dockerdata-nfs nfs auto,nofail,noatime,nolock,intr,tcp,actimeo=1800 0 0" | sudo tee -a /etc/fstab