diff options
Diffstat (limited to 'docs/oom_user_guide.rst')
-rw-r--r-- | docs/oom_user_guide.rst | 57 |
1 files changed, 29 insertions, 28 deletions
diff --git a/docs/oom_user_guide.rst b/docs/oom_user_guide.rst index 3a707e25ea..2059251928 100644 --- a/docs/oom_user_guide.rst +++ b/docs/oom_user_guide.rst @@ -204,7 +204,7 @@ deployment:: To install a specific version of a single ONAP component (`so` in this example) with the given release name enter:: - > helm deploy so onap/so --version 8.0.0 --set global.masterPassword=password --set global.flavor=unlimited --namespace onap + > helm deploy so onap/so --version 9.0.0 --set global.masterPassword=password --set global.flavor=unlimited --namespace onap .. note:: The dependent components should be installed for component being installed @@ -332,7 +332,7 @@ file: dependencies: <...> - name: so - version: ~8.0.0 + version: ~9.0.0 repository: '@local' condition: so.enabled <...> @@ -444,23 +444,24 @@ the portal and then simply access now the new ssl-encrypted URL: | Alternatives Considered: - - Kubernetes port forwarding was considered but discarded as it would require - the end user to run a script that opens up port forwarding tunnels to each of - the pods that provides a portal application widget. + - Kubernetes port forwarding was considered but discarded as it would + require the end user to run a script that opens up port forwarding tunnels + to each of the pods that provides a portal application widget. - Reverting to a VNC server similar to what was deployed in the Amsterdam - release was also considered but there were many issues with resolution, lack - of volume mount, /etc/hosts dynamic update, file upload that were a tall order - to solve in time for the Beijing release. + release was also considered but there were many issues with resolution, + lack of volume mount, /etc/hosts dynamic update, file upload that were + a tall order to solve in time for the Beijing release. Observations: - - If you are not using floating IPs in your Kubernetes deployment and directly attaching - a public IP address (i.e. by using your public provider network) to your K8S Node - VMs' network interface, then the output of 'kubectl -n onap get services | grep "portal-app"' + - If you are not using floating IPs in your Kubernetes deployment and + directly attaching a public IP address (i.e. by using your public provider + network) to your K8S Node VMs' network interface, then the output of + 'kubectl -n onap get services | grep "portal-app"' will show your public IP instead of the private network's IP. Therefore, - you can grab this public IP directly (as compared to trying to find the floating - IP first) and map this IP in /etc/hosts. + you can grab this public IP directly (as compared to trying to find the + floating IP first) and map this IP in /etc/hosts. .. figure:: oomLogoV2-Monitor.png :align: right @@ -561,15 +562,15 @@ Below is the example for the same:: > helm list NAME REVISION UPDATED STATUS CHART APP VERSION NAMESPACE - dev 1 Wed Oct 14 13:49:52 2020 DEPLOYED onap-8.0.0 Honolulu onap - dev-cassandra 5 Thu Oct 15 14:45:34 2020 DEPLOYED cassandra-8.0.0 onap - dev-contrib 1 Wed Oct 14 13:52:53 2020 DEPLOYED contrib-8.0.0 onap - dev-mariadb-galera 1 Wed Oct 14 13:55:56 2020 DEPLOYED mariadb-galera-8.0.0 onap + dev 1 Wed Oct 14 13:49:52 2020 DEPLOYED onap-9.0.0 Istanbul onap + dev-cassandra 5 Thu Oct 15 14:45:34 2020 DEPLOYED cassandra-9.0.0 onap + dev-contrib 1 Wed Oct 14 13:52:53 2020 DEPLOYED contrib-9.0.0 onap + dev-mariadb-galera 1 Wed Oct 14 13:55:56 2020 DEPLOYED mariadb-galera-9.0.0 onap Here the Name column shows the RELEASE NAME, In our case we want to try the scale operation on cassandra, thus the RELEASE NAME would be dev-cassandra. -Now we need to obtain the chart name for casssandra. Use the below +Now we need to obtain the chart name for cassandra. Use the below command to get the chart name:: > helm search cassandra @@ -578,21 +579,21 @@ Below is the example for the same:: > helm search cassandra NAME CHART VERSION APP VERSION DESCRIPTION - local/cassandra 8.0.0 ONAP cassandra - local/portal-cassandra 8.0.0 Portal cassandra - local/aaf-cass 8.0.0 ONAP AAF cassandra - local/sdc-cs 8.0.0 ONAP Service Design and Creation Cassandra + local/cassandra 9.0.0 ONAP cassandra + local/portal-cassandra 9.0.0 Portal cassandra + local/aaf-cass 9.0.0 ONAP AAF cassandra + local/sdc-cs 9.0.0 ONAP Service Design and Creation Cassandra Here the Name column shows the chart name. As we want to try the scale -operation for cassandra, thus the correponding chart name is local/cassandra +operation for cassandra, thus the corresponding chart name is local/cassandra Now we have both the command's arguments, thus we can perform the -scale opeartion for cassandra as follows:: +scale operation for cassandra as follows:: > helm upgrade dev-cassandra local/cassandra --set replicaCount=3 -Using this command we can scale up or scale down the cassadra db instances. +Using this command we can scale up or scale down the cassandra db instances. The ONAP components use Kubernetes provided facilities to build clustered, @@ -643,7 +644,7 @@ Prior to doing an upgrade, determine of the status of the deployed charts:: > helm list NAME REVISION UPDATED STATUS CHART NAMESPACE - so 1 Mon Feb 5 10:05:22 2020 DEPLOYED so-8.0.0 onap + so 1 Mon Feb 5 10:05:22 2020 DEPLOYED so-9.0.0 onap When upgrading a cluster a parameter controls the minimum size of the cluster during the upgrade while another parameter controls the maximum number of nodes @@ -680,7 +681,7 @@ To fetch release history enter:: > helm history so REVISION UPDATED STATUS CHART DESCRIPTION 1 Mon Feb 5 10:05:22 2020 SUPERSEDED so-8.0.0 Install complete - 2 Mon Feb 5 10:10:55 2020 DEPLOYED so-8.0.1 Upgrade complete + 2 Mon Feb 5 10:10:55 2020 DEPLOYED so-9.0.0 Upgrade complete Unfortunately, not all upgrades are successful. In recognition of this the lineup of pods within an ONAP deployment is tagged such that an administrator @@ -703,7 +704,7 @@ For example, to roll-back back to previous system revision enter:: > helm history so REVISION UPDATED STATUS CHART DESCRIPTION 1 Mon Feb 5 10:05:22 2020 SUPERSEDED so-8.0.0 Install complete - 2 Mon Feb 5 10:10:55 2020 SUPERSEDED so-8.0.1 Upgrade complete + 2 Mon Feb 5 10:10:55 2020 SUPERSEDED so-9.0.0 Upgrade complete 3 Mon Feb 5 10:14:32 2020 DEPLOYED so-8.0.0 Rollback to 1 .. note:: |