summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorVijay Venkatesh Kumar <vv770d@att.com>2022-10-26 13:49:38 -0400
committerVijay Venkatesh Kumar <vv770d@att.com>2022-10-26 13:50:12 -0400
commit39479a850807e840ff08d1fcf4b01a04ba334e05 (patch)
tree2c8b20629a4f36c5efba7a61510bce3669b736d7
parentdcfec1a95f0a4f605ac64833b59e39a0b584daa9 (diff)
DCAE Kohn rls note updates
Third pass Removed all cloudify install steps + Misc fixes for issues noted on https://gerrit.onap.org/r/c/dcaegen2/+/131680 Change-Id: I825717d7ef42cc7b008c613956ca20b8393f5366 Signed-off-by: Vijay Venkatesh Kumar <vv770d@att.com> Issue-ID: DCAEGEN2-3236 Signed-off-by: Vijay Venkatesh Kumar <vv770d@att.com>
-rw-r--r--docs/sections/architecture.rst6
-rw-r--r--docs/sections/images/R11_architecture_diagram.png (renamed from docs/sections/images/R11_architecture_diagram.PNG)bin96735 -> 96735 bytes
-rw-r--r--docs/sections/services/datalake-handler/index.rst1
-rw-r--r--docs/sections/services/datalake-handler/installation.rst135
-rw-r--r--docs/sections/services/dfc/index.rst1
-rw-r--r--docs/sections/services/dfc/installation.rst85
-rw-r--r--docs/sections/services/kpi-computation-ms/index.rst1
-rw-r--r--docs/sections/services/kpi-computation-ms/installation.rst95
-rw-r--r--docs/sections/services/pm-mapper/index.rst1
-rw-r--r--docs/sections/services/pm-mapper/installation.rst62
-rw-r--r--docs/sections/services/slice-analysis-ms/index.rst1
-rw-r--r--docs/sections/services/slice-analysis-ms/installation.rst80
-rw-r--r--docs/sections/services/son-handler/index.rst1
-rw-r--r--docs/sections/services/son-handler/installation.rst156
-rw-r--r--docs/sections/services/tcagen2-docker/installation.rst79
-rw-r--r--docs/sections/services/ves-http/index.rst1
-rw-r--r--docs/sections/services/ves-http/installation.rst157
-rw-r--r--docs/sections/services/ves-hv/index.rst1
18 files changed, 3 insertions, 860 deletions
diff --git a/docs/sections/architecture.rst b/docs/sections/architecture.rst
index 370acac0..11271e98 100644
--- a/docs/sections/architecture.rst
+++ b/docs/sections/architecture.rst
@@ -1,4 +1,4 @@
- .. This work is licensed under a Creative Commons Attribution 4.0 International License.
+.. This work is licensed under a Creative Commons Attribution 4.0 International License.
.. http://creativecommons.org/licenses/by/4.0
.. _architecture:
@@ -8,12 +8,12 @@ Architecture
DCAE project provides intelligence for ONAP to support automation (via open-loop and CL) by performing network data collections, analytics & correlation and trigger actionable rootcause events.
-Prior to Jakarta release, DCAE architecture comprised of DCAE Platform and DCAE Service components; DCAE Platform supported the functions to deploy, host and perform LCM applications of Service components. With Jakarta release, these DCAE Platform components centered around Cloudify has been removed and all Microservice orchestration and lifecycle management are supported primarily through Helm/Kubernetes.
+Prior to Jakarta release, DCAE architecture comprised of DCAE Platform and DCAE Service components; DCAE Platform supported the functions to deploy, host and perform LCM applications of Service components. With Jakarta release, these DCAE Platform components centered around Cloudify have been removed and all Microservice orchestration and lifecycle management are supported through Helm/Kubernetes.
The DCAE services components includes all the microservices - collectors, analytics and event processor which supports active data-flow and processing as required by ONAP usecases. These Service components are the functional entities that realize the various
data collection needs, event processors for data standardization, analytics that assess collected data, and various auxiliary microservices that assist automated closed loop flows.
-The architecture of DCAE with Helm transformation is more flexible, microsservice oriented and supports model based component design and deployment through DCAE-MOD. Also with migration to helm, DCAE microservice deployments can be handled independently, dependencies are captured under its helm charts.
+The architecture of DCAE with Helm transformation is more flexible, microservice oriented and supports model based component design and deployment through DCAE-MOD. Also with migration to helm, DCAE microservice deployments can be handled independently, dependencies are captured under its helm charts.
Under the Cloudify/DCAE Platform architecture, DCAE Service components relied on Consul's distributed K-V to manage and store component configuration. With Platform simplification under DCAE Transformation initiative for Kohn, Consul dependency has been removed across all DCAE service components. All Microservice configuration are currently resolved through files mounted via Configmap created part of dcae-services helm chart deployment.
diff --git a/docs/sections/images/R11_architecture_diagram.PNG b/docs/sections/images/R11_architecture_diagram.png
index d84fc7ba..d84fc7ba 100644
--- a/docs/sections/images/R11_architecture_diagram.PNG
+++ b/docs/sections/images/R11_architecture_diagram.png
Binary files differ
diff --git a/docs/sections/services/datalake-handler/index.rst b/docs/sections/services/datalake-handler/index.rst
index 3fade9c6..e4f1c905 100644
--- a/docs/sections/services/datalake-handler/index.rst
+++ b/docs/sections/services/datalake-handler/index.rst
@@ -27,7 +27,6 @@ DataLake-Handler MS Installation Steps and Configurations
.. toctree::
:maxdepth: 1
- ./installation.rst
./installation-helm.rst
DataLake-Handler MS Admin UI User Guide
diff --git a/docs/sections/services/datalake-handler/installation.rst b/docs/sections/services/datalake-handler/installation.rst
deleted file mode 100644
index e9215c29..00000000
--- a/docs/sections/services/datalake-handler/installation.rst
+++ /dev/null
@@ -1,135 +0,0 @@
-.. This work is licensed under a Creative Commons Attribution 4.0 International License.
-.. http://creativecommons.org/licenses/by/4.0
-.. _dl-installation:
-
-Deployment Steps
-
-################
-DL-handler consists of three pods- the feeder, admin UI and des. It can be deployed by using cloudify blueprint. Datalake can be easily deployed through DCAE cloudify manager. The following steps guides you launch Datalake though cloudify manager.
-
-Pre-requisite
--------------
-Make sure dcae postgres is properly deployed and functional.
-An external database, such as Elasticsearch and MongoDB is deployed. Install mongodb through the following command.
-
- #docker run -itd --restart=always --name dl-mongo -p 27017:27017 mongo
-
-For DES service deployment, presto service is deployed. Here is a sample how presto deploy in the environment.
- Build a presto image:
- The package of presto version we are using is v0.0.2:presto-v0.0.2.tar.gz
-
- #docker build -t presto:v0.0.2 .
- #docker tag presto:v0.0.2 registry.baidubce.com/onap/presto:v0.0.2
- #docker push registry.baidubce.com/onap/presto:v0.0.2
-
- Note: Replace the repository path with your own repository.
-
- Install presto service:
-
- #kubectl -n onap run dl-presto --image=registry.baidubce.com/onap/presto:v0.0.2 --env="MongoDB_IP=192.168.235.11" --env="MongoDB_PORT=27017"
- #kubectl -n onap expose deployment dl-presto --port=9000 --target-port=9000 --type=NodePort
-
- Note: MonoDB_IP and Mongo_PORT you can replace this two values with your own configuration.
-
-After datalake getting deployed, the admin UI can be used to configure the sink database address and credentials.
-
-Log-in to the DCAE Bootstrap POD
---------------------------------
-
-First, we should find the bootstrap pod name through the following command and make sure that DCAE coudify manager is properly deployed.
- .. image :: ./images/bootstrap-pod.png
-
-Login to the DCAE bootstrap pod through the following command.
- .. code-block :: bash
-
- #kubectl exec -it <DCAE bootstrap pod> /bin/bash -n onap
-
-Validate Blueprint
-------------------
-Before the blueprints uploading to Cloudify manager, the blueprints shoule be validated first through the following command.
- .. code-block :: bash
-
- #cfy blueprint validate /bluerints/k8s-datalake-feeder.yaml
- #cfy blueprint validate /blueprints/k8s-datalake-admin-ui.yaml
- #cfy blueprint validate /blueprints/k8s-datalake-des.yaml
-
-Upload the Blueprint to Cloudify Manager.
------------------------------------------
-After validating, we can start to proceed blueprints uploading.
- .. code-block :: bash
-
- #cfy blueprint upload -b dl-feeder /bluerints/k8s-datalake-feeder.yaml
- #cfy blueprint upload -b dl-admin-ui /blueprints/k8s-datalake-admin-ui.yaml
- #cfy blueprint upload -b des /blueprints/k8s-datalake-des.yaml
-
-Verify Uploaded Blueprints
---------------------------
-Using "cfy blueprint list" to verify your work.
- .. code-block :: bash
-
- #cfy blueprint list
-
-You can see the following returned message to show the blueprints have been correctly uploaded.
- .. image :: ./images/blueprint-list.png
-
-
-Verify Plugin Versions
-----------------------
-If the version of the plugin used is different, update the blueprint import to match.
- .. code-block :: bash
-
- #cfy plugins list
-
-Create Deployment
------------------
-Here we are going to create deployments for both feeder and admin UI.
- .. code-block :: bash
-
- #cfy deployments create -b dl-feeder feeder-deploy
- #cfy deployments create -b dl-admin-ui admin-ui-deploy
- #cfy deployments create -b des des
-
-Launch Service
---------------
-Next, we are going to launch the datalake.
- .. code-block :: bash
-
- #cfy executions start -d feeder-deploy install
- #cfy executions start -d admin-ui-deploy install
- #cfy executions start -d des install
-
-
-Verify the Deployment Result
-----------------------------
-The following command can be used to list the datalake logs.
-
- .. code-block :: bash
-
- #kubectl logs <datalake-pod> -n onap
-
-The output should looks like.
- .. image :: ./images/feeder-log.png
-
-The des output should looks like.
- .. image :: ./des-log.png
-
-If you find any Java exception from log, make sure that the external database and datalake configuration are properly configured.
-Admin UI can be used to configure the external database configuration.
-
-
-Uninstall
----------
-Uninstall running component and delete deployment
- .. code-block :: bash
-
- #cfy uninstall feeder-deploy
- #cfy uninstall admin-ui-deploy
- #cfy uninstall des
-
-Delete Blueprint
-----------------
- .. code-block :: bash
-
- #cfy blueprints delete dl-feeder
- #cfy blueprints delett dl-admin-ui
- #cfy blueprints delete des
diff --git a/docs/sections/services/dfc/index.rst b/docs/sections/services/dfc/index.rst
index ed9234ae..56056379 100644
--- a/docs/sections/services/dfc/index.rst
+++ b/docs/sections/services/dfc/index.rst
@@ -13,7 +13,6 @@ DataFile Collector(DFC)
./architecture.rst
./delivery.rst
./logging.rst
- ./installation.rst
./installation-helm.rst
./certificates.rst
./configuration.rst
diff --git a/docs/sections/services/dfc/installation.rst b/docs/sections/services/dfc/installation.rst
deleted file mode 100644
index 27d1ac5a..00000000
--- a/docs/sections/services/dfc/installation.rst
+++ /dev/null
@@ -1,85 +0,0 @@
-.. This work is licensed under a Creative Commons Attribution 4.0 International License.
-.. http://creativecommons.org/licenses/by/4.0
-.. _dfc-installation:
-
-
-Installation
-============
-
-An environment suitable for running docker containers is recommended.
-
-Using Cloudify deployment
--------------------------
-
-The first possibility is to use blueprints and cfy commands.
-
-Deployment Prerequisite/dependencies
-^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
-
-Make sure that **cfy** is installed and configured to work with the Cloudify deployment.
-
-Make sure the Message Router and Data Router are running.
-
-Deployment steps
-^^^^^^^^^^^^^^^^
-
-1. Execute bash on the cloudify manager kubernetes pod.
-
- ``kubectl -n onap exec -it <dev-dcaegen2-dcae-cloudify-manager> bash``
-
-2. Download the dfc `blueprint`_.
-
-.. _blueprint: https://gerrit.onap.org/r/gitweb?p=dcaegen2/platform/blueprints.git;a=blob;f=blueprints/reference_templates/k8s-datafile-collector.yaml-template;h=17d2aedec131154b4f5f84a08a099b0364b1e627;hb=HEAD
-
-3. Run Cloudify Install command to install dfc.
-
- ``cfy install <dfc-blueprint-path>``
-
-Sample output:
-
- ``cfy install k8s-datafile.yaml``
-
-Run '*cfy events list -e 37da3f5f-a06b-4ce8-84d3-8b64ccd81c33'* to retrieve the execution's events/logs.
-
-Validation
-^^^^^^^^^^
-
-curl <dcaegen2-dcae-healthcheck> and check if datafile-collector is in *'ready'* state.
-
-Standalone deployment of a container
-------------------------------------
-
-DFC is delivered as a docker container based on openjdk:8-jre-alpine. The
-host or VM that will run this container must have the docker application
-loaded and available to the userID that will be running the DFC container.
-
-Also required is a working DMAAP/MR and DMAAP/DR environment. DFC
-subscribes to DMAAP/MR fileReady event as JSON messages and publishes the downloaded files to the DMAAP/DR.
-
-Installation
-^^^^^^^^^^^^
-
-The following command will download the Frankfurt version of the datafile image from
-nexus and launch it in the container named "datafile":
-
- ``docker run -d -p 8100:8100 -p 8433:8433 nexus3.onap.org:10001/onap/org.onap.dcaegen2.collectors.datafile.datafile-app-server:1.3.0``
-
-For another version, it is possible to replace the tag '1.2.3' with any version that seems suitable (including latest).
-Available images are visible following this `link`_.
-
-.. _link: https://nexus3.onap.org/#browse/search=keyword%3D*datafile*
-
-Another option is to pull the image first, and then run the image's container with the image ID:
-
- ``docker pull nexus3.onap.org:10001/onap/org.onap.dcaegen2.collectors.datafile.datafile-app-server:latest``
-
- ``docker images | grep 'datafile'``
-
- ``docker run -d -p 8100:8100 -p 8433:8433 <image ID>``
-
-The grep command will display the images corresponding to DFC. There can be several due to remotely or locally built
-image, and also to different tags, i.e. different versions.
-
-Certifcates
-^^^^^^^^^^^
-.. _page: ./certifcates.rst
diff --git a/docs/sections/services/kpi-computation-ms/index.rst b/docs/sections/services/kpi-computation-ms/index.rst
index 5bbefa1f..e52e26db 100644
--- a/docs/sections/services/kpi-computation-ms/index.rst
+++ b/docs/sections/services/kpi-computation-ms/index.rst
@@ -23,7 +23,6 @@ Kpi Computation MS Installation Steps
.. toctree::
:maxdepth: 1
- ./installation.rst
./installation-helm.rst
Kpi Computation MS Configurations
diff --git a/docs/sections/services/kpi-computation-ms/installation.rst b/docs/sections/services/kpi-computation-ms/installation.rst
deleted file mode 100644
index eda1ae82..00000000
--- a/docs/sections/services/kpi-computation-ms/installation.rst
+++ /dev/null
@@ -1,95 +0,0 @@
-.. This work is licensed under a Creative Commons Attribution 4.0 International License.
-.. http://creativecommons.org/licenses/by/4.0
-.. _kpi-installation:
-
-
-Installation
-============
-
-Kpi Computation MS can be deployed using cloudify blueprint using bootstrap container of an existing DCAE deployment.
-
-Deployment Pre-requisites
-~~~~~~~~~~~~~~~~~~~~~~~~~
-- DCAE and DMaaP pods should be up and running.
-
-- PM mapper service should be running.
-
-- Make sure that cfy is installed and configured to work with the Cloudify deployment.
-
-
-Deployment steps
-~~~~~~~~~~~~~~~~
-Execute bash on the bootstrap Kubernetes pod.
-
- kubectl -n onap exec -it <dcaegen2-dcae-bootstrap> bash
-
-Validate Blueprint
-------------------
-Before the blueprints uploading to Cloudify manager, the blueprints shoule be validated first through the following command.
- .. code-block :: bash
-
- #cfy blueprint validate /bluerints/k8s-kpi-ms.yaml
-
-Upload the Blueprint to Cloudify Manager.
------------------------------------------
-After validating, we can start to proceed blueprints uploading.
- .. code-block :: bash
-
- #cfy blueprint upload -b kpi-ms /bluerints/k8s-kpi-ms.yaml
-
-Verify Uploaded Blueprints
---------------------------
-Using "cfy blueprint list" to verify your work.
- .. code-block :: bash
-
- #cfy blueprint list
-
-You can see the following returned message to show the blueprints have been correctly uploaded.
- .. image :: ./blueprint-list.png
-
-
-Verify Plugin Versions
-----------------------
-If the version of the plugin used is different, update the blueprint import to match.
- .. code-block :: bash
-
- #cfy plugins list
-
-Create Deployment
------------------
-Here we are going to create deployments for both feeder and admin UI.
- .. code-block :: bash
-
- #cfy deployments create -b kpi-ms kpi-ms
-
-Launch Service
---------------
-Next, we are going to launch the KPI-MS.
- .. code-block :: bash
-
- #cfy executions start -d kpi-ms install
-
-
-Verify the Deployment Result
-----------------------------
-The following command can be used to list the kpi-ms logs.
-
- .. code-block :: bash
-
- #kubectl logs <kpi-pod> -n onap
-
-The output should looks like.
- .. image :: ./kpi-log.PNG
-
-Uninstall
----------
-Uninstall running component and delete deployment
- .. code-block :: bash
-
- #cfy uninstall kpi-ms
-
-Delete Blueprint
-----------------
- .. code-block :: bash
-
- #cfy blueprints delete kpi-ms \ No newline at end of file
diff --git a/docs/sections/services/pm-mapper/index.rst b/docs/sections/services/pm-mapper/index.rst
index d8834b18..97712762 100644
--- a/docs/sections/services/pm-mapper/index.rst
+++ b/docs/sections/services/pm-mapper/index.rst
@@ -13,7 +13,6 @@
./architecture.rst
./delivery.rst
./logging.rst
- ./installation.rst
./installation-helm.rst
./configuration.rst
./troubleshooting.rst
diff --git a/docs/sections/services/pm-mapper/installation.rst b/docs/sections/services/pm-mapper/installation.rst
deleted file mode 100644
index cf7f972e..00000000
--- a/docs/sections/services/pm-mapper/installation.rst
+++ /dev/null
@@ -1,62 +0,0 @@
-.. This work is licensed under a Creative Commons Attribution 4.0 International License.
-.. http://creativecommons.org/licenses/by/4.0
-.. _pm-mapper-installation:
-
-Cloudify Installation
-=====================
-
-PM mapper is a microservice that will be configured and instantiated through Cloudify Manager, either through the user
-interface or the command line tool. During instantiation, the PM Mapper will fetch its configuration through the Config Binding Service. Steps to deploy using the CLI tool are shown below.
-
-Deployment Prerequisite/dependencies
-^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
-
- - DCAE and DMaaP pods should be up and running.
- - DMaaP Bus Controller post install jobs should have completed successfully (executed as part of an OOM install).
- - Make sure that cfy is installed and configured to work with the Cloudify deployment.
-
-Deployment steps
-^^^^^^^^^^^^^^^^
-
-Enter the Cloudify Manager kuberenetes pod
-
- - Download the PM Mapper blueprint onto the pod, this can be found in:
-
- https://git.onap.org/dcaegen2/services/pm-mapper/tree/dpo/blueprints/k8s-pm-mapper.yaml
-
- - Create inputs.yaml
-
- Configuration of the service consists of generating an inputs file (YAML) which will be used as part of the
- Cloudify install. The PM-Mapper blueprints were designed with sane defaults for the majority of the fields.
- Below you will find some examples of fields which can be configured, and some of the fields
- which must be configured. The full list of configurable parameters can be seen within the blueprint file under
- "inputs".
-
- .. csv-table::
- :widths: auto
- :delim: ;
- :header: Property , Sample Value , Description , Required
-
- client_id ; dcae@dcae.onap.org ; Information about the AAF user must be provided to enable publishing to authenticated topics. ; Yes
- client_password ; <dcae_password> ; This is the password for the given user e.g. The <dcae_password> is dcae@dcae.onap.org's password. ; Yes
- enable_http ; true ; By default, the PM-Mapper will only allow inbound queries over HTTPS. However, it is possible to configure it to enable HTTP also. ; No
- tag_version ; nexus3.onap.org:10001/onap/org.onap.dcaegen2.services.pm-mapper:1.0.1 ; The tag of the Docker image will be used when deploying the PM-Mapper. ; No
- pm-mapper-filter ; {"filters": [{"pmDefVsn":"targetVersion", "nfType":"targetNodeType", "vendor":"targetVendor","measTypes":["targetMeasType"]}]} ; The default behavior of the PM-Mapper is to map all measType in the received PM XML files, however, it's possible to provide filtering configuration which will reduce the VES event to the counters that the designer has expressed interest in. ; No
-
- Example inputs.yaml
-
- .. code-block:: yaml
-
- client_id: dcae@dcae.onap.org
- client_password: <dcae_password>
- enable_http: false
- tag_version: nexus3.onap.org:10001/onap/org.onap.dcaegen2.services.pm-mapper:latest
- pm-mapper-filter: {"filters": []}
-
-
-
- - Create deployment
-
- .. code-block:: bash
-
- cfy install --blueprint-id pm-mapper --deployment-id pm-mapper -i inputs.yaml k8s-pm-mapper.yaml
diff --git a/docs/sections/services/slice-analysis-ms/index.rst b/docs/sections/services/slice-analysis-ms/index.rst
index 53e5c643..2b7a2905 100644
--- a/docs/sections/services/slice-analysis-ms/index.rst
+++ b/docs/sections/services/slice-analysis-ms/index.rst
@@ -47,7 +47,6 @@ Slice Analysis MS Installation Steps
.. toctree::
:maxdepth: 1
- ./installation.rst
./installation-helm.rst
diff --git a/docs/sections/services/slice-analysis-ms/installation.rst b/docs/sections/services/slice-analysis-ms/installation.rst
deleted file mode 100644
index 531ff4ca..00000000
--- a/docs/sections/services/slice-analysis-ms/installation.rst
+++ /dev/null
@@ -1,80 +0,0 @@
-.. This work is licensed under a Creative Commons Attribution 4.0 International License.
-.. http://creativecommons.org/licenses/by/4.0
-.. _sliceanalysis-installation:
-
-
-Installation
-============
-
-Slice Analysis MS can be deployed by OOM. Make sure to enable "dcae-slice-analysis-ms" in oom/kubernetes/dcaegen2-services/values.yaml first.
-
-Deployment Pre-requisites
-~~~~~~~~~~~~~~~~~~~~~~~~~
-- DCAE and DMaaP pods should be up and running.
-
-- DMaaP Bus Controller PostInstalls job should have completed successfully (executed as part of an OOM install).
-
-- PM mapper service should be running.
-
-- Config DB service should be running.
-
-- Make sure that cfy is installed and configured to work with the Cloudify deployment.
-
-
-Deployment steps
-~~~~~~~~~~~~~~~~
-1. Execute bash on the bootstrap Kubernetes pod.
-
- kubectl -n onap exec -it <dcaegen2-dcae-bootstrap> bash
-
-2. Go to the /blueprints directory.
-
- Check that the tag_version in the slice-analysis-ms blueprint is correct for the release
- of ONAP that it is being installed on see Nexus link below for slice-analysis-ms for tag_versions.
- Nexus link: https://nexus3.onap.org/#browse/browse:docker.public:v2%2Fonap%2Forg.onap.dcaegen2.services.components.slice-analysis-ms%2Ftags
-
-3. Create an input file.
-
-4. Run the Cloudify install command to install the slice-analysis-ms with the blueprint and the newly created input file k8s-slice-input.yaml.
-
- $ cfy install k8s-slice-analysis-ms.yaml -i k8s-slice-input.yaml --blueprint-id sliceanalysisms
-
- Details of the sample output are available at:
- https://wiki.onap.org/pages/viewpage.action?pageId=92998809.
-
-5. To un-deploy
-
- $ cfy uninstall sliceanalysisms
-
-
-Application configurations
-~~~~~~~~~~~~~~~~~~~~~~~~~~
-+-------------------------------+------------------------------------------------+
-|Configuration | Description |
-+===============================+================================================+
-|samples | Minimum number of samples to be present |
-| | for analysis |
-+-------------------------------+------------------------------------------------+
-|minimumPercentageChange | Minimum percentage of configuration change |
-| | above which control loop should be triggered |
-+-------------------------------+------------------------------------------------+
-|initialDelaySeconds | Initial delay in milliseconds for the consumer |
-| | thread to start after the application startup |
-+-------------------------------+------------------------------------------------+
-|config_db | Host where the config DB application is running|
-+-------------------------------+------------------------------------------------+
-|performance_management_topicurl| Dmaap Topic URL to which PM data are posted |
-| | by network functions |
-+-------------------------------+------------------------------------------------+
-|dcae_cl_topic_url | Dmaap topic to which onset message to trigger |
-| | the control loop are posted |
-+-------------------------------+------------------------------------------------+
-|dcae_cl_response_topic_url | Dmaap topic URL to which Policy posts the |
-| | message after successful control loop trigger |
-+-------------------------------+------------------------------------------------+
-|intelligent_slicing_topic_url | Dmaap topic URL to which ML MS posts the |
-| | messages |
-+-------------------------------+------------------------------------------------+
-|dmaap_polling_interval | Dmaap Polling interval in milliseconds |
-+-------------------------------+------------------------------------------------+
-
diff --git a/docs/sections/services/son-handler/index.rst b/docs/sections/services/son-handler/index.rst
index 3d1f0f13..64e0ebc7 100644
--- a/docs/sections/services/son-handler/index.rst
+++ b/docs/sections/services/son-handler/index.rst
@@ -29,6 +29,5 @@ SON-Handler MS Installation Steps, Configurations, Troubleshooting Tips and Logg
.. toctree::
:maxdepth: 1
- ./installation.rst
./installation-helm.rst
./son_handler_troubleshooting.rst
diff --git a/docs/sections/services/son-handler/installation.rst b/docs/sections/services/son-handler/installation.rst
deleted file mode 100644
index a91ab196..00000000
--- a/docs/sections/services/son-handler/installation.rst
+++ /dev/null
@@ -1,156 +0,0 @@
-.. This work is licensed under a Creative Commons Attribution 4.0 International License.
-.. http://creativecommons.org/licenses/by/4.0
-.. _sonhandler-installation:
-
-
-Installation
-============
-
-SON handler microservice can be deployed using cloudify blueprint using bootstrap container of an existing DCAE deployment
-
-Deployment Prerequisites
-~~~~~~~~~~~~~~~~~~~~~~~~
-
-- SON-Handler service requires DMAAP and Policy components to be functional.
-
-- SON-hadler service requires the following dmaap topics to be present in the running DMAAP instance :
-
- 1.PCI-NOTIF-TOPIC-NGHBR-LIST-CHANGE-INFO
-
- 2.unauthenticated.SEC_FAULT_OUTPUT
-
- 3.unauthenticated.SEC_MEASUREMENT_OUTPUT
-
- 4.DCAE_CL_RSP
-
-- Policy model required for SON-handler service should be created and pushed to policy component. Steps for creating and pushing the policy model:
-
- 1.Login to PDP container and execute
-
- .. code-block:: bash
-
- kubectl exec -ti --namespace onap policy-pdp-0 bash
-
- 2.Create policy model
-
- .. code-block:: bash
-
- curl -k -v --silent -X PUT --header 'Content-Type: application/json' --header 'Accept: text/plain' --header 'ClientAuth: cHl0aG9uOnRlc3Q=' --header 'Authorization: Basic dGVzdHBkcDphbHBoYTEyMw==' --header 'Environment: TEST' -d '{
- "policyName": "com.PCIMS_CONFIG_POLICY",
- "configBody": "{ \"PCI_NEIGHBOR_CHANGE_CLUSTER_TIMEOUT_IN_SECS\":60, \"PCI_MODCONFIG_POLICY_NAME\":\"ControlLoop-vPCI-fb41f388-a5f2-11e8-98d0-529269fb1459\", \"PCI_OPTMIZATION_ALGO_CATEGORY_IN_OOF\":\"OOF-PCI-OPTIMIZATION\", \"PCI_SDNR_TARGET_NAME\":\"SDNR\" }",
- "policyType": "Config", "attributes" : { "matching" : { "key1" : "value1" } },
- "policyConfigType": "Base",
- "onapName": "DCAE",
- "configName": "PCIMS_CONFIG_POLICY",
- "configBodyType": "JSON" }' 'https://pdp:8081/pdp/api/createPolicy'
-
- 3.Push policy model
-
- .. code-block:: bash
-
- curl -k -v --silent -X PUT --header 'Content-Type: application/json' --header 'Accept: text/plain' --header 'ClientAuth: cHl0aG9uOnRlc3Q=' --header 'Authorization: Basic dGVzdHBkcDphbHBoYTEyMw==' --header 'Environment: TEST' -d '{
- "policyName": "com.PCIMS_CONFIG_POLICY",
- "policyType": "Base"}' 'https://pdp:8081/pdp/api/pushPolicy'
-
- 4.Verify config policy is present
-
- .. code-block:: bash
-
- curl -k -v --silent -X POST --header 'Content-Type: application/json' --header 'Accept: application/json' --header 'ClientAuth: cHl0aG9uOnRlc3Q=' --header 'Authorization: Basic dGVzdHBkcDphbHBoYTEyMw==' --header 'Environment: TEST' -d '{ "configName": "PCIMS_CONFIG_POLICY", "policyName": "com.Config_PCIMS_CONFIG_POLICY1*", "requestID":"e65cc45a-9efb-11e8-98d0-529269ffa459" }' 'https://pdp:8081/pdp/api/getConfig'
-
-Deployment steps
-~~~~~~~~~~~~~~~~
-1.Using DCAE Dashboard
- - Login to DCAE Dashboard (https://{k8s-nodeip}:30418/ccsdk-app/login_external.htm)
- - Go to Inventory --> Blueprints
- - Click on Deploy Action for son-handler blueprint
- - Override the value of 'tag_version' to 'nexus3.onap.org:10001/onap/org.onap.dcaegen2.services.son-handler:2.1.5' and click deploy.
- - Deployment logs can be viewed under Deployments section
-
-2.Using cloudify commands
-
-- Login to the bootstrap container
- kubectl exec -ti --namespace onap <bootstrap pod name> bash
-- The blueprint can be found under dcaegen2 blueprint repo and is part of bootstrap container. (https://gerrit.onap.org/r/dcaegen2/platform/blueprints)
-- Deploy the microservice into the cloudify using the following command
- cfy install -d sonhms -b sonhms <blueprint file path>
-- Deployment status of the microservice can be found from kubernetes pods status (MS will be deployed as a k8s pod in the kubernetes environment under the same namespace as the DCAE environment).
- kubectl get pods --namespace onap
-- To uninstall the microservice
- cfy uninstall sonhms
-- To delete the blueprint from the cloudify instance
- cfy blueprints delete sonhms
-
-
-Application Configurations
---------------------------
-+-------------------------------+------------------------------------------------+
-|Configuration | Description |
-+===============================+================================================+
-|Streams_subscribes | Dmaap topics that the MS will consume messages |
-+-------------------------------+------------------------------------------------+
-|Streams_publishes | Dmaap topics that the MS will publish messages |
-+-------------------------------+------------------------------------------------+
-|postgres.host | Host where the postgres database is running |
-+-------------------------------+------------------------------------------------+
-|postgres.port | Host where the postgres database is running |
-+-------------------------------+------------------------------------------------+
-|postgres.username | Postgres username |
-+-------------------------------+------------------------------------------------+
-|postgres.password | Postgres password |
-+-------------------------------+------------------------------------------------+
-|sonhandler.pollingInterval | Polling Interval for consuming dmaap messages |
-+-------------------------------+------------------------------------------------+
-|sonhandler.pollingTimeout | Polling timeout for consuming dmaap messages |
-+-------------------------------+------------------------------------------------+
-|sonhandler.numSolutions | Number for solutions for OOF optimization |
-+-------------------------------+------------------------------------------------+
-|sonhandler.minCollision | Minimum collision criteria to trigger OOF |
-+-------------------------------+------------------------------------------------+
-|sonhandler.minConfusion | Minimum confusion criteria to trigger OOF |
-+-------------------------------+------------------------------------------------+
-|sonhandler.maximumClusters | Maximum number of clusters MS can process |
-+-------------------------------+------------------------------------------------+
-|sonhandler.badThreshold | Bad threshold for Handover success rate |
-+-------------------------------+------------------------------------------------+
-|sonhandler.poorThreshold | Poor threshold for Handover success rate |
-+-------------------------------+------------------------------------------------+
-|sonhandler.namespace | Namespace where MS is going to be deployed |
-+-------------------------------+------------------------------------------------+
-|sonhandler.sourceId | Source ID of the Microservice (to OOF) |
-+-------------------------------+------------------------------------------------+
-|sonhandler.dmaap.server | Location of message routers |
-+-------------------------------+------------------------------------------------+
-|sonhandler.bufferTime | Buffer time for MS to wait for notifications |
-+-------------------------------+------------------------------------------------+
-|sonhandler.cg | DMAAP Consumer group for subscription |
-+-------------------------------+------------------------------------------------+
-|sonhandler.cid | DMAAP Consumer id for subcription |
-+-------------------------------+------------------------------------------------+
-|sonhandler.configDbService | Location of config DB (protocol, host & port) |
-+-------------------------------+------------------------------------------------+
-|sonhandler.oof.service | Location of OOF (protocol, host & port) |
-+-------------------------------+------------------------------------------------+
-|sonhandler.optimizers | Optimizer to trigger in OOF |
-+-------------------------------+------------------------------------------------+
-|sonhandler.poorCountThreshold | Threshold for number of times poorThreshold |
-| | can be recorded for the cell |
-+-------------------------------+------------------------------------------------+
-|sonhandler.badCountThreshold | Threshold for number of times badThreshold can |
-| | be recorded for the cell |
-+-------------------------------+------------------------------------------------+
-|sonhandler. | Timer for oof triggered count in minutes |
-|oofTriggerCountTimer | |
-+-------------------------------+------------------------------------------------+
-|sonhandler.policyRespTimer | Timer to wait for notification from policy |
-+-------------------------------+------------------------------------------------+
-|sonhandler. | Maximum number of negative acknowledgements |
-|policyNegativeAckThreshold | from policy for a given cell |
-+-------------------------------+------------------------------------------------+
-|sonhandler. | Time interval to trigger OOF with fixed pci |
-|policyFixedPciTimeInterval | cells |
-+-------------------------------+------------------------------------------------+
-|sonhandler.nfNamingCode | Parameter to filter FM and PM notifications |
-| | coming from ves |
-+-------------------------------+------------------------------------------------+
-
diff --git a/docs/sections/services/tcagen2-docker/installation.rst b/docs/sections/services/tcagen2-docker/installation.rst
deleted file mode 100644
index 450c7037..00000000
--- a/docs/sections/services/tcagen2-docker/installation.rst
+++ /dev/null
@@ -1,79 +0,0 @@
-.. This work is licensed under a Creative Commons Attribution 4.0 International License.
-.. http://creativecommons.org/licenses/by/4.0
-.. _tcagen2-installation:
-
-
-Installation
-============
-
-TCA-gen2 is a microservice that will be configured and instantiated through Cloudify Manager.TCA-gen2 will be deployed by DCAE deployment among the bootstrapped services. This is more to facilitate automated deployment of ONAP regression test cases required services. During instantiation, the TCA-gen2 will fetch its configuration through the Config Binding Service. Steps to deploy using the CLI tool are shown below.
-
-Deployment Prerequisite/dependencies
-^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
-
- - DCAE and DMaaP pods should be up and running.
- - MongoDB should be up and running
- - Make sure that cfy is installed and configured to work with the Cloudify deployment.
-
-Deployment steps
-^^^^^^^^^^^^^^^^
-
-Following are steps if manual deployment/undeployment is required. Steps to deploy are below
-
-
-Enter the Cloudify Manager kuberenetes pod
-
- .. note::
- For doing this, follow the below steps
-
- * First get the bootstrap pod name by running run this: kubectl get pods -n onap | grep bootstrap
- * Then login to bootstrap pod by running this: kubectl exec -it <bootstrap pod> bash -n onap
-
- - Tca-gen2 blueprint directory (/blueprints/k8s-tcagen2.yaml). The blueprint is also maintained in gerrit and can be downloaded from https://git.onap.org/dcaegen2/platform/blueprints/tree/blueprints/k8s-tcagen2.yaml
-
- - Create input file required for deployment
-
- Configuration of the service consists of generating an inputs file (YAML) which will be used as part of the
- Cloudify install. The tca-gen2 blueprints was designed with known defaults for the majority of the fields.
-
- Below you will find examples of fields which can be configured, and some of the fields
- which must be configured. An input file is loaded into bootstrap container (/inputs/k8s-tcagen2-inputs.yaml).
-
-
- .. csv-table::
- :widths: auto
- :delim: ;
- :header: Property , Sample Value , Description , Required
-
- tca_handle_in_subscribe_url ; http://message-router:3904/events/unauthenticated.TCAGEN2_OUTPUT/; DMaap topic to publish CL event output ; No
- tca_handle_in_subscribe_url ; http://message-router:3904/events/unauthenticated.VES_MEASUREMENT_OUTPUT/; DMaap topic to subscribe VES measurement feeds ; No
- tag_version ; nexus3.onap.org:10001/onap/org.onap.dcaegen2.analytics.tca-gen2.dcae-analytics-tca-web:1.0.1 ; The tag of the Docker image will be used when deploying the tca-gen2. ; No
-
- Example inputs.yaml
-
- .. code-block:: yaml
-
- tag_version: nexus3.onap.org:10001/onap/org.onap.dcaegen2.analytics.tca-gen2.dcae-analytics-tca-web:1.0.1
- tca_handle_in_subscribe_url: "http://message-router:3904/events/unauthenticated.VES_MEASUREMENT_OUTPUT/"
- tca_handle_out_publish_url: "http://message-router:3904/events/unauthenticated.TCAGEN2_OUTPUT/"
-
-
- - Create deployment
-
- .. code-block:: bash
-
- cfy install --blueprint-id tcagen2 --deployment-id tcagen2 -i /inputs/k8s-tcagen2-inputs.yaml /blueprints/k8s-tcagen2.yaml
-
-
-
-To undeploy TCA-gen2, steps are shown below
-
-- Uninstall running TCA-gen2 and delete deployment
- .. code-block:: bash
-
- cfy uninstall tcagen2
-- Delete blueprint
- .. code-block:: bash
-
- cfy blueprints delete tcagen2
-
diff --git a/docs/sections/services/ves-http/index.rst b/docs/sections/services/ves-http/index.rst
index 7f444247..e0897a2d 100644
--- a/docs/sections/services/ves-http/index.rst
+++ b/docs/sections/services/ves-http/index.rst
@@ -22,7 +22,6 @@ VES Collector (HTTP) overview and functions
./architecture.rst
./configuration.rst
./delivery.rst
- ./installation.rst
./installation-helm.rst
./tls-authentication.rst
./stnd-defined-validation.rst
diff --git a/docs/sections/services/ves-http/installation.rst b/docs/sections/services/ves-http/installation.rst
deleted file mode 100644
index 156427af..00000000
--- a/docs/sections/services/ves-http/installation.rst
+++ /dev/null
@@ -1,157 +0,0 @@
-.. This work is licensed under a Creative Commons Attribution 4.0 International License.
-.. http://creativecommons.org/licenses/by/4.0
-.. _ves-installation:
-
-
-VES Collector Cloudify Installation
-===================================
-
-VESCollector is installed via cloudify blueprint by DCAE bootstrap process on typical ONAP installation.
-As the service is containerized, it can be started on stand-alone mode also.
-
-
-To run VES Collector container on standalone mode, following parameters are required
-
- ``docker run -d -p 8080:8080/tcp -p 8443:8443/tcp -P -e DMAAPHOST='10.0.11.1' nexus.onap.org:10001/onap/org.onap.dcaegen2.collectors.ves.vescollector:1.7.9``
-
-
-DMAAPHOST is required for standalone; for normal platform installed instance the publish URL are obtained from Consul. Below parameters are exposed for DCAE platform (cloudify) deployed instance
-
-
-- COLLECTOR_IP
-- DMAAPHOST - should contain an address to DMaaP, so that event publishing can work
-- CONFIG_BINDING_SERVICE - should be a name of CBS
-- CONFIG_BINDING_SERVICE_SERVICE_PORT - should be a http port of CBS
-- HOSTNAME - should be a name of VESCollector application as it is registered in CBS catalog
-
-These parameters can be configured either by passing command line option during `docker run` call or by specifying environment variables named after command line option name
-
-
-Authentication Support
-----------------------
-
-VES Collector support following authentication types
-
- * *auth.method=noAuth* default option - no security (http)
- * *auth.method=certBasicAuth* is used to enable mutual TLS authentication or/and basic HTTPs authentication
-
-The blueprint is same for both deployments - based on the input configuration, VESCollector can be set for required authentication type.
-Default ONAP deployed VESCollector is configured for "certBasicAuth".
-
-If VESCollector instance need to be deployed with authentication disabled, follow below setup
-
-
-- Execute into Bootstrap POD using kubectl command
- .. note::
- For doing this, follow the below steps
-
- * First get the bootstrap pod name by running this: kubectl get pods -n onap | grep bootstrap
- * Then login to bootstrap pod by running this: kubectl exec -it <bootstrap pod> -n onap -- bash
-
-- VES blueprint is available under /blueprints directory ``k8s-ves.yaml``. A corresponding input file is also pre-loaded into bootstrap pod under /inputs/k8s-ves-inputs.yaml
-
-- Deploy blueprint
- .. code-block:: bash
-
- cfy install -b ves-http -d ves-http -i /inputs/k8s-ves-inputs.yaml /blueprints/k8s-ves.yaml
-
-To undeploy ves-http, steps are noted below
-
-- Uninstall running ves-http and delete deployment
- .. code-block:: bash
-
- cfy uninstall ves-http
-
-The deployment uninstall will also delete the blueprint. In some case you might notice 400 error reported indicating active deployment exist such as below
-** An error occurred on the server: 400: Can't delete blueprint ves-http - There exist deployments for this blueprint; Deployments ids: ves-http**
-
-In this case blueprint can be deleted explicitly using this command.
-
- .. code-block:: bash
-
- cfy blueprint delete ves-http
-
-External repo schema files from OOM connection to VES collector
--------------------------------------------------------------------
-In order to not use schema files bundled in VES Collector image but schema files defined in `OOM <https://gerrit.onap.org/r/gitweb?p=oom.git;a=tree;f=kubernetes/dcaegen2/resources/external>`_ repository and installed with dcaegen2 module, follow below setup.
-
-- Execute into Bootstrap POD using kubectl command
- .. note::
- For doing this, follow the below steps
-
- * First get the bootstrap pod name by running this: kubectl get pods -n onap | grep bootstrap
- * Then login to bootstrap pod by running this: kubectl exec -it <bootstrap pod> -n onap -- bash
-
-- VES blueprint is available under /blueprints directory ``k8s-ves.yaml``. A corresponding input file is also pre-loaded into bootstrap pod under /inputs/k8s-ves-inputs.yaml
-
-- Edit ``k8s-ves.yaml`` blueprint by adding section below ``docker_config:`` tag:
- .. code-block:: bash
-
- volumes:
- - container:
- bind: /opt/app/VESCollector/etc/externalRepo/3gpp/rep/sa5/MnS/blob/SA88-Rel16/OpenAPI
- config_volume:
- name: dcae-external-repo-configmap-sa88-rel16
- - container:
- bind: /opt/app/VESCollector/etc/externalRepo/
- config_volume:
- name: dcae-external-repo-configmap-schema-map
-
-- After all ``docker_config:`` section in blueprint should looks like:
- .. code-block:: bash
-
- docker_config:
- healthcheck:
- endpoint: /healthcheck
- interval: 15s
- timeout: 1s
- type: http
- volumes:
- - container:
- bind: /opt/app/VESCollector/etc/externalRepo/3gpp/rep/sa5/MnS/blob/SA88-Rel16/OpenAPI
- config_volume:
- name: dcae-external-repo-configmap-sa88-rel16
- - container:
- bind: /opt/app/VESCollector/etc/externalRepo/
- config_volume:
- name: dcae-external-repo-configmap-schema-map
-
-.. note::
-
- To undeploy ves-http if it is deployed, steps are noted below
-
- Uninstall running ves-http and delete deployment
- .. code-block:: bash
-
- cfy uninstall ves-http
-
- The deployment uninstall will also delete the blueprint. In some case you might notice 400 error reported indicating active deployment exist such as below
- ** An error occurred on the server: 400: Can't delete blueprint ves-http - There exist deployments for this blueprint; Deployments ids: ves-http**
-
- In this case blueprint can be deleted explicitly using this command.
-
- .. code-block:: bash
-
- cfy blueprint delete ves-http
-
-To deploy modified ves-http, steps are noted below
-
-- Load blueprint:
- .. code-block:: bash
-
- cfy blueprints upload -b ves-http /blueprints/k8s-ves.yaml
-
-- Deploy blueprint
- .. code-block:: bash
-
- cfy install -b ves-http -d ves-http -i /inputs/k8s-ves-inputs.yaml /blueprints/k8s-ves.yaml
-
-Using external TLS certificates obtained using CMP v2 protocol
---------------------------------------------------------------
-
-In order to use the X.509 certificates obtained from the CMP v2 server (so called "operator`s certificates"), refer to the following description:
-
-.. toctree::
- :maxdepth: 1
-
- Enabling TLS with external x.509 certificates <../../tls_enablement>
diff --git a/docs/sections/services/ves-hv/index.rst b/docs/sections/services/ves-hv/index.rst
index ec0623ba..f6209cff 100644
--- a/docs/sections/services/ves-hv/index.rst
+++ b/docs/sections/services/ves-hv/index.rst
@@ -31,7 +31,6 @@ High Volume VES Collector overview and functions
design
repositories
deployment
- installation
installation-helm
run-time-configuration
HV-VES Offered APIs <../../apis/ves-hv/index>