From edaf8b1d69fa97862bd246dc61c1e4e77a01b5be Mon Sep 17 00:00:00 2001 From: Vijay VK Date: Thu, 23 May 2019 22:02:50 +0100 Subject: Fix doc issues Change-Id: I711a40a64a1d47b50f0923ba67b02ad121ace5b4 Signed-off-by: VENKATESH KUMAR Issue-ID: DCAEGEN2-1505 --- docs/sections/apis/SDK.rst | 4 +- docs/sections/architecture.rst | 38 +++-- docs/sections/consumedapis.rst | 15 +- docs/sections/healthcheck.rst | 2 +- docs/sections/installation_MS_ondemand.rst | 2 +- docs/sections/logging.rst | 27 ++-- .../services/bbs-event-processor/installation.rst | 1 + docs/sections/services/mapper/installation.rst | 6 +- docs/sections/services/restconf/installation.rst | 17 ++- .../sections/services/son-handler/installation.rst | 170 +++++++++++---------- .../services/tca-cdap/development_info.rst | 18 ++- docs/sections/services/tca-cdap/index.rst | 8 +- docs/sections/services/ves-http/architecture.rst | 8 +- docs/sections/services/ves-http/installation.rst | 4 + .../services/ves-http/tls-authentication.rst | 4 +- 15 files changed, 170 insertions(+), 154 deletions(-) (limited to 'docs/sections') diff --git a/docs/sections/apis/SDK.rst b/docs/sections/apis/SDK.rst index 567b8a7f..b118ad92 100644 --- a/docs/sections/apis/SDK.rst +++ b/docs/sections/apis/SDK.rst @@ -10,14 +10,14 @@ DCAE SDK .. Overview -======== +-------- DCAE SDK contains utilities and clients which may be used for fetching configuration from CBS, consuming messages from DMaaP, interacting with A&AI, etc. SDK is written in Java. Introduction -============ +------------ SDK Maven dependencies (modules). diff --git a/docs/sections/architecture.rst b/docs/sections/architecture.rst index 92ed40f6..9daa7070 100644 --- a/docs/sections/architecture.rst +++ b/docs/sections/architecture.rst @@ -5,15 +5,30 @@ Architecture ============ -Data Collection Analytics and Events (DCAE) is the primary data collection and analysis system of ONAP. DCAE architecture comprises of DCAE Platform and DCAE Service components making DCAE flexible, elastic, and expansive enough for supporting the potentially infinite number of ways of constructing intelligent and automated control loops on distributed and heterogeneous infrastructure. - -DCAE Platform supports the functions to deploy, host and perform LCM applications of Service components. DCAE Platform components enable model driven deployment of service components and middleware infrastructures that service components depend upon, such as special storage and computation platforms. When triggered by an invocation call (such as CLAMP or via DCAE Dashboard), DCAE Platform follows the TOSCA model of the control loop that is specified by the triggering call, interacts with the underlying networking and computing infrastructure such as OpenSatck installations and Kubernetes clusters to deploy and configure the virtual apparatus (i.e. the collectors, the analytics, and auxiliary microservices) that are needed to form the control loop, at locations that requested. DCAE Platform also provisions DMaaP topics and manages the distribution scopes of the topics following the prescription of the control loop model by interacting with controlling function of DMaaP. - -DCAE Service components are the functional entities that realize the collection and analytics needs of ONAP control loops. They include the collectors for various data collection needs, event processors for data standardization, analytics that assess collected data, and various auxiliary microservices that assist data collection and analytics, and support other ONAP functions. Service components and DMaaP buses form the "data plane" for DCAE, where DCAE collected data is transported among different DCAE service components. - -DCAE service components configuration are stored under Key-Value store service, embodied by a Consul cluster. During deployment, DCAE platform (via Cloudify plugin) stores service component configuration under Consul for each deployment/instance (identified by ServiceComponentName). All DCAE components during startup will acess these configuration through ConfigBindingService api's to load deployment configuration and watch for any subsequent update. - -DCAE components use Consul's distributed K-V store service to distribute and manage component configurations where each key is based on the unique identity of a DCAE component, and the value is the configuration for the corresponding component. DCAE platform creates and updates the K-V pairs based on information provided as part of the control loop blueprint, or received from other ONAP components such as Policy Framework and CLAMP. Either through periodically polling or proactive pushing, the DCAE components get the configuration updates in realtime and apply the configuration updates. DCAE Platform also offers dynamic template resolution for configuration parameters that are dynamic and only known by the DCAE platform, such as dynamically provisioned DMaaP topics. This approach standardizes component deployment and configuration management for DCAE service components in multi-site deployment. +Data Collection Analytics and Events (DCAE) is the primary data collection and analysis system of ONAP. DCAE architecture comprises of DCAE Platform and +DCAE Service components making DCAE flexible, elastic, and expansive enough for supporting the potentially infinite number of ways of constructing intelligent +and automated control loops on distributed and heterogeneous infrastructure. + +DCAE Platform supports the functions to deploy, host and perform LCM applications of Service components. DCAE Platform components enable model driven deployment of +service components and middleware infrastructures that service components depend upon, such as special storage and computation platforms. When triggered by an +invocation call (such as CLAMP or via DCAE Dashboard), DCAE Platform follows the TOSCA model of the control loop that is specified by the triggering call, +interacts with the underlying networking and computing infrastructure such as OpenSatck installations and Kubernetes clusters to deploy and configure the virtual +apparatus (i.e. the collectors, the analytics, and auxiliary microservices) that are needed to form the control loop, at locations that requested. +DCAE Platform also provisions DMaaP topics and manages the distribution scopes of the topics following the prescription of the control loop model by interacting +with controlling function of DMaaP. + +DCAE Service components are the functional entities that realize the collection and analytics needs of ONAP control loops. They include the collectors for various +data collection needs, event processors for data standardization, analytics that assess collected data, and various auxiliary microservices that assist data +collection and analytics, and support other ONAP functions. Service components and DMaaP buses form the "data plane" for DCAE, where DCAE collected data is +transported among different DCAE service components. + +DCAE use Consul's distributed K-V store service to manage component configurations where each key is based on the unique identity of a +DCAE component (identified by ServiceComponentName), and the value is the configuration for the corresponding component. The K-V store for each service components is +created during deployment. DCAE platform creates and updates the K-V pairs based on information provided as part of the control loop blueprint deployment, or through + a notification/trigger received from other ONAP components such as Policy Framework and CLAMP. Either through periodically polling or proactive pushing, the DCAE + components get the configuration updates in realtime and apply the configuration updates. DCAE Platform also offers dynamic template resolution for configuration +parameters that are dynamic and only known by the DCAE platform, such as dynamically provisioned DMaaP topics. This approach standardizes component deployment and +configuration management for DCAE service components in multi-site deployment. DCAE R4 Components @@ -24,7 +39,7 @@ The following list displays the details of what are included in ONAP DCAE R4. A - DCAE Platform - Core Platform - Cloudify Manager: TOSCA model executor. Materializes TOSCA models of control loop, or Blueprints, into properly configured and managed virtual DCAE functional components. - - Plugins (K8s, Dmaap, Policy, Clamp, Pg) + - Plugins (K8S, Dmaap, Policy, Clamp, Postgres) - Extended Platform - Configuration Binding Service: Agent for service component configuration fetching; providing configuration parameter resolution. - Deployment Handler: API for triggering control loop deployment based on control loop's TOSCA model. @@ -56,7 +71,6 @@ The following list displays the details of what are included in ONAP DCAE R4. A - BBS-EventProcessor Service - The figure below shows the DCAE R4 architecture and how the components work with each other. The components on the right constitute the Platform/controller components which are statically deployed. The components on the right represent the services which can be both deployed statically or dynamically (via CLAMP) .. image:: images/R4_architecture_diagram.png @@ -69,7 +83,7 @@ Because DCAE service components are deployed on-demand following the control loo For R4, ONAP supports deployment via OOM Helm Chart method and Heat deployment support is discontinued. DCAE Platform components are deployed via Helm charts - this includes Cloudify Manager, ConfigBinding service, ServiceChange Handler, Policy Handler and Inventory. Once DCAE platform components are up and running, rest of DCAE service components required for ONAP flow are deployed via bootstrap POD, which invokes Cloudify Manager API with Blueprints for various DCAE components that are needed for the built-in collections and control loops flow support. -To keep the ONAP footprint minimal, only minmial set MS (required for ONAP Integration usecases) are deployed via bootstrap pod. Rest of service blueprints are available for operator to deploy on-demand as required. +To keep the ONAP footprint minimal, only minimal set of MS (required for ONAP Integration usecases) are deployed via bootstrap pod. Rest of service blueprints are available for operator to deploy on-demand as required. The PNDA platform service is an optional component that can be installed when using the OOM Helm Chart installation method on Openstack based Kubernetes infrastructure. diff --git a/docs/sections/consumedapis.rst b/docs/sections/consumedapis.rst index 17850c7e..637ac7b2 100644 --- a/docs/sections/consumedapis.rst +++ b/docs/sections/consumedapis.rst @@ -15,13 +15,8 @@ Consumed APIs DCAEGEN2 Components making following API calls into other ONAP components. -DMaaP Message Router -* https://docs.onap.org/en/latest/submodules/dmaap/messagerouter/messageservice.git/docs/offeredapis/offeredapis.html -DMaaP Data Router -* https://docs.onap.org/en/latest/submodules/dmaap/datarouter.git/docs/offeredapis.html -Policy -* https://docs.onap.org/en/latest/submodules/policy/engine.git/docs/platform/offeredapis.html -SDC -* https://docs.onap.org/en/latest/submodules/sdc.git/docs/offeredapis.html -A&AI -* https://docs.onap.org/en/latest/submodules/aai/aai-common.git/docs/platform/offeredapis.html \ No newline at end of file +* `DMaaP Message Router `_ +* `DMaaP Data Router `_ +* `Policy `_ +* `SDC `_ +* `A&AI `_ \ No newline at end of file diff --git a/docs/sections/healthcheck.rst b/docs/sections/healthcheck.rst index 9fec3a80..dea0807b 100644 --- a/docs/sections/healthcheck.rst +++ b/docs/sections/healthcheck.rst @@ -26,7 +26,7 @@ blueprints after the initial DCAE installation. The healthcheck service is exposed as a Kubernetes ClusterIP Service named `dcae-healthcheck`. The service can be queried for status as shown below. -.. code-block::json +.. code-block:: json $ curl dcae-healthcheck { diff --git a/docs/sections/installation_MS_ondemand.rst b/docs/sections/installation_MS_ondemand.rst index ee5e639e..ce32764e 100644 --- a/docs/sections/installation_MS_ondemand.rst +++ b/docs/sections/installation_MS_ondemand.rst @@ -25,7 +25,7 @@ Following additional MS are available for on-demand deployment as necessary for Mapper MS Installation <./services/mapper/installation> DFC MS Installation <./services/dfc/installation> - Heartbeat MS Installation <./services/heartbeat/installation> + Heartbeat MS Installation <./services/heartbeat-ms/installation> PM-Mapper MS Installation <./services/pm-mapper/installation> BBS EventProcessor MS Installation <./services/bbs-event-processor/installation> Son-Handler MS Installation <./services/son-handler/installation> diff --git a/docs/sections/logging.rst b/docs/sections/logging.rst index a6da9518..efd60cca 100644 --- a/docs/sections/logging.rst +++ b/docs/sections/logging.rst @@ -4,27 +4,18 @@ Logging ======= -DCAE logging is available in several levels. - -Platform VM Logging -------------------- -1. DCAE bootstrap VM: - * /var/log directory containing various system logs including cloud init logs. - * /tmp/dcae2_install.log file provided installation logs. - * **docker logs** command for DCAE bootstrap container logs. -2. Cloudify Manager VM: - * /var/log directory containing various system logs including cloud init logs. - * Cloudify Manager GUI provides viewing access to Cloudify's operation logs. -3. Consul cluster: - * /var/log directory containing various system logs including cloud init logs. - * Consul GUI provides viewing access to Consul registered platform and service components healthcheck logs. -4. Docker hosts - * /var/log directory containing various system logs including cloud init logs. - * **docker logs** command for Docker container logs. +DCAE logging is available in several levels; most DCAE Components are complaint with EELF logging standard and generates debug, audit, metric logging. + + +Platform Components Logging +--------------------------- +As all the platform components are containered and deployed under K8S as pod; corresponding log information can be accessed using ``kubectl get logs -n onap `` + +More detailed audit/debug logs can be found within the pod. Component Logging ----------------- -In general the logs of service component can be accessed under the /opt/log directory of the container, either the Docker container or the VM. Their deployment logs can be found at the deployment engine and deployment location, e.g. Cloudify Manager, CDAP, and Docker hosts. +Please refer to individual service component webpage for more information. In general the logs of service component can be accessed using ``kubectl get logs -n onap `` diff --git a/docs/sections/services/bbs-event-processor/installation.rst b/docs/sections/services/bbs-event-processor/installation.rst index 242f9f9b..92d2e83f 100644 --- a/docs/sections/services/bbs-event-processor/installation.rst +++ b/docs/sections/services/bbs-event-processor/installation.rst @@ -40,6 +40,7 @@ The following docker-compose-yaml file shows a default configuration. The file c LOGGING_LEVEL_ORG_ONAP_BBS: TRACE For Dublin release, it will be a DCAE component that can dynamically be deployed via Cloudify blueprint installation. + Steps to deploy are shown below - Transfer blueprint component file in DCAE bootstrap POD under /blueprints directory. Blueprint can be found in diff --git a/docs/sections/services/mapper/installation.rst b/docs/sections/services/mapper/installation.rst index 7559f4a5..a83229ce 100644 --- a/docs/sections/services/mapper/installation.rst +++ b/docs/sections/services/mapper/installation.rst @@ -15,7 +15,7 @@ VES-Mapper can be deployed individually though it will throw errors if it can't **Blueprint/model/image** VES-Mapper blueprint is available @ -https://gerrit.onap.org/r/gitweb?p=dcaegen2/services/mapper.git;a=blob;f=UniversalVesAdapter/dpo/blueprints/k8s-vesmapper.yaml-template.yaml;h=refs/head/master +https://git.onap.org/dcaegen2/services/mapper/tree/UniversalVesAdapter/dpo/blueprints/k8s-vesmapper.yaml-template.yaml?h=dublin VES-Mapper docker image is available in Nexus repo @ `nexus3.onap.org:10001/onap/org.onap.dcaegen2.services.mapper.vesadapter.universalvesadaptor:1.0.0-SNAPSHOT `_ @@ -33,9 +33,7 @@ VES-Mapper docker image is available in Nexus repo @ `nexus3.onap.org:10001/onap update these ``urls`` as per your DMaaP configurations in the blueprint. -*b. Verify the Smooks mapping files in the blueprint as per the usecase. Blueprint contains default mapping file for each supported collector ( SNMP Collector and RESTConf collector currently) which may serve the purpose for the usecase. - -``mapping-files`` in ``collectors`` contains the contents of the mapping file. +*b. Verify the Smooks mapping configuration in the blueprint as per the usecase. Blueprint contains default mapping for each supported collector ( SNMP Collector and RESTConf collector currently) which may serve the purpose for the usecase. The ``mapping-files`` in ``collectors`` contains the contents of the mapping file. *c. Upload the blueprint in the DCAE's Cloudify instance* diff --git a/docs/sections/services/restconf/installation.rst b/docs/sections/services/restconf/installation.rst index 2f369873..d475ef7a 100644 --- a/docs/sections/services/restconf/installation.rst +++ b/docs/sections/services/restconf/installation.rst @@ -1,18 +1,21 @@ Installation ============ -Sample docker run command could be - -.. code-block:: bash - docker run onap/org.onap.dcaegen2.collectors.restconfcollector +Standalone docker run command + .. code-block:: bash + + docker run onap/org.onap.dcaegen2.collectors.restconfcollector + +For Dublin release, RESTConf collector will be a DCAE component that can dynamically be deployed via Cloudify blueprint installation. + -For Dublin release, it will be a DCAE component that can dynamically be deployed via Cloudify blueprint installation. Steps to deploy are shown below -- Transfer blueprint component file in DCAE bootstrap POD under /blueprints directory. Blueprint can be found in +- Enter the Bootstrap POD using kubectl - https://git.onap.org/dcaegen2/collectors/restconf/tree/dpo/blueprints/k8s-rcc-policy.yaml-template?h=master +- Transfer blueprint component file in DCAE bootstrap POD under /blueprints directory. Blueprint can be found in + https://git.onap.org/dcaegen2/collectors/restconf/tree/dpo/blueprints/k8s-rcc-policy.yaml-template?h=dublin -- Enter the Bootstrap POD - Validate blueprint .. code-block:: bash diff --git a/docs/sections/services/son-handler/installation.rst b/docs/sections/services/son-handler/installation.rst index f529bc4a..edfcb39f 100644 --- a/docs/sections/services/son-handler/installation.rst +++ b/docs/sections/services/son-handler/installation.rst @@ -10,38 +10,48 @@ Deployment Prerequisites - SON-hadler service requires the following dmaap topics to be present in the running DMAAP instance : - 1.PCI-NOTIF-TOPIC-NGHBR-LIST-CHANGE-INFO - - 2.unauthenticated.SEC_FAULT_OUTPUT - - 3.unauthenticated.SEC_MEASUREMENT_OUTPUT - - 4.DCAE_CL_RSP - -- Policy model required for SON-handler service should be created and pushed to policy component.Steps for creating and pushing the policy model: - 1.Login to PDP container and execute - kubectl exec -ti --namespace onap policy-pdp-0 bash - 2.Create policy model - curl -k -v --silent -X PUT --header 'Content-Type: application/json' --header 'Accept: text/plain' --header 'ClientAuth: cHl0aG9uOnRlc3Q=' --header 'Authorization: Basic dGVzdHBkcDphbHBoYTEyMw==' --header 'Environment: TEST' -d '{ - "policyName": "com.PCIMS_CONFIG_POLICY", - "configBody": "{ \"PCI_NEIGHBOR_CHANGE_CLUSTER_TIMEOUT_IN_SECS\":60, \"PCI_MODCONFIG_POLICY_NAME\":\"ControlLoop-vPCI-fb41f388-a5f2-11e8-98d0-529269fb1459\", \"PCI_OPTMIZATION_ALGO_CATEGORY_IN_OOF\":\"OOF-PCI-OPTIMIZATION\", \"PCI_SDNR_TARGET_NAME\":\"SDNR\" }", - "policyType": "Config", - "attributes" : { "matching" : { "key1" : "value1" } }, - "policyConfigType": "Base", - "onapName": "DCAE", - "configName": "PCIMS_CONFIG_POLICY", - "configBodyType": "JSON" -}' 'https://pdp:8081/pdp/api/createPolicy' - - 3.Push policy model - curl -k -v --silent -X PUT --header 'Content-Type: application/json' --header 'Accept: text/plain' --header 'ClientAuth: cHl0aG9uOnRlc3Q=' --header 'Authorization: Basic dGVzdHBkcDphbHBoYTEyMw==' --header 'Environment: TEST' -d '{ - "policyName": "com.PCIMS_CONFIG_POLICY", - "policyType": "Base" -}' 'https://pdp:8081/pdp/api/pushPolicy' - - 4.Verify config policy is present - - curl -k -v --silent -X POST --header 'Content-Type: application/json' --header 'Accept: application/json' --header 'ClientAuth: cHl0aG9uOnRlc3Q=' --header 'Authorization: Basic dGVzdHBkcDphbHBoYTEyMw==' --header 'Environment: TEST' -d '{ "configName": "PCIMS_CONFIG_POLICY", "policyName": "com.Config_PCIMS_CONFIG_POLICY1*", "requestID":"e65cc45a-9efb-11e8-98d0-529269ffa459" }' 'https://pdp:8081/pdp/api/getConfig' + 1.PCI-NOTIF-TOPIC-NGHBR-LIST-CHANGE-INFO + + 2.unauthenticated.SEC_FAULT_OUTPUT + + 3.unauthenticated.SEC_MEASUREMENT_OUTPUT + + 4.DCAE_CL_RSP + +- Policy model required for SON-handler service should be created and pushed to policy component. Steps for creating and pushing the policy model: + + 1.Login to PDP container and execute + + .. code-block:: bash + + kubectl exec -ti --namespace onap policy-pdp-0 bash + + 2.Create policy model + + .. code-block:: bash + + curl -k -v --silent -X PUT --header 'Content-Type: application/json' --header 'Accept: text/plain' --header 'ClientAuth: cHl0aG9uOnRlc3Q=' --header 'Authorization: Basic dGVzdHBkcDphbHBoYTEyMw==' --header 'Environment: TEST' -d '{ + "policyName": "com.PCIMS_CONFIG_POLICY", + "configBody": "{ \"PCI_NEIGHBOR_CHANGE_CLUSTER_TIMEOUT_IN_SECS\":60, \"PCI_MODCONFIG_POLICY_NAME\":\"ControlLoop-vPCI-fb41f388-a5f2-11e8-98d0-529269fb1459\", \"PCI_OPTMIZATION_ALGO_CATEGORY_IN_OOF\":\"OOF-PCI-OPTIMIZATION\", \"PCI_SDNR_TARGET_NAME\":\"SDNR\" }", + "policyType": "Config", "attributes" : { "matching" : { "key1" : "value1" } }, + "policyConfigType": "Base", + "onapName": "DCAE", + "configName": "PCIMS_CONFIG_POLICY", + "configBodyType": "JSON" }' 'https://pdp:8081/pdp/api/createPolicy' + + 3.Push policy model + + .. code-block:: bash + + curl -k -v --silent -X PUT --header 'Content-Type: application/json' --header 'Accept: text/plain' --header 'ClientAuth: cHl0aG9uOnRlc3Q=' --header 'Authorization: Basic dGVzdHBkcDphbHBoYTEyMw==' --header 'Environment: TEST' -d '{ + "policyName": "com.PCIMS_CONFIG_POLICY", + "policyType": "Base"}' 'https://pdp:8081/pdp/api/pushPolicy' + + 4.Verify config policy is present + + .. code-block:: bash + + curl -k -v --silent -X POST --header 'Content-Type: application/json' --header 'Accept: application/json' --header 'ClientAuth: cHl0aG9uOnRlc3Q=' --header 'Authorization: Basic dGVzdHBkcDphbHBoYTEyMw==' --header 'Environment: TEST' -d '{ "configName": "PCIMS_CONFIG_POLICY", "policyName": "com.Config_PCIMS_CONFIG_POLICY1*", "requestID":"e65cc45a-9efb-11e8-98d0-529269ffa459" }' 'https://pdp:8081/pdp/api/getConfig' Deployment steps ~~~~~~~~~~~~~~~~ @@ -60,51 +70,53 @@ Deployment steps Application Configurations -------------------------- - -Streams_subscribes Dmaap topics that the MS will consume messages - -Streams_publishes Dmaap topics that the MS will publish messages - -postgres.host Host where the postgres database is running - -postgres.port Host where the postgres database is running - -postgres.username Postgres username - -postgres.password Postgres password - -sonhandler.pollingInterval Polling Interval for consuming dmaap messages - -sonhandler.pollingTimeout Polling timeout for consuming dmaap messages - -sonhandler.numSolutions Number for solutions for OOF optimization - -sonhandler.minCollision Minimum collision criteria to trigger OOF - -sonhandler.minConfusion Minimum confusion criteria to trigger OOF - -sonhandler.maximumClusters Maximum number of clusters MS can process - -sonhandler.badThreshold Bad threshold for Handover success rate - -sonhandler.poorThreshold Poor threshold for Handover success rate - -sonhandler.namespace Namespace where MS is going to be deployed - -sonhandler.sourceId Source ID of the Microservice (Required for Sending request to OOF) - -sonhandler.dmaap.server Location of message routers - -sonhandler.bufferTime Buffer time for MS to wait for more notifications when the optimization criteria is not met - -sonhandler.cg Consumer group for the MS to consume message from dmaap - -sonhandler.cid Consumer ID for the MS to consume message from dmaap - -sonhandler.configDbService Location of the config DB (protocol, host & port) - -sonhandler.oof.service Location of OOF (protocol, host & port) - -sonhandler.optimizers Optimizer to trigger in OOF - ++-------------------------------+------------------------------------------------+ +|Configuration | Description | ++===============================+================================================+ +|Streams_subscribes | Dmaap topics that the MS will consume messages | ++-------------------------------+------------------------------------------------+ +|Streams_publishes | Dmaap topics that the MS will publish messages | ++-------------------------------+------------------------------------------------+ +|postgres.host | Host where the postgres database is running | ++-------------------------------+------------------------------------------------+ +|postgres.port | Host where the postgres database is running | ++-------------------------------+------------------------------------------------+ +|postgres.username | Postgres username | ++-------------------------------+------------------------------------------------+ +|postgres.password | Postgres password | ++-------------------------------+------------------------------------------------+ +|sonhandler.pollingInterval | Polling Interval for consuming dmaap messages | ++-------------------------------+------------------------------------------------+ +|sonhandler.pollingTimeout | Polling timeout for consuming dmaap messages | ++-------------------------------+------------------------------------------------+ +|sonhandler.numSolutions | Number for solutions for OOF optimization | ++-------------------------------+------------------------------------------------+ +|sonhandler.minCollision | Minimum collision criteria to trigger OOF | ++-------------------------------+------------------------------------------------+ +|sonhandler.minConfusion | Minimum confusion criteria to trigger OOF | ++-------------------------------+------------------------------------------------+ +|sonhandler.maximumClusters | Maximum number of clusters MS can process | ++-------------------------------+------------------------------------------------+ +|sonhandler.badThreshold | Bad threshold for Handover success rate | ++-------------------------------+------------------------------------------------+ +|sonhandler.poorThreshold | Poor threshold for Handover success rate | ++-------------------------------+------------------------------------------------+ +|sonhandler.namespace | Namespace where MS is going to be deployed | ++-------------------------------+------------------------------------------------+ +|sonhandler.sourceId | Source ID of the Microservice (to OOF) | ++-------------------------------+------------------------------------------------+ +|sonhandler.dmaap.server | Location of message routers | ++-------------------------------+------------------------------------------------+ +|sonhandler.bufferTime | Buffer time for MS to wait for notifications | ++-------------------------------+------------------------------------------------+ +|sonhandler.cg | DMAAP Consumer group for subscription | ++-------------------------------+------------------------------------------------+ +|sonhandler.cid | DMAAP Consumer id for subcription | ++-------------------------------+------------------------------------------------+ +|sonhandler.configDbService | Location of config DB (protocol, host & port) | ++-------------------------------+------------------------------------------------+ +|sonhandler.oof.service | Location of OOF (protocol, host & port) | ++-------------------------------+------------------------------------------------+ +|sonhandler.optimizers | Optimizer to trigger in OOF | ++-------------------------------+------------------------------------------------+ diff --git a/docs/sections/services/tca-cdap/development_info.rst b/docs/sections/services/tca-cdap/development_info.rst index afb240ef..fe48c267 100644 --- a/docs/sections/services/tca-cdap/development_info.rst +++ b/docs/sections/services/tca-cdap/development_info.rst @@ -6,16 +6,16 @@ To build just the TCA component, run the following maven command `mvn clean install` -Maven GroupId: -============== +Maven GroupId +------------- org.onap.dcaegen2.analytics.tca -Maven Parent ArtifactId: ----------------- +Maven Parent ArtifactId +----------------------- dcae-analytics -Maven Children Artifacts: +Maven Children Artifacts ------------------------ 1. dcae-analytics-test: Common test code for all DCAE Analytics Modules 2. dcae-analytics-model: Contains models (e.g. Common Event Format) which are common to DCAE Analytics @@ -29,7 +29,10 @@ Maven Children Artifacts: API Endpoints -============= +------------- + +For deployment into CDAP, following API's can be used to deploy TCA application. + # create namespace curl -X PUT http://:11015/v3/namespaces/cdap_tca_hi_lo @@ -60,8 +63,7 @@ curl http://:11015/v3/namespaces/cdap_tca_hi_lo/apps/dcae-tca/flo TCA CDAP Container -================= - +------------------ If new jar is generated, corresponding version should be updated into https://git.onap.org/dcaegen2/deployments/tree/tca-cdap-container. Following files should be revised diff --git a/docs/sections/services/tca-cdap/index.rst b/docs/sections/services/tca-cdap/index.rst index 9d184f5b..b8bc2ef8 100644 --- a/docs/sections/services/tca-cdap/index.rst +++ b/docs/sections/services/tca-cdap/index.rst @@ -1,15 +1,11 @@ .. This work is licensed under a Creative Commons Attribution 4.0 International License. .. http://creativecommons.org/licenses/by/4.0 -================================= +================================== Threshold Crossing Analytics (TCA) -================================= +================================== -.. contents:: - :depth: 3 -.. - Overview ======== diff --git a/docs/sections/services/ves-http/architecture.rst b/docs/sections/services/ves-http/architecture.rst index 29077afe..41f3e8f1 100644 --- a/docs/sections/services/ves-http/architecture.rst +++ b/docs/sections/services/ves-http/architecture.rst @@ -7,13 +7,13 @@ VES Architecture .. image:: ./ves-deployarch.png VES Processing Flow -=================== +------------------- .. image:: ./VES-processingFlow.png VES Schema Validation -===================== +--------------------- VES Collector is configured to support below VES Version; the corresponding API uses VES schema definition for event validation. @@ -30,7 +30,7 @@ Schema definition files are contained within VES collector gerrit repo - https:/ Features Supported -================== +------------------ - VES collector deployed as docker containers - Acknowledgement to sender with appropriate response code (both successful and failure) - Authentication of the events posted to collector (support 4 types of authentication setting) @@ -45,7 +45,7 @@ The collector can receive events via standard HTTP port (8080) or secure port (8 Dynamic configuration fed into Collector via DCAEPlatform -========================================================= +--------------------------------------------------------- - Outbound Dmaap/UEB topic - Schema version to be validated against diff --git a/docs/sections/services/ves-http/installation.rst b/docs/sections/services/ves-http/installation.rst index 9a627df3..b39b007d 100644 --- a/docs/sections/services/ves-http/installation.rst +++ b/docs/sections/services/ves-http/installation.rst @@ -4,6 +4,10 @@ Installation ============ +VESCollector is installed via cloudify blueprint by DCAE bootstrap process on typical ONAP installation. +As the service is containerized, it can be started on stand-alone mode also. + + To run VES Collector container on standalone mode, following parameters are required ``docker run -d -p 8080:8080/tcp -p 8443:8443/tcp -P -e DMAAPHOST='10.0.11.1' nexus.onap.org:10001/onap/org.onap.dcaegen2.collectors.ves.vescollector:1.3.2`` diff --git a/docs/sections/services/ves-http/tls-authentication.rst b/docs/sections/services/ves-http/tls-authentication.rst index b5226dc5..ad8d8d22 100644 --- a/docs/sections/services/ves-http/tls-authentication.rst +++ b/docs/sections/services/ves-http/tls-authentication.rst @@ -1,8 +1,8 @@ .. This work is licensed under a Creative Commons Attribution 4.0 International License. .. http://creativecommons.org/licenses/by/4.0 -Certificates as authentication method for PNFs/VNFs -=================================================== +Authentication Types +==================== VES supports mutual TLS authentication via X.509 certificates. If VES is deployed via docker image then VES configuration can be modified by editing */opt/app/VESCollector/etc/collector.properties* which is present on the docker container. VES detects changes made to the mentioned file automatically and restarts the application. -- cgit 1.2.3-korg