summaryrefslogtreecommitdiffstats
path: root/docs/sections/services/datalake-handler
diff options
context:
space:
mode:
Diffstat (limited to 'docs/sections/services/datalake-handler')
-rw-r--r--docs/sections/services/datalake-handler/index.rst1
-rw-r--r--docs/sections/services/datalake-handler/installation.rst135
2 files changed, 0 insertions, 136 deletions
diff --git a/docs/sections/services/datalake-handler/index.rst b/docs/sections/services/datalake-handler/index.rst
index 3fade9c6..e4f1c905 100644
--- a/docs/sections/services/datalake-handler/index.rst
+++ b/docs/sections/services/datalake-handler/index.rst
@@ -27,7 +27,6 @@ DataLake-Handler MS Installation Steps and Configurations
.. toctree::
:maxdepth: 1
- ./installation.rst
./installation-helm.rst
DataLake-Handler MS Admin UI User Guide
diff --git a/docs/sections/services/datalake-handler/installation.rst b/docs/sections/services/datalake-handler/installation.rst
deleted file mode 100644
index e9215c29..00000000
--- a/docs/sections/services/datalake-handler/installation.rst
+++ /dev/null
@@ -1,135 +0,0 @@
-.. This work is licensed under a Creative Commons Attribution 4.0 International License.
-.. http://creativecommons.org/licenses/by/4.0
-.. _dl-installation:
-
-Deployment Steps
-
-################
-DL-handler consists of three pods- the feeder, admin UI and des. It can be deployed by using cloudify blueprint. Datalake can be easily deployed through DCAE cloudify manager. The following steps guides you launch Datalake though cloudify manager.
-
-Pre-requisite
--------------
-Make sure dcae postgres is properly deployed and functional.
-An external database, such as Elasticsearch and MongoDB is deployed. Install mongodb through the following command.
-
- #docker run -itd --restart=always --name dl-mongo -p 27017:27017 mongo
-
-For DES service deployment, presto service is deployed. Here is a sample how presto deploy in the environment.
- Build a presto image:
- The package of presto version we are using is v0.0.2:presto-v0.0.2.tar.gz
-
- #docker build -t presto:v0.0.2 .
- #docker tag presto:v0.0.2 registry.baidubce.com/onap/presto:v0.0.2
- #docker push registry.baidubce.com/onap/presto:v0.0.2
-
- Note: Replace the repository path with your own repository.
-
- Install presto service:
-
- #kubectl -n onap run dl-presto --image=registry.baidubce.com/onap/presto:v0.0.2 --env="MongoDB_IP=192.168.235.11" --env="MongoDB_PORT=27017"
- #kubectl -n onap expose deployment dl-presto --port=9000 --target-port=9000 --type=NodePort
-
- Note: MonoDB_IP and Mongo_PORT you can replace this two values with your own configuration.
-
-After datalake getting deployed, the admin UI can be used to configure the sink database address and credentials.
-
-Log-in to the DCAE Bootstrap POD
---------------------------------
-
-First, we should find the bootstrap pod name through the following command and make sure that DCAE coudify manager is properly deployed.
- .. image :: ./images/bootstrap-pod.png
-
-Login to the DCAE bootstrap pod through the following command.
- .. code-block :: bash
-
- #kubectl exec -it <DCAE bootstrap pod> /bin/bash -n onap
-
-Validate Blueprint
-------------------
-Before the blueprints uploading to Cloudify manager, the blueprints shoule be validated first through the following command.
- .. code-block :: bash
-
- #cfy blueprint validate /bluerints/k8s-datalake-feeder.yaml
- #cfy blueprint validate /blueprints/k8s-datalake-admin-ui.yaml
- #cfy blueprint validate /blueprints/k8s-datalake-des.yaml
-
-Upload the Blueprint to Cloudify Manager.
------------------------------------------
-After validating, we can start to proceed blueprints uploading.
- .. code-block :: bash
-
- #cfy blueprint upload -b dl-feeder /bluerints/k8s-datalake-feeder.yaml
- #cfy blueprint upload -b dl-admin-ui /blueprints/k8s-datalake-admin-ui.yaml
- #cfy blueprint upload -b des /blueprints/k8s-datalake-des.yaml
-
-Verify Uploaded Blueprints
---------------------------
-Using "cfy blueprint list" to verify your work.
- .. code-block :: bash
-
- #cfy blueprint list
-
-You can see the following returned message to show the blueprints have been correctly uploaded.
- .. image :: ./images/blueprint-list.png
-
-
-Verify Plugin Versions
-----------------------
-If the version of the plugin used is different, update the blueprint import to match.
- .. code-block :: bash
-
- #cfy plugins list
-
-Create Deployment
------------------
-Here we are going to create deployments for both feeder and admin UI.
- .. code-block :: bash
-
- #cfy deployments create -b dl-feeder feeder-deploy
- #cfy deployments create -b dl-admin-ui admin-ui-deploy
- #cfy deployments create -b des des
-
-Launch Service
---------------
-Next, we are going to launch the datalake.
- .. code-block :: bash
-
- #cfy executions start -d feeder-deploy install
- #cfy executions start -d admin-ui-deploy install
- #cfy executions start -d des install
-
-
-Verify the Deployment Result
-----------------------------
-The following command can be used to list the datalake logs.
-
- .. code-block :: bash
-
- #kubectl logs <datalake-pod> -n onap
-
-The output should looks like.
- .. image :: ./images/feeder-log.png
-
-The des output should looks like.
- .. image :: ./des-log.png
-
-If you find any Java exception from log, make sure that the external database and datalake configuration are properly configured.
-Admin UI can be used to configure the external database configuration.
-
-
-Uninstall
----------
-Uninstall running component and delete deployment
- .. code-block :: bash
-
- #cfy uninstall feeder-deploy
- #cfy uninstall admin-ui-deploy
- #cfy uninstall des
-
-Delete Blueprint
-----------------
- .. code-block :: bash
-
- #cfy blueprints delete dl-feeder
- #cfy blueprints delett dl-admin-ui
- #cfy blueprints delete des