aboutsummaryrefslogtreecommitdiffstats
path: root/vnfs
diff options
context:
space:
mode:
authorDileep Ranganathan <dileep.ranganathan@intel.com>2019-06-14 00:09:33 -0700
committerGary Wu <gary.wu@futurewei.com>2019-06-14 15:07:09 +0000
commit1bbe6436eeb56a8a091fc15ca3a023fcba31d24e (patch)
tree86c67a9b4cacf467ae7c98e078270092d9cb7177 /vnfs
parent8c634d536d17f0a8414b0a2bc7b2f924ea565222 (diff)
Updated README for Distributed Analytics
Updated readme for distributed analytics. Change-Id: I675c8b943df119b57638d893089e6384c4fcf5bb Issue-ID: ONAPARC-280 Signed-off-by: Dileep Ranganathan <dileep.ranganathan@intel.com>
Diffstat (limited to 'vnfs')
-rw-r--r--vnfs/DAaaS/README.md69
1 files changed, 55 insertions, 14 deletions
diff --git a/vnfs/DAaaS/README.md b/vnfs/DAaaS/README.md
index 2b3c0d6a..4b6fcf50 100644
--- a/vnfs/DAaaS/README.md
+++ b/vnfs/DAaaS/README.md
@@ -11,20 +11,28 @@
```bash
git clone https://github.com/onap/demo.git
DA_WORKING_DIR=$PWD/demo/vnfs/DAaaS/deploy
-cd $DA_WORKING_DIR
```
#### Install Rook-Ceph for Persistent Storage
Note: This is unusual but Flex volume path can be different than the default value. values.yaml has the most common flexvolume path configured. In case of errors related to flexvolume please refer to the https://rook.io/docs/rook/v0.9/flexvolume.html#configuring-the-flexvolume-path to find the appropriate flexvolume-path and set it in values.yaml
```bash
-cd 00-init/rook-ceph
+cd $DA_WORKING_DIR/00-init/rook-ceph
helm install -n rook . -f values.yaml --namespace=rook-ceph-system
```
Check for the status of the pods in rook-ceph namespace. Once all pods are in Ready state move on to the next section.
+
+```bash
+$ kubectl get pods -n rook-ceph-system
+NAME READY STATUS RESTARTS AGE
+rook-ceph-agent-9wszf 1/1 Running 0 121s
+rook-ceph-agent-xnbt8 1/1 Running 0 121s
+rook-ceph-operator-bc77d6d75-ltwww 1/1 Running 0 158s
+rook-discover-bvj65 1/1 Running 0 133s
+rook-discover-nbfrp 1/1 Running 0 133s
+```
```bash
$ kubectl -n rook-ceph get pod
NAME READY STATUS RESTARTS AGE
-rook-ceph-agent-4zkg8 1/1 Running 0 140s
rook-ceph-mgr-a-d9dcf5748-5s9ft 1/1 Running 0 77s
rook-ceph-mon-a-7d8f675889-nw5pl 1/1 Running 0 105s
rook-ceph-mon-b-856fdd5cb9-5h2qk 1/1 Running 0 94s
@@ -33,7 +41,7 @@ rook-ceph-osd-0-7cbbbf749f-j8fsd 1/1 Running 0 25s
rook-ceph-osd-1-7f67f9646d-44p7v 1/1 Running 0 25s
rook-ceph-osd-2-6cd4b776ff-v4d68 1/1 Running 0 25s
rook-ceph-osd-prepare-vx2rz 0/2 Completed 0 60s
-rook-discover-dhkb8 1/1 Running 0 140s
+rook-ceph-tools-5bd5cdb949-j68kk 1/1 Running 0 53s
```
#### Install Operator package
@@ -63,22 +71,55 @@ helm install -n cp . -f values.yaml --namespace=edge1
Custom Collectd
===============
-Build the image and set the image name with tag to COLLECTD_IMAGE_NAME
-Push the image to docker registry using the command
-docker push dcr.default.svc.local:32000/${COLLECTD_IMAGE_NAME}
-Edit the values.yaml and change the image name as the value of COLLECTD_IMAGE_NAME
-place the collectd.conf in $DA_WORKING_DIR/collection/charts/collectd/resources/config directory
+1. Build the custom collectd image
+2. Set COLLECTD_IMAGE_NAME with appropriate image_repository:tag
+3. Push the image to docker registry using the command
+4. docker push ${COLLECTD_IMAGE_NAME}
+5. Edit the values.yaml and change the image repository and tag using
+ COLLECTD_IMAGE_NAME appropriately.
+6. Place the collectd.conf in
+ $DA_WORKING_DIR/collection/charts/collectd/resources/config
-cd $DA_WORKING_DIR/collection
-helm install -n cp . -f values.yaml --namespace=edge1
+7. cd $DA_WORKING_DIR/collection
+8. helm install -n cp . -f values.yaml --namespace=edge1
```
#### Verify Collection package
+* Check if all pods are up in edge1 namespace
+* Check the prometheus UI using port-forwarding port 9090 (default for prometheus service)
```
-TODO
-1. Check if all pods are up uin edge1 namespace
-2. Check the prometheus UI using the port 30090
+$ kubectl get pods -n edge1
+NAME READY STATUS RESTARTS AGE
+cp-cadvisor-8rk2b 1/1 Running 0 15s
+cp-cadvisor-nsjr6 1/1 Running 0 15s
+cp-collectd-h5krd 1/1 Running 0 23s
+cp-collectd-jc9m2 1/1 Running 0 23s
+cp-prometheus-node-exporter-blc6p 1/1 Running 0 17s
+cp-prometheus-node-exporter-qbvdx 1/1 Running 0 17s
+prometheus-cp-prometheus-prometheus-0 4/4 Running 1 33s
+
+$ kubectl get svc -n edge1
+NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S)
+cadvisor NodePort 10.43.53.122 <none> 80:30091/TCP
+collectd ClusterIP 10.43.222.34 <none> 9103/TCP
+cp13-prometheus-node-exporter ClusterIP 10.43.17.242 <none> 9100/TCP
+cp13-prometheus-prometheus NodePort 10.43.26.155 <none> 9090:30090/TCP
+prometheus-operated ClusterIP None <none> 9090/TCP
```
+
+#### Install Minio Model repository
+* Prerequisite: Dynamic storage provisioner needs to be enabled. Either rook-ceph ($DA_WORKING_DIR/00-init) or another alternate provisioner needs to be enabled.
+```bash
+cd $DA_WORKING_DIR/minio
+
+Edit the values.yaml to set the credentials to access the minio UI.
+Default values are
+accessKey: "onapdaas"
+secretKey: "onapsecretdaas"
+
+helm install -n minio . -f values.yaml --namespace=edge1
+```
+
#### Onboard an Inference Application
```
TODO