summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--TOSCA/Helm/onap-blueprint.yaml10
-rw-r--r--docs/oom_user_guide.rst6
-rw-r--r--kubernetes/aai/charts/aai-babel/values.yaml2
-rw-r--r--kubernetes/aai/charts/aai-champ/values.yaml3
-rw-r--r--kubernetes/aai/charts/aai-gizmo/values.yaml4
-rw-r--r--kubernetes/aai/charts/aai-hbase/templates/service.yaml2
-rw-r--r--kubernetes/aai/charts/aai-modelloader/templates/service.yaml10
-rw-r--r--kubernetes/aai/charts/aai-modelloader/values.yaml3
-rw-r--r--kubernetes/aai/charts/aai-search-data/templates/service.yaml6
-rw-r--r--kubernetes/aai/charts/aai-search-data/values.yaml2
-rw-r--r--kubernetes/aai/charts/aai-sparky-be/templates/service.yaml4
-rw-r--r--kubernetes/aai/charts/aai-sparky-be/values.yaml4
-rw-r--r--kubernetes/aai/templates/service.yaml16
-rw-r--r--kubernetes/aai/values.yaml11
-rw-r--r--kubernetes/clamp/values.yaml4
-rw-r--r--kubernetes/cli/values.yaml2
-rw-r--r--kubernetes/common/dgbuilder/templates/deployment.yaml8
-rw-r--r--kubernetes/consul/resources/config/consul-agent-config/aaf-sms-health.json17
-rw-r--r--kubernetes/consul/resources/config/consul-agent-config/mr-dmaap-health.json10
-rw-r--r--kubernetes/consul/resources/config/consul-agent-config/mr-health.json29
-rw-r--r--kubernetes/consul/resources/config/consul-agent-config/mr-kafka-health.json10
-rw-r--r--kubernetes/consul/resources/config/consul-agent-config/mr-zookeeper-health.json10
-rwxr-xr-xkubernetes/consul/resources/config/consul-agent-config/scripts/mr-kafka-health.sh2
-rwxr-xr-xkubernetes/consul/resources/config/consul-agent-config/scripts/mr-zookeeper-health.sh4
-rw-r--r--kubernetes/esr/charts/esr-gui/values.yaml2
-rw-r--r--kubernetes/esr/values.yaml2
-rw-r--r--kubernetes/log/charts/log-elasticsearch/values.yaml4
-rw-r--r--kubernetes/log/charts/log-kibana/values.yaml4
-rw-r--r--kubernetes/multicloud/charts/multicloud-ocata/values.yaml2
-rw-r--r--kubernetes/multicloud/charts/multicloud-vio/values.yaml2
-rw-r--r--kubernetes/multicloud/charts/multicloud-windriver/values.yaml2
-rw-r--r--kubernetes/multicloud/values.yaml2
-rw-r--r--kubernetes/nbi/templates/deployment.yaml4
-rw-r--r--kubernetes/nbi/values.yaml2
-rw-r--r--kubernetes/onap/resources/environments/dev.yaml2
-rw-r--r--kubernetes/onap/resources/environments/disable-allcharts.yaml2
-rw-r--r--kubernetes/policy/charts/drools/resources/config/opt/policy/config/drools/base.conf4
-rw-r--r--kubernetes/policy/charts/drools/resources/config/opt/policy/config/drools/drools-tweaks.sh60
-rw-r--r--kubernetes/policy/charts/drools/templates/statefulset.yaml4
-rw-r--r--kubernetes/policy/charts/drools/values.yaml3
-rw-r--r--kubernetes/policy/charts/pdp/resources/config/pe/pdplp.conf4
-rw-r--r--kubernetes/policy/resources/config/pe/paplp.conf4
-rw-r--r--kubernetes/portal/charts/portal-mariadb/values.yaml4
43 files changed, 184 insertions, 108 deletions
diff --git a/TOSCA/Helm/onap-blueprint.yaml b/TOSCA/Helm/onap-blueprint.yaml
index 3ac76fa8b1..c5b699b5be 100644
--- a/TOSCA/Helm/onap-blueprint.yaml
+++ b/TOSCA/Helm/onap-blueprint.yaml
@@ -347,3 +347,13 @@ node_templates:
relationships:
- type: cloudify.relationships.connected_to
target: k8s_env
+
+ nbi:
+ type: cloudify.nodes.Helm
+ properties:
+ args:
+ - { get_input: namespace }
+ - nbi
+ relationships:
+ - type: cloudify.relationships.connected_to
+ target: k8s_env \ No newline at end of file
diff --git a/docs/oom_user_guide.rst b/docs/oom_user_guide.rst
index bb8ff0b27c..fbd3d9ea97 100644
--- a/docs/oom_user_guide.rst
+++ b/docs/oom_user_guide.rst
@@ -196,7 +196,7 @@ Many technologies are used across the projects resulting in significant
operational complexity and an inability to apply global parameters across the
entire ONAP deployment. OOM solves this problem by introducing a common
configuration technology, Helm charts, that provide a hierarchical
-configuration configuration with the ability to override values with higher
+configuration with the ability to override values with higher
level charts or command line options.
The structure of the configuration of ONAP is shown in the following diagram.
@@ -420,10 +420,10 @@ to monitor the real-time health of an ONAP deployment:
- a set of liveness probes which feed into the Kubernetes manager which
are described in the Heal section.
-Within ONAP Consul is the monitoring system of choice and deployed by OOM in two parts:
+Within ONAP, Consul is the monitoring system of choice and deployed by OOM in two parts:
- a three-way, centralized Consul server cluster is deployed as a highly
- available monitor of all of the ONAP components,and
+ available monitor of all of the ONAP components, and
- a number of Consul agents.
The Consul server provides a user interface that allows a user to graphically
diff --git a/kubernetes/aai/charts/aai-babel/values.yaml b/kubernetes/aai/charts/aai-babel/values.yaml
index 1e74e35cf9..7118d1669b 100644
--- a/kubernetes/aai/charts/aai-babel/values.yaml
+++ b/kubernetes/aai/charts/aai-babel/values.yaml
@@ -53,7 +53,7 @@ readiness:
service:
type: NodePort
portName: babel
- externalPort: 79
+ externalPort: 9516
internalPort: 9516
nodePort: 79
diff --git a/kubernetes/aai/charts/aai-champ/values.yaml b/kubernetes/aai/charts/aai-champ/values.yaml
index 7849418ccd..13cc9e77c0 100644
--- a/kubernetes/aai/charts/aai-champ/values.yaml
+++ b/kubernetes/aai/charts/aai-champ/values.yaml
@@ -52,8 +52,7 @@ readiness:
service:
type: NodePort
- name: champ
- portName: champ
+ portName: aai-champ
internalPort: 9522
nodePort: 78
diff --git a/kubernetes/aai/charts/aai-gizmo/values.yaml b/kubernetes/aai/charts/aai-gizmo/values.yaml
index 0e982d8fc6..1ea924fc13 100644
--- a/kubernetes/aai/charts/aai-gizmo/values.yaml
+++ b/kubernetes/aai/charts/aai-gizmo/values.yaml
@@ -50,8 +50,8 @@ readiness:
service:
type: NodePort
- name: crud-service
- portName: crud-service
+ name: aai-crud-service
+ portName: aai-crud-service
internalPort: 9520
nodePort: 68
diff --git a/kubernetes/aai/charts/aai-hbase/templates/service.yaml b/kubernetes/aai/charts/aai-hbase/templates/service.yaml
index c503e6ca6a..53c5e5f979 100644
--- a/kubernetes/aai/charts/aai-hbase/templates/service.yaml
+++ b/kubernetes/aai/charts/aai-hbase/templates/service.yaml
@@ -1,7 +1,7 @@
apiVersion: v1
kind: Service
metadata:
- name: {{ include "common.name" . }}
+ name: {{ include "common.servicename" . }}
namespace: {{ include "common.namespace" . }}
labels:
app: {{ include "common.name" . }}
diff --git a/kubernetes/aai/charts/aai-modelloader/templates/service.yaml b/kubernetes/aai/charts/aai-modelloader/templates/service.yaml
index bd59edb94f..4bfe7879e7 100644
--- a/kubernetes/aai/charts/aai-modelloader/templates/service.yaml
+++ b/kubernetes/aai/charts/aai-modelloader/templates/service.yaml
@@ -1,7 +1,7 @@
apiVersion: v1
kind: Service
metadata:
- name: {{ include "common.name" . }}
+ name: {{ include "common.servicename" . }}
namespace: {{ include "common.namespace" . }}
labels:
app: {{ include "common.name" . }}
@@ -14,15 +14,15 @@ spec:
{{if eq .Values.service.type "NodePort" -}}
- port: {{ .Values.service.internalPort }}
nodePort: {{ .Values.global.nodePortPrefix | default .Values.nodePortPrefix }}{{ .Values.service.nodePort }}
- name: {{ .Values.service.name }}
+ name: {{ .Values.service.portName }}
- port: {{ .Values.service.internalPort2 }}
nodePort: {{ .Values.global.nodePortPrefix | default .Values.nodePortPrefix }}{{ .Values.service.nodePort2 }}
- name: {{ .Values.service.name }}2
+ name: {{ .Values.service.portName2 }}
{{- else -}}
- port: {{ .Values.service.internalPort }}
- name: {{ .Values.service.name }}
+ name: {{ .Values.service.portName }}
- port: {{ .Values.service.internalPort2 }}
- name: {{ .Values.service.name }}2
+ name: {{ .Values.service.portName2 }}
{{- end}}
selector:
app: {{ include "common.name" . }}
diff --git a/kubernetes/aai/charts/aai-modelloader/values.yaml b/kubernetes/aai/charts/aai-modelloader/values.yaml
index a5095ade51..676fc6fdfc 100644
--- a/kubernetes/aai/charts/aai-modelloader/values.yaml
+++ b/kubernetes/aai/charts/aai-modelloader/values.yaml
@@ -35,10 +35,11 @@ readiness:
service:
type: NodePort
- name: aai-modelloader
+ portName: aai-modelloader
externalPort: 8080
internalPort: 8080
nodePort: 10
+ portName2: aai-modelloader-ssl
externalPort2: 8443
internalPort2: 8443
nodePort2: 29
diff --git a/kubernetes/aai/charts/aai-search-data/templates/service.yaml b/kubernetes/aai/charts/aai-search-data/templates/service.yaml
index 41bc163696..e342a9a688 100644
--- a/kubernetes/aai/charts/aai-search-data/templates/service.yaml
+++ b/kubernetes/aai/charts/aai-search-data/templates/service.yaml
@@ -1,7 +1,7 @@
apiVersion: v1
kind: Service
metadata:
- name: {{ include "common.name" . }}
+ name: {{ include "common.servicename" . }}
namespace: {{ include "common.namespace" . }}
labels:
app: {{ include "common.name" . }}
@@ -14,10 +14,10 @@ spec:
{{if eq .Values.service.type "NodePort" -}}
- port: {{ .Values.service.internalPort }}
nodePort: {{ .Values.global.nodePortPrefix | default .Values.nodePortPrefix }}{{ .Values.service.nodePort }}
- name: {{ .Values.service.name }}
+ name: {{ .Values.service.portName }}
{{- else -}}
- port: {{ .Values.service.internalPort }}
- name: {{ .Values.service.name }}
+ name: {{ .Values.service.portName }}
{{- end}}
selector:
app: {{ include "common.name" . }}
diff --git a/kubernetes/aai/charts/aai-search-data/values.yaml b/kubernetes/aai/charts/aai-search-data/values.yaml
index 4e47d37134..c6f42bd38b 100644
--- a/kubernetes/aai/charts/aai-search-data/values.yaml
+++ b/kubernetes/aai/charts/aai-search-data/values.yaml
@@ -41,7 +41,7 @@ readiness:
service:
type: ClusterIP
- name: aai-search-data
+ portName: aai-search-data
internalPort: 9509
ingress:
diff --git a/kubernetes/aai/charts/aai-sparky-be/templates/service.yaml b/kubernetes/aai/charts/aai-sparky-be/templates/service.yaml
index e97935aedd..e342a9a688 100644
--- a/kubernetes/aai/charts/aai-sparky-be/templates/service.yaml
+++ b/kubernetes/aai/charts/aai-sparky-be/templates/service.yaml
@@ -14,10 +14,10 @@ spec:
{{if eq .Values.service.type "NodePort" -}}
- port: {{ .Values.service.internalPort }}
nodePort: {{ .Values.global.nodePortPrefix | default .Values.nodePortPrefix }}{{ .Values.service.nodePort }}
- name: {{ .Values.service.name }}
+ name: {{ .Values.service.portName }}
{{- else -}}
- port: {{ .Values.service.internalPort }}
- name: {{ .Values.service.name }}
+ name: {{ .Values.service.portName }}
{{- end}}
selector:
app: {{ include "common.name" . }}
diff --git a/kubernetes/aai/charts/aai-sparky-be/values.yaml b/kubernetes/aai/charts/aai-sparky-be/values.yaml
index 91c4026381..b9ef4ca85d 100644
--- a/kubernetes/aai/charts/aai-sparky-be/values.yaml
+++ b/kubernetes/aai/charts/aai-sparky-be/values.yaml
@@ -4,7 +4,7 @@
global: # global defaults
nodePortPrefix: 302
aai:
- serviceName: aai-aai
+ serviceName: aai
aaiElasticsearch:
serviceName: aai-elasticsearch
gizmo:
@@ -62,7 +62,7 @@ readiness:
service:
type: ClusterIP
- name: aai-sparky-be
+ portName: aai-sparky-be
internalPort: 9517
internalPort2: 8000
diff --git a/kubernetes/aai/templates/service.yaml b/kubernetes/aai/templates/service.yaml
index d21ac7ed87..a2f96a843e 100644
--- a/kubernetes/aai/templates/service.yaml
+++ b/kubernetes/aai/templates/service.yaml
@@ -1,7 +1,7 @@
apiVersion: v1
kind: Service
metadata:
- name: {{ include "common.name" . }}
+ name: {{ include "common.servicename" . }}
namespace: {{ include "common.namespace" . }}
labels:
app: {{ include "common.name" . }}
@@ -11,23 +11,23 @@ metadata:
spec:
ports:
{{if eq .Values.service.type "NodePort" -}}
- - name: {{ .Values.service.name }}
- port: {{ .Values.service.internalPort }}
+ - name: {{ .Values.service.portName }}
+ port: {{ .Values.service.externalPort }}
targetPort: {{ .Values.service.internalPort }}
nodePort: {{ .Values.global.nodePortPrefix | default .Values.nodePortPrefix }}{{ .Values.service.nodePort }}
- - name: {{ .Values.service.name }}2
- port: {{ .Values.service.internalPort2 }}
+ - name: {{ .Values.service.portName2 }}
+ port: {{ .Values.service.externalPort2 }}
targetPort: {{ .Values.service.internalPort2 }}
nodePort: {{ .Values.global.nodePortPrefix | default .Values.nodePortPrefix }}{{ .Values.service.nodePort2 }}
{{- else -}}
- port: {{ .Values.service.externalPort }}
targetPort: {{ .Values.service.internalPort }}
- name: {{ .Values.service.name }}
+ name: {{ .Values.service.portName }}
- port: {{ .Values.service.externalPort2 }}
targetPort: {{ .Values.service.internalPort2 }}
- name: {{ .Values.service.name }}2
+ name: {{ .Values.service.portName }}
{{- end}}
type: {{ .Values.service.type }}
selector:
app: {{ include "common.name" . }}
- clusterIP: {{ .Values.config.aaiServiceClusterIp }}
+ clusterIP: {{ .Values.service.aaiServiceClusterIp }}
diff --git a/kubernetes/aai/values.yaml b/kubernetes/aai/values.yaml
index 153379b06c..2b0180e53b 100644
--- a/kubernetes/aai/values.yaml
+++ b/kubernetes/aai/values.yaml
@@ -29,7 +29,7 @@ global: # global defaults
serviceName: aai-cassandra
replicas: 3
aai:
- serviceName: aai-aai
+ serviceName: aai
babel:
serviceName: aai-babel
champ:
@@ -64,9 +64,6 @@ debugEnabled: false
# application configuration
config:
- # POLICY hotfix - Note this must be temporary
- # See https://jira.onap.org/browse/POLICY-510
- aaiServiceClusterIp:
logstashServiceName: log-ls
logstashPort: 5044
@@ -91,13 +88,17 @@ readiness:
service:
type: NodePort
- name: aai
+ portName: aai
externalPort: 8080
internalPort: 8080
nodePort: 32
+ portName2: aai-ssl
externalPort2: 8443
internalPort2: 8443
nodePort2: 33
+ # POLICY hotfix - Note this must be temporary
+ # See https://jira.onap.org/browse/POLICY-510
+ aaiServiceClusterIp:
ingress:
enabled: false
diff --git a/kubernetes/clamp/values.yaml b/kubernetes/clamp/values.yaml
index 71752fcfc6..751e4a0c11 100644
--- a/kubernetes/clamp/values.yaml
+++ b/kubernetes/clamp/values.yaml
@@ -48,8 +48,8 @@ config:
"clamp.config.sdc.serviceUsername": "clamp",
"clamp.config.sdc.servicePassword": "b7acccda32b98c5bb7acccda32b98c5b05D511BD6D93626E90D18E9D24D9B78CD34C7EE8012F0A189A28763E82271E50A5D4EC10C7D93E06E0A2D27CAE66B981",
"clamp.config.files.sdcController": "file:/opt/clamp/sdc-controllers-config.json",
- "clamp.config.dcae.inventory.url": "http://dcaegen2.{{ include "common.namespace" . }}:8080",
- "clamp.config.dcae.dispatcher.url": "http://dcaegen2.{{ include "common.namespace" . }}:8080",
+ "clamp.config.dcae.inventory.url": "http://inventory.{{ include "common.namespace" . }}:8080",
+ "clamp.config.dcae.dispatcher.url": "http://deployment-handler.{{ include "common.namespace" . }}:8443",
"clamp.config.policy.pdpUrl1": "https://pdp.{{ include "common.namespace" . }}:9091/pdp/ , testpdp, alpha123",
"clamp.config.policy.pdpUrl2": "https://pdp.{{ include "common.namespace" . }}:9091/pdp/ , testpdp, alpha123",
"clamp.config.policy.papUrl": "https://pap.{{ include "common.namespace" . }}:8443/pap/ , testpap, alpha123",
diff --git a/kubernetes/cli/values.yaml b/kubernetes/cli/values.yaml
index dd034f7b76..bd52ced25f 100644
--- a/kubernetes/cli/values.yaml
+++ b/kubernetes/cli/values.yaml
@@ -59,7 +59,7 @@ service:
internalPort: "80"
internalPort1: 8080
nodePort: "60"
- nodePort1: "61"
+ nodePort1: "71"
ingress:
enabled: false
diff --git a/kubernetes/common/dgbuilder/templates/deployment.yaml b/kubernetes/common/dgbuilder/templates/deployment.yaml
index b9c90f2681..328e058f5c 100644
--- a/kubernetes/common/dgbuilder/templates/deployment.yaml
+++ b/kubernetes/common/dgbuilder/templates/deployment.yaml
@@ -60,16 +60,16 @@ spec:
mountPath: /opt/app/application.properties
subPath: application.properties
- name: config
- mountPath: /opt/onap/sdnc/dgbuilder/releases/sdnc1.0/conf/svclogic.properties
+ mountPath: /opt/onap/ccsdk/dgbuilder/releases/sdnc1.0/conf/svclogic.properties
subPath: svclogic.properties
- name: config
- mountPath: /opt/onap/sdnc/dgbuilder/svclogic/svclogic.properties
+ mountPath: /opt/onap/ccsdk/dgbuilder/svclogic/svclogic.properties
subPath: svclogic.properties
- name: scripts
- mountPath: /opt/onap/sdnc/dgbuilder/createReleaseDir.sh
+ mountPath: /opt/onap/ccsdk/dgbuilder/createReleaseDir.sh
subPath: createReleaseDir.sh
- name: scripts
- mountPath: /opt/onap/sdnc/dgbuilder/releases/sdnc1.0/customSettings.js
+ mountPath: /opt/onap/ccsdk/dgbuilder/releases/sdnc1.0/customSettings.js
subPath: customSettings.js
resources:
{{ toYaml .Values.resources | indent 12 }}
diff --git a/kubernetes/consul/resources/config/consul-agent-config/aaf-sms-health.json b/kubernetes/consul/resources/config/consul-agent-config/aaf-sms-health.json
new file mode 100644
index 0000000000..965732da5d
--- /dev/null
+++ b/kubernetes/consul/resources/config/consul-agent-config/aaf-sms-health.json
@@ -0,0 +1,17 @@
+{
+ "service": {
+ "name": "Health Check: Secret Management Service (sms)",
+ "check":[
+ {
+ "id" : "aaf-sms-health",
+ "name": "SMS Health Check",
+ "http": "https://aaf-sms.{{ .Release.Namespace }}:10443/v1/sms/healthcheck",
+ "tls_skip_verify": true,
+ "method": "GET",
+ "interval": "20s",
+ "timeout": "5s"
+ }
+ ]
+ }
+}
+
diff --git a/kubernetes/consul/resources/config/consul-agent-config/mr-dmaap-health.json b/kubernetes/consul/resources/config/consul-agent-config/mr-dmaap-health.json
deleted file mode 100644
index 386e226168..0000000000
--- a/kubernetes/consul/resources/config/consul-agent-config/mr-dmaap-health.json
+++ /dev/null
@@ -1,10 +0,0 @@
-{
- "service": {
- "name": "Health Check: Message Router - DMaaP",
- "check": {
- "http": "http://message-router:3904/topics",
- "interval": "30s",
- "timeout": "1s"
- }
- }
-}
diff --git a/kubernetes/consul/resources/config/consul-agent-config/mr-health.json b/kubernetes/consul/resources/config/consul-agent-config/mr-health.json
new file mode 100644
index 0000000000..d8c056f006
--- /dev/null
+++ b/kubernetes/consul/resources/config/consul-agent-config/mr-health.json
@@ -0,0 +1,29 @@
+{
+ "service": {
+ "name": "Health Check: DMaaP",
+ "checks":[
+ {
+ "id": "dmaap",
+ "name": "Health Check: Message Router",
+ "http": "http://message-router:3904/topics",
+ "tls_skip_verify": true,
+ "interval": "30s",
+ "timeout": "1s"
+ },
+ {
+ "id": "mr-zookeeper",
+ "name": "Health Check: Message Router - ZooKeeper",
+ "script": "/consul/scripts/mr-zookeeper-health.sh",
+ "interval": "10s",
+ "timeout": "5s"
+ },
+ {
+ "id": "mr-kafka",
+ "name": "Health Check: Message Router - Kafka",
+ "script": "/consul/scripts/mr-kafka-health.sh",
+ "interval": "30s",
+ "timeout": "5s"
+ }
+ ]
+ }
+}
diff --git a/kubernetes/consul/resources/config/consul-agent-config/mr-kafka-health.json b/kubernetes/consul/resources/config/consul-agent-config/mr-kafka-health.json
deleted file mode 100644
index df3b190726..0000000000
--- a/kubernetes/consul/resources/config/consul-agent-config/mr-kafka-health.json
+++ /dev/null
@@ -1,10 +0,0 @@
-{
- "service": {
- "name": "Health Check: Message Router - Kafka",
- "check": {
- "script": "/consul/scripts/mr-kafka-health.sh",
- "interval": "30s",
- "timeout": "1s"
- }
- }
-}
diff --git a/kubernetes/consul/resources/config/consul-agent-config/mr-zookeeper-health.json b/kubernetes/consul/resources/config/consul-agent-config/mr-zookeeper-health.json
deleted file mode 100644
index 36d295c1ef..0000000000
--- a/kubernetes/consul/resources/config/consul-agent-config/mr-zookeeper-health.json
+++ /dev/null
@@ -1,10 +0,0 @@
-{
- "service": {
- "name": "Health Check: Message Router - ZooKeeper",
- "check": {
- "script": "/consul/scripts/mr-zookeeper-health.sh",
- "interval": "30s",
- "timeout": "1s"
- }
- }
-}
diff --git a/kubernetes/consul/resources/config/consul-agent-config/scripts/mr-kafka-health.sh b/kubernetes/consul/resources/config/consul-agent-config/scripts/mr-kafka-health.sh
index fd4587551d..4be0a24851 100755
--- a/kubernetes/consul/resources/config/consul-agent-config/scripts/mr-kafka-health.sh
+++ b/kubernetes/consul/resources/config/consul-agent-config/scripts/mr-kafka-health.sh
@@ -1,4 +1,4 @@
-kafkapod=$(/consul/bin/kubectl -n {{ include "common.namespace" . }} get pod | grep -o "message-router-kafka-[^[:space:]]*")
+kafkapod=$(/consul/bin/kubectl -n {{ include "common.namespace" . }} get pod | grep -o "[^[:space:]]*-message-router-kafka-[^[:space:]]*")
if [ -n "$kafkapod" ]; then
if /consul/bin/kubectl -n {{ include "common.namespace" . }} exec -it $kafkapod -- ps ef | grep -i kafka; then
echo Success. Kafka process is running. 2>&1
diff --git a/kubernetes/consul/resources/config/consul-agent-config/scripts/mr-zookeeper-health.sh b/kubernetes/consul/resources/config/consul-agent-config/scripts/mr-zookeeper-health.sh
index fb34057bfc..f8f40b9f83 100755
--- a/kubernetes/consul/resources/config/consul-agent-config/scripts/mr-zookeeper-health.sh
+++ b/kubernetes/consul/resources/config/consul-agent-config/scripts/mr-zookeeper-health.sh
@@ -1,6 +1,6 @@
-zkpod=$(/consul/bin/kubectl -n {{ include "common.namespace" . }} get pod | grep -o "message-router-zookeeper-[^[:space:]]*")
+zkpod=$(/consul/bin/kubectl -n {{ include "common.namespace" . }} get pod | grep -o "[^[:space:]]*-message-router-zookeeper-[^[:space:]]*")
if [ -n "$zkpod" ]; then
- if /consul/bin/kubectl -n {{ include "common.namespace" . }} exec -it $zkpod -- ps ef | grep -i zookeeper; then
+ if /consul/bin/kubectl -n {{ include "common.namespace" . }} exec -it $zkpod -- ps aux | grep -i zookeeper; then
echo Success. Zookeeper process is running. 2>&1
exit 0
else
diff --git a/kubernetes/esr/charts/esr-gui/values.yaml b/kubernetes/esr/charts/esr-gui/values.yaml
index 23ac080ab1..610d9eb17b 100644
--- a/kubernetes/esr/charts/esr-gui/values.yaml
+++ b/kubernetes/esr/charts/esr-gui/values.yaml
@@ -23,7 +23,7 @@ global:
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/aai/esr-gui:1.1.0-SNAPSHOT
+image: onap/aai/esr-gui:latest
pullPolicy: Always
msbaddr: msb-iag.{{ include "common.namespace" . }}:80
diff --git a/kubernetes/esr/values.yaml b/kubernetes/esr/values.yaml
index 750bcffa20..baa2155375 100644
--- a/kubernetes/esr/values.yaml
+++ b/kubernetes/esr/values.yaml
@@ -27,7 +27,7 @@ subChartsOnly:
# application image
repository: nexus3.onap.org:10001
-image: onap/aai/esr-server:1.1.0-SNAPSHOT
+image: onap/aai/esr-server:latest
pullPolicy: Always
msbaddr: msb-iag.{{ include "common.namespace" . }}:80
diff --git a/kubernetes/log/charts/log-elasticsearch/values.yaml b/kubernetes/log/charts/log-elasticsearch/values.yaml
index 1b5c14a57c..9e87b99af6 100644
--- a/kubernetes/log/charts/log-elasticsearch/values.yaml
+++ b/kubernetes/log/charts/log-elasticsearch/values.yaml
@@ -48,14 +48,14 @@ affinity: {}
# probe configuration parameters
liveness:
- initialDelaySeconds: 10
+ initialDelaySeconds: 120
periodSeconds: 10
# necessary to disable liveness probe when setting breakpoints
# in debugger so K8s doesn't restart unresponsive container
enabled: true
readiness:
- initialDelaySeconds: 10
+ initialDelaySeconds: 120
periodSeconds: 10
## Persist data to a persitent volume
diff --git a/kubernetes/log/charts/log-kibana/values.yaml b/kubernetes/log/charts/log-kibana/values.yaml
index a133eac1a6..032b60c4b1 100644
--- a/kubernetes/log/charts/log-kibana/values.yaml
+++ b/kubernetes/log/charts/log-kibana/values.yaml
@@ -51,7 +51,7 @@ affinity: {}
# probe configuration parameters
liveness:
- initialDelaySeconds: 180
+ initialDelaySeconds: 300
periodSeconds: 10
timeoutSeconds: 1
# necessary to disable liveness probe when setting breakpoints
@@ -59,7 +59,7 @@ liveness:
enabled: true
readiness:
- initialDelaySeconds: 180
+ initialDelaySeconds: 300
periodSeconds: 10
timeoutSeconds: 1
diff --git a/kubernetes/multicloud/charts/multicloud-ocata/values.yaml b/kubernetes/multicloud/charts/multicloud-ocata/values.yaml
index 4dff580136..593757c5a2 100644
--- a/kubernetes/multicloud/charts/multicloud-ocata/values.yaml
+++ b/kubernetes/multicloud/charts/multicloud-ocata/values.yaml
@@ -32,7 +32,7 @@ config:
msbPort: 80
aai:
port: 8443
- schemaVersion: v11
+ schemaVersion: v13
username: AAI
password: AAI
diff --git a/kubernetes/multicloud/charts/multicloud-vio/values.yaml b/kubernetes/multicloud/charts/multicloud-vio/values.yaml
index a7507e0da2..3b206cc20d 100644
--- a/kubernetes/multicloud/charts/multicloud-vio/values.yaml
+++ b/kubernetes/multicloud/charts/multicloud-vio/values.yaml
@@ -32,7 +32,7 @@ config:
msbPort: 80
aai:
port: 8443
- schemaVersion: v11
+ schemaVersion: v13
username: AAI
password: AAI
diff --git a/kubernetes/multicloud/charts/multicloud-windriver/values.yaml b/kubernetes/multicloud/charts/multicloud-windriver/values.yaml
index 820a6d1199..ef3833ee5d 100644
--- a/kubernetes/multicloud/charts/multicloud-windriver/values.yaml
+++ b/kubernetes/multicloud/charts/multicloud-windriver/values.yaml
@@ -32,7 +32,7 @@ config:
msbPort: 80
aai:
port: 8443
- schemaVersion: v11
+ schemaVersion: v13
username: AAI
password: AAI
diff --git a/kubernetes/multicloud/values.yaml b/kubernetes/multicloud/values.yaml
index 016f64ff46..5d742cbc48 100644
--- a/kubernetes/multicloud/values.yaml
+++ b/kubernetes/multicloud/values.yaml
@@ -36,7 +36,7 @@ config:
logstashPort: 5044
aai:
port: 8443
- schemaVersion: v11
+ schemaVersion: v13
username: AAI
password: AAI
diff --git a/kubernetes/nbi/templates/deployment.yaml b/kubernetes/nbi/templates/deployment.yaml
index 42d610f351..06572492fd 100644
--- a/kubernetes/nbi/templates/deployment.yaml
+++ b/kubernetes/nbi/templates/deployment.yaml
@@ -76,13 +76,13 @@ spec:
- name: NBI_URL
value: "http://nbi.{{ include "common.namespace" . }}:8080/nbi/api/v1"
- name: SDC_HOST
- value: "http://sdc-fe.{{ include "common.namespace" . }}:8080"
+ value: "http://sdc-be.{{ include "common.namespace" . }}:8080"
- name: SDC_HEADER_ECOMPINSTANCEID
value: {{ .Values.config.ecompInstanceId }}
- name: SDC_HEADER_AUTHORIZATION
value: {{ .Values.sdc_authorization }}
- name: AAI_HOST
- value: "http://aai.{{ include "common.namespace" . }}:8443"
+ value: "https://aai.{{ include "common.namespace" . }}:8443"
- name: AAI_HEADER_AUTHORIZATION
value: {{ .Values.aai_authorization }}
- name: SO_HOST
diff --git a/kubernetes/nbi/values.yaml b/kubernetes/nbi/values.yaml
index 5557edf697..709e580d27 100644
--- a/kubernetes/nbi/values.yaml
+++ b/kubernetes/nbi/values.yaml
@@ -97,7 +97,7 @@ service:
type: NodePort
portName: api
name: nbi
- nodePort: 64
+ nodePort: 74
internalPort: 8080
ingress:
diff --git a/kubernetes/onap/resources/environments/dev.yaml b/kubernetes/onap/resources/environments/dev.yaml
index 9b10012879..7a988a532f 100644
--- a/kubernetes/onap/resources/environments/dev.yaml
+++ b/kubernetes/onap/resources/environments/dev.yaml
@@ -78,6 +78,8 @@ msb:
enabled: false
multicloud:
enabled: false
+nbi:
+ enabled: false
policy:
enabled: false
portal:
diff --git a/kubernetes/onap/resources/environments/disable-allcharts.yaml b/kubernetes/onap/resources/environments/disable-allcharts.yaml
index 0f669beaab..8b66dab319 100644
--- a/kubernetes/onap/resources/environments/disable-allcharts.yaml
+++ b/kubernetes/onap/resources/environments/disable-allcharts.yaml
@@ -51,6 +51,8 @@ msb:
enabled: false
multicloud:
enabled: false
+nbi:
+ enabled: false
oof:
enabled: false
policy:
diff --git a/kubernetes/policy/charts/drools/resources/config/opt/policy/config/drools/base.conf b/kubernetes/policy/charts/drools/resources/config/opt/policy/config/drools/base.conf
index 9202b0ada7..1236176514 100644
--- a/kubernetes/policy/charts/drools/resources/config/opt/policy/config/drools/base.conf
+++ b/kubernetes/policy/charts/drools/resources/config/opt/policy/config/drools/base.conf
@@ -78,13 +78,13 @@ DMAAP_SERVERS=message-router
# AAI
-AAI_URL=https://aai.api.simpledemo.openecomp.org:8443
+AAI_URL=https://aai.{{.Release.Namespace}}:8443
AAI_USERNAME=POLICY
AAI_PASSWORD=POLICY
# MSO
-SO_URL=http://mso:8080/ecomp/mso/infra
+SO_URL=http://so.{{.Release.Namespace}}:8080/ecomp/mso/infra
SO_USERNAME=InfraPortalClient
SO_PASSWORD=password1$
diff --git a/kubernetes/policy/charts/drools/resources/config/opt/policy/config/drools/drools-tweaks.sh b/kubernetes/policy/charts/drools/resources/config/opt/policy/config/drools/drools-tweaks.sh
index 54700078c5..218cd8aead 100644
--- a/kubernetes/policy/charts/drools/resources/config/opt/policy/config/drools/drools-tweaks.sh
+++ b/kubernetes/policy/charts/drools/resources/config/opt/policy/config/drools/drools-tweaks.sh
@@ -1,4 +1,4 @@
-#! /bin/bash
+#! /bin/bash -xv
# Copyright © 2017-2018 Amdocs, Bell Canada, AT&T
#
@@ -15,8 +15,58 @@
# limitations under the License.
-${POLICY_HOME}/bin/features enable healthcheck
-${POLICY_HOME}/bin/features enable pooling-dmaap
-${POLICY_HOME}/bin/features enable distributed-locking
+"${POLICY_HOME}"/bin/features enable healthcheck
+"${POLICY_HOME}"/bin/features enable pooling-dmaap
+"${POLICY_HOME}"/bin/features enable distributed-locking
-${POLICY_HOME}/bin/db-migrator -s pooling -o upgrade
+"${POLICY_HOME}"/bin/db-migrator -s pooling -o upgrade
+
+# make sure the PDPD-CONFIGURATION anonymous topic is created
+# so not to lose any configuration updates
+
+echo
+echo "creating PDPD-CONFIGURATION topic"
+echo
+
+curl --silent --connect-timeout 60 -X POST --header "Content-Type: application/json" -d "{}" http://message-router:3904/events/PDPD-CONFIGURATION
+
+echo
+echo "removing PDPD-CONFIGURATION topic dummy message"
+echo
+
+curl --silent --connect-timeout 60 -X GET http://message-router:3904/events/PDPD-CONFIGURATION/1/1?timeout=15000
+
+# for resiliency/scalability scenarios, check to see
+# if there's an amsterdam artifact already deployed
+# by brmsgw. If so, update the amsterdam controller
+# coordinates. In the future, a more sophisticated
+# solution will be put in place, that will required
+# coordination among policy components.
+
+echo
+echo "checking if there are amsterdam policies already deployed .."
+echo
+
+AMSTERDAM_VERSION=$(curl --silent --connect-timeout 60 -X GET "http://nexus:8081/nexus/service/local/artifact/maven/resolve?r=releases&g=org.onap.policy-engine.drools.amsterdam&a=policy-amsterdam-rules&v=RELEASE" | grep -Po "(?<=<version>).*(?=</version>)")
+
+if [[ -z ${AMSTERDAM_VERSION} ]]; then
+ echo "no amsterdam policies have been found .."
+ exit 0
+fi
+
+echo
+echo "The latest deployed amsterdam artifact in nexus has version ${AMSTERDAM_VERSION}"
+echo
+
+sed -i.INSTALL -e "s/^rules.artifactId=.*/rules.artifactId=policy-amsterdam-rules/g" \
+ -e "s/^rules.groupId=.*/rules.groupId=org.onap.policy-engine.drools.amsterdam/g" \
+ -e "s/^rules.version=.*/rules.version=${AMSTERDAM_VERSION}/g" "${POLICY_HOME}"/config/amsterdam-controller.properties
+
+echo
+echo "amsterdam controller will be started brained with maven coordinates:"
+echo
+
+grep "^rules" "${POLICY_HOME}"/config/amsterdam-controller.properties
+
+echo
+echo
diff --git a/kubernetes/policy/charts/drools/templates/statefulset.yaml b/kubernetes/policy/charts/drools/templates/statefulset.yaml
index 68b415e989..71844f108d 100644
--- a/kubernetes/policy/charts/drools/templates/statefulset.yaml
+++ b/kubernetes/policy/charts/drools/templates/statefulset.yaml
@@ -51,10 +51,6 @@ spec:
image: "{{ .Values.global.readinessRepository }}/{{ .Values.global.readinessImage }}"
imagePullPolicy: {{ .Values.global.pullPolicy | default .Values.pullPolicy }}
name: {{ include "common.name" . }}-readiness
- hostAliases:
- - ip: "{{ .Values.config.aaiServiceClusterIp }}"
- hostnames:
- - "aai.api.simpledemo.openecomp.org"
containers:
- name: {{ include "common.name" . }}
image: "{{ include "common.repository" . }}/{{ .Values.image }}"
diff --git a/kubernetes/policy/charts/drools/values.yaml b/kubernetes/policy/charts/drools/values.yaml
index 9ce804032e..ca62f2b01d 100644
--- a/kubernetes/policy/charts/drools/values.yaml
+++ b/kubernetes/policy/charts/drools/values.yaml
@@ -36,10 +36,9 @@ debugEnabled: false
# application configuration
config:
nexusPort: 8081
- aaiServiceClusterIp: 10.43.255.254
# default number of instances
-replicaCount: 4
+replicaCount: 2
nodeSelector: {}
diff --git a/kubernetes/policy/charts/pdp/resources/config/pe/pdplp.conf b/kubernetes/policy/charts/pdp/resources/config/pe/pdplp.conf
index f5f4060aad..7f894fd4a9 100644
--- a/kubernetes/policy/charts/pdp/resources/config/pe/pdplp.conf
+++ b/kubernetes/policy/charts/pdp/resources/config/pe/pdplp.conf
@@ -18,8 +18,8 @@ LOGPARSER_X_MS_MB=1024
LOGPARSER_X_MX_MB=1024
SERVER=http://{{ include "common.servicename" . }}:{{.Values.service.externalPort}}/pdp/
-LOGPATH=/opt/app/policy/servers/pdp/logs/pdp-rest.log
-PARSERLOGPATH=IntegrityMonitor.log
+LOGPATH=/var/log/onap/policy/pdpx/pdp-rest.log
+PARSERLOGPATH=/opt/app/policy/servers/pdplp/bin/IntegrityMonitor.log
node_type=logparser
# the java property is RESOURCE_NAME (uppercase), but the conf parameter is lowercase
diff --git a/kubernetes/policy/resources/config/pe/paplp.conf b/kubernetes/policy/resources/config/pe/paplp.conf
index 205497c35d..3d59d818c9 100644
--- a/kubernetes/policy/resources/config/pe/paplp.conf
+++ b/kubernetes/policy/resources/config/pe/paplp.conf
@@ -18,8 +18,8 @@ LOGPARSER_X_MS_MB=1024
LOGPARSER_X_MX_MB=1024
SERVER=http://{{ include "common.servicename" . }}:{{.Values.service.externalPort2}}/pap/
-LOGPATH=/opt/app/policy/servers/pap/logs/pap-rest.log
-PARSERLOGPATH=IntegrityMonitor.log
+LOGPATH=/var/log/onap/policy/pap/pap-rest.log
+PARSERLOGPATH=/opt/app/policy/servers/paplp/bin/IntegrityMonitor.log
node_type=logparser
# the java property is RESOURCE_NAME (uppercase), but the conf parameter is lowercase
diff --git a/kubernetes/portal/charts/portal-mariadb/values.yaml b/kubernetes/portal/charts/portal-mariadb/values.yaml
index ae5849eb27..fc4836bf07 100644
--- a/kubernetes/portal/charts/portal-mariadb/values.yaml
+++ b/kubernetes/portal/charts/portal-mariadb/values.yaml
@@ -77,14 +77,14 @@ affinity: {}
# probe configuration parameters
liveness:
- initialDelaySeconds: 300
+ initialDelaySeconds: 450
periodSeconds: 10
# necessary to disable liveness probe when setting breakpoints
# in debugger so K8s doesn't restart unresponsive container
enabled: true
readiness:
- initialDelaySeconds: 300
+ initialDelaySeconds: 450
periodSeconds: 10
## Persist data to a persitent volume