aboutsummaryrefslogtreecommitdiffstats
path: root/docker_files/__MsgRtrApi.properties
blob: 47293a72a207b20a67073a62abcab2524d8666c4 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
###############################################################################
##
## Cambria API Server config
##
##	- Default values are shown as commented settings.
##

###############################################################################
##
## HTTP service
##
##		- 3904 is standard as of 7/29/14.
#
## Zookeeper Connection
##
##	Both Cambria and Kafka make use of Zookeeper.
##
config.zk.servers=zookeeper:2181
#config.zk.servers={{ ip.docker0 }}:2181
#10.0.11.1:2181
#10.208.128.229:2181
#config.zk.root=/fe3c/cambria/config


###############################################################################
##
## Kafka Connection
##
##	Items below are passed through to Kafka's producer and consumer
##	configurations (after removing "kafka.")
##	if you want to change request.required.acks it can take this one value
#kafka.metadata.broker.list=localhost:9092,localhost:9093
kafka.metadata.broker.list=kafka:9092
#kafka.metadata.broker.list={{ ip.docker0 }}:9092
#10.0.11.1:9092
#10.208.128.229:9092
##kafka.request.required.acks=-1
#kafka.client.zookeeper=${config.zk.servers}
consumer.timeout.ms=100
zookeeper.connection.timeout.ms=6000
zookeeper.session.timeout.ms=6000
zookeeper.sync.time.ms=2000
auto.commit.interval.ms=1000
fetch.message.max.bytes =1000000
auto.commit.enable=false


###############################################################################
##
##	Secured Config
##
##	Some data stored in the config system is sensitive -- API keys and secrets,
##	for example. to protect it, we use an encryption layer for this section
##	of the config.
##
## The key is a base64 encode AES key. This must be created/configured for
## each installation.
#cambria.secureConfig.key=
##
## The initialization vector is a 16 byte value specific to the secured store.
## This must be created/configured for each installation.
#cambria.secureConfig.iv=

## Southfield Sandbox
cambria.secureConfig.key=b/7ouTn9FfEw2PQwL0ov/Q==
cambria.secureConfig.iv=wR9xP5k5vbz/xD0LmtqQLw==
authentication.adminSecret=fe3cCompound
#cambria.secureConfig.key[pc569h]=YT3XPyxEmKCTLI2NK+Sjbw==
#cambria.secureConfig.iv[pc569h]=rMm2jhR3yVnU+u2V9Ugu3Q==


###############################################################################
##
## Consumer Caching
##
##	Kafka expects live connections from the consumer to the broker, which
##	obviously doesn't work over connectionless HTTP requests. The Cambria
##	server proxies HTTP requests into Kafka consumer sessions that are kept
##	around for later re-use. Not doing so is costly for setup per request,
##	which would substantially impact a high volume consumer's performance.
##
##	This complicates Cambria server failover, because we often need server
##	A to close its connection before server B brings up the replacement.    
##

## The consumer cache is normally enabled.
#cambria.consumer.cache.enabled=true

## Cached consumers are cleaned up after a period of disuse. The server inspects
## consumers every sweepFreqSeconds and will clean up any connections that are
## dormant for touchFreqMs.
#cambria.consumer.cache.sweepFreqSeconds=15
#cambria.consumer.cache.touchFreqMs=120000

## The cache is managed through ZK. The default value for the ZK connection
## string is the same as config.zk.servers.
#cambria.consumer.cache.zkConnect=${config.zk.servers}

##
## Shared cache information is associated with this node's name. The default
## name is the hostname plus the HTTP service port this host runs on. (The
## hostname is determined via InetAddress.getLocalHost ().getCanonicalHostName(),
## which is not always adequate.) You can set this value explicitly here.
##
#cambria.api.node.identifier=<use-something-unique-to-this-instance>

###############################################################################
##
## Metrics Reporting
##
##	This server can report its metrics periodically on a topic.
##
#metrics.send.cambria.enabled=true
#metrics.send.cambria.topic=cambria.apinode.metrics                                  #msgrtr.apinode.metrics.dmaap 
#metrics.send.cambria.sendEverySeconds=60

cambria.consumer.cache.zkBasePath=/fe3c/cambria/consumerCache

##############################################################################
#100mb
maxcontentlength=10000


##############################################################################
#AAF Properties
msgRtr.namespace.aaf=org.openecomp.dcae.dmaap.mtnje2.mr.topic
msgRtr.topicfactory.aaf=org.openecomp.dcae.dmaap.topicFactory|:org.openecomp.dcae.dmaap.mtnje2.mr.topic:
enforced.topic.name.AAF=org.openecomp
forceAAF=false
transidUEBtopicreqd=false
defaultNSforUEB=org.openecomp.dmaap.mr.ueb
##############################################################################
#Mirror Maker Agent
msgRtr.mirrormakeradmin.aaf=org.openecomp.dmaap.mr.dev.mirrormaker|*|admin
msgRtr.mirrormakeruser.aaf=org.openecomp.dmaap.mr.dev.mirrormaker|*|user
msgRtr.mirrormakeruser.aaf.create=org.openecomp.dmaap.mr.dev.topicFactory|:org.openecomp.dmaap.mr.dev.topic:
msgRtr.mirrormaker.timeout=15000
msgRtr.mirrormaker.topic=org.openecomp.dmaap.mr.prod.mm.agent
msgRtr.mirrormaker.consumergroup=mmagentserver
msgRtr.mirrormaker.consumerid=1