diff options
Diffstat (limited to 'src/main/java/com/att')
5 files changed, 45 insertions, 11 deletions
diff --git a/src/main/java/com/att/nsa/cambria/backends/kafka/KafkaConsumer.java b/src/main/java/com/att/nsa/cambria/backends/kafka/KafkaConsumer.java index c0ba42d..692f093 100644 --- a/src/main/java/com/att/nsa/cambria/backends/kafka/KafkaConsumer.java +++ b/src/main/java/com/att/nsa/cambria/backends/kafka/KafkaConsumer.java @@ -142,7 +142,7 @@ public class KafkaConsumer implements Consumer { } catch (kafka.consumer.ConsumerTimeoutException x) { log.error(fLogTag + ": ConsumerTimeoutException in Kafka consumer; returning null & Exception at nextMessage() : " + x); } catch (java.lang.IllegalStateException x) { - log.error(fLogTag + ": Illegal state exception in Kafka consumer; dropping stream. " + x.getMessage()); + log.error(fLogTag + ": Error found next() at : " + x); } return null; diff --git a/src/main/java/com/att/nsa/cambria/backends/kafka/KafkaConsumerCache.java b/src/main/java/com/att/nsa/cambria/backends/kafka/KafkaConsumerCache.java index ceb16c6..178d143 100644 --- a/src/main/java/com/att/nsa/cambria/backends/kafka/KafkaConsumerCache.java +++ b/src/main/java/com/att/nsa/cambria/backends/kafka/KafkaConsumerCache.java @@ -288,9 +288,9 @@ public class KafkaConsumerCache { curatorConsumerCache.close(); log.info("Curator client closed"); } catch (ZkInterruptedException e) { - log.warn("Curator client close interrupted: " + e.getMessage()); + log.error("Curator client close interrupted: " + e); } catch (IOException e) { - log.warn("Error while closing curator PathChildrenCache for KafkaConsumerCache" + e.getMessage()); + log.error("Error while closing curator PathChildrenCache for KafkaConsumerCache" + e); } curatorConsumerCache = null; @@ -400,8 +400,7 @@ public class KafkaConsumerCache { try { curator.delete().guaranteed().forPath(fBaseZkPath + "/" + key); } catch (NoNodeException e) { - log.warn("A consumer was deleted from " + fApiId - + "'s cache, but no Cambria API node had ownership of it in ZooKeeper"); + log.error("Exception at : " + e); } catch (Exception e) { log.debug("Unexpected exception while deleting consumer: " + e.getMessage()); } diff --git a/src/main/java/com/att/nsa/cambria/beans/DMaaPKafkaMetaBroker.java b/src/main/java/com/att/nsa/cambria/beans/DMaaPKafkaMetaBroker.java index 9d53ef2..e7d777e 100644 --- a/src/main/java/com/att/nsa/cambria/beans/DMaaPKafkaMetaBroker.java +++ b/src/main/java/com/att/nsa/cambria/beans/DMaaPKafkaMetaBroker.java @@ -170,9 +170,7 @@ public class DMaaPKafkaMetaBroker implements Broker { // make Kafka aware of the // topic creation. (Otherwise, the topic is only partially // created in ZK.) - zkClient = new ZkClient(ConfigurationReader.getMainZookeeperConnectionString(), 10000, 10000, - ZKStringSerializer$.MODULE$); - + zkClient = ZkClientFactory.createZkClient(); log.info("Zookeeper client loaded successfully. Creating topic."); AdminUtils.createTopic(zkClient, topic, partitions, replicas, new Properties()); } catch (kafka.common.TopicExistsException e) { @@ -217,8 +215,7 @@ public class DMaaPKafkaMetaBroker implements Broker { // Kafka aware of the // topic creation. (Otherwise, the topic is only partially created // in ZK.) - zkClient = new ZkClient(ConfigurationReader.getMainZookeeperConnectionString(), 10000, 10000, - ZKStringSerializer$.MODULE$); + zkClient = ZkClientFactory.createZkClient(); log.info("Zookeeper client loaded successfully. Deleting topic."); AdminUtils.deleteTopic(zkClient, topic); @@ -245,6 +242,8 @@ public class DMaaPKafkaMetaBroker implements Broker { // throw new UnsupportedOperationException ( "We can't programmatically // delete Kafka topics yet." ); } + + //private final rrNvReadable fSettings; private final ZkClient fZk; diff --git a/src/main/java/com/att/nsa/cambria/beans/ZkClientFactory.java b/src/main/java/com/att/nsa/cambria/beans/ZkClientFactory.java new file mode 100644 index 0000000..2aedb95 --- /dev/null +++ b/src/main/java/com/att/nsa/cambria/beans/ZkClientFactory.java @@ -0,0 +1,36 @@ +/*- + * ============LICENSE_START======================================================= + * ONAP Policy Engine + * ================================================================================ + * Copyright (C) 2017 AT&T Intellectual Property. All rights reserved. + * ================================================================================ + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + * ============LICENSE_END========================================================= + */ +package com.att.nsa.cambria.beans; + +import org.I0Itec.zkclient.ZkClient; + +import com.att.nsa.cambria.utils.ConfigurationReader; + +import kafka.utils.ZKStringSerializer$; + +public class ZkClientFactory { + + public static ZkClient createZkClient(){ + return new ZkClient(ConfigurationReader.getMainZookeeperConnectionString(), 10000, 10000, + ZKStringSerializer$.MODULE$); + + } + +} diff --git a/src/main/java/com/att/nsa/cambria/service/impl/MMServiceImpl.java b/src/main/java/com/att/nsa/cambria/service/impl/MMServiceImpl.java index 4f6a9a1..7fe1eac 100644 --- a/src/main/java/com/att/nsa/cambria/service/impl/MMServiceImpl.java +++ b/src/main/java/com/att/nsa/cambria/service/impl/MMServiceImpl.java @@ -214,7 +214,7 @@ public class MMServiceImpl implements MMService { try { coes.write(baos); } catch (Exception ex) { - + LOG.error("Exception at subscribe(.......): " + ex); } c.commitOffsets(); |