diff options
author | mpriyank <priyank.maheshwari@est.tech> | 2024-01-23 15:48:07 +0000 |
---|---|---|
committer | mpriyank <priyank.maheshwari@est.tech> | 2024-01-25 16:02:25 +0000 |
commit | 33f7b3ef7b4f32ed33abd0890c71b48de2a17625 (patch) | |
tree | 853793452dd986024ac45dc428ce590b924e2c7d /cps-service/src/test | |
parent | 52a7c2b2685ad63154e347782f0ff5c16f87ef53 (diff) |
CloudEvents support for cps-core
- Moving cloud events dependencies to the cps-service module instead of
cps-ncmp-service
- Testware has moved and new tests also introduced
- EventsPublisher also moved
Issue-ID: CPS-2040
Change-Id: Ibafb15e7e9efbdbe1e00e2b4f0da820bbcead004
Signed-off-by: mpriyank <priyank.maheshwari@est.tech>
Diffstat (limited to 'cps-service/src/test')
-rw-r--r-- | cps-service/src/test/groovy/org/onap/cps/events/EventsPublisherSpec.groovy | 194 |
1 files changed, 194 insertions, 0 deletions
diff --git a/cps-service/src/test/groovy/org/onap/cps/events/EventsPublisherSpec.groovy b/cps-service/src/test/groovy/org/onap/cps/events/EventsPublisherSpec.groovy new file mode 100644 index 0000000000..fe67287dec --- /dev/null +++ b/cps-service/src/test/groovy/org/onap/cps/events/EventsPublisherSpec.groovy @@ -0,0 +1,194 @@ +/* + * ============LICENSE_START======================================================= + * Copyright (C) 2024 Nordix Foundation + * ================================================================================ + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + * + * SPDX-License-Identifier: Apache-2.0 + * ============LICENSE_END========================================================= + */ + +package org.onap.cps.events + +import ch.qos.logback.classic.Level +import ch.qos.logback.classic.Logger +import ch.qos.logback.classic.spi.ILoggingEvent +import ch.qos.logback.core.read.ListAppender +import io.cloudevents.CloudEvent +import org.apache.kafka.clients.producer.ProducerRecord +import org.apache.kafka.clients.producer.RecordMetadata +import org.apache.kafka.common.TopicPartition +import org.apache.kafka.common.header.Headers +import org.apache.kafka.common.header.internals.RecordHeader +import org.apache.kafka.common.header.internals.RecordHeaders +import org.junit.jupiter.api.AfterEach +import org.junit.jupiter.api.BeforeEach +import org.slf4j.LoggerFactory +import org.springframework.kafka.core.KafkaTemplate +import org.springframework.kafka.support.SendResult +import org.springframework.util.SerializationUtils +import spock.lang.Specification + +import java.util.concurrent.CompletableFuture + +class EventsPublisherSpec extends Specification { + + def legacyKafkaTemplateMock = Mock(KafkaTemplate) + def mockCloudEventKafkaTemplate = Mock(KafkaTemplate) + def logger = Spy(ListAppender<ILoggingEvent>) + + void setup() { + def setupLogger = ((Logger) LoggerFactory.getLogger(EventsPublisher.class)) + setupLogger.setLevel(Level.DEBUG) + setupLogger.addAppender(logger) + logger.start() + } + + void cleanup() { + ((Logger) LoggerFactory.getLogger(EventsPublisher.class)).detachAndStopAllAppenders() + } + + def objectUnderTest = new EventsPublisher(legacyKafkaTemplateMock, mockCloudEventKafkaTemplate) + + def 'Publish Cloud Event'() { + given: 'a successfully published event' + def eventFuture = CompletableFuture.completedFuture( + new SendResult( + new ProducerRecord('some-topic', 'some-value'), + new RecordMetadata(new TopicPartition('some-topic', 0), 0, 0, 0, 0, 0) + ) + ) + def someCloudEvent = Mock(CloudEvent) + 1 * mockCloudEventKafkaTemplate.send('some-topic', 'some-event-key', someCloudEvent) >> eventFuture + when: 'publishing the cloud event' + objectUnderTest.publishCloudEvent('some-topic', 'some-event-key', someCloudEvent) + then: 'the correct debug message is logged' + def lastLoggingEvent = logger.list[0] + assert lastLoggingEvent.level == Level.DEBUG + assert lastLoggingEvent.formattedMessage.contains('Successfully published event') + } + + def 'Publish Cloud Event with Exception'() { + given: 'a failed event' + def eventFutureWithFailure = new CompletableFuture<SendResult<String, String>>() + eventFutureWithFailure.completeExceptionally(new RuntimeException('some exception')) + def someCloudEvent = Mock(CloudEvent) + 1 * mockCloudEventKafkaTemplate.send('some-topic', 'some-event-key', someCloudEvent) >> eventFutureWithFailure + when: 'publishing the cloud event' + objectUnderTest.publishCloudEvent('some-topic', 'some-event-key', someCloudEvent) + then: 'the correct error message is logged' + def lastLoggingEvent = logger.list[0] + assert lastLoggingEvent.level == Level.ERROR + assert lastLoggingEvent.formattedMessage.contains('Unable to publish event') + } + + def 'Publish Legacy Event'() { + given: 'a successfully published event' + def eventFuture = CompletableFuture.completedFuture( + new SendResult( + new ProducerRecord('some-topic', 'some-value'), + new RecordMetadata(new TopicPartition('some-topic', 0), 0, 0, 0, 0, 0) + ) + ) + def someEvent = Mock(Object) + 1 * legacyKafkaTemplateMock.send('some-topic', 'some-event-key', someEvent) >> eventFuture + when: 'publishing the cloud event' + objectUnderTest.publishEvent('some-topic', 'some-event-key', someEvent) + then: 'the correct debug message is logged' + def lastLoggingEvent = logger.list[0] + assert lastLoggingEvent.level == Level.DEBUG + assert lastLoggingEvent.formattedMessage.contains('Successfully published event') + } + + def 'Publish Legacy Event with Headers as Map'() { + given: 'a successfully published event' + def sampleEventHeaders = ['k1': SerializationUtils.serialize('v1')] + def eventFuture = CompletableFuture.completedFuture( + new SendResult( + new ProducerRecord('some-topic', 'some-value'), + new RecordMetadata(new TopicPartition('some-topic', 0), 0, 0, 0, 0, 0) + ) + ) + def someEvent = Mock(Object.class) + when: 'publishing the legacy event' + objectUnderTest.publishEvent('some-topic', 'some-event-key', sampleEventHeaders, someEvent) + then: 'event is published' + 1 * legacyKafkaTemplateMock.send(_) >> eventFuture + and: 'the correct debug message is logged' + def lastLoggingEvent = logger.list[0] + assert lastLoggingEvent.level == Level.DEBUG + assert lastLoggingEvent.formattedMessage.contains('Successfully published event') + } + + def 'Publish Legacy Event with Record Headers'() { + given: 'a successfully published event' + def sampleEventHeaders = new RecordHeaders([new RecordHeader('k1', SerializationUtils.serialize('v1'))]) + def sampleProducerRecord = new ProducerRecord('some-topic', null, 'some-key', 'some-value', sampleEventHeaders) + def eventFuture = CompletableFuture.completedFuture( + new SendResult( + sampleProducerRecord, + new RecordMetadata(new TopicPartition('some-topic', 0), 0, 0, 0, 0, 0) + ) + ) + def someEvent = Mock(Object.class) + when: 'publishing the legacy event' + objectUnderTest.publishEvent('some-topic', 'some-event-key', sampleEventHeaders, someEvent) + then: 'event is published' + 1 * legacyKafkaTemplateMock.send(_) >> eventFuture + and: 'the correct debug message is logged' + def lastLoggingEvent = logger.list[0] + assert lastLoggingEvent.level == Level.DEBUG + assert lastLoggingEvent.formattedMessage.contains('Successfully published event') + } + + def 'Handle Legacy Event Callback'() { + given: 'an event is successfully published' + def eventFuture = CompletableFuture.completedFuture( + new SendResult( + new ProducerRecord('some-topic', 'some-value'), + new RecordMetadata(new TopicPartition('some-topic', 0), 0, 0, 0, 0, 0) + ) + ) + when: 'handling legacy event callback' + objectUnderTest.handleLegacyEventCallback('some-topic', eventFuture) + then: 'the correct debug message is logged' + def lastLoggingEvent = logger.list[0] + assert lastLoggingEvent.level == Level.DEBUG + assert lastLoggingEvent.formattedMessage.contains('Successfully published event') + } + + def 'Handle Legacy Event Callback with Exception'() { + given: 'a failure to publish an event' + def eventFutureWithFailure = new CompletableFuture<SendResult<String, String>>() + eventFutureWithFailure.completeExceptionally(new RuntimeException('some exception')) + when: 'handling legacy event callback' + objectUnderTest.handleLegacyEventCallback('some-topic', eventFutureWithFailure) + then: 'the correct error message is logged' + def lastLoggingEvent = logger.list[0] + assert lastLoggingEvent.level == Level.ERROR + assert lastLoggingEvent.formattedMessage.contains('Unable to publish event') + } + + def 'Convert to kafka headers'() { + given: 'Few key value pairs' + def someKeyValue = ['key1': 'value1', 'key2': 'value2'] + when: 'we convert to headers' + def headers = objectUnderTest.convertToKafkaHeaders(someKeyValue) + then: 'it is correctly converted' + assert headers instanceof Headers + and: 'also has correct values' + assert headers[0].key() == 'key1' + assert headers[1].key() == 'key2' + } + +}
\ No newline at end of file |