diff options
author | Vijay Venkatesh Kumar <vv770d@att.com> | 2019-01-17 13:30:45 +0000 |
---|---|---|
committer | Gerrit Code Review <gerrit@onap.org> | 2019-01-17 13:30:45 +0000 |
commit | 142235dd389cf085a670aadcffdd9252ea1b69e2 (patch) | |
tree | fc6268a421362fe96d2d57467b8e96fad9060faf | |
parent | 0632c60c2a0eb0800a709e3e2e2917f7e25c3e36 (diff) | |
parent | cf9b6e19f41aa8c689db281154d3cae373697d8d (diff) |
Merge "Update HV-VES docs troubleshooting guide"
-rw-r--r-- | docs/sections/services/ves-hv/troubleshooting.rst | 727 |
1 files changed, 505 insertions, 222 deletions
diff --git a/docs/sections/services/ves-hv/troubleshooting.rst b/docs/sections/services/ves-hv/troubleshooting.rst index 722e64c3..3defe7f1 100644 --- a/docs/sections/services/ves-hv/troubleshooting.rst +++ b/docs/sections/services/ves-hv/troubleshooting.rst @@ -6,6 +6,34 @@ Troubleshooting =============== +**NOTE** + +According to **ONAP** logging policy, **HV-VES** logs contain all required markers as well as service and client specific Mapped Diagnostic Context (later referred as MDC) + +Default console log pattern: + +:: + + | %date{"yyyy-MM-dd'T'HH:mm:ss.SSSXXX", UTC}\t| %thread\t| %highlight(%-5level)\t| %msg\t| %marker\t| %rootException\t| %mdc\t| %thread + +A sample, fully qualified message implementing this pattern: + +:: + + | 2018-12-18T13:12:44.369Z | p.dcae.collectors.veshv.impl.socket.NettyTcpServer | DEBUG | Client connection request received | ENTRY | | RequestID=d7762b18-854c-4b8c-84aa-95762c6f8e62, InstanceID=9b9799ca-33a5-4f61-ba33-5c7bf7e72d07, InvocationID=b13d34ba-e1cd-4816-acda-706415308107, PartnerName=C=PL, ST=DL, L=Wroclaw, O=Nokia, OU=MANO, CN=dcaegen2-hvves-client, StatusCode=INPROGRESS, ClientIPAddress=192.168.0.9, ServerFQDN=a4ca8f96c7e5 | reactor-tcp-nio-2 + + +For simplicity, all log messages in this section are shortened to contain only: + * logger name + * log level + * message + +Error and warning logs contain also: + * exception message + * stack trace + +**Do not rely on exact log messages or their presence, as they are often subject to change.** + Deployment/Installation errors ------------------------------ @@ -13,37 +41,39 @@ Deployment/Installation errors :: - Unexpected error when parsing command line arguments - usage: java org.onap.dcae.collectors.veshv.main.MainKt - Required parameters: p, c - -c,--config-url <arg> URL of ves configuration on consul - -d,--first-request-delay <arg> Delay of first request to consul in - seconds - -H,--health-check-api-port <arg> Health check rest api listen port - -I,--request-interval <arg> Interval of consul configuration - requests in seconds - -i,--idle-timeout-sec <arg> Idle timeout for remote hosts. After - given time without any data exchange - the - connection might be closed. - -k,--key-store <arg> Key store in PKCS12 format - -kp,--key-store-password <arg> Key store password - -l,--ssl-disable Disable SSL encryption - -m,--max-payload-size <arg> Maximum supported payload size in - bytes - -p,--listen-port <arg> Listen port - -t,--trust-store <arg> File with trusted certificate bundle - in PKCS12 format - -tp,--trust-store-password <arg> Trust store password - -u,--dummy If present will start in dummy mode - (dummy external services) - All parameters can be specified as environment variables using - upper-snake-case full name with prefix `VESHV_`. + Unexpected error when parsing command line arguments + usage: java org.onap.dcae.collectors.veshv.main.MainKt + Required parameters: s, p, c + -c,--config-url <arg> URL of ves configuration on consul + -d,--first-request-delay <arg> Delay of first request to consul in + seconds + -H,--health-check-api-port <arg> Health check rest api listen port + -I,--request-interval <arg> Interval of consul configuration + requests in seconds + -i,--idle-timeout-sec <arg> Idle timeout for remote hosts. After + given time without any data exchange + the + connection might be closed. + -k,--key-store <arg> Key store in PKCS12 format + -kp,--key-store-password <arg> Key store password + -l,--ssl-disable Disable SSL encryption + -m,--max-payload-size <arg> Maximum supported payload size in + bytes + -p,--listen-port <arg> Listen port + -s,--kafka-bootstrap-servers <arg> Comma-separated Kafka bootstrap + servers in <host>:<port> format + -t,--trust-store <arg> File with trusted certificate bundle + in PKCS12 format + -tp,--trust-store-password <arg> Trust store password + -u,--dummy If present will start in dummy mode + (dummy external services) + All parameters can be specified as environment variables using + upper-snake-case full name with prefix `VESHV_`. This log message is printed when you do not specify the required parameters (via command line, or in environment variables). As described in the above log message, there are a few required parameters: -**listen port**, **config url** and either **trust store password** and **key store password** if you want to use SSL, or only **ssl disable** if not. +**listen port**, **config url**, **kafka-bootstrap-servers** and either **trust store password** and **key store password** if you want to use SSL, or only **ssl disable** if not. To get rid of this error, specify the required parameters. For example: @@ -51,7 +81,7 @@ To get rid of this error, specify the required parameters. For example: :: - <hv-ves run command> --listen-port 6061 --config-url http://consul-url/key-path --key-store-password password --trust-store-password password + <hv-ves run command> --listen-port 6061 --config-url http://consul-url/key-path --kafka-bootstrap-servers message-router-kafka:9092 --key-store-password password --trust-store-password password - By defining environment variables: @@ -59,6 +89,7 @@ To get rid of this error, specify the required parameters. For example: export VESHV_LISTEN_PORT=6061 export VESHV_CONFIG_URL=http://consul-url/key-path + export VESHV_KAFKA_BOOTSTRAP_SERVERS=message-router-kafka:9092 export VESHV_KEY_STORE_PASSWORD=password export VESHV_TRUST_STORE_PASSWORD=password @@ -75,39 +106,86 @@ Configuration errors :: - ap.dcae.collectors.veshv.impl.adapters.HttpAdapter | 2018-10-16T13:13:01.155Z | ERROR | Failed to get resource on path: http://localhost:8500/v1/hv/veshv-config (Connection refused: localhost/127.0.0.1:8500) | | reactor-http-client-epoll-8 - ap.dcae.collectors.veshv.impl.adapters.HttpAdapter | 2018-10-16T13:13:01.155Z | DEBUG | Nested exception: | java.net.ConnectException: Connection refused - ... 10 common frames omitted - Wrapped by: io.netty.channel.AbstractChannel$AnnotatedConnectException: Connection refused: localhost/127.0.0.1:8500 - at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) - at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717) - at io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect(NioSocketChannel.java:327) - at io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:340) - at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:616) - at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:563) - at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:480) - at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:442) + | ap.dcae.collectors.veshv.impl.adapters.HttpAdapter | ERROR | Failed to get resource on path: http://invalid-host:8500/v1/kv/veshv-config?raw=true (consul-server1: Temporary failure in name resolution) + | ap.dcae.collectors.veshv.impl.adapters.HttpAdapter | DEBUG | Nested exception: | | java.net.UnknownHostException: consul-server1: Temporary failure in name resolution + at java.base/java.net.Inet4AddressImpl.lookupAllHostAddr(Native Method) + at java.base/java.net.InetAddress$PlatformNameService.lookupAllHostAddr(InetAddress.java:929) + at java.base/java.net.InetAddress.getAddressesFromNameService(InetAddress.java:1515) + at java.base/java.net.InetAddress$NameServiceAddresses.get(InetAddress.java:848) + at java.base/java.net.InetAddress.getAllByName0(InetAddress.java:1505) + at java.base/java.net.InetAddress.getAllByName(InetAddress.java:1364) + at java.base/java.net.InetAddress.getAllByName(InetAddress.java:1298) + at java.base/java.net.InetAddress.getByName(InetAddress.java:1248) + at io.netty.util.internal.SocketUtils$8.run(SocketUtils.java:146) + at io.netty.util.internal.SocketUtils$8.run(SocketUtils.java:143) + at java.base/java.security.AccessController.doPrivileged(Native Method) + at io.netty.util.internal.SocketUtils.addressByName(SocketUtils.java:143) + at io.netty.resolver.DefaultNameResolver.doResolve(DefaultNameResolver.java:43) + at io.netty.resolver.SimpleNameResolver.resolve(SimpleNameResolver.java:63) + at io.netty.resolver.SimpleNameResolver.resolve(SimpleNameResolver.java:55) + at io.netty.resolver.InetSocketAddressResolver.doResolve(InetSocketAddressResolver.java:57) + at io.netty.resolver.InetSocketAddressResolver.doResolve(InetSocketAddressResolver.java:32) + at io.netty.resolver.AbstractAddressResolver.resolve(AbstractAddressResolver.java:108) + at io.netty.bootstrap.Bootstrap.doResolveAndConnect0(Bootstrap.java:208) + at io.netty.bootstrap.Bootstrap.access$000(Bootstrap.java:49) + at io.netty.bootstrap.Bootstrap$1.operationComplete(Bootstrap.java:188) + at io.netty.bootstrap.Bootstrap$1.operationComplete(Bootstrap.java:174) + at io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:511) + at io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:485) + at io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:424) + at io.netty.util.concurrent.DefaultPromise.trySuccess(DefaultPromise.java:103) + at io.netty.channel.DefaultChannelPromise.trySuccess(DefaultChannelPromise.java:84) + at io.netty.channel.AbstractChannel$AbstractUnsafe.safeSetSuccess(AbstractChannel.java:978) + at io.netty.channel.AbstractChannel$AbstractUnsafe.register0(AbstractChannel.java:512) + at io.netty.channel.AbstractChannel$AbstractUnsafe.access$200(AbstractChannel.java:423) + at io.netty.channel.AbstractChannel$AbstractUnsafe$1.run(AbstractChannel.java:482) + at io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java:163) + at io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:404) + at io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:315) at io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:884) - at java.lang.Thread.run(Thread.java:748) - | reactor-http-client-epoll-8 - rs.veshv.impl.adapters.ConsulConfigurationProvider | 2018-10-16T13:13:01.163Z | WARN | Could not get fresh configuration | java.net.ConnectException: Connection refused - ... 10 common frames omitted - Wrapped by: io.netty.channel.AbstractChannel$AnnotatedConnectException: Connection refused: localhost/127.0.0.1:8500 - at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) - at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717) - at io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect(NioSocketChannel.java:327) - at io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:340) - at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:616) - at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:563) - at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:480) - at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:442) + at java.base/java.lang.Thread.run(Thread.java:834) + | rs.veshv.impl.adapters.ConsulConfigurationProvider | WARN | Could not load fresh configuration | java.net.UnknownHostException: consul-server1: Temporary failure in name resolution + at java.base/java.net.Inet4AddressImpl.lookupAllHostAddr(Native Method) + at java.base/java.net.InetAddress$PlatformNameService.lookupAllHostAddr(InetAddress.java:929) + at java.base/java.net.InetAddress.getAddressesFromNameService(InetAddress.java:1515) + at java.base/java.net.InetAddress$NameServiceAddresses.get(InetAddress.java:848) + at java.base/java.net.InetAddress.getAllByName0(InetAddress.java:1505) + at java.base/java.net.InetAddress.getAllByName(InetAddress.java:1364) + at java.base/java.net.InetAddress.getAllByName(InetAddress.java:1298) + at java.base/java.net.InetAddress.getByName(InetAddress.java:1248) + at io.netty.util.internal.SocketUtils$8.run(SocketUtils.java:146) + at io.netty.util.internal.SocketUtils$8.run(SocketUtils.java:143) + at java.base/java.security.AccessController.doPrivileged(Native Method) + at io.netty.util.internal.SocketUtils.addressByName(SocketUtils.java:143) + at io.netty.resolver.DefaultNameResolver.doResolve(DefaultNameResolver.java:43) + at io.netty.resolver.SimpleNameResolver.resolve(SimpleNameResolver.java:63) + at io.netty.resolver.SimpleNameResolver.resolve(SimpleNameResolver.java:55) + at io.netty.resolver.InetSocketAddressResolver.doResolve(InetSocketAddressResolver.java:57) + at io.netty.resolver.InetSocketAddressResolver.doResolve(InetSocketAddressResolver.java:32) + at io.netty.resolver.AbstractAddressResolver.resolve(AbstractAddressResolver.java:108) + at io.netty.bootstrap.Bootstrap.doResolveAndConnect0(Bootstrap.java:208) + at io.netty.bootstrap.Bootstrap.access$000(Bootstrap.java:49) + at io.netty.bootstrap.Bootstrap$1.operationComplete(Bootstrap.java:188) + at io.netty.bootstrap.Bootstrap$1.operationComplete(Bootstrap.java:174) + at io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:511) + at io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:485) + at io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:424) + at io.netty.util.concurrent.DefaultPromise.trySuccess(DefaultPromise.java:103) + at io.netty.channel.DefaultChannelPromise.trySuccess(DefaultChannelPromise.java:84) + at io.netty.channel.AbstractChannel$AbstractUnsafe.safeSetSuccess(AbstractChannel.java:978) + at io.netty.channel.AbstractChannel$AbstractUnsafe.register0(AbstractChannel.java:512) + at io.netty.channel.AbstractChannel$AbstractUnsafe.access$200(AbstractChannel.java:423) + at io.netty.channel.AbstractChannel$AbstractUnsafe$1.run(AbstractChannel.java:482) + at io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java:163) + at io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:404) + at io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:315) at io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:884) - at java.lang.Thread.run(Thread.java:748) - | reactor-http-client-epoll-8 + at java.base/java.lang.Thread.run(Thread.java:834) + | ors.veshv.healthcheck.factory.HealthCheckApiServer | DEBUG | HV-VES status: OUT_OF_SERVICE, Consul configuration not available. Retrying. -The above three logs indicate that **HV-VES** cannot connect to the Consul service under url given in **--consul-url** parameter. +The above three logs indicate that **HV-VES** cannot connect to the Consul service under url given in ``--consul-url`` parameter. Make sure Consul is up and running and the **ip + port** combination is correct. ==== @@ -116,17 +194,73 @@ Make sure Consul is up and running and the **ip + port** combination is correct. :: - ap.dcae.collectors.veshv.impl.adapters.HttpAdapter | 2018-10-16T13:52:20.585Z | ERROR | Failed to get resource on path: http://consul:8500/v1/kv/veshv-config (HTTP request failed with code: 404. - Failing URI: /v1/kv/veshv-config) | | reactor-http-nio-1 - ap.dcae.collectors.veshv.impl.adapters.HttpAdapter | 2018-10-16T13:52:20.586Z | DEBUG | Nested exception: | reactor.ipc.netty.http.client.HttpClientException: HTTP request failed with code: 404. - Failing URI: /v1/kv/veshv-config - | reactor-http-nio-1 - rs.veshv.impl.adapters.ConsulConfigurationProvider | 2018-10-16T13:52:20.591Z | WARN | Could not get fresh configuration | reactor.ipc.netty.http.client.HttpClientException: HTTP request failed with code: 404. - Failing URI: /v1/kv/veshv-config - | reactor-http-nio-1 + | ap.dcae.collectors.veshv.impl.adapters.HttpAdapter | ERROR | Failed to get resource on path: http://consul-server:8500/v1/kv/invalid-resource?raw=true (http://consul-server:8500/v1/kv/invalid-resource?raw=true 404 Not Found) + | ap.dcae.collectors.veshv.impl.adapters.HttpAdapter | DEBUG | Nested exception: | java.lang.IllegalStateException: http://consul-server:8500/v1/kv/invalid-resource?raw=true 404 Not Found + at org.onap.dcae.collectors.veshv.impl.adapters.HttpAdapter$get$2.apply(HttpAdapter.kt:46) + at org.onap.dcae.collectors.veshv.impl.adapters.HttpAdapter$get$2.apply(HttpAdapter.kt:34) + at reactor.netty.http.client.HttpClientFinalizer.lambda$responseSingle$7(HttpClientFinalizer.java:95) + at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:118) + at reactor.core.publisher.FluxRetryPredicate$RetryPredicateSubscriber.onNext(FluxRetryPredicate.java:81) + at reactor.core.publisher.MonoCreate$DefaultMonoSink.success(MonoCreate.java:147) + at reactor.netty.http.client.HttpClientConnect$HttpObserver.onStateChange(HttpClientConnect.java:383) + at reactor.netty.resources.PooledConnectionProvider$DisposableAcquire.onStateChange(PooledConnectionProvider.java:501) + at reactor.netty.resources.PooledConnectionProvider$PooledConnection.onStateChange(PooledConnectionProvider.java:443) + at reactor.netty.http.client.HttpClientOperations.onInboundNext(HttpClientOperations.java:494) + at reactor.netty.channel.ChannelOperationsHandler.channelRead(ChannelOperationsHandler.java:141) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) + at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) + at io.netty.channel.CombinedChannelDuplexHandler$DelegatingChannelHandlerContext.fireChannelRead(CombinedChannelDuplexHandler.java:438) + at io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:310) + at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:284) + at io.netty.channel.CombinedChannelDuplexHandler.channelRead(CombinedChannelDuplexHandler.java:253) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) + at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) + at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1434) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) + at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:965) + at io.netty.channel.epoll.AbstractEpollStreamChannel$EpollStreamUnsafe.epollInReady(AbstractEpollStreamChannel.java:808) + at io.netty.channel.epoll.EpollEventLoop.processReady(EpollEventLoop.java:410) + at io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:310) + at io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:884) + at java.base/java.lang.Thread.run(Thread.java:834) + | rs.veshv.impl.adapters.ConsulConfigurationProvider | WARN | Could not load fresh configuration | java.lang.IllegalStateException: http://consul-server:8500/v1/kv/invalid-resource?raw=true 404 Not Found + at org.onap.dcae.collectors.veshv.impl.adapters.HttpAdapter$get$2.apply(HttpAdapter.kt:46) + at org.onap.dcae.collectors.veshv.impl.adapters.HttpAdapter$get$2.apply(HttpAdapter.kt:34) + at reactor.netty.http.client.HttpClientFinalizer.lambda$responseSingle$7(HttpClientFinalizer.java:95) + at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:118) + at reactor.core.publisher.FluxRetryPredicate$RetryPredicateSubscriber.onNext(FluxRetryPredicate.java:81) + at reactor.core.publisher.MonoCreate$DefaultMonoSink.success(MonoCreate.java:147) + at reactor.netty.http.client.HttpClientConnect$HttpObserver.onStateChange(HttpClientConnect.java:383) + at reactor.netty.resources.PooledConnectionProvider$DisposableAcquire.onStateChange(PooledConnectionProvider.java:501) + at reactor.netty.resources.PooledConnectionProvider$PooledConnection.onStateChange(PooledConnectionProvider.java:443) + at reactor.netty.http.client.HttpClientOperations.onInboundNext(HttpClientOperations.java:494) + at reactor.netty.channel.ChannelOperationsHandler.channelRead(ChannelOperationsHandler.java:141) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) + at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) + at io.netty.channel.CombinedChannelDuplexHandler$DelegatingChannelHandlerContext.fireChannelRead(CombinedChannelDuplexHandler.java:438) + at io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:310) + at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:284) + at io.netty.channel.CombinedChannelDuplexHandler.channelRead(CombinedChannelDuplexHandler.java:253) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) + at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) + at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1434) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) + at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:965) + at io.netty.channel.epoll.AbstractEpollStreamChannel$EpollStreamUnsafe.epollInReady(AbstractEpollStreamChannel.java:808) + at io.netty.channel.epoll.EpollEventLoop.processReady(EpollEventLoop.java:410) + at io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:310) + at io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:884) + at java.base/java.lang.Thread.run(Thread.java:834) + | ors.veshv.healthcheck.factory.HealthCheckApiServer | DEBUG | HV-VES status: OUT_OF_SERVICE, Consul configuration not available. Retrying. -**HV-VES** logs this information when connected to Consul, but cannot find any json configuration under given key which in this case is **veshv-config**. +**HV-VES** logs this information when connected to Consul, but cannot find any JSON configuration under given key which in this case is **invalid-resource**. For more information, see :ref:`run_time_configuration` ==== @@ -135,36 +269,40 @@ For more information, see :ref:`run_time_configuration` :: - rs.veshv.impl.adapters.ConsulConfigurationProvider | 2018-10-16T14:06:14.792Z | INFO | Obtained new configuration from consul: - { - "invalidKey": "value" - } | | reactor-http-nio-1 - rs.veshv.impl.adapters.ConsulConfigurationProvider | 2018-10-16T14:06:14.796Z | WARN | Could not get fresh configuration | java.lang.NullPointerException: null - at org.glassfish.json.JsonObjectBuilderImpl$JsonObjectImpl.getString(JsonObjectBuilderImpl.java:257) - at org.onap.dcae.collectors.veshv.impl.adapters.ConsulConfigurationProvider.createCollectorConfiguration(ConsulConfigurationProvider.kt:113) - at org.onap.dcae.collectors.veshv.impl.adapters.ConsulConfigurationProvider.access$createCollectorConfiguration(ConsulConfigurationProvider.kt:44) - at org.onap.dcae.collectors.veshv.impl.adapters.ConsulConfigurationProvider$invoke$6.invoke(ConsulConfigurationProvider.kt:80) - at org.onap.dcae.collectors.veshv.impl.adapters.ConsulConfigurationProvider$invoke$6.invoke(ConsulConfigurationProvider.kt:44) + | rs.veshv.impl.adapters.ConsulConfigurationProvider | INFO | Obtained new configuration from consul: + { "invalidKey": "value" } + | 2018-12-20T15:38:14.543Z | rs.veshv.impl.adapters.ConsulConfigurationProvider | WARN | Could not load fresh configuration | org.onap.dcae.collectors.veshv.impl.adapters.ParsingException: Failed to parse consul configuration + at org.onap.dcae.collectors.veshv.impl.adapters.ConsulConfigurationProvider.createCollectorConfiguration(ConsulConfigurationProvider.kt:125) + at org.onap.dcae.collectors.veshv.impl.adapters.ConsulConfigurationProvider.access$createCollectorConfiguration(ConsulConfigurationProvider.kt:48) + at org.onap.dcae.collectors.veshv.impl.adapters.ConsulConfigurationProvider$invoke$4.invoke(ConsulConfigurationProvider.kt:80) + at org.onap.dcae.collectors.veshv.impl.adapters.ConsulConfigurationProvider$invoke$4.invoke(ConsulConfigurationProvider.kt:48) at org.onap.dcae.collectors.veshv.impl.adapters.ConsulConfigurationProvider$sam$java_util_function_Function$0.apply(ConsulConfigurationProvider.kt) at reactor.core.publisher.FluxMap$MapSubscriber.onNext(FluxMap.java:100) - at reactor.core.publisher.FluxMap$MapSubscriber.onNext(FluxMap.java:108) - at reactor.core.publisher.FluxFlatMap$FlatMapMain.tryEmitScalar(FluxFlatMap.java:432) - at reactor.core.publisher.FluxFlatMap$FlatMapMain.onNext(FluxFlatMap.java:366) - at reactor.core.publisher.FluxMap$MapSubscriber.onNext(FluxMap.java:108) - at reactor.core.publisher.FluxMap$MapSubscriber.onNext(FluxMap.java:108) - at reactor.core.publisher.FluxFlatMap$FlatMapMain.tryEmit(FluxFlatMap.java:484) - at reactor.core.publisher.FluxFlatMap$FlatMapInner.onNext(FluxFlatMap.java:916) - at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onNext(FluxMapFuseable.java:115) - at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1083) + at reactor.core.publisher.FluxMap$MapSubscriber.onNext(FluxMap.java:114) + at reactor.core.publisher.FluxFlatMap$FlatMapMain.tryEmitScalar(FluxFlatMap.java:449) + at reactor.core.publisher.FluxFlatMap$FlatMapMain.onNext(FluxFlatMap.java:384) + at reactor.core.publisher.FluxConcatMap$ConcatMapImmediate.innerNext(FluxConcatMap.java:275) + at reactor.core.publisher.FluxConcatMap$ConcatMapInner.onNext(FluxConcatMap.java:849) + at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onNext(FluxMapFuseable.java:121) + at reactor.core.publisher.FluxPeekFuseable$PeekFuseableSubscriber.onNext(FluxPeekFuseable.java:204) + at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1476) at reactor.core.publisher.MonoFlatMap$FlatMapInner.onNext(MonoFlatMap.java:241) - at reactor.core.publisher.MonoNext$NextSubscriber.onNext(MonoNext.java:76) - at reactor.core.publisher.FluxMap$MapSubscriber.onNext(FluxMap.java:108) - at reactor.core.publisher.FluxFilter$FilterSubscriber.onNext(FluxFilter.java:97) - at reactor.ipc.netty.channel.FluxReceive.drainReceiver(FluxReceive.java:213) - at reactor.ipc.netty.channel.FluxReceive.onInboundNext(FluxReceive.java:329) - at reactor.ipc.netty.channel.ChannelOperations.onInboundNext(ChannelOperations.java:311) - at reactor.ipc.netty.http.client.HttpClientOperations.onInboundNext(HttpClientOperations.java:605) - at reactor.ipc.netty.channel.ChannelOperationsHandler.channelRead(ChannelOperationsHandler.java:138) + at reactor.core.publisher.FluxDoFinally$DoFinallySubscriber.onNext(FluxDoFinally.java:123) + at reactor.core.publisher.FluxHandle$HandleSubscriber.onNext(FluxHandle.java:113) + at reactor.core.publisher.FluxMapFuseable$MapFuseableConditionalSubscriber.onNext(FluxMapFuseable.java:287) + at reactor.core.publisher.FluxUsing$UsingFuseableSubscriber.onNext(FluxUsing.java:350) + at reactor.core.publisher.FluxFilterFuseable$FilterFuseableSubscriber.onNext(FluxFilterFuseable.java:113) + at reactor.core.publisher.FluxPeekFuseable$PeekFuseableConditionalSubscriber.onNext(FluxPeekFuseable.java:486) + at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1476) + at reactor.core.publisher.MonoReduceSeed$ReduceSeedSubscriber.onComplete(MonoReduceSeed.java:156) + at reactor.core.publisher.FluxMap$MapSubscriber.onComplete(FluxMap.java:136) + at reactor.netty.channel.FluxReceive.terminateReceiver(FluxReceive.java:378) + at reactor.netty.channel.FluxReceive.drainReceiver(FluxReceive.java:202) + at reactor.netty.channel.FluxReceive.onInboundComplete(FluxReceive.java:343) + at reactor.netty.channel.ChannelOperations.onInboundComplete(ChannelOperations.java:325) + at reactor.netty.channel.ChannelOperations.terminate(ChannelOperations.java:372) + at reactor.netty.http.client.HttpClientOperations.onInboundNext(HttpClientOperations.java:522) + at reactor.netty.channel.ChannelOperationsHandler.channelRead(ChannelOperationsHandler.java:141) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) @@ -179,14 +317,12 @@ For more information, see :ref:`run_time_configuration` at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:965) - at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:163) - at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:628) - at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:563) - at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:480) - at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:442) + at io.netty.channel.epoll.AbstractEpollStreamChannel$EpollStreamUnsafe.epollInReady(AbstractEpollStreamChannel.java:808) + at io.netty.channel.epoll.EpollEventLoop.processReady(EpollEventLoop.java:410) + at io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:310) at io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:884) - at java.lang.Thread.run(Thread.java:748) - | reactor-http-nio-1 + at java.base/java.lang.Thread.run(Thread.java:834) + | ors.veshv.healthcheck.factory.HealthCheckApiServer | DEBUG | HV-VES status: OUT_OF_SERVICE, Consul configuration not available. Retrying. This log is printed when you upload a configuration in an invalid format (for example, with missing fields). In the first log you can see that configuration on Consul is: @@ -204,106 +340,169 @@ For more information on **Consul configuration**, see :ref:`run_time_configurati Message handling errors ----------------------- -**Handling messages when invalid kafka url is specified** +**Handling messages when invalid Kafka url is specified** :: - | reactor-tcp-server-epoll-6 - org.apache.kafka.clients.ClientUtils | 2018-10-19T08:29:36.917Z | WARN | Removing server invalid-kafka:9093 from bootstrap.servers as DNS resolution failed for invalid-kafka | | reactor-tcp-server-epoll-6 - org.apache.kafka.clients.producer.KafkaProducer | 2018-10-19T08:29:36.918Z | INFO | [Producer clientId=producer-1] Closing the Kafka producer with timeoutMillis = 0 ms. | | reactor-tcp-server-epoll-6 - org.onap.dcae.collectors.veshv.impl.VesHvCollector | 2018-10-19T08:29:36.962Z | WARN | Error while handling message stream: org.apache.kafka.common.KafkaException (Failed to construct kafka producer) | | reactor-tcp-server-epoll-6 - org.onap.dcae.collectors.veshv.impl.VesHvCollector | 2018-10-19T08:29:36.966Z | DEBUG | Detailed stack trace | org.apache.kafka.common.config.ConfigException: No resolvable bootstrap urls given in bootstrap.servers - at org.apache.kafka.clients.ClientUtils.parseAndValidateAddresses(ClientUtils.java:64) - at org.apache.kafka.clients.producer.KafkaProducer.<init>(KafkaProducer.java:396) - ... 24 common frames omitted - Wrapped by: org.apache.kafka.common.KafkaException: Failed to construct kafka producer - at org.apache.kafka.clients.producer.KafkaProducer.<init>(KafkaProducer.java:441) - at org.apache.kafka.clients.producer.KafkaProducer.<init>(KafkaProducer.java:285) + | p.dcae.collectors.veshv.impl.socket.NettyTcpServer | DEBUG | Client connection request received + | p.dcae.collectors.veshv.impl.socket.NettyTcpServer | INFO | Handling new connection + | org.apache.kafka.clients.ClientUtils | WARN | Removing server invalid-kafka-host:9092 from bootstrap.servers as DNS resolution failed for invalid-kafka-host + | org.onap.dcae.collectors.veshv.impl.VesHvCollector | WARN | Error while handling message stream: org.apache.kafka.common.KafkaException (Failed to construct kafka producer) + | org.onap.dcae.collectors.veshv.impl.VesHvCollector | DEBUG | Detailed stack trace | org.apache.kafka.common.config.ConfigException: No resolvable bootstrap urls given in bootstrap.servers + at org.apache.kafka.clients.ClientUtils.parseAndValidateAddresses(ClientUtils.java:66) + at org.apache.kafka.clients.producer.KafkaProducer.<init>(KafkaProducer.java:406) + ... 49 common frames omitted + Wrapped by: org.apache.kafka.common.KafkaException: Failed to construct kafka producer + at org.apache.kafka.clients.producer.KafkaProducer.<init>(KafkaProducer.java:457) + at org.apache.kafka.clients.producer.KafkaProducer.<init>(KafkaProducer.java:289) at reactor.kafka.sender.internals.ProducerFactory.createProducer(ProducerFactory.java:33) - at reactor.kafka.sender.internals.DefaultKafkaSender.lambda$new$0(DefaultKafkaSender.java:90) - at reactor.core.publisher.MonoCallable.subscribe(MonoCallable.java:57) + at reactor.kafka.sender.internals.DefaultKafkaSender.lambda$new$0(DefaultKafkaSender.java:96) + at reactor.core.publisher.MonoCallable.subscribe(MonoCallable.java:56) at reactor.core.publisher.MonoPeekFuseable.subscribe(MonoPeekFuseable.java:74) - at reactor.core.publisher.Mono.subscribe(Mono.java:3088) + at reactor.core.publisher.Mono.subscribe(Mono.java:3590) at reactor.core.publisher.MonoProcessor.add(MonoProcessor.java:531) at reactor.core.publisher.MonoProcessor.subscribe(MonoProcessor.java:444) at reactor.core.publisher.MonoFlatMapMany.subscribe(MonoFlatMapMany.java:49) - at reactor.core.publisher.FluxPeek.subscribe(FluxPeek.java:80) - at reactor.core.publisher.FluxFilter.subscribe(FluxFilter.java:52) + at reactor.core.publisher.FluxPeek.subscribe(FluxPeek.java:83) at reactor.core.publisher.FluxMap.subscribe(FluxMap.java:62) - at reactor.core.publisher.FluxDefer.subscribe(FluxDefer.java:55) at reactor.core.publisher.FluxPeek.subscribe(FluxPeek.java:83) - at reactor.core.publisher.FluxDoFinally.subscribe(FluxDoFinally.java:73) + at reactor.core.publisher.FluxDefer.subscribe(FluxDefer.java:54) + at reactor.core.publisher.FluxPeek.subscribe(FluxPeek.java:83) at reactor.core.publisher.FluxOnErrorResume.subscribe(FluxOnErrorResume.java:47) + at reactor.core.publisher.FluxDoFinally.subscribe(FluxDoFinally.java:73) at reactor.core.publisher.MonoIgnoreElements.subscribe(MonoIgnoreElements.java:37) - at reactor.ipc.netty.channel.ChannelOperations.applyHandler(ChannelOperations.java:381) - at reactor.ipc.netty.channel.ChannelOperations.onHandlerStart(ChannelOperations.java:296) - at io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java:163) - at io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:404) - at io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:315) + at reactor.netty.tcp.TcpServerHandle.onStateChange(TcpServerHandle.java:64) + at reactor.netty.tcp.TcpServerBind$ChildObserver.onStateChange(TcpServerBind.java:226) + at reactor.netty.channel.ChannelOperationsHandler.channelActive(ChannelOperationsHandler.java:112) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelActive(AbstractChannelHandlerContext.java:213) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelActive(AbstractChannelHandlerContext.java:199) + at io.netty.channel.AbstractChannelHandlerContext.fireChannelActive(AbstractChannelHandlerContext.java:192) + at reactor.netty.tcp.SslProvider$SslReadHandler.userEventTriggered(SslProvider.java:720) + at io.netty.channel.AbstractChannelHandlerContext.invokeUserEventTriggered(AbstractChannelHandlerContext.java:329) + at io.netty.channel.AbstractChannelHandlerContext.invokeUserEventTriggered(AbstractChannelHandlerContext.java:315) + at io.netty.channel.AbstractChannelHandlerContext.fireUserEventTriggered(AbstractChannelHandlerContext.java:307) + at io.netty.handler.ssl.SslHandler.setHandshakeSuccess(SslHandler.java:1530) + at io.netty.handler.ssl.SslHandler.wrapNonAppData(SslHandler.java:937) + at io.netty.handler.ssl.SslHandler.unwrap(SslHandler.java:1360) + at io.netty.handler.ssl.SslHandler.decodeJdkCompatible(SslHandler.java:1199) + at io.netty.handler.ssl.SslHandler.decode(SslHandler.java:1243) + at io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:489) + at io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:428) + at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:265) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) + at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) + at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1434) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) + at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:965) + at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:163) + at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:628) + at io.netty.channel.nio.NioEventLoop.processSelectedKeysPlain(NioEventLoop.java:528) + at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:482) + at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:442) at io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:884) - at java.lang.Thread.run(Thread.java:748) - | reactor-tcp-server-epoll-6 - p.dcae.collectors.veshv.impl.socket.NettyTcpServer | 2018-10-19T08:29:36.971Z | INFO | Connection from /172.26.0.6:55574 has been closed | | reactor-tcp-server-epoll-6 - - -**HV-VES** responds with the above when it handles a message and currently applied configuration has invalid Kafka bootstrap server defined. -The configuration read from Consul in this case: + at java.base/java.lang.Thread.run(Thread.java:834) + | org.onap.dcae.collectors.veshv.impl.VesHvCollector | DEBUG | Released buffer memory after handling message stream -.. code-block:: json - - { - "dmaap.kafkaBootstrapServers": "invalid-kafka:9093", - "collector.routing": [ - { - "fromDomain": "perf3gpp", - "toTopic": "HV_VES_PERF3GPP" - }, - { - "fromDomain": "heartbeat", - "toTopic": "HV_VES_HEARTBEAT" - } - ] - } - -where **invalid-kafka:9093** is not an existing **ip+port** combination. +**HV-VES** responds with the above when it handles a message and specified DmaaP MR Kafka bootstrap server is invalid. +Restart with different ``--kafka-bootstrap-servers`` command line option value is required. +For more information, see: :ref:`deployment` ==== -**First creation of topics on kafka** +**Kafka service became unavailable after producer has been created** +**HV-VES** lazily creates Kafka consumer after first successfully handled event. +If Kafka service becomes unreachable after consumer initialization, it is removed from bootstrap.servers list on next connection. -On the first try of creating and publishing to a given kafka topic, **HV-VES** logs the following warnings and creates the requested topics anyway. +Following information is logged: :: - org.apache.kafka.clients.NetworkClient | 2018-10-22T10:11:53.396Z | WARN | [Producer clientId=producer-1] Error while fetching metadata with correlation id 1 : {HV_VES_PERF3GPP=LEADER_NOT_AVAILABLE} | | kafka-producer-network-thread | producer-1 - org.apache.kafka.clients.NetworkClient | 2018-10-22T10:11:53.524Z | WARN | [Producer clientId=producer-1] Error while fetching metadata with correlation id 3 : {HV_VES_PERF3GPP=LEADER_NOT_AVAILABLE} | | kafka-producer-network-thread | producer-1 - org.apache.kafka.clients.NetworkClient | 2018-10-22T10:11:53.636Z | WARN | [Producer clientId=producer-1] Error while fetching metadata with correlation id 4 : {HV_VES_PERF3GPP=LEADER_NOT_AVAILABLE} | | kafka-producer-network-thread | producer-1 - org.apache.kafka.clients.NetworkClient | 2018-10-22T10:11:53.749Z | WARN | [Producer clientId=producer-1] Error while fetching metadata with correlation id 5 : {HV_VES_PERF3GPP=LEADER_NOT_AVAILABLE} | | kafka-producer-network-thread | producer-1 - -==== - -**Kafka service became unavailable after producer for a given topic was successfully created** - - -After receiving a **Ves Common Event**, **HV-VES** creates a producer for a given topic and keeps it for the whole lifetime of an application. -If Kafka service becomes unreachable after the producer creation, you will see the following logs when trying to establish another connection with the Kafka server. - -:: + | org.apache.kafka.clients.NetworkClient | WARN | [Producer clientId=producer-1] Connection to node 1001 could not be established. Broker may not be available. + | org.apache.kafka.clients.NetworkClient | WARN | [Producer clientId=producer-1] Connection to node 1001 could not be established. Broker may not be available. + | org.apache.kafka.clients.NetworkClient | WARN | [Producer clientId=producer-1] Connection to node 1001 could not be established. Broker may not be available. + | org.apache.kafka.clients.NetworkClient | WARN | [Producer clientId=producer-1] Connection to node 1001 could not be established. Broker may not be available. + | org.apache.kafka.clients.NetworkClient | WARN | [Producer clientId=producer-1] Error connecting to node message-router-kafka:9092 (id: 1001 rack: null) | | java.nio.channels.UnresolvedAddressException: null + at java.base/sun.nio.ch.Net.checkAddress(Net.java:130) + at java.base/sun.nio.ch.SocketChannelImpl.connect(SocketChannelImpl.java:675) + at org.apache.kafka.common.network.Selector.doConnect(Selector.java:233) + ... 9 common frames omitted + Wrapped by: java.io.IOException: Can't resolve address: message-router-kafka:9092 + at org.apache.kafka.common.network.Selector.doConnect(Selector.java:235) + at org.apache.kafka.common.network.Selector.connect(Selector.java:214) + at org.apache.kafka.clients.NetworkClient.initiateConnect(NetworkClient.java:864) + at org.apache.kafka.clients.NetworkClient.access$700(NetworkClient.java:64) + at org.apache.kafka.clients.NetworkClient$DefaultMetadataUpdater.maybeUpdate(NetworkClient.java:1035) + at org.apache.kafka.clients.NetworkClient$DefaultMetadataUpdater.maybeUpdate(NetworkClient.java:920) + at org.apache.kafka.clients.NetworkClient.poll(NetworkClient.java:508) + at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:239) + at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:163) + at java.base/java.lang.Thread.run(Thread.java:834) + | | kafka-producer-network-thread | producer-1 + | p.dcae.collectors.veshv.impl.socket.NettyTcpServer | INFO | Handling new connection + | org.apache.kafka.clients.ClientUtils | WARN | Removing server message-router-kafka:9092 from bootstrap.servers as DNS resolution failed for message-router-kafka + | org.onap.dcae.collectors.veshv.impl.VesHvCollector | WARN | Error while handling message stream: org.apache.kafka.common.KafkaException (Failed to construct kafka producer) + | org.onap.dcae.collectors.veshv.impl.VesHvCollector | DEBUG | Detailed stack trace + at org.apache.kafka.clients.ClientUtils.parseAndValidateAddresses(ClientUtils.java:66) + at org.apache.kafka.clients.producer.KafkaProducer.<init>(KafkaProducer.java:406) + ... 48 common frames omitted + Wrapped by: org.apache.kafka.common.KafkaException: Failed to construct kafka producer + at org.apache.kafka.clients.producer.KafkaProducer.<init>(KafkaProducer.java:457) + at org.apache.kafka.clients.producer.KafkaProducer.<init>(KafkaProducer.java:289) + at reactor.kafka.sender.internals.ProducerFactory.createProducer(ProducerFactory.java:33) + at reactor.kafka.sender.internals.DefaultKafkaSender.lambda$new$0(DefaultKafkaSender.java:96) + at reactor.core.publisher.MonoCallable.subscribe(MonoCallable.java:56) + at reactor.core.publisher.MonoPeekFuseable.subscribe(MonoPeekFuseable.java:74) + at reactor.core.publisher.Mono.subscribe(Mono.java:3590) + at reactor.core.publisher.MonoProcessor.add(MonoProcessor.java:531) + at reactor.core.publisher.MonoProcessor.subscribe(MonoProcessor.java:444) + at reactor.core.publisher.MonoFlatMapMany.subscribe(MonoFlatMapMany.java:49) + at reactor.core.publisher.FluxPeek.subscribe(FluxPeek.java:83) + at reactor.core.publisher.FluxMap.subscribe(FluxMap.java:62) + at reactor.core.publisher.FluxPeek.subscribe(FluxPeek.java:83) + at reactor.core.publisher.FluxDefer.subscribe(FluxDefer.java:54) + at reactor.core.publisher.FluxPeek.subscribe(FluxPeek.java:83) + at reactor.core.publisher.FluxOnErrorResume.subscribe(FluxOnErrorResume.java:47) + at reactor.core.publisher.FluxDoFinally.subscribe(FluxDoFinally.java:73) + at reactor.core.publisher.MonoIgnoreElements.subscribe(MonoIgnoreElements.java:37) + at reactor.netty.tcp.TcpServerHandle.onStateChange(TcpServerHandle.java:64) + at reactor.netty.tcp.TcpServerBind$ChildObserver.onStateChange(TcpServerBind.java:226) + at reactor.netty.channel.ChannelOperationsHandler.channelActive(ChannelOperationsHandler.java:112) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelActive(AbstractChannelHandlerContext.java:213) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelActive(AbstractChannelHandlerContext.java:199) + at io.netty.channel.AbstractChannelHandlerContext.fireChannelActive(AbstractChannelHandlerContext.java:192) + at reactor.netty.tcp.SslProvider$SslReadHandler.userEventTriggered(SslProvider.java:720) + at io.netty.channel.AbstractChannelHandlerContext.invokeUserEventTriggered(AbstractChannelHandlerContext.java:329) + at io.netty.channel.AbstractChannelHandlerContext.invokeUserEventTriggered(AbstractChannelHandlerContext.java:315) + at io.netty.channel.AbstractChannelHandlerContext.fireUserEventTriggered(AbstractChannelHandlerContext.java:307) + at io.netty.handler.ssl.SslHandler.setHandshakeSuccess(SslHandler.java:1530) + at io.netty.handler.ssl.SslHandler.unwrap(SslHandler.java:1368) + at io.netty.handler.ssl.SslHandler.decodeJdkCompatible(SslHandler.java:1199) + at io.netty.handler.ssl.SslHandler.decode(SslHandler.java:1243) + at io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:489) + at io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:428) + at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:265) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) + at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) + at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1434) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) + at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:965) + at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:163) + at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:628) + at io.netty.channel.nio.NioEventLoop.processSelectedKeysPlain(NioEventLoop.java:528) + at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:482) + at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:442) + at io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:884) + at java.base/java.lang.Thread.run(Thread.java:834) + | org.onap.dcae.collectors.veshv.impl.VesHvCollector | DEBUG | Released buffer memory after handling message stream - org.apache.kafka.clients.NetworkClient | 2018-10-22T10:04:08.604Z | WARN | [Producer clientId=producer-1] Connection to node 1001 could not be established. Broker may not be available. | | kafka-producer-network-thread | producer-1 - org.apache.kafka.clients.NetworkClient | 2018-10-22T10:04:11.896Z | WARN | [Producer clientId=producer-1] Connection to node 1001 could not be established. Broker may not be available. | | kafka-producer-network-thread | producer-1 - org.apache.kafka.clients.NetworkClient | 2018-10-22T10:04:14.968Z | WARN | [Producer clientId=producer-1] Connection to node 1001 could not be established. Broker may not be available. | | kafka-producer-network-thread | producer-1 - org.apache.kafka.clients.NetworkClient | 2018-10-22T10:04:18.040Z | WARN | [Producer clientId=producer-1] Connection to node 1001 could not be established. Broker may not be available. | | kafka-producer-network-thread | producer-1 - org.apache.kafka.clients.NetworkClient | 2018-10-22T10:04:21.111Z | WARN | [Producer clientId=producer-1] Connection to node 1001 could not be established. Broker may not be available. | | kafka-producer-network-thread | producer-1 - reactor.kafka.sender.internals.DefaultKafkaSender | 2018-10-22T10:04:23.519Z | ERROR | error {} | org.apache.kafka.common.errors.TimeoutException: Expiring 1 record(s) for HV_VES_PERF3GPP-0: 30050 ms has passed since batch creation plus linger time - | kafka-producer-network-thread | producer-1 - cae.collectors.veshv.impl.adapters.kafka.KafkaSink | 2018-10-22T10:04:23.522Z | WARN | Failed to send message to Kafka | org.apache.kafka.common.errors.TimeoutException: Expiring 1 record(s) for HV_VES_PERF3GPP-0: 30050 ms has passed since batch creation plus linger time - | single-1 - org.onap.dcae.collectors.veshv.impl.VesHvCollector | 2018-10-22T10:04:23.528Z | WARN | Error while handling message stream: org.apache.kafka.common.errors.TimeoutException (Expiring 1 record(s) for HV_VES_PERF3GPP-0: 30050 ms has passed since batch creation plus linger time) | | single-1 -To resolve this issue, **HV-VES** restart is required. +To resolve this issue, you can either wait for that Kafka service to be available, or just like in previous paragraph, restart **HV-VES** with different value of ``--kafka-bootstrap-servers`` option. ==== @@ -311,37 +510,96 @@ To resolve this issue, **HV-VES** restart is required. :: - g.onap.dcae.collectors.veshv.impl.VesHvCollector | 2018-10-19T08:53:18.349Z | WARN | Error while handling message stream: org.onap.dcae.collectors.veshv.impl.wire.WireFrameException (PayloadSizeExceeded: payload size exceeds the limit (1048576 bytes)) | | single-1 - org.onap.dcae.collectors.veshv.impl.VesHvCollector | 2018-10-19T08:53:18.349Z | DEBUG | Detailed stack trace | org.onap.dcae.collectors.veshv.impl.wire.WireFrameException: PayloadSizeExceeded: payload size exceeds the limit (1048576 bytes) - at org.onap.dcae.collectors.veshv.impl.wire.WireChunkDecoder$onError$1$1.invoke(WireChunkDecoder.kt:67) - at org.onap.dcae.collectors.veshv.impl.wire.WireChunkDecoder$onError$1$1.invoke(WireChunkDecoder.kt:38) - at arrow.effects.IO$Companion$invoke$1.invoke(IO.kt:28) - at arrow.effects.IO$Companion$invoke$1.invoke(IO.kt:22) - at arrow.effects.IORunLoop.step(IORunLoop.kt:50) - at arrow.effects.IO.unsafeRunTimed(IO.kt:109) - at arrow.effects.IO.unsafeRunSync(IO.kt:106) - at org.onap.dcae.collectors.veshv.impl.wire.WireChunkDecoder$generateFrames$1.accept(WireChunkDecoder.kt:61) - at org.onap.dcae.collectors.veshv.impl.wire.WireChunkDecoder$generateFrames$1.accept(WireChunkDecoder.kt:38) + | p.dcae.collectors.veshv.impl.socket.NettyTcpServer | DEBUG | Client connection request received + | p.dcae.collectors.veshv.impl.socket.NettyTcpServer | INFO | Handling new connection + | p.dcae.collectors.veshv.impl.wire.WireChunkDecoder | TRACE | Got message with total size of 16384 B + | p.dcae.collectors.veshv.impl.wire.WireChunkDecoder | WARN | Error while handling message stream: org.onap.dcae.collectors.veshv.impl.wire.WireFrameException (PayloadSizeExceeded: payload size exceeds the limit (1048576 bytes)) + | p.dcae.collectors.veshv.impl.wire.WireChunkDecoder | DEBUG | Detailed stack trace | org.onap.dcae.collectors.veshv.impl.wire.WireFrameException: PayloadSizeExceeded: payload size exceeds the limit (1048576 bytes) + at org.onap.dcae.collectors.veshv.impl.wire.WireChunkDecoder$onError$1$1.invoke(WireChunkDecoder.kt:72) + at org.onap.dcae.collectors.veshv.impl.wire.WireChunkDecoder$onError$1$1.invoke(WireChunkDecoder.kt:41) + at arrow.effects.IO$Companion$invoke$1.invoke(IO.kt:33) + at arrow.effects.IO$Companion$invoke$1.invoke(IO.kt:27) + at arrow.effects.IORunLoop.step(IORunLoop.kt:49) + at arrow.effects.IO.unsafeRunTimed(IO.kt:115) + at arrow.effects.IO.unsafeRunSync(IO.kt:112) + at org.onap.dcae.collectors.veshv.impl.wire.WireChunkDecoder$generateFrames$1.accept(WireChunkDecoder.kt:66) + at org.onap.dcae.collectors.veshv.impl.wire.WireChunkDecoder$generateFrames$1.accept(WireChunkDecoder.kt:41) + at reactor.core.publisher.FluxGenerate.lambda$new$1(FluxGenerate.java:56) + at reactor.core.publisher.FluxGenerate$GenerateSubscription.slowPath(FluxGenerate.java:262) + at reactor.core.publisher.FluxGenerate$GenerateSubscription.request(FluxGenerate.java:204) + at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.set(Operators.java:1849) + at reactor.core.publisher.FluxOnErrorResume$ResumeSubscriber.onSubscribe(FluxOnErrorResume.java:68) + at reactor.core.publisher.FluxGenerate.subscribe(FluxGenerate.java:83) + at reactor.core.publisher.FluxOnErrorResume.subscribe(FluxOnErrorResume.java:47) + at reactor.core.publisher.FluxDoFinally.subscribe(FluxDoFinally.java:73) + at reactor.core.publisher.FluxDefer.subscribe(FluxDefer.java:54) + at reactor.core.publisher.Flux.subscribe(Flux.java:7734) + at reactor.core.publisher.FluxConcatMap$ConcatMapImmediate.drain(FluxConcatMap.java:442) + at reactor.core.publisher.FluxConcatMap$ConcatMapImmediate.onNext(FluxConcatMap.java:244) + at reactor.core.publisher.FluxPeek$PeekSubscriber.onNext(FluxPeek.java:192) + at reactor.core.publisher.FluxPeek$PeekSubscriber.onNext(FluxPeek.java:192) + at reactor.core.publisher.FluxMap$MapSubscriber.onNext(FluxMap.java:114) + at reactor.netty.channel.FluxReceive.drainReceiver(FluxReceive.java:211) + at reactor.netty.channel.FluxReceive.onInboundNext(FluxReceive.java:327) + at reactor.netty.channel.ChannelOperations.onInboundNext(ChannelOperations.java:310) + at reactor.netty.channel.ChannelOperationsHandler.channelRead(ChannelOperationsHandler.java:141) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) + at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) + at io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:286) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) + at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) + at io.netty.handler.ssl.SslHandler.unwrap(SslHandler.java:1429) + at io.netty.handler.ssl.SslHandler.decodeJdkCompatible(SslHandler.java:1199) + at io.netty.handler.ssl.SslHandler.decode(SslHandler.java:1243) + at io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:489) + at io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:428) + at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:265) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) + at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) + at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1434) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) + at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:965) + at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:163) + at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:628) + at io.netty.channel.nio.NioEventLoop.processSelectedKeysPlain(NioEventLoop.java:528) + at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:482) + at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:442) + at io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:884) + at java.base/java.lang.Thread.run(Thread.java:834) + | org.onap.dcae.collectors.veshv.impl.VesHvCollector | WARN | Error while handling message stream: org.onap.dcae.collectors.veshv.impl.wire.WireFrameException (PayloadSizeExceeded: payload size exceeds the limit (1048576 bytes)) + | org.onap.dcae.collectors.veshv.impl.VesHvCollector | DEBUG | Detailed stack trace | org.onap.dcae.collectors.veshv.impl.wire.WireFrameException: PayloadSizeExceeded: payload size exceeds the limit (1048576 bytes) + at org.onap.dcae.collectors.veshv.impl.wire.WireChunkDecoder$onError$1$1.invoke(WireChunkDecoder.kt:72) + at org.onap.dcae.collectors.veshv.impl.wire.WireChunkDecoder$onError$1$1.invoke(WireChunkDecoder.kt:41) + at arrow.effects.IO$Companion$invoke$1.invoke(IO.kt:33) + at arrow.effects.IO$Companion$invoke$1.invoke(IO.kt:27) + at arrow.effects.IORunLoop.step(IORunLoop.kt:49) + at arrow.effects.IO.unsafeRunTimed(IO.kt:115) + at arrow.effects.IO.unsafeRunSync(IO.kt:112) + at org.onap.dcae.collectors.veshv.impl.wire.WireChunkDecoder$generateFrames$1.accept(WireChunkDecoder.kt:66) + at org.onap.dcae.collectors.veshv.impl.wire.WireChunkDecoder$generateFrames$1.accept(WireChunkDecoder.kt:41) at reactor.core.publisher.FluxGenerate.lambda$new$1(FluxGenerate.java:56) at reactor.core.publisher.FluxGenerate$GenerateSubscription.slowPath(FluxGenerate.java:262) at reactor.core.publisher.FluxGenerate$GenerateSubscription.request(FluxGenerate.java:204) - at reactor.core.publisher.FluxPeekFuseable$PeekFuseableSubscriber.request(FluxPeekFuseable.java:138) - at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.set(Operators.java:1454) - at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.onSubscribe(Operators.java:1328) - at reactor.core.publisher.FluxPeekFuseable$PeekFuseableSubscriber.onSubscribe(FluxPeekFuseable.java:172) + at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.set(Operators.java:1849) + at reactor.core.publisher.FluxOnErrorResume$ResumeSubscriber.onSubscribe(FluxOnErrorResume.java:68) at reactor.core.publisher.FluxGenerate.subscribe(FluxGenerate.java:83) - at reactor.core.publisher.FluxPeekFuseable.subscribe(FluxPeekFuseable.java:86) - at reactor.core.publisher.FluxDefer.subscribe(FluxDefer.java:55) - at reactor.core.publisher.Flux.subscribe(Flux.java:6877) - at reactor.core.publisher.FluxConcatMap$ConcatMapImmediate.drain(FluxConcatMap.java:418) - at reactor.core.publisher.FluxConcatMap$ConcatMapImmediate.onNext(FluxConcatMap.java:241) - at reactor.core.publisher.FluxPeek$PeekSubscriber.onNext(FluxPeek.java:185) - at reactor.core.publisher.FluxPeek$PeekSubscriber.onNext(FluxPeek.java:185) - at reactor.core.publisher.FluxMap$MapSubscriber.onNext(FluxMap.java:108) - at reactor.ipc.netty.channel.FluxReceive.drainReceiver(FluxReceive.java:213) - at reactor.ipc.netty.channel.FluxReceive.onInboundNext(FluxReceive.java:329) - at reactor.ipc.netty.channel.ChannelOperations.onInboundNext(ChannelOperations.java:311) - at reactor.ipc.netty.channel.ChannelOperationsHandler.channelRead(ChannelOperationsHandler.java:138) + at reactor.core.publisher.FluxOnErrorResume.subscribe(FluxOnErrorResume.java:47) + at reactor.core.publisher.FluxDoFinally.subscribe(FluxDoFinally.java:73) + at reactor.core.publisher.FluxDefer.subscribe(FluxDefer.java:54) + at reactor.core.publisher.Flux.subscribe(Flux.java:7734) + at reactor.core.publisher.FluxConcatMap$ConcatMapImmediate.drain(FluxConcatMap.java:442) + at reactor.core.publisher.FluxConcatMap$ConcatMapImmediate.onNext(FluxConcatMap.java:244) + at reactor.core.publisher.FluxPeek$PeekSubscriber.onNext(FluxPeek.java:192) + at reactor.core.publisher.FluxPeek$PeekSubscriber.onNext(FluxPeek.java:192) + at reactor.core.publisher.FluxMap$MapSubscriber.onNext(FluxMap.java:114) + at reactor.netty.channel.FluxReceive.drainReceiver(FluxReceive.java:211) + at reactor.netty.channel.FluxReceive.onInboundNext(FluxReceive.java:327) + at reactor.netty.channel.ChannelOperations.onInboundNext(ChannelOperations.java:310) + at reactor.netty.channel.ChannelOperationsHandler.channelRead(ChannelOperationsHandler.java:141) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) @@ -349,41 +607,69 @@ To resolve this issue, **HV-VES** restart is required. at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) + at io.netty.handler.ssl.SslHandler.unwrap(SslHandler.java:1429) + at io.netty.handler.ssl.SslHandler.decodeJdkCompatible(SslHandler.java:1199) + at io.netty.handler.ssl.SslHandler.decode(SslHandler.java:1243) + at io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:489) + at io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:428) + at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:265) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) + at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) + at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:340) at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1434) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:362) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348) at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:965) - at io.netty.channel.epoll.AbstractEpollStreamChannel$EpollStreamUnsafe.epollInReady(AbstractEpollStreamChannel.java:808) - at io.netty.channel.epoll.EpollEventLoop.processReady(EpollEventLoop.java:410) - at io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:310) + at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:163) + at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:628) + at io.netty.channel.nio.NioEventLoop.processSelectedKeysPlain(NioEventLoop.java:528) + at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:482) + at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:442) at io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:884) - at java.lang.Thread.run(Thread.java:748) - | single-1 - p.dcae.collectors.veshv.impl.socket.NettyTcpServer | 2018-10-19T08:53:18.351Z | INFO | Connection from /172.26.0.6:56924 has been closed | | single-1 - + at java.base/java.lang.Thread.run(Thread.java:834) + | org.onap.dcae.collectors.veshv.impl.VesHvCollector | DEBUG | Released buffer memory after handling message stream The above log is printed when the message payload size is too big. **HV-VES** does not handle messages that exceed specified payload size. Default value is **1048576 bytes (1MiB)**, but it can be configured via cmd or by environment variables. +==== +**Invalid GPB data** -==== +Messages with invalid Google Protocol Buffers data encoded are omitted. **HV-VES** responds as follows: + +:: + + | p.dcae.collectors.veshv.impl.socket.NettyTcpServer | DEBUG | Client connection request received + | p.dcae.collectors.veshv.impl.socket.NettyTcpServer | INFO | Handling new connection + | p.dcae.collectors.veshv.impl.wire.WireChunkDecoder | TRACE | Got message with total size of 28 B + | p.dcae.collectors.veshv.impl.wire.WireChunkDecoder | TRACE | Wire payload size: 16 B + | org.onap.dcae.collectors.veshv.impl.VesHvCollector | TRACE | Wire frame header is valid + | org.onap.dcae.collectors.veshv.impl.VesHvCollector | WARN | Failed to decode ves event header, reason: Protocol message tag had invalid wire type. + | p.dcae.collectors.veshv.impl.wire.WireChunkDecoder | TRACE | End of data in current TCP buffer -**Other invalid messages** +==== +**Invalid Wire Frame** -Messages with **invalid wire frame** or **invalid gpb** data are ommitted and **HV-VES** only logs the connection-related logs as follows: +Messages with invalid Wire Frame, just like those containing invalid GPB data, will be dropped. The exact reason can be found in logs. :: - p.dcae.collectors.veshv.impl.socket.NettyTcpServer | 2018-10-19T09:03:03.345Z | INFO | Handling connection from /172.26.0.6:57432 | | reactor-tcp-server-epoll-5 - p.dcae.collectors.veshv.impl.socket.NettyTcpServer | 2018-10-19T09:04:03.351Z | INFO | Idle timeout of 60 s reached. Closing connection from /172.26.0.6:57432... | | reactor-tcp-server-epoll-5 - p.dcae.collectors.veshv.impl.socket.NettyTcpServer | 2018-10-19T09:04:03.353Z | INFO | Connection from /172.26.0.6:57432 has been closed | | reactor-tcp-server-epoll-5 - p.dcae.collectors.veshv.impl.socket.NettyTcpServer | 2018-10-19T09:04:03.354Z | DEBUG | Channel (/172.26.0.6:57432) closed successfully. | | reactor-tcp-server-epoll-5 + | p.dcae.collectors.veshv.impl.socket.NettyTcpServer | DEBUG | Client connection request received + | p.dcae.collectors.veshv.impl.socket.NettyTcpServer | INFO | Handling new connection + | p.dcae.collectors.veshv.impl.wire.WireChunkDecoder | TRACE | Got message with total size of 322 B + | p.dcae.collectors.veshv.impl.wire.WireChunkDecoder | TRACE | Wire payload size: 310 B + | org.onap.dcae.collectors.veshv.impl.VesHvCollector | WARN | Invalid wire frame header, reason: Invalid major version in wire frame header. Expected 1 but was 2 + | p.dcae.collectors.veshv.impl.wire.WireChunkDecoder | TRACE | End of data in current TCP buffer + + +==== For more information, see the :ref:`hv_ves_behaviors` section. + Authorization related errors ---------------------------- @@ -394,7 +680,7 @@ Authorization related errors :: - org.onap.dcae.collectors.veshv.main | 2018-10-22T06:51:54.191Z | ERROR | Failed to start a server | java.io.FileNotFoundException: /etc/ves-hv/server.p12 (No such file or directory) + | org.onap.dcae.collectors.veshv.main | ERROR | Failed to start a server | java.io.FileNotFoundException: /etc/ves-hv/server.p12 (No such file or directory) at java.io.FileInputStream.open0(Native Method) at java.io.FileInputStream.open(FileInputStream.java:195) at java.io.FileInputStream.<init>(FileInputStream.java:138) @@ -421,11 +707,10 @@ Authorization related errors at arrow.effects.IO.unsafeRunSync(IO.kt:106) at org.onap.dcae.collectors.veshv.utils.arrow.EffectsKt.unsafeRunEitherSync(effects.kt:50) at org.onap.dcae.collectors.veshv.main.MainKt.main(main.kt:41) - | main The above error is logged when key store is not provided. Similarly, when trust store is not provided, **/etc/ves-hv/trust.p12** file missing is logged. -**server.p12** and **trust.p12** are default names of key and trust stores. They can be changed by specifying **--trust-store** or **--key-store** command line arguments on deployment. +**server.p12** and **trust.p12** are default names of key and trust stores. They can be changed by specifying ``--trust-store`` or ``--key-store`` command line arguments on deployment. ==== @@ -433,7 +718,7 @@ The above error is logged when key store is not provided. Similarly, when trust :: - org.onap.dcae.collectors.veshv.main | 2018-10-22T06:59:24.039Z | ERROR | Failed to start a server | java.security.UnrecoverableKeyException: failed to decrypt safe contents entry: javax.crypto.BadPaddingException: Given final block not properly padded. Such issues can arise if a bad key is used during decryption. + | org.onap.dcae.collectors.veshv.main | ERROR | Failed to start a server | java.security.UnrecoverableKeyException: failed to decrypt safe contents entry: javax.crypto.BadPaddingException: Given final block not properly padded. Such issues can arise if a bad key is used during decryption. ... 23 common frames omitted Wrapped by: java.io.IOException: keystore password was incorrect at sun.security.pkcs12.PKCS12KeyStore.engineLoad(PKCS12KeyStore.java:2059) @@ -459,7 +744,6 @@ The above error is logged when key store is not provided. Similarly, when trust at arrow.effects.IO.unsafeRunSync(IO.kt:106) at org.onap.dcae.collectors.veshv.utils.arrow.EffectsKt.unsafeRunEitherSync(effects.kt:50) at org.onap.dcae.collectors.veshv.main.MainKt.main(main.kt:41) - | main Key or trust store password provided in configuration is invalid. @@ -470,7 +754,7 @@ Key or trust store password provided in configuration is invalid. :: - org.onap.dcae.collectors.veshv.main | 2018-10-22T09:11:38.200Z | ERROR | Failed to start a server | java.io.IOException: DerInputStream.getLength(): lengthTag=111, too big. + | org.onap.dcae.collectors.veshv.main | ERROR | Failed to start a server | java.io.IOException: DerInputStream.getLength(): lengthTag=111, too big. at sun.security.util.DerInputStream.getLength(DerInputStream.java:599) at sun.security.util.DerValue.init(DerValue.java:391) at sun.security.util.DerValue.<init>(DerValue.java:332) @@ -498,7 +782,6 @@ Key or trust store password provided in configuration is invalid. at arrow.effects.IO.unsafeRunSync(IO.kt:106) at org.onap.dcae.collectors.veshv.utils.arrow.EffectsKt.unsafeRunEitherSync(effects.kt:50) at org.onap.dcae.collectors.veshv.main.MainKt.main(main.kt:41) - | main The above is logged when provided keystore has invalid or corrupted content. This log also appears when you try to use key store/trust store in archive format other than **PKCS12** (the only supported by **HV-VES** store type). |