Open
Description
We get different errors and warnings, that bloat our logstash logs. Getting rid of them would be best, but i am troubleshooting the cause of them here, because the logstash plugin is the one handling these errors/warnings.
I understand this is from the microsoft library, but does this not mean that the azure event hub plugin is not correctly executed or set up?
[2021-03-12T07:36:33,625][ERROR][com.microsoft.azure.eventprocessorhost.PumpManager][....-pipeline][719a950a821fba7e0946da2859ca0b00f6a05393f387cb6d34063dbca9f45466] host logstash-1f1d8eb8-824e-4c31-b669-21ce265682e0: 0: throwing away zombie pump
[2021-03-12T07:36:33,625][ERROR][com.microsoft.azure.eventprocessorhost.PumpManager][....-pipeline][719a950a821fba7e0946da2859ca0b00f6a05393f387cb6d34063dbca9f45466] host logstash-1f1d8eb8-824e-4c31-b669-21ce265682e0: 1: throwing away zombie pump
We get a transient storage failures, but logstash catches them as errors, but it should ignored them since they are just info level.
[2021-03-12T07:57:44,097][ERROR][logstash.inputs.azure.errornotificationhandler][....-pipeline][719a950a821fba7e0946da2859ca0b00f6a05393f387cb6d34063dbca9f45466] Error with Event Processor Host. {:host_name=>"logstash-1f1d8eb8-824e-4c31-b669-21ce265682e0", :action=>"Renewing Lease", :exception=>"com.microsoft.azure.storage.StorageException: The client could not finish the operation within specified maximum execution timeout."}
[2021-03-12T07:57:44,097][INFO ][com.microsoft.azure.eventprocessorhost.PartitionPump][....-pipeline][719a950a821fba7e0946da2859ca0b00f6a05393f387cb6d34063dbca9f45466] host logstash-1f1d8eb8-824e-4c31-b669-21ce265682e0: 0: Transient failure renewing lease
com.microsoft.azure.storage.StorageException: The client could not finish the operation within specified maximum execution timeout.
at com.microsoft.azure.storage.core.ExecutionEngine.executeWithRetry(ExecutionEngine.java:243) ~[azure-storage-8.0.0.jar:?]
at com.microsoft.azure.storage.blob.CloudBlob.renewLease(CloudBlob.java:2682) ~[azure-storage-8.0.0.jar:?]
at com.microsoft.azure.eventprocessorhost.AzureStorageCheckpointLeaseManager.renewLeaseInternal(AzureStorageCheckpointLeaseManager.java:514) ~[azure-eventhubs-eph-2.4.0.jar:?]
at com.microsoft.azure.eventprocessorhost.AzureStorageCheckpointLeaseManager.renewLease(AzureStorageCheckpointLeaseManager.java:497) ~[azure-eventhubs-eph-2.4.0.jar:?]
at com.microsoft.azure.eventprocessorhost.PartitionPump.leaseRenewer(PartitionPump.java:418) ~[azure-eventhubs-eph-2.4.0.jar:?]
at com.microsoft.azure.eventprocessorhost.PartitionPump.lambda$scheduleLeaseRenewer$11(PartitionPump.java:167) ~[azure-eventhubs-eph-2.4.0.jar:?]
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) [?:?]
at java.util.concurrent.FutureTask.run(FutureTask.java:264) [?:?]
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304) [?:?]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?]
at java.lang.Thread.run(Thread.java:834) [?:?]
Caused by: java.util.concurrent.TimeoutException: The client could not finish the operation within specified maximum execution timeout.
at com.microsoft.azure.storage.core.ExecutionEngine.executeWithRetry(ExecutionEngine.java:242) ~[azure-storage-8.0.0.jar:?]
... 11 more
We are getting these very periodic. Either would be nice to increase the timeout or skip logs like this. They really polute the log files.
[2021-03-12T02:19:05,992][WARN ][com.microsoft.azure.eventprocessorhost.AzureStorageCheckpointLeaseManager][....-pipeline][18c4ca558b08b3b9aa48d566c294be6f1dfa7b464da81c7150942bc50be35373] host logstash-9e4cdcf4-2a76-44fe-a046-e09b75713118: 0: Failure updating checkpoint
com.microsoft.azure.storage.StorageException: The client could not finish the operation within specified maximum execution timeout.
at com.microsoft.azure.storage.core.ExecutionEngine.executeWithRetry(ExecutionEngine.java:243) ~[azure-storage-8.0.0.jar:?]
at com.microsoft.azure.storage.blob.CloudBlob.renewLease(CloudBlob.java:2682) ~[azure-storage-8.0.0.jar:?]
at com.microsoft.azure.eventprocessorhost.AzureStorageCheckpointLeaseManager.renewLeaseInternal(AzureStorageCheckpointLeaseManager.java:514) ~[azure-eventhubs-eph-2.4.0.jar:?]
at com.microsoft.azure.eventprocessorhost.AzureStorageCheckpointLeaseManager.updateLeaseInternal(AzureStorageCheckpointLeaseManager.java:591) ~[azure-eventhubs-eph-2.4.0.jar:?]
at com.microsoft.azure.eventprocessorhost.AzureStorageCheckpointLeaseManager.updateCheckpoint(AzureStorageCheckpointLeaseManager.java:176) [azure-eventhubs-eph-2.4.0.jar:?]
at com.microsoft.azure.eventprocessorhost.PartitionContext.checkpoint(PartitionContext.java:198) [azure-eventhubs-eph-2.4.0.jar:?]
at com.microsoft.azure.eventprocessorhost.PartitionContext.checkpoint(PartitionContext.java:177) [azure-eventhubs-eph-2.4.0.jar:?]
at jdk.internal.reflect.GeneratedMethodAccessor91.invoke(Unknown Source) ~[?:?]
at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?]
at java.lang.reflect.Method.invoke(Method.java:566) ~[?:?]
at org.jruby.javasupport.JavaMethod.invokeDirectWithExceptionHandling(JavaMethod.java:456) [jruby-complete-9.2.13.0.jar:?]
at org.jruby.javasupport.JavaMethod.invokeDirect(JavaMethod.java:317) [jruby-complete-9.2.13.0.jar:?]
at org.jruby.java.invokers.InstanceMethodInvoker.call(InstanceMethodInvoker.java:42) [jruby-complete-9.2.13.0.jar:?]
We also get these epoch errors. Is this error somehow created by multiple threads in the azure event hub plugin?
[2021-03-09T20:37:50,554][ERROR][logstash.inputs.azure.processor][....-pipeline]
[ba2387301f3557a552d831c2e777dae8f69c245695332ed7c788bdc3c28769be] Event Hub: some-hub, Partition: 1 experienced an error com.microsoft.azure.eventhubs.ReceiverDisconnectedException: Receiver 'nil' with a higher epoch '79' already exists. Receiver 'nil' with epoch 78 cannot be created.
Make sure you are creating receiver with increasing epoch value to ensure connectivity, or ensure all old epoch receivers are closed or disconnected.
TrackingId:4eca56f7-0d3d-435f-a5e8-d37d3f18ddc2_B43, SystemTracker:some-hub:eventhub:some-hub~32766, Timestamp:2021-03-09T20:37:50 Reference:961669b3-7b1b-4a3f-b90c-9f76e0fc1b54, TrackingId:625575eb-b666-4ac9-ac33-e292ec77a9f9_B43, SystemTracker:some-hub:eventhub:some-hub~32766|$default, Timestamp:2021-03-09T20:37:50 TrackingId:bf8ef3cf525b45d98af89b561e240988_G25, SystemTracker:gateway5, Timestamp:2021-03-09T20:37:50, errorContext[NS: some-hub.servicebus.windows.net, PATH: some-hub/ConsumerGroups/$Default/Partitions/1, REFERENCE_ID: ab89da_988_G25_1615322270154, PREFETCH_COUNT: 300, LINK_CREDIT: 0, PREFETCH_Q_LEN: 0])
Metadata
Metadata
Assignees
Labels
No labels