[JBoss JIRA] (ISPN-7008) IllegalLifeCycleStateException from BlockingRunnable.isReady() escapes to originator
by Tristan Tarrant (Jira)
[ https://issues.jboss.org/browse/ISPN-7008?page=com.atlassian.jira.plugin.... ]
Tristan Tarrant updated ISPN-7008:
----------------------------------
Fix Version/s: 10.0.1.Final
(was: 10.0.0.Final)
> IllegalLifeCycleStateException from BlockingRunnable.isReady() escapes to originator
> ------------------------------------------------------------------------------------
>
> Key: ISPN-7008
> URL: https://issues.jboss.org/browse/ISPN-7008
> Project: Infinispan
> Issue Type: Bug
> Components: Core
> Affects Versions: 8.2.4.Final, 9.0.0.Alpha4
> Reporter: Dan Berindei
> Priority: Major
> Fix For: 10.0.1.Final
>
>
> The ISPN-5539 fix caught {{IllegalLifecycleStateException}} in {{BlockingRunnable.run()}} and sent a {{CacheNotFoundResponse}} to the originator instead. However, exceptions thrown from {{BlockingRunnable.isReady()}} are not caught, and are instead sent back to the originator. An example stack trace is available in the WFLY-7029 description.
> In this case, the exception was thrown directly when the inbound invocation handler tried to submit the command to the remote commands executor. But I see a related problem in {{BlockingTaskAwareExecutorServiceImpl.ControllerThread.run()}}, which stops with the first exception from {{isReady()}}. For {{IllegalLifecycleStateException}} it's probably ok, but for other exceptions it should continue processing new tasks.
--
This message was sent by Atlassian Jira
(v7.13.8#713008)
6 years, 5 months
[JBoss JIRA] (ISPN-7022) TwoWaySplitAndMergeTest.testSplitAndMerge2 fails
by Tristan Tarrant (Jira)
[ https://issues.jboss.org/browse/ISPN-7022?page=com.atlassian.jira.plugin.... ]
Tristan Tarrant updated ISPN-7022:
----------------------------------
Fix Version/s: 10.0.1.Final
(was: 10.0.0.Final)
> TwoWaySplitAndMergeTest.testSplitAndMerge2 fails
> ------------------------------------------------
>
> Key: ISPN-7022
> URL: https://issues.jboss.org/browse/ISPN-7022
> Project: Infinispan
> Issue Type: Bug
> Affects Versions: 9.2.1.Final
> Reporter: Sebastian Laskawiec
> Assignee: Dan Berindei
> Priority: Major
> Fix For: 10.0.1.Final
>
>
> http://ci.infinispan.org/viewLog.html?buildId=42202&tab=buildResultsDiv&b...
> {noformat}
> java.lang.AssertionError:
> at org.testng.AssertJUnit.fail(AssertJUnit.java:59)
> at org.testng.AssertJUnit.assertTrue(AssertJUnit.java:24)
> at org.infinispan.test.AbstractInfinispanTest.eventually(AbstractInfinispanTest.java:235)
> at org.infinispan.test.AbstractInfinispanTest.eventually(AbstractInfinispanTest.java:217)
> at org.infinispan.test.AbstractInfinispanTest.eventually(AbstractInfinispanTest.java:193)
> at org.infinispan.test.AbstractInfinispanTest.eventually(AbstractInfinispanTest.java:527)
> at org.infinispan.partitionhandling.BasePartitionHandlingTest.access$600(BasePartitionHandlingTest.java:44)
> at org.infinispan.partitionhandling.BasePartitionHandlingTest$Partition.assertAvailabilityMode(BasePartitionHandlingTest.java:328)
> at org.infinispan.partitionhandling.BasePartitionHandlingTest$Partition.assertDegradedMode(BasePartitionHandlingTest.java:265)
> at org.infinispan.partitionhandling.TwoWaySplitAndMergeTest.testSplitAndMerge(TwoWaySplitAndMergeTest.java:74)
> at org.infinispan.partitionhandling.TwoWaySplitAndMergeTest.testSplitAndMerge2(TwoWaySplitAndMergeTest.java:26)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:497)
> at org.testng.internal.MethodInvocationHelper.invokeMethod(MethodInvocationHelper.java:84)
> at org.testng.internal.Invoker.invokeMethod(Invoker.java:714)
> at org.testng.internal.Invoker.invokeTestMethod(Invoker.java:901)
> at org.testng.internal.Invoker.invokeTestMethods(Invoker.java:1231)
> at org.testng.internal.TestMethodWorker.invokeTestMethods(TestMethodWorker.java:127)
> at org.testng.internal.TestMethodWorker.run(TestMethodWorker.java:111)
> at org.testng.TestRunner.privateRun(TestRunner.java:767)
> at org.testng.TestRunner.run(TestRunner.java:617)
> at org.testng.SuiteRunner.runTest(SuiteRunner.java:348)
> at org.testng.SuiteRunner.access$000(SuiteRunner.java:38)
> at org.testng.SuiteRunner$SuiteWorker.run(SuiteRunner.java:382)
> at org.testng.internal.thread.ThreadUtil$2.call(ThreadUtil.java:64)
> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> at java.lang.Thread.run(Thread.java:745)
> ------- Stdout: -------
> 13:26:54,221 ERROR (testng-DelayedAvailabilityUpdateTest) [InvocationContextInterceptor] ISPN000136: Error executing command GetKeyValueCommand, writing keys []
> org.infinispan.partitionhandling.AvailabilityException: ISPN000306: Key 'MagicKey#k1Existing{1420/A1FBD015/225@DelayedAvailabilityUpdateTest-NodeT-8452}' is not available. Not all owners are in this partition
> at org.infinispan.partitionhandling.impl.PartitionHandlingManagerImpl.doCheck(PartitionHandlingManagerImpl.java:250)
> at org.infinispan.partitionhandling.impl.PartitionHandlingManagerImpl.checkRead(PartitionHandlingManagerImpl.java:102)
> at org.infinispan.partitionhandling.impl.PartitionHandlingInterceptor.postOperationPartitionCheck(PartitionHandlingInterceptor.java:188)
> at org.infinispan.partitionhandling.impl.PartitionHandlingInterceptor.lambda$handleDataReadCommand$412(PartitionHandlingInterceptor.java:145)
> at org.infinispan.partitionhandling.impl.PartitionHandlingInterceptor$$Lambda$1639/1963537113.handle(Unknown Source)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.invokeReturnHandlerSync(BaseAsyncInvocationContext.java:493)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.invokeReturnHandlersSync(BaseAsyncInvocationContext.java:125)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.forkInvocationSync(BaseAsyncInvocationContext.java:99)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.invokeForkAndHandlerSync(BaseAsyncInvocationContext.java:474)[TestSuiteProgress] Test failed: org.infinispan.partitionhandling.OptimisticTxPartitionAndMergeDuringCommitTest.testOriginatorIsolatedPartition
> 13:26:54,221 ERROR (testng-OptimisticTxPartitionAndMergeDuringCommitTest) [TestSuiteProgress] Test failed: org.infinispan.partitionhandling.OptimisticTxPartitionAndMergeDuringCommitTest.testOriginatorIsolatedPartition
> java.lang.AssertionError: There are pending transactions!
> at org.testng.AssertJUnit.fail(AssertJUnit.java:59)
> at org.testng.AssertJUnit.assertTrue(AssertJUnit.java:24)
> at org.infinispan.test.AbstractInfinispanTest.eventually(AbstractInfinispanTest.java:235)
> at org.infinispan.test.AbstractInfinispanTest.eventually(AbstractInfinispanTest.java:531)
> at org.infinispan.test.MultipleCacheManagersTest.assertNoTransactions(MultipleCacheManagersTest.java:794)
> at org.infinispan.partitionhandling.BaseTxPartitionAndMergeTest.finalAsserts(BaseTxPartitionAndMergeTest.java:96)
> at org.infinispan.partitionhandling.BaseOptimisticTxPartitionAndMergeTest.doTest(BaseOptimisticTxPartitionAndMergeTest.java:77)
> at org.infinispan.partitionhandling.OptimisticTxPartitionAndMergeDuringCommitTest.testOriginatorIsolatedPartition(OptimisticTxPartitionAndMergeDuringCommitTest.java:37)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)[TestSuiteProgress] Tests succeeded: 6704, failed: 1, skipped: 0
> 13:26:54,222 ERROR (testng-DelayedAvailabilityUpdateTest) [InvocationContextInterceptor] ISPN000136: Error executing command GetKeyValueCommand, writing keys []
> org.infinispan.partitionhandling.AvailabilityException: ISPN000306: Key 'MagicKey#k2Existing{1421/89F8758B/114@DelayedAvailabilityUpdateTest-NodeQ-17336}' is not available. Not all owners are in this partition
> at org.infinispan.partitionhandling.impl.PartitionHandlingInterceptor.postOperationPartitionCheck(PartitionHandlingInterceptor.java:197)
> at org.infinispan.partitionhandling.impl.PartitionHandlingInterceptor.lambda$handleDataReadCommand$412(PartitionHandlingInterceptor.java:145)
> at org.infinispan.partitionhandling.impl.PartitionHandlingInterceptor$$Lambda$1639/1963537113.handle(Unknown Source)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.invokeReturnHandlerSync(BaseAsyncInvocationContext.java:493)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.invokeReturnHandlersSync(BaseAsyncInvocationContext.java:125)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.forkInvocationSync(BaseAsyncInvocationContext.java:99)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.invokeForkAndHandlerSync(BaseAsyncInvocationContext.java:474)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.afterVisit(BaseAsyncInvocationContext.java:463)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.invokeInterceptorsSync(BaseAsyncInvocationContext.java:322)13:26:54,222 ERROR (testng-DelayedAvailabilityUpdateTest) [InvocationContextInterceptor] ISPN000136: Error executing command GetKeyValueCommand, writing keys []
> org.infinispan.partitionhandling.AvailabilityException: ISPN000306: Key 'MagicKey#k2Existing{1421/89F8758B/114@DelayedAvailabilityUpdateTest-NodeQ-17336}' is not available. Not all owners are in this partition
> at org.infinispan.partitionhandling.impl.PartitionHandlingManagerImpl.doCheck(PartitionHandlingManagerImpl.java:250)
> at org.infinispan.partitionhandling.impl.PartitionHandlingManagerImpl.checkRead(PartitionHandlingManagerImpl.java:102)
> at org.infinispan.partitionhandling.impl.PartitionHandlingInterceptor.postOperationPartitionCheck(PartitionHandlingInterceptor.java:188)
> at org.infinispan.partitionhandling.impl.PartitionHandlingInterceptor.lambda$handleDataReadCommand$412(PartitionHandlingInterceptor.java:145)
> at org.infinispan.partitionhandling.impl.PartitionHandlingInterceptor$$Lambda$1639/1963537113.handle(Unknown Source)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.invokeReturnHandlerSync(BaseAsyncInvocationContext.java:493)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.invokeReturnHandlersSync(BaseAsyncInvocationContext.java:125)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.forkInvocationSync(BaseAsyncInvocationContext.java:99)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.invokeForkAndHandlerSync(BaseAsyncInvocationContext.java:474)13:26:49,964 ERROR (transport-thread-PessimisticTxPartitionAndMergeDuringPrepareTest-NodeG-p44001-t2) [CLUSTER] [Context=pes-cache]ISPN000319: After merge (or coordinator change), cache still hasn't recovered all its data and must stay in degraded mode. Current members are [PessimisticTxPartitionAndMergeDuringPrepareTest-NodeG-37404, PessimisticTxPartitionAndMergeDuringPrepareTest-NodeH-18972], lost members are [PessimisticTxPartitionAndMergeDuringPrepareTest-NodeE-27252, PessimisticTxPartitionAndMergeDuringPrepareTest-NodeF-47835], stable members are [PessimisticTxPartitionAndMergeDuringPrepareTest-NodeE-27252, PessimisticTxPartitionAndMergeDuringPrepareTest-NodeF-47835, PessimisticTxPartitionAndMergeDuringPrepareTest-NodeG-37404, PessimisticTxPartitionAndMergeDuringPrepareTest-NodeH-18972]
> [TestSuiteProgress] Test failed: org.infinispan.partitionhandling.TwoWaySplitAndMergeTest.testSplitAndMerge2
> 13:26:55,721 ERROR (testng-TwoWaySplitAndMergeTest) [TestSuiteProgress] Test failed: org.infinispan.partitionhandling.TwoWaySplitAndMergeTest.testSplitAndMerge2
> java.lang.AssertionError:
> at org.testng.AssertJUnit.fail(AssertJUnit.java:59)
> at org.testng.AssertJUnit.assertTrue(AssertJUnit.java:24)
> at org.infinispan.test.AbstractInfinispanTest.eventually(AbstractInfinispanTest.java:235)
> at org.infinispan.test.AbstractInfinispanTest.eventually(AbstractInfinispanTest.java:217)
> at org.infinispan.test.AbstractInfinispanTest.eventually(AbstractInfinispanTest.java:193)
> at org.infinispan.test.AbstractInfinispanTest.eventually(AbstractInfinispanTest.java:527)
> at org.infinispan.partitionhandling.BasePartitionHandlingTest.access$600(BasePartitionHandlingTest.java:44)
> at org.infinispan.partitionhandling.BasePartitionHandlingTest$Partition.assertAvailabilityMode(BasePartitionHandlingTest.java:328)
> at org.infinispan.partitionhandling.BasePartitionHandlingTest$Partition.assertDegradedMode(BasePartitionHandlingTest.java:265)13:26:49,964 ERROR (transport-thread-PessimisticTxPartitionAndMergeDuringPrepareTest-NodeG-p44001-t3) [CLUSTER] [Context=___defaultcache]ISPN000319: After merge (or coordinator change), cache still hasn't recovered all its data and must stay in degraded mode. Current members are [PessimisticTxPartitionAndMergeDuringPrepareTest-NodeG-37404, PessimisticTxPartitionAndMergeDuringPrepareTest-NodeH-18972], lost members are [PessimisticTxPartitionAndMergeDuringPrepareTest-NodeE-27252, PessimisticTxPartitionAndMergeDuringPrepareTest-NodeF-47835], stable members are [PessimisticTxPartitionAndMergeDuringPrepareTest-NodeE-27252, PessimisticTxPartitionAndMergeDuringPrepareTest-NodeF-47835, PessimisticTxPartitionAndMergeDuringPrepareTest-NodeG-37404, PessimisticTxPartitionAndMergeDuringPrepareTest-NodeH-18972]
> [TestSuiteProgress] Tests succeeded: 6704, failed: 2, skipped: 0
> 13:26:55,095 ERROR (testng-OptimisticTxPartitionAndMergeDuringPrepareTest) [InvocationContextInterceptor] ISPN000136: Error executing command ClearCommand, writing keys []
> org.infinispan.partitionhandling.AvailabilityException: ISPN000307: Cannot clear when the cluster is partitioned
> at org.infinispan.partitionhandling.impl.PartitionHandlingManagerImpl.checkClear(PartitionHandlingManagerImpl.java:108)
> at org.infinispan.partitionhandling.impl.PartitionHandlingInterceptor.visitClearCommand(PartitionHandlingInterceptor.java:95)
> at org.infinispan.partitionhandling.impl.PartitionHandlingInterceptor.visitClearCommand(PartitionHandlingInterceptor.java:38)
> at org.infinispan.commands.write.ClearCommand.acceptVisitor(ClearCommand.java:44)
> at org.infinispan.interceptors.DDAsyncInterceptor.visitCommand(DDAsyncInterceptor.java:53)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.invokeInterceptorsSync(BaseAsyncInvocationContext.java:307)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.forkInvocationSync(BaseAsyncInvocationContext.java:98)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.invokeForkAndHandlerSync(BaseAsyncInvocationContext.java:474)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.afterVisit(BaseAsyncInvocationContext.java:463)
> {noformat}
--
This message was sent by Atlassian Jira
(v7.13.8#713008)
6 years, 5 months
[JBoss JIRA] (ISPN-6470) CapacityFactorsFunctionalTest.testCapacityFactors random failures
by Tristan Tarrant (Jira)
[ https://issues.jboss.org/browse/ISPN-6470?page=com.atlassian.jira.plugin.... ]
Tristan Tarrant updated ISPN-6470:
----------------------------------
Fix Version/s: 10.0.1.Final
(was: 10.0.0.Final)
> CapacityFactorsFunctionalTest.testCapacityFactors random failures
> ------------------------------------------------------------------
>
> Key: ISPN-6470
> URL: https://issues.jboss.org/browse/ISPN-6470
> Project: Infinispan
> Issue Type: Bug
> Components: Test Suite - Core
> Affects Versions: 8.2.1.Final
> Reporter: Dan Berindei
> Assignee: Dan Berindei
> Priority: Critical
> Labels: testsuite_stability
> Fix For: 10.0.1.Final
>
>
> When nodes have different capacity factors, {{SyncConsistentHashFactory}} sometimes has trouble assigning the segments correctly:
> {noformat}
> java.lang.AssertionError: Expected between:<20.0> and:<50.0> but was:<52.0>
> at org.testng.AssertJUnit.fail(AssertJUnit.java:59)
> at org.infinispan.test.TestingUtil.assertBetween(TestingUtil.java:1455)
> at org.infinispan.distribution.ch.CapacityFactorsFunctionalTest.assertOwned(CapacityFactorsFunctionalTest.java:96)
> at org.infinispan.distribution.ch.CapacityFactorsFunctionalTest.testCapacityFactors(CapacityFactorsFunctionalTest.java:65)
> {noformat}
--
This message was sent by Atlassian Jira
(v7.13.8#713008)
6 years, 5 months
[JBoss JIRA] (ISPN-6367) HotRod client unmarshalling ignores configured classloader
by Tristan Tarrant (Jira)
[ https://issues.jboss.org/browse/ISPN-6367?page=com.atlassian.jira.plugin.... ]
Tristan Tarrant updated ISPN-6367:
----------------------------------
Fix Version/s: 10.0.1.Final
(was: 10.0.0.Final)
> HotRod client unmarshalling ignores configured classloader
> ----------------------------------------------------------
>
> Key: ISPN-6367
> URL: https://issues.jboss.org/browse/ISPN-6367
> Project: Infinispan
> Issue Type: Bug
> Components: Remote Protocols, Server
> Affects Versions: 8.2.0.Final
> Reporter: Dan Berindei
> Priority: Major
> Fix For: 10.0.1.Final
>
>
> The {{RemoteCacheManager}}'s {{ConfigurationBuilder}} allows you to set a classloader, but the client ignores that classloader when unmarshalling values received from the server.
> This leads to problems when the client doesn't have a flat classpath (e.g. in an OSGi or JBoss Modules-based container).
--
This message was sent by Atlassian Jira
(v7.13.8#713008)
6 years, 5 months
[JBoss JIRA] (ISPN-6336) NonIndexedClusteredDummyInMemoryStoreQueryDslConditionsTest random failures
by Tristan Tarrant (Jira)
[ https://issues.jboss.org/browse/ISPN-6336?page=com.atlassian.jira.plugin.... ]
Tristan Tarrant updated ISPN-6336:
----------------------------------
Fix Version/s: 10.0.1.Final
(was: 10.0.0.Final)
> NonIndexedClusteredDummyInMemoryStoreQueryDslConditionsTest random failures
> ---------------------------------------------------------------------------
>
> Key: ISPN-6336
> URL: https://issues.jboss.org/browse/ISPN-6336
> Project: Infinispan
> Issue Type: Bug
> Reporter: Gustavo Fernandes
> Assignee: Nistor Adrian
> Priority: Major
> Labels: testsuite_stability
> Fix For: 10.0.1.Final
>
>
> Picked this failure when running the query testsuite in loop. It's somehow hard to reproduce.
> The presence of the Byteman test runner in the stacktrace looks suspicious, since it is not supposed to run this test.
> {code}
> Tests run: 2907, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 51.21 sec <<< FAILURE! - in TestSuite
> testSampleDomainQuery16(org.infinispan.query.dsl.embedded.NonIndexedClusteredDummyInMemoryStoreQueryDslConditionsTest) Time elapsed: 0.01 sec <<< FAILURE!
> java.lang.AssertionError: expected:<50> but was:<0>
> at org.junit.Assert.fail(Assert.java:88)
> at org.junit.Assert.failNotEquals(Assert.java:743)
> at org.junit.Assert.assertEquals(Assert.java:118)
> at org.junit.Assert.assertEquals(Assert.java:555)
> at org.junit.Assert.assertEquals(Assert.java:542)
> at org.infinispan.query.dsl.embedded.QueryDslConditionsTest.testSampleDomainQuery16(QueryDslConditionsTest.java:1366)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:497)
> at org.testng.internal.MethodInvocationHelper.invokeMethod(MethodInvocationHelper.java:84)
> at org.testng.internal.MethodInvocationHelper$1.runTestMethod(MethodInvocationHelper.java:200)
> at org.jboss.byteman.contrib.bmunit.BMNGAbstractRunner.run(BMNGAbstractRunner.java:58)
> at sun.reflect.GeneratedMethodAccessor72.invoke(Unknown Source)
> at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:497)
> at org.testng.internal.MethodInvocationHelper.invokeHookable(MethodInvocationHelper.java:212)
> at org.testng.internal.Invoker.invokeMethod(Invoker.java:707)
> at org.testng.internal.Invoker.invokeTestMethod(Invoker.java:901)
> at org.testng.internal.Invoker.invokeTestMethods(Invoker.java:1231)
> at org.testng.internal.TestMethodWorker.invokeTestMethods(TestMethodWorker.java:127)
> at org.testng.internal.TestMethodWorker.run(TestMethodWorker.java:111)
> at org.testng.TestRunner.privateRun(TestRunner.java:767)
> at org.testng.TestRunner.run(TestRunner.java:617)
> at org.testng.SuiteRunner.runTest(SuiteRunner.java:348)
> at org.testng.SuiteRunner.access$000(SuiteRunner.java:38)
> at org.testng.SuiteRunner$SuiteWorker.run(SuiteRunner.java:382)
> at org.testng.internal.thread.ThreadUtil$2.call(ThreadUtil.java:64)
> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> at java.lang.Thread.run(Thread.java:745)
> {code}
--
This message was sent by Atlassian Jira
(v7.13.8#713008)
6 years, 5 months
[JBoss JIRA] (ISPN-7159) CompleteShutdownDistRetryTest.testRetryAfterCompleteShutdown random failures
by Tristan Tarrant (Jira)
[ https://issues.jboss.org/browse/ISPN-7159?page=com.atlassian.jira.plugin.... ]
Tristan Tarrant updated ISPN-7159:
----------------------------------
Fix Version/s: 10.0.1.Final
(was: 10.0.0.Final)
> CompleteShutdownDistRetryTest.testRetryAfterCompleteShutdown random failures
> ----------------------------------------------------------------------------
>
> Key: ISPN-7159
> URL: https://issues.jboss.org/browse/ISPN-7159
> Project: Infinispan
> Issue Type: Bug
> Components: Test Suite - Server
> Affects Versions: 9.0.0.Alpha4
> Reporter: Dan Berindei
> Priority: Critical
> Labels: testsuite_stability
> Fix For: 10.0.1.Final
>
>
> {noformat}
> java.lang.AssertionError: expected:<two> but was:<null>
> at org.testng.AssertJUnit.assertEquals(AssertJUnit.java:101)
> at org.testng.AssertJUnit.assertEquals(AssertJUnit.java:108)
> at org.infinispan.client.hotrod.retry.CompleteShutdownDistRetryTest.testRetryAfterCompleteShutdown(CompleteShutdownDistRetryTest.java:58)
> {noformat}
> I think the problem is that killing node 0 can make node 2's key move from node 2 to node 1, and then it's lost after node 1 is also killed.
--
This message was sent by Atlassian Jira
(v7.13.8#713008)
6 years, 5 months
[JBoss JIRA] (ISPN-7139) Consistent prefix in property names of Hot Rod client configurations
by Tristan Tarrant (Jira)
[ https://issues.jboss.org/browse/ISPN-7139?page=com.atlassian.jira.plugin.... ]
Tristan Tarrant updated ISPN-7139:
----------------------------------
Fix Version/s: 10.0.1.Final
(was: 10.0.0.Final)
> Consistent prefix in property names of Hot Rod client configurations
> --------------------------------------------------------------------
>
> Key: ISPN-7139
> URL: https://issues.jboss.org/browse/ISPN-7139
> Project: Infinispan
> Issue Type: Enhancement
> Components: Configuration, Remote Protocols
> Reporter: Sanne GRINOVERO
> Priority: Major
> Fix For: 10.0.1.Final
>
>
> Most Hot Rod configuration properties use the prefix {{infinispan.client.hotrod.}} which makes it easy to recognize them.
> Some properties don't use the prefix though, such as all those related to connection pooling, e.g. {{maxActive}}, {{maxTotal}}, ..
> This makes it hard to read Hot Rod specific configuration properties from alternative sources. In particular we'd like to embed such properties in an Hibernate configuration file (for Hibernate OGM) but the fact that some properties need special handling forces us to keep a list of these to have special care, rather than being able to use a prefix like it is done in most such cases.
> It should be easy to prefix all the properties, and deprecate the names.
--
This message was sent by Atlassian Jira
(v7.13.8#713008)
6 years, 5 months
[JBoss JIRA] (ISPN-7092) Failed to broadcast asynchronous command
by Tristan Tarrant (Jira)
[ https://issues.jboss.org/browse/ISPN-7092?page=com.atlassian.jira.plugin.... ]
Tristan Tarrant updated ISPN-7092:
----------------------------------
Fix Version/s: 10.0.1.Final
(was: 10.0.0.Final)
> Failed to broadcast asynchronous command
> ----------------------------------------
>
> Key: ISPN-7092
> URL: https://issues.jboss.org/browse/ISPN-7092
> Project: Infinispan
> Issue Type: Bug
> Components: Core
> Affects Versions: 8.2.4.Final
> Reporter: Radoslav Husar
> Assignee: Dan Berindei
> Priority: Minor
> Labels: infinispan_clean_shutdown
> Fix For: 10.0.1.Final
>
>
> Occured on server at the end of server stopping.
> Scenarios affected:
> eap-7x-failover-ejb-ejbremote-jvmkill-dist-sync
> eap-7x-failover-ejb-ejbremote-jvmkill-repl-async
> eap-7x-failover-ejb-ejbremote-jvmkill-repl-sync
> eap-7x-failover-ejb-ejbremote-shutdown-dist-async
> eap-7x-failover-ejb-ejbremote-shutdown-dist-sync-tcpStack
> eap-7x-failover-ejb-ejbremote-shutdown-repl-sync
> eap-7x-failover-ejb-ejbremote-undeploy-repl-sync
> eap-7x-failover-ejb-ejbservlet-shutdown-repl-async
> eap-7x-failover-ejb-ejbstateless-jvmkill-repl-async
> eap-7x-failover-http-granular-shutdown-repl-sync
> eap-7x-failover-http-session-shutdown-repl-async
> eap-7x-failover-http-session-shutdown-repl-async-tcpStack
> eap-7x-failover-http-session-shutdown-repl-sync
> eap-7x-failover-jvmkill-ha-ss
> WARN message has two variants.
> *Variant 1*:
> {code}
> 15:01:15,157 INFO [org.infinispan.remoting.transport.jgroups.JGroupsTransport] (MSC service thread 1-3) ISPN000080: Disconnecting JGroups channel web
> [JBossINF] [0m[0m15:01:15,157 INFO [org.infinispan.remoting.transport.jgroups.JGroupsTransport] (MSC service thread 1-3) ISPN000082: Stopping the RpcDispatcher for channel web
> [JBossINF] [0m[0m15:01:15,158 INFO [org.jboss.as.connector.deployers.jdbc] (MSC service thread 1-3) WFLYJCA0019: Stopped Driver service with driver-name = h2
> [JBossINF] [0m[33m15:01:15,158 WARN [org.infinispan.topology.CacheTopologyControlCommand] (remote-thread--p7-t13) ISPN000071: Caught exception when handling command CacheTopologyControlCommand{cache=clusterbench-ee7.ear.clusterbench-ee7-web-default.war, type=REBALANCE_CONFIRM, sender=dev214, joinInfo=null, topologyId=35, rebalanceId=0, currentCH=null, pendingCH=null, availabilityMode=null, actualMembers=null, throwable=null, viewId=12}: org.infinispan.commons.CacheException: Failed to broadcast asynchronous command: CacheTopologyControlCommand{cache=clusterbench-ee7.ear.clusterbench-ee7-web-default.war, type=STABLE_TOPOLOGY_UPDATE, sender=dev213, joinInfo=null, topologyId=36, rebalanceId=17, currentCH=DefaultConsistentHash{ns=256, owners = (2)[dev214: 124+132, dev215: 132+124]}, pendingCH=null, availabilityMode=null, actualMembers=[dev214, dev215], throwable=null, viewId=12}
> [JBossINF] at org.infinispan.topology.ClusterTopologyManagerImpl.executeOnClusterAsync(ClusterTopologyManagerImpl.java:595)
> [JBossINF] at org.infinispan.topology.ClusterTopologyManagerImpl.broadcastStableTopologyUpdate(ClusterTopologyManagerImpl.java:615)
> [JBossINF] at org.infinispan.topology.ClusterCacheStatus.startQueuedRebalance(ClusterCacheStatus.java:633)
> [JBossINF] at org.infinispan.topology.ClusterCacheStatus.endRebalance(ClusterCacheStatus.java:373)
> [JBossINF] at org.infinispan.topology.ClusterCacheStatus.doConfirmRebalance(ClusterCacheStatus.java:310)
> [JBossINF] at org.infinispan.topology.ClusterTopologyManagerImpl.handleRebalanceCompleted(ClusterTopologyManagerImpl.java:263)
> [JBossINF] at org.infinispan.topology.CacheTopologyControlCommand.doPerform(CacheTopologyControlCommand.java:176)
> [JBossINF] at org.infinispan.topology.CacheTopologyControlCommand.perform(CacheTopologyControlCommand.java:153)
> [JBossINF] at org.infinispan.remoting.inboundhandler.GlobalInboundInvocationHandler$2.run(GlobalInboundInvocationHandler.java:159)
> [JBossINF] at org.infinispan.util.concurrent.BlockingTaskAwareExecutorServiceImpl$RunnableWrapper.run(BlockingTaskAwareExecutorServiceImpl.java:199)
> [JBossINF] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> [JBossINF] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> [JBossINF] at org.jboss.as.clustering.infinispan.ClassLoaderThreadFactory.lambda$newThread$0(ClassLoaderThreadFactory.java:48)
> [JBossINF] at java.lang.Thread.run(Thread.java:745)
> [JBossINF] Caused by: org.infinispan.commons.CacheException: java.lang.IllegalStateException: channel is not connected
> [JBossINF] at org.infinispan.remoting.transport.jgroups.CommandAwareRpcDispatcher.rethrowAsCacheException(CommandAwareRpcDispatcher.java:158)
> [JBossINF] at org.infinispan.remoting.transport.jgroups.CommandAwareRpcDispatcher.invokeRemoteCommands(CommandAwareRpcDispatcher.java:138)
> [JBossINF] at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotelyAsync(JGroupsTransport.java:578)
> [JBossINF] at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:532)
> [JBossINF] at org.infinispan.topology.ClusterTopologyManagerImpl.executeOnClusterAsync(ClusterTopologyManagerImpl.java:592)
> [JBossINF] ... 13 more
> [JBossINF] Caused by: java.lang.IllegalStateException: channel is not connected
> [JBossINF] at org.jgroups.blocks.MessageDispatcher$ProtocolAdapter.down(MessageDispatcher.java:682)
> [JBossINF] at org.jgroups.blocks.RequestCorrelator.sendRequest(RequestCorrelator.java:172)
> [JBossINF] at org.jgroups.blocks.GroupRequest.sendRequest(GroupRequest.java:325)
> [JBossINF] at org.jgroups.blocks.GroupRequest.sendRequest(GroupRequest.java:76)
> [JBossINF] at org.jgroups.blocks.Request.execute(Request.java:67)
> [JBossINF] at org.jgroups.blocks.MessageDispatcher.cast(MessageDispatcher.java:373)
> [JBossINF] at org.jgroups.blocks.MessageDispatcher.cast(MessageDispatcher.java:386)
> [JBossINF] at org.infinispan.remoting.transport.jgroups.CommandAwareRpcDispatcher.processCalls(CommandAwareRpcDispatcher.java:407)
> [JBossINF] at org.infinispan.remoting.transport.jgroups.CommandAwareRpcDispatcher.invokeRemoteCommands(CommandAwareRpcDispatcher.java:133)
> [JBossINF] ... 16 more
> {code}
> *Variant 2*:
> {code}
> 15:53:27,821 INFO [org.infinispan.CLUSTER] (remote-thread--p6-t19) ISPN000336: Finished cluster-wide rebalance for cache routing, topology id = 33
> [JBossINF] [0m[0m15:53:27,829 INFO [org.jboss.as.connector.subsystems.datasources] (MSC service thread 1-8) WFLYJCA0010: Unbound data source [java:jboss/datasources/ExampleDS]
> [JBossINF] [0m[0m15:53:27,830 INFO [org.infinispan.remoting.transport.jgroups.JGroupsTransport] (MSC service thread 1-5) ISPN000080: Disconnecting JGroups channel web
> [JBossINF] [0m[0m15:53:27,830 INFO [org.infinispan.remoting.transport.jgroups.JGroupsTransport] (MSC service thread 1-5) ISPN000082: Stopping the RpcDispatcher for channel web
> [JBossINF] [0m[0m15:53:27,821 INFO [org.infinispan.CLUSTER] (remote-thread--p6-t14) ISPN000336: Finished cluster-wide rebalance for cache clusterbench-ee7.ear.clusterbench-ee7-web-default.war, topology id = 39
> [JBossINF] [0m[0m15:53:27,834 INFO [org.jboss.as.connector.deployers.jdbc] (MSC service thread 1-3) WFLYJCA0019: Stopped Driver service with driver-name = h2
> [JBossINF] [0m[33m15:53:27,832 WARN [org.infinispan.topology.CacheTopologyControlCommand] (remote-thread--p6-t14) ISPN000071: Caught exception when handling command CacheTopologyControlCommand{cache=clusterbench-ee7.ear.clusterbench-ee7-web-default.war, type=REBALANCE_CONFIRM, sender=dev215, joinInfo=null, topologyId=39, rebalanceId=0, currentCH=null, pendingCH=null, availabilityMode=null, actualMembers=null, throwable=null, viewId=12}: org.infinispan.commons.CacheException: Failed to broadcast asynchronous command: CacheTopologyControlCommand{cache=clusterbench-ee7.ear.clusterbench-ee7-web-default.war, type=CH_UPDATE, sender=dev213, joinInfo=null, topologyId=40, rebalanceId=21, currentCH=DefaultConsistentHash{ns=256, owners = (2)[dev214: 129+127, dev215: 127+129]}, pendingCH=null, availabilityMode=null, actualMembers=[dev214, dev215], throwable=null, viewId=12}
> [JBossINF] at org.infinispan.topology.ClusterTopologyManagerImpl.executeOnClusterAsync(ClusterTopologyManagerImpl.java:595)
> [JBossINF] at org.infinispan.topology.ClusterTopologyManagerImpl.broadcastTopologyUpdate(ClusterTopologyManagerImpl.java:606)
> [JBossINF] at org.infinispan.topology.ClusterCacheStatus.endRebalance(ClusterCacheStatus.java:369)
> [JBossINF] at org.infinispan.topology.ClusterCacheStatus.doConfirmRebalance(ClusterCacheStatus.java:310)
> [JBossINF] at org.infinispan.topology.ClusterTopologyManagerImpl.handleRebalanceCompleted(ClusterTopologyManagerImpl.java:263)
> [JBossINF] at org.infinispan.topology.CacheTopologyControlCommand.doPerform(CacheTopologyControlCommand.java:176)
> [JBossINF] at org.infinispan.topology.CacheTopologyControlCommand.perform(CacheTopologyControlCommand.java:153)
> [JBossINF] at org.infinispan.remoting.inboundhandler.GlobalInboundInvocationHandler$2.run(GlobalInboundInvocationHandler.java:159)
> [JBossINF] at org.infinispan.util.concurrent.BlockingTaskAwareExecutorServiceImpl$RunnableWrapper.run(BlockingTaskAwareExecutorServiceImpl.java:199)
> [JBossINF] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> [JBossINF] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> [JBossINF] at org.jboss.as.clustering.infinispan.ClassLoaderThreadFactory.lambda$newThread$0(ClassLoaderThreadFactory.java:48)
> [JBossINF] at java.lang.Thread.run(Thread.java:745)
> [JBossINF] Caused by: java.lang.NullPointerException
> [JBossINF] at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotelyAsync(JGroupsTransport.java:611)
> [JBossINF] at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:532)
> [JBossINF] at org.infinispan.topology.ClusterTopologyManagerImpl.executeOnClusterAsync(ClusterTopologyManagerImpl.java:592)
> [JBossINF] ... 12 more
> {code}
> Link to server log:
> http://jenkins.mw.lab.eng.bos.redhat.com/hudson/job/eap-7x-failover-ejb-e...
--
This message was sent by Atlassian Jira
(v7.13.8#713008)
6 years, 5 months