[JBoss JIRA] (ISPN-9112) getCache() should not wait for the initial state transfer by default
by Adrian Nistor (JIRA)
[ https://issues.jboss.org/browse/ISPN-9112?page=com.atlassian.jira.plugin.... ]
Adrian Nistor commented on ISPN-9112:
-------------------------------------
This has a high chance of breaking query and the infinispan index manager. Fingers crossed ! :)
> getCache() should not wait for the initial state transfer by default
> --------------------------------------------------------------------
>
> Key: ISPN-9112
> URL: https://issues.jboss.org/browse/ISPN-9112
> Project: Infinispan
> Issue Type: Enhancement
> Components: Configuration, Core
> Affects Versions: 9.2.1.Final
> Reporter: Dan Berindei
> Assignee: Dan Berindei
> Fix For: 9.3.0.Beta1
>
>
> {{DefaultCacheManager.getCache(name)}} normally blocks until the local node becomes a full member of the consistent hash and the rebalance has finished. This can cause problems when rebalancing is disabled, either explicitly by the administrator, or implicitly because the cache is in degraded mode.
> Blocking can be disabled with {{builder.clustering().stateTransfer().awaitInitialTransfer(false)}} ({{await-initial-transfer="false"}} in XML}}, but we should make the non-blocking behaviour the default. We should also deprecate the method and attribute to enable the blocking behaviour, and instead add a method to {{AdvancedCache}} to wait for the initial state transfer.
--
This message was sent by Atlassian JIRA
(v7.5.0#75005)
6 years, 8 months
[JBoss JIRA] (ISPN-7800) Cluster always in Degraded Mode
by Dan Berindei (JIRA)
[ https://issues.jboss.org/browse/ISPN-7800?page=com.atlassian.jira.plugin.... ]
Dan Berindei closed ISPN-7800.
------------------------------
Resolution: Duplicate Issue
Duplicate of ISPN-5290.
I also created ISPN-9112 to not wait for the initial transfer by default, so the servers can restart without the timeout error.
> Cluster always in Degraded Mode
> -------------------------------
>
> Key: ISPN-7800
> URL: https://issues.jboss.org/browse/ISPN-7800
> Project: Infinispan
> Issue Type: Bug
> Affects Versions: 8.2.6.Final, 9.0.0.Final
> Reporter: Pedro Ruivo
>
> Scenario:
> * 3 nodes, server mode with Partition handling enabled
> * 2 nodes are killed and bring back online
> * the nodes are unable to merge and the cluster remains in degraded mode.
> I suspect that the FORK channel/protocol is the culprit since the heartbeat command is never handled in the joiner node, but the coordinator receives a {{CacheNotFoundResponse}} quickly (i.e. without timeout). The request is received and "delivered" but never reaches Infinispan.
> When starting node 1 (logs from coordinator):
> {code}
> Received new cluster view: 5, isCoordinator = true, old status = COORDINATOR
> Received new cluster view: 5, isCoordinator = true, old status = COORDINATOR
> //hearbeat sent, ClusterTopologyManagerImpl.confirmMembersAvailable();
> Responses: value=CacheNotFoundResponse, received=true, suspected=false
> Node node01-47572 left while updating cache members
> //the view is not handled
> {code}
> When I started node 2:
> {code}
> Received new cluster view: 6, isCoordinator = true, old status = COORDINATOR
> Updating cluster members for all the caches. New list is [node03-48579, node01-47572, node02-32959]
> //hearbeat sent, ClusterTopologyManagerImpl.confirmMembersAvailable();
> Responses: Responses{
> node01-47572: value=SuccessfulResponse{responseValue=true} , received=true, suspected=false
> node02-32959: value=CacheNotFoundResponse, received=true, suspected=false}
> Node node02-32959 left while updating cache members
> //the view is not handled
> {code}
> It is always reproducible. The configuration is
> {code:xml}
> <replicated-cache name="default" mode="SYNC" batching="true">
> <partition-handling enabled="true"/>
> <locking isolation="REPEATABLE_READ"/>
> <state-transfer enabled="false"/>
> {code}
--
This message was sent by Atlassian JIRA
(v7.5.0#75005)
6 years, 8 months
[JBoss JIRA] (ISPN-9112) getCache() should not wait for the initial state transfer by default
by Dan Berindei (JIRA)
Dan Berindei created ISPN-9112:
----------------------------------
Summary: getCache() should not wait for the initial state transfer by default
Key: ISPN-9112
URL: https://issues.jboss.org/browse/ISPN-9112
Project: Infinispan
Issue Type: Enhancement
Components: Configuration, Core
Affects Versions: 9.2.1.Final
Reporter: Dan Berindei
Assignee: Dan Berindei
Fix For: 9.3.0.Beta1
{{DefaultCacheManager.getCache(name)}} normally blocks until the local node becomes a full member of the consistent hash and the rebalance has finished. This can cause problems when rebalancing is disabled, either explicitly by the administrator, or implicitly because the cache is in degraded mode.
Blocking can be disabled with {{builder.clustering().stateTransfer().awaitInitialTransfer(false)}} ({{await-initial-transfer="false"}} in XML}}, but we should make the non-blocking behaviour the default. We should also deprecate the method and attribute to enable the blocking behaviour, and instead add a method to {{AdvancedCache}} to wait for the initial state transfer.
--
This message was sent by Atlassian JIRA
(v7.5.0#75005)
6 years, 8 months
[JBoss JIRA] (ISPN-9112) getCache() should not wait for the initial state transfer by default
by Dan Berindei (JIRA)
[ https://issues.jboss.org/browse/ISPN-9112?page=com.atlassian.jira.plugin.... ]
Dan Berindei updated ISPN-9112:
-------------------------------
Status: Open (was: New)
> getCache() should not wait for the initial state transfer by default
> --------------------------------------------------------------------
>
> Key: ISPN-9112
> URL: https://issues.jboss.org/browse/ISPN-9112
> Project: Infinispan
> Issue Type: Enhancement
> Components: Configuration, Core
> Affects Versions: 9.2.1.Final
> Reporter: Dan Berindei
> Assignee: Dan Berindei
> Fix For: 9.3.0.Beta1
>
>
> {{DefaultCacheManager.getCache(name)}} normally blocks until the local node becomes a full member of the consistent hash and the rebalance has finished. This can cause problems when rebalancing is disabled, either explicitly by the administrator, or implicitly because the cache is in degraded mode.
> Blocking can be disabled with {{builder.clustering().stateTransfer().awaitInitialTransfer(false)}} ({{await-initial-transfer="false"}} in XML}}, but we should make the non-blocking behaviour the default. We should also deprecate the method and attribute to enable the blocking behaviour, and instead add a method to {{AdvancedCache}} to wait for the initial state transfer.
--
This message was sent by Atlassian JIRA
(v7.5.0#75005)
6 years, 8 months
[JBoss JIRA] (ISPN-7439) InitialClusterSizeTest can hang during teardown
by Bela Ban (JIRA)
[ https://issues.jboss.org/browse/ISPN-7439?page=com.atlassian.jira.plugin.... ]
Bela Ban commented on ISPN-7439:
--------------------------------
Can you retest with https://issues.jboss.org/browse/JGRP-2262 (snapshot of JGroups) fixed? I suggest a smallish {{max_join_attempts}} attribute in GMS.
> InitialClusterSizeTest can hang during teardown
> -----------------------------------------------
>
> Key: ISPN-7439
> URL: https://issues.jboss.org/browse/ISPN-7439
> Project: Infinispan
> Issue Type: Bug
> Components: Core, Test Suite - Core
> Affects Versions: 9.0.0.Beta2
> Reporter: Dan Berindei
> Assignee: Pedro Ruivo
> Labels: testsuite_stability
>
> Test {{testInitialClusterSizeFail}} expects the nodes to time out in {{JGroupsTransport.waitForInitialNodes()}}, but in at least one case the timeout didn't happen. The test then tried to shut down the cache managers, but it hanged because another thread was holding the {{GlobalComponentRegistry}} lock:
> {noformat}
> "testng-InitialClusterSizeTest" #13 prio=5 os_prio=0 tid=0x00007f1874d1f000 nid=0x1778 waiting for monitor entry [0x00007f181bafc000]
> java.lang.Thread.State: BLOCKED (on object monitor)
> at org.infinispan.factories.GlobalComponentRegistry.stop(GlobalComponentRegistry.java:280)
> - waiting to lock <0x0000000093c7afe0> (a org.infinispan.factories.GlobalComponentRegistry)
> at org.infinispan.manager.DefaultCacheManager.stop(DefaultCacheManager.java:701)
> - locked <0x000000008a005b80> (a org.infinispan.manager.DefaultCacheManager)
> at org.infinispan.test.TestingUtil.killCacheManagers(TestingUtil.java:656)
> at org.infinispan.test.MultipleCacheManagersTest.clearContent(MultipleCacheManagersTest.java:138)
> at sun.reflect.GeneratedMethodAccessor175.invoke(Unknown Source)
> at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:497)
> at org.testng.internal.MethodInvocationHelper.invokeMethod(MethodInvocationHelper.java:84)
> at org.testng.internal.Invoker.invokeConfigurationMethod(Invoker.java:564)
> at org.testng.internal.Invoker.invokeConfigurations(Invoker.java:213)
> at org.testng.internal.Invoker.invokeMethod(Invoker.java:786)
> at org.testng.internal.Invoker.invokeTestMethod(Invoker.java:901)
> at org.testng.internal.Invoker.invokeTestMethods(Invoker.java:1231)
> at org.testng.internal.TestMethodWorker.invokeTestMethods(TestMethodWorker.java:127)
> at org.testng.internal.TestMethodWorker.run(TestMethodWorker.java:111)
> at org.testng.TestRunner.privateRun(TestRunner.java:767)
> at org.testng.TestRunner.run(TestRunner.java:617)
> at org.testng.SuiteRunner.runTest(SuiteRunner.java:348)
> at org.testng.SuiteRunner.access$000(SuiteRunner.java:38)
> at org.testng.SuiteRunner$SuiteWorker.run(SuiteRunner.java:382)
> at org.testng.internal.thread.ThreadUtil$2.call(ThreadUtil.java:64)
> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> at java.lang.Thread.run(Thread.java:745)
> "ForkThread-4,InitialClusterSizeTest" #167842 prio=5 os_prio=0 tid=0x00007f1824163800 nid=0x3316 waiting on condition [0x00007f17e62b9000]
> java.lang.Thread.State: TIMED_WAITING (sleeping)
> at java.lang.Thread.sleep(Native Method)
> at org.jgroups.util.Util.sleep(Util.java:1818)
> at org.jgroups.protocols.pbcast.ClientGmsImpl.firstOfAllClients(ClientGmsImpl.java:181)
> at org.jgroups.protocols.pbcast.ClientGmsImpl.joinInternal(ClientGmsImpl.java:97)
> at org.jgroups.protocols.pbcast.ClientGmsImpl.join(ClientGmsImpl.java:41)
> at org.jgroups.protocols.pbcast.GMS.down(GMS.java:1066)
> at org.jgroups.protocols.tom.TOA.down(TOA.java:73)
> at org.jgroups.protocols.FlowControl.down(FlowControl.java:302)
> at org.jgroups.protocols.RSVP.down(RSVP.java:102)
> at org.jgroups.stack.ProtocolStack.down(ProtocolStack.java:900)
> at org.jgroups.JChannel.down(JChannel.java:644)
> at org.jgroups.JChannel._connect(JChannel.java:873)
> at org.jgroups.JChannel.connect(JChannel.java:369)
> - locked <0x0000000093c7aea0> (a org.jgroups.JChannel)
> at org.jgroups.JChannel.connect(JChannel.java:360)
> - locked <0x0000000093c7aea0> (a org.jgroups.JChannel)
> at org.infinispan.remoting.transport.jgroups.JGroupsTransport.startJGroupsChannelIfNeeded(JGroupsTransport.java:221)
> at org.infinispan.remoting.transport.jgroups.JGroupsTransport.start(JGroupsTransport.java:211)
> at sun.reflect.GeneratedMethodAccessor109.invoke(Unknown Source)
> at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:497)
> at org.infinispan.commons.util.ReflectionUtil.invokeAccessibly(ReflectionUtil.java:168)
> at org.infinispan.factories.AbstractComponentRegistry$PrioritizedMethod.invoke(AbstractComponentRegistry.java:867)
> at org.infinispan.factories.AbstractComponentRegistry.invokeStartMethods(AbstractComponentRegistry.java:633)
> at org.infinispan.factories.AbstractComponentRegistry.internalStart(AbstractComponentRegistry.java:622)
> at org.infinispan.factories.AbstractComponentRegistry.start(AbstractComponentRegistry.java:547)
> - locked <0x0000000093c7afe0> (a org.infinispan.factories.GlobalComponentRegistry)
> at org.infinispan.factories.GlobalComponentRegistry.start(GlobalComponentRegistry.java:244)
> - locked <0x0000000093c7afe0> (a org.infinispan.factories.GlobalComponentRegistry)
> at org.infinispan.manager.DefaultCacheManager.start(DefaultCacheManager.java:666)
> at org.infinispan.remoting.transport.InitialClusterSizeTest.lambda$testInitialClusterSize$799(InitialClusterSizeTest.java:47)
> at org.infinispan.remoting.transport.InitialClusterSizeTest$$Lambda$2092/593962598.run(Unknown Source)
> at org.infinispan.test.AbstractInfinispanTest$RunnableWrapper.run(AbstractInfinispanTest.java:510)
> at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> at java.lang.Thread.run(Thread.java:745)
> "ForkThread-2,InitialClusterSizeTest" #167840 prio=5 os_prio=0 tid=0x00007f1824164800 nid=0x3314 waiting on condition [0x00007f17eec40000]
> java.lang.Thread.State: TIMED_WAITING (sleeping)
> at java.lang.Thread.sleep(Native Method)
> at org.jgroups.util.Util.sleep(Util.java:1818)
> at org.jgroups.protocols.pbcast.ClientGmsImpl.firstOfAllClients(ClientGmsImpl.java:181)
> at org.jgroups.protocols.pbcast.ClientGmsImpl.joinInternal(ClientGmsImpl.java:97)
> at org.jgroups.protocols.pbcast.ClientGmsImpl.join(ClientGmsImpl.java:41)
> at org.jgroups.protocols.pbcast.GMS.down(GMS.java:1066)
> at org.jgroups.protocols.tom.TOA.down(TOA.java:73)
> at org.jgroups.protocols.FlowControl.down(FlowControl.java:302)
> at org.jgroups.protocols.RSVP.down(RSVP.java:102)
> at org.jgroups.stack.ProtocolStack.down(ProtocolStack.java:900)
> at org.jgroups.JChannel.down(JChannel.java:644)
> at org.jgroups.JChannel._connect(JChannel.java:873)
> at org.jgroups.JChannel.connect(JChannel.java:369)
> - locked <0x0000000093c7b7f0> (a org.jgroups.JChannel)
> at org.jgroups.JChannel.connect(JChannel.java:360)
> - locked <0x0000000093c7b7f0> (a org.jgroups.JChannel)
> at org.infinispan.remoting.transport.jgroups.JGroupsTransport.startJGroupsChannelIfNeeded(JGroupsTransport.java:221)
> at org.infinispan.remoting.transport.jgroups.JGroupsTransport.start(JGroupsTransport.java:211)
> at sun.reflect.GeneratedMethodAccessor109.invoke(Unknown Source)
> at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:497)
> at org.infinispan.commons.util.ReflectionUtil.invokeAccessibly(ReflectionUtil.java:168)
> at org.infinispan.factories.AbstractComponentRegistry$PrioritizedMethod.invoke(AbstractComponentRegistry.java:867)
> at org.infinispan.factories.AbstractComponentRegistry.invokeStartMethods(AbstractComponentRegistry.java:633)
> at org.infinispan.factories.AbstractComponentRegistry.internalStart(AbstractComponentRegistry.java:622)
> at org.infinispan.factories.AbstractComponentRegistry.start(AbstractComponentRegistry.java:547)
> - locked <0x0000000093c7b930> (a org.infinispan.factories.GlobalComponentRegistry)
> at org.infinispan.factories.GlobalComponentRegistry.start(GlobalComponentRegistry.java:244)
> - locked <0x0000000093c7b930> (a org.infinispan.factories.GlobalComponentRegistry)
> at org.infinispan.manager.DefaultCacheManager.start(DefaultCacheManager.java:666)
> at org.infinispan.remoting.transport.InitialClusterSizeTest.lambda$testInitialClusterSize$799(InitialClusterSizeTest.java:47)
> at org.infinispan.remoting.transport.InitialClusterSizeTest$$Lambda$2092/593962598.run(Unknown Source)
> at org.infinispan.test.AbstractInfinispanTest$RunnableWrapper.run(AbstractInfinispanTest.java:510)
> at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> at java.lang.Thread.run(Thread.java:745)
> {noformat}
> See the full thread dump here: http://ci.infinispan.org/viewLog.html?buildId=49393&buildTypeId=bt9
--
This message was sent by Atlassian JIRA
(v7.5.0#75005)
6 years, 8 months
[JBoss JIRA] (ISPN-7022) TwoWaySplitAndMergeTest.testSplitAndMerge2 fails
by Dan Berindei (JIRA)
[ https://issues.jboss.org/browse/ISPN-7022?page=com.atlassian.jira.plugin.... ]
Dan Berindei commented on ISPN-7022:
------------------------------------
Looking closer, none of the recent failures have the stack trace in the description. Most of them are caused by ISPN-9099, only one of them is different:
{noformat}
[ERROR] testSplitAndMerge3[DIST_SYNC](org.infinispan.partitionhandling.TwoWaySplitAndMergeTest) Time elapsed: 0.031 s <<< FAILURE!
java.lang.AssertionError: Sets differ: expected [TwoWaySplitAndMergeTest[DIST_SYNC]-NodeC-27237, TwoWaySplitAndMergeTest[DIST_SYNC]-NodeD-27548, TwoWaySplitAndMergeTest[DIST_SYNC]-NodeB-35865, TwoWaySplitAndMergeTest[DIST_SYNC]-NodeA-35488] but got [TwoWaySplitAndMergeTest[DIST_SYNC]-NodeC-27237, TwoWaySplitAndMergeTest[DIST_SYNC]-NodeB-35865, TwoWaySplitAndMergeTest[DIST_SYNC]-NodeA-35488]
at org.testng.Assert.fail(Assert.java:94)
at org.testng.Assert.assertEquals(Assert.java:808)
at org.testng.Assert.assertEquals(Assert.java:786)
at org.infinispan.partitionhandling.TwoWaySplitAndMergeTest.testSplitAndMerge(TwoWaySplitAndMergeTest.java:69)
at org.infinispan.partitionhandling.TwoWaySplitAndMergeTest.testSplitAndMerge3(TwoWaySplitAndMergeTest.java:39)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.testng.internal.MethodInvocationHelper.invokeMethod(MethodInvocationHelper.java:85)
at org.testng.internal.Invoker.invokeMethod(Invoker.java:639)
at org.testng.internal.Invoker.invokeTestMethod(Invoker.java:816)
at org.testng.internal.Invoker.invokeTestMethods(Invoker.java:1124)
at org.testng.internal.TestMethodWorker.invokeTestMethods(TestMethodWorker.java:125)
at org.testng.internal.TestMethodWorker.run(TestMethodWorker.java:108)
at org.testng.TestRunner.privateRun(TestRunner.java:774)
at org.testng.TestRunner.run(TestRunner.java:624)
at org.testng.SuiteRunner.runTest(SuiteRunner.java:359)
at org.testng.SuiteRunner.access$000(SuiteRunner.java:39)
at org.testng.SuiteRunner$SuiteWorker.run(SuiteRunner.java:393)
at org.testng.internal.thread.ThreadUtil$2.call(ThreadUtil.java:64)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
{noformat}
Stable topology updates are sent asynchronously after the {{NO_REBALANCE}} topology update is sent at the end of the rebalance, so it's always a good idea to use {{eventuallyEquals}} when it comes to the stable topology.
> TwoWaySplitAndMergeTest.testSplitAndMerge2 fails
> ------------------------------------------------
>
> Key: ISPN-7022
> URL: https://issues.jboss.org/browse/ISPN-7022
> Project: Infinispan
> Issue Type: Bug
> Affects Versions: 9.2.1.Final
> Reporter: Sebastian Łaskawiec
> Assignee: Dan Berindei
> Fix For: 9.3.0.Beta1
>
>
> http://ci.infinispan.org/viewLog.html?buildId=42202&tab=buildResultsDiv&b...
> {noformat}
> java.lang.AssertionError:
> at org.testng.AssertJUnit.fail(AssertJUnit.java:59)
> at org.testng.AssertJUnit.assertTrue(AssertJUnit.java:24)
> at org.infinispan.test.AbstractInfinispanTest.eventually(AbstractInfinispanTest.java:235)
> at org.infinispan.test.AbstractInfinispanTest.eventually(AbstractInfinispanTest.java:217)
> at org.infinispan.test.AbstractInfinispanTest.eventually(AbstractInfinispanTest.java:193)
> at org.infinispan.test.AbstractInfinispanTest.eventually(AbstractInfinispanTest.java:527)
> at org.infinispan.partitionhandling.BasePartitionHandlingTest.access$600(BasePartitionHandlingTest.java:44)
> at org.infinispan.partitionhandling.BasePartitionHandlingTest$Partition.assertAvailabilityMode(BasePartitionHandlingTest.java:328)
> at org.infinispan.partitionhandling.BasePartitionHandlingTest$Partition.assertDegradedMode(BasePartitionHandlingTest.java:265)
> at org.infinispan.partitionhandling.TwoWaySplitAndMergeTest.testSplitAndMerge(TwoWaySplitAndMergeTest.java:74)
> at org.infinispan.partitionhandling.TwoWaySplitAndMergeTest.testSplitAndMerge2(TwoWaySplitAndMergeTest.java:26)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:497)
> at org.testng.internal.MethodInvocationHelper.invokeMethod(MethodInvocationHelper.java:84)
> at org.testng.internal.Invoker.invokeMethod(Invoker.java:714)
> at org.testng.internal.Invoker.invokeTestMethod(Invoker.java:901)
> at org.testng.internal.Invoker.invokeTestMethods(Invoker.java:1231)
> at org.testng.internal.TestMethodWorker.invokeTestMethods(TestMethodWorker.java:127)
> at org.testng.internal.TestMethodWorker.run(TestMethodWorker.java:111)
> at org.testng.TestRunner.privateRun(TestRunner.java:767)
> at org.testng.TestRunner.run(TestRunner.java:617)
> at org.testng.SuiteRunner.runTest(SuiteRunner.java:348)
> at org.testng.SuiteRunner.access$000(SuiteRunner.java:38)
> at org.testng.SuiteRunner$SuiteWorker.run(SuiteRunner.java:382)
> at org.testng.internal.thread.ThreadUtil$2.call(ThreadUtil.java:64)
> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> at java.lang.Thread.run(Thread.java:745)
> ------- Stdout: -------
> 13:26:54,221 ERROR (testng-DelayedAvailabilityUpdateTest) [InvocationContextInterceptor] ISPN000136: Error executing command GetKeyValueCommand, writing keys []
> org.infinispan.partitionhandling.AvailabilityException: ISPN000306: Key 'MagicKey#k1Existing{1420/A1FBD015/225@DelayedAvailabilityUpdateTest-NodeT-8452}' is not available. Not all owners are in this partition
> at org.infinispan.partitionhandling.impl.PartitionHandlingManagerImpl.doCheck(PartitionHandlingManagerImpl.java:250)
> at org.infinispan.partitionhandling.impl.PartitionHandlingManagerImpl.checkRead(PartitionHandlingManagerImpl.java:102)
> at org.infinispan.partitionhandling.impl.PartitionHandlingInterceptor.postOperationPartitionCheck(PartitionHandlingInterceptor.java:188)
> at org.infinispan.partitionhandling.impl.PartitionHandlingInterceptor.lambda$handleDataReadCommand$412(PartitionHandlingInterceptor.java:145)
> at org.infinispan.partitionhandling.impl.PartitionHandlingInterceptor$$Lambda$1639/1963537113.handle(Unknown Source)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.invokeReturnHandlerSync(BaseAsyncInvocationContext.java:493)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.invokeReturnHandlersSync(BaseAsyncInvocationContext.java:125)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.forkInvocationSync(BaseAsyncInvocationContext.java:99)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.invokeForkAndHandlerSync(BaseAsyncInvocationContext.java:474)[TestSuiteProgress] Test failed: org.infinispan.partitionhandling.OptimisticTxPartitionAndMergeDuringCommitTest.testOriginatorIsolatedPartition
> 13:26:54,221 ERROR (testng-OptimisticTxPartitionAndMergeDuringCommitTest) [TestSuiteProgress] Test failed: org.infinispan.partitionhandling.OptimisticTxPartitionAndMergeDuringCommitTest.testOriginatorIsolatedPartition
> java.lang.AssertionError: There are pending transactions!
> at org.testng.AssertJUnit.fail(AssertJUnit.java:59)
> at org.testng.AssertJUnit.assertTrue(AssertJUnit.java:24)
> at org.infinispan.test.AbstractInfinispanTest.eventually(AbstractInfinispanTest.java:235)
> at org.infinispan.test.AbstractInfinispanTest.eventually(AbstractInfinispanTest.java:531)
> at org.infinispan.test.MultipleCacheManagersTest.assertNoTransactions(MultipleCacheManagersTest.java:794)
> at org.infinispan.partitionhandling.BaseTxPartitionAndMergeTest.finalAsserts(BaseTxPartitionAndMergeTest.java:96)
> at org.infinispan.partitionhandling.BaseOptimisticTxPartitionAndMergeTest.doTest(BaseOptimisticTxPartitionAndMergeTest.java:77)
> at org.infinispan.partitionhandling.OptimisticTxPartitionAndMergeDuringCommitTest.testOriginatorIsolatedPartition(OptimisticTxPartitionAndMergeDuringCommitTest.java:37)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)[TestSuiteProgress] Tests succeeded: 6704, failed: 1, skipped: 0
> 13:26:54,222 ERROR (testng-DelayedAvailabilityUpdateTest) [InvocationContextInterceptor] ISPN000136: Error executing command GetKeyValueCommand, writing keys []
> org.infinispan.partitionhandling.AvailabilityException: ISPN000306: Key 'MagicKey#k2Existing{1421/89F8758B/114@DelayedAvailabilityUpdateTest-NodeQ-17336}' is not available. Not all owners are in this partition
> at org.infinispan.partitionhandling.impl.PartitionHandlingInterceptor.postOperationPartitionCheck(PartitionHandlingInterceptor.java:197)
> at org.infinispan.partitionhandling.impl.PartitionHandlingInterceptor.lambda$handleDataReadCommand$412(PartitionHandlingInterceptor.java:145)
> at org.infinispan.partitionhandling.impl.PartitionHandlingInterceptor$$Lambda$1639/1963537113.handle(Unknown Source)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.invokeReturnHandlerSync(BaseAsyncInvocationContext.java:493)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.invokeReturnHandlersSync(BaseAsyncInvocationContext.java:125)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.forkInvocationSync(BaseAsyncInvocationContext.java:99)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.invokeForkAndHandlerSync(BaseAsyncInvocationContext.java:474)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.afterVisit(BaseAsyncInvocationContext.java:463)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.invokeInterceptorsSync(BaseAsyncInvocationContext.java:322)13:26:54,222 ERROR (testng-DelayedAvailabilityUpdateTest) [InvocationContextInterceptor] ISPN000136: Error executing command GetKeyValueCommand, writing keys []
> org.infinispan.partitionhandling.AvailabilityException: ISPN000306: Key 'MagicKey#k2Existing{1421/89F8758B/114@DelayedAvailabilityUpdateTest-NodeQ-17336}' is not available. Not all owners are in this partition
> at org.infinispan.partitionhandling.impl.PartitionHandlingManagerImpl.doCheck(PartitionHandlingManagerImpl.java:250)
> at org.infinispan.partitionhandling.impl.PartitionHandlingManagerImpl.checkRead(PartitionHandlingManagerImpl.java:102)
> at org.infinispan.partitionhandling.impl.PartitionHandlingInterceptor.postOperationPartitionCheck(PartitionHandlingInterceptor.java:188)
> at org.infinispan.partitionhandling.impl.PartitionHandlingInterceptor.lambda$handleDataReadCommand$412(PartitionHandlingInterceptor.java:145)
> at org.infinispan.partitionhandling.impl.PartitionHandlingInterceptor$$Lambda$1639/1963537113.handle(Unknown Source)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.invokeReturnHandlerSync(BaseAsyncInvocationContext.java:493)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.invokeReturnHandlersSync(BaseAsyncInvocationContext.java:125)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.forkInvocationSync(BaseAsyncInvocationContext.java:99)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.invokeForkAndHandlerSync(BaseAsyncInvocationContext.java:474)13:26:49,964 ERROR (transport-thread-PessimisticTxPartitionAndMergeDuringPrepareTest-NodeG-p44001-t2) [CLUSTER] [Context=pes-cache]ISPN000319: After merge (or coordinator change), cache still hasn't recovered all its data and must stay in degraded mode. Current members are [PessimisticTxPartitionAndMergeDuringPrepareTest-NodeG-37404, PessimisticTxPartitionAndMergeDuringPrepareTest-NodeH-18972], lost members are [PessimisticTxPartitionAndMergeDuringPrepareTest-NodeE-27252, PessimisticTxPartitionAndMergeDuringPrepareTest-NodeF-47835], stable members are [PessimisticTxPartitionAndMergeDuringPrepareTest-NodeE-27252, PessimisticTxPartitionAndMergeDuringPrepareTest-NodeF-47835, PessimisticTxPartitionAndMergeDuringPrepareTest-NodeG-37404, PessimisticTxPartitionAndMergeDuringPrepareTest-NodeH-18972]
> [TestSuiteProgress] Test failed: org.infinispan.partitionhandling.TwoWaySplitAndMergeTest.testSplitAndMerge2
> 13:26:55,721 ERROR (testng-TwoWaySplitAndMergeTest) [TestSuiteProgress] Test failed: org.infinispan.partitionhandling.TwoWaySplitAndMergeTest.testSplitAndMerge2
> java.lang.AssertionError:
> at org.testng.AssertJUnit.fail(AssertJUnit.java:59)
> at org.testng.AssertJUnit.assertTrue(AssertJUnit.java:24)
> at org.infinispan.test.AbstractInfinispanTest.eventually(AbstractInfinispanTest.java:235)
> at org.infinispan.test.AbstractInfinispanTest.eventually(AbstractInfinispanTest.java:217)
> at org.infinispan.test.AbstractInfinispanTest.eventually(AbstractInfinispanTest.java:193)
> at org.infinispan.test.AbstractInfinispanTest.eventually(AbstractInfinispanTest.java:527)
> at org.infinispan.partitionhandling.BasePartitionHandlingTest.access$600(BasePartitionHandlingTest.java:44)
> at org.infinispan.partitionhandling.BasePartitionHandlingTest$Partition.assertAvailabilityMode(BasePartitionHandlingTest.java:328)
> at org.infinispan.partitionhandling.BasePartitionHandlingTest$Partition.assertDegradedMode(BasePartitionHandlingTest.java:265)13:26:49,964 ERROR (transport-thread-PessimisticTxPartitionAndMergeDuringPrepareTest-NodeG-p44001-t3) [CLUSTER] [Context=___defaultcache]ISPN000319: After merge (or coordinator change), cache still hasn't recovered all its data and must stay in degraded mode. Current members are [PessimisticTxPartitionAndMergeDuringPrepareTest-NodeG-37404, PessimisticTxPartitionAndMergeDuringPrepareTest-NodeH-18972], lost members are [PessimisticTxPartitionAndMergeDuringPrepareTest-NodeE-27252, PessimisticTxPartitionAndMergeDuringPrepareTest-NodeF-47835], stable members are [PessimisticTxPartitionAndMergeDuringPrepareTest-NodeE-27252, PessimisticTxPartitionAndMergeDuringPrepareTest-NodeF-47835, PessimisticTxPartitionAndMergeDuringPrepareTest-NodeG-37404, PessimisticTxPartitionAndMergeDuringPrepareTest-NodeH-18972]
> [TestSuiteProgress] Tests succeeded: 6704, failed: 2, skipped: 0
> 13:26:55,095 ERROR (testng-OptimisticTxPartitionAndMergeDuringPrepareTest) [InvocationContextInterceptor] ISPN000136: Error executing command ClearCommand, writing keys []
> org.infinispan.partitionhandling.AvailabilityException: ISPN000307: Cannot clear when the cluster is partitioned
> at org.infinispan.partitionhandling.impl.PartitionHandlingManagerImpl.checkClear(PartitionHandlingManagerImpl.java:108)
> at org.infinispan.partitionhandling.impl.PartitionHandlingInterceptor.visitClearCommand(PartitionHandlingInterceptor.java:95)
> at org.infinispan.partitionhandling.impl.PartitionHandlingInterceptor.visitClearCommand(PartitionHandlingInterceptor.java:38)
> at org.infinispan.commands.write.ClearCommand.acceptVisitor(ClearCommand.java:44)
> at org.infinispan.interceptors.DDAsyncInterceptor.visitCommand(DDAsyncInterceptor.java:53)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.invokeInterceptorsSync(BaseAsyncInvocationContext.java:307)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.forkInvocationSync(BaseAsyncInvocationContext.java:98)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.invokeForkAndHandlerSync(BaseAsyncInvocationContext.java:474)
> at org.infinispan.interceptors.impl.BaseAsyncInvocationContext.afterVisit(BaseAsyncInvocationContext.java:463)
> {noformat}
--
This message was sent by Atlassian JIRA
(v7.5.0#75005)
6 years, 8 months
[JBoss JIRA] (ISPN-8969) CacheManagerAdminTest.testClusterCacheTest random failure
by Tristan Tarrant (JIRA)
[ https://issues.jboss.org/browse/ISPN-8969?page=com.atlassian.jira.plugin.... ]
Tristan Tarrant updated ISPN-8969:
----------------------------------
Status: Open (was: New)
> CacheManagerAdminTest.testClusterCacheTest random failure
> ---------------------------------------------------------
>
> Key: ISPN-8969
> URL: https://issues.jboss.org/browse/ISPN-8969
> Project: Infinispan
> Issue Type: Bug
> Components: Test Suite - Core
> Affects Versions: 9.2.0.Final
> Reporter: William Burns
> Assignee: Tristan Tarrant
>
> https://ci.infinispan.org/job/Infinispan/job/master/510/testReport/junit/...
> Stack trace in case if the history is lost:
> {code}
> java.lang.AssertionError: Cache 'a' should NOT be present on org.infinispan.manager.DefaultCacheManager@592285b5@Address:CacheManagerAdminTest-NodeC-58305
> at org.infinispan.manager.CacheManagerAdminTest.checkCacheExistenceAcrossCluster(CacheManagerAdminTest.java:72)
> at org.infinispan.manager.CacheManagerAdminTest.testClusterCacheTest(CacheManagerAdminTest.java:61)
> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> at java.lang.Thread.run(Thread.java:748)
> ... Removed 19 stack frames
> {code}
--
This message was sent by Atlassian JIRA
(v7.5.0#75005)
6 years, 8 months