[JBoss JIRA] (ISPN-5526) Replication: The DELTA_WRITE flag should force a remote get during state transfer
by Ryan Emerson (JIRA)
[ https://issues.jboss.org/browse/ISPN-5526?page=com.atlassian.jira.plugin.... ]
Ryan Emerson reassigned ISPN-5526:
----------------------------------
Assignee: Ryan Emerson (was: Dan Berindei)
> Replication: The DELTA_WRITE flag should force a remote get during state transfer
> ---------------------------------------------------------------------------------
>
> Key: ISPN-5526
> URL: https://issues.jboss.org/browse/ISPN-5526
> Project: Infinispan
> Issue Type: Bug
> Affects Versions: 5.2.10.Final
> Reporter: Dennis Reed
> Assignee: Ryan Emerson
> Priority: Critical
>
> Same issue as ISPN-3184, but for repl caches in Infinispan 5.2.x.
> (ISPN-3184 only fixed dist caches, since repl uses the same code in 5.3+).
> AtomicHashMap and FineGrainedAtomicHashMap, as well as custom DeltaAware implementations, use PutKeyValueCommands with the DELTA_WRITE flag to execute incremental updates. These commands need the previous value of the entry in order to work.
> If a node is joining and it receives a PutKeyValueCommand with the DELTA_WRITE flag before it has received the value of the affected key, it should do a remote get to retrieve the previous value and apply the change on top of that value, just like we do for conditional commands. Not doing so leads to data loss.
--
This message was sent by Atlassian JIRA
(v6.3.15#6346)
9 years, 7 months
[JBoss JIRA] (ISPN-5526) Replication: The DELTA_WRITE flag should force a remote get during state transfer
by RH Bugzilla Integration (JIRA)
[ https://issues.jboss.org/browse/ISPN-5526?page=com.atlassian.jira.plugin.... ]
RH Bugzilla Integration commented on ISPN-5526:
-----------------------------------------------
Ryan Emerson <remerson(a)redhat.com> changed the Status of [bug 1228780|https://bugzilla.redhat.com/show_bug.cgi?id=1228780] from NEW to ASSIGNED
> Replication: The DELTA_WRITE flag should force a remote get during state transfer
> ---------------------------------------------------------------------------------
>
> Key: ISPN-5526
> URL: https://issues.jboss.org/browse/ISPN-5526
> Project: Infinispan
> Issue Type: Bug
> Affects Versions: 5.2.10.Final
> Reporter: Dennis Reed
> Assignee: Dan Berindei
> Priority: Critical
>
> Same issue as ISPN-3184, but for repl caches in Infinispan 5.2.x.
> (ISPN-3184 only fixed dist caches, since repl uses the same code in 5.3+).
> AtomicHashMap and FineGrainedAtomicHashMap, as well as custom DeltaAware implementations, use PutKeyValueCommands with the DELTA_WRITE flag to execute incremental updates. These commands need the previous value of the entry in order to work.
> If a node is joining and it receives a PutKeyValueCommand with the DELTA_WRITE flag before it has received the value of the affected key, it should do a remote get to retrieve the previous value and apply the change on top of that value, just like we do for conditional commands. Not doing so leads to data loss.
--
This message was sent by Atlassian JIRA
(v6.3.15#6346)
9 years, 7 months
[JBoss JIRA] (ISPN-5534) CreateCacheCommand doesn't properly wait for the cache to stabilize
by Dan Berindei (JIRA)
[ https://issues.jboss.org/browse/ISPN-5534?page=com.atlassian.jira.plugin.... ]
Dan Berindei updated ISPN-5534:
-------------------------------
Description:
Because of ISPN-5533, {{CreateCacheCommand.perform()}} waits for the cache to stabilize with a certain number of members. However, the wait condition is wrong: it will exit the loop when either the number of members is correct *or* there is no rebalance in progress.
{code}
while (stm.getCacheTopology().getMembers().size() != expectedSize && stm.getCacheTopology().getPendingCH() != null) {
{code}
should be
{code}
while (stm.getCacheTopology().getMembers().size() != expectedSize || stm.getCacheTopology().getPendingCH() != null) {
{code}
Combined with ISPN-5533, this is causing random failures in the M/R tests, e.g.
{noformat}
23:17:45,353 ERROR (testng-DistributedFourNodesMapReduceTest:) [UnitTestTestNGListener] Test testInvokeMapReduceOnAllKeysWithResultCache(org.infinispan.distexec.mapreduce.DistributedFourNodesMapReduceTest) failed.
java.lang.AssertionError: key 'community' does not have count 2 but 4
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.assertTrue(Assert.java:41)
at org.infinispan.distexec.mapreduce.BaseWordCountMapReduceTest.verifyResults(BaseWordCountMapReduceTest.java:256)
at org.infinispan.distexec.mapreduce.BaseWordCountMapReduceTest.verifyResults(BaseWordCountMapReduceTest.java:380)
at org.infinispan.distexec.mapreduce.BaseWordCountMapReduceTest.testInvokeMapReduceOnAllKeysWithResultCache(BaseWordCountMapReduceTest.java:175)
{noformat}
was:
Because of ISPN-5533, {{CreateCacheCommand.perform()}} waits for the cache to stabilize with a certain number of members. However, the wait condition is wrong: it will exit the loop when either the number of members is correct *or* there is no rebalance in progress.
{code}
while (stm.getCacheTopology().getMembers().size() != expectedSize || stm.getCacheTopology().getPendingCH() != null) {
{code}
Combined with ISPN-5533, this is causing random failures in the M/R tests, e.g.
{noformat}
23:17:45,353 ERROR (testng-DistributedFourNodesMapReduceTest:) [UnitTestTestNGListener] Test testInvokeMapReduceOnAllKeysWithResultCache(org.infinispan.distexec.mapreduce.DistributedFourNodesMapReduceTest) failed.
java.lang.AssertionError: key 'community' does not have count 2 but 4
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.assertTrue(Assert.java:41)
at org.infinispan.distexec.mapreduce.BaseWordCountMapReduceTest.verifyResults(BaseWordCountMapReduceTest.java:256)
at org.infinispan.distexec.mapreduce.BaseWordCountMapReduceTest.verifyResults(BaseWordCountMapReduceTest.java:380)
at org.infinispan.distexec.mapreduce.BaseWordCountMapReduceTest.testInvokeMapReduceOnAllKeysWithResultCache(BaseWordCountMapReduceTest.java:175)
{noformat}
> CreateCacheCommand doesn't properly wait for the cache to stabilize
> -------------------------------------------------------------------
>
> Key: ISPN-5534
> URL: https://issues.jboss.org/browse/ISPN-5534
> Project: Infinispan
> Issue Type: Bug
> Components: Core
> Affects Versions: 7.2.2.Final, 8.0.0.Alpha1
> Reporter: Dan Berindei
> Assignee: Dan Berindei
> Priority: Blocker
> Labels: testsuite_stability
> Fix For: 8.0.0.Alpha2
>
>
> Because of ISPN-5533, {{CreateCacheCommand.perform()}} waits for the cache to stabilize with a certain number of members. However, the wait condition is wrong: it will exit the loop when either the number of members is correct *or* there is no rebalance in progress.
> {code}
> while (stm.getCacheTopology().getMembers().size() != expectedSize && stm.getCacheTopology().getPendingCH() != null) {
> {code}
> should be
> {code}
> while (stm.getCacheTopology().getMembers().size() != expectedSize || stm.getCacheTopology().getPendingCH() != null) {
> {code}
> Combined with ISPN-5533, this is causing random failures in the M/R tests, e.g.
> {noformat}
> 23:17:45,353 ERROR (testng-DistributedFourNodesMapReduceTest:) [UnitTestTestNGListener] Test testInvokeMapReduceOnAllKeysWithResultCache(org.infinispan.distexec.mapreduce.DistributedFourNodesMapReduceTest) failed.
> java.lang.AssertionError: key 'community' does not have count 2 but 4
> at org.junit.Assert.fail(Assert.java:88)
> at org.junit.Assert.assertTrue(Assert.java:41)
> at org.infinispan.distexec.mapreduce.BaseWordCountMapReduceTest.verifyResults(BaseWordCountMapReduceTest.java:256)
> at org.infinispan.distexec.mapreduce.BaseWordCountMapReduceTest.verifyResults(BaseWordCountMapReduceTest.java:380)
> at org.infinispan.distexec.mapreduce.BaseWordCountMapReduceTest.testInvokeMapReduceOnAllKeysWithResultCache(BaseWordCountMapReduceTest.java:175)
> {noformat}
--
This message was sent by Atlassian JIRA
(v6.3.15#6346)
9 years, 7 months
[JBoss JIRA] (ISPN-5534) CreateCacheCommand doesn't properly wait for the cache to stabilize
by Dan Berindei (JIRA)
Dan Berindei created ISPN-5534:
----------------------------------
Summary: CreateCacheCommand doesn't properly wait for the cache to stabilize
Key: ISPN-5534
URL: https://issues.jboss.org/browse/ISPN-5534
Project: Infinispan
Issue Type: Bug
Components: Core
Affects Versions: 8.0.0.Alpha1, 7.2.2.Final
Reporter: Dan Berindei
Assignee: Dan Berindei
Priority: Blocker
Fix For: 8.0.0.Alpha2
Because of ISPN-5533, {{CreateCacheCommand.perform()}} waits for the cache to stabilize with a certain number of members. However, the wait condition is wrong: it will exit the loop when either the number of members is correct *or* there is no rebalance in progress.
{code}
while (stm.getCacheTopology().getMembers().size() != expectedSize || stm.getCacheTopology().getPendingCH() != null) {
{code}
Combined with ISPN-5533, this is causing random failures in the M/R tests, e.g.
{noformat}
23:17:45,353 ERROR (testng-DistributedFourNodesMapReduceTest:) [UnitTestTestNGListener] Test testInvokeMapReduceOnAllKeysWithResultCache(org.infinispan.distexec.mapreduce.DistributedFourNodesMapReduceTest) failed.
java.lang.AssertionError: key 'community' does not have count 2 but 4
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.assertTrue(Assert.java:41)
at org.infinispan.distexec.mapreduce.BaseWordCountMapReduceTest.verifyResults(BaseWordCountMapReduceTest.java:256)
at org.infinispan.distexec.mapreduce.BaseWordCountMapReduceTest.verifyResults(BaseWordCountMapReduceTest.java:380)
at org.infinispan.distexec.mapreduce.BaseWordCountMapReduceTest.testInvokeMapReduceOnAllKeysWithResultCache(BaseWordCountMapReduceTest.java:175)
{noformat}
--
This message was sent by Atlassian JIRA
(v6.3.15#6346)
9 years, 7 months
[JBoss JIRA] (ISPN-5533) M/R DeltaAwareList can add duplicate values because of topology changes
by Dan Berindei (JIRA)
Dan Berindei created ISPN-5533:
----------------------------------
Summary: M/R DeltaAwareList can add duplicate values because of topology changes
Key: ISPN-5533
URL: https://issues.jboss.org/browse/ISPN-5533
Project: Infinispan
Issue Type: Bug
Components: Core, Distributed Execution and Map/Reduce
Affects Versions: 8.0.0.Alpha1, 7.2.2.Final
Reporter: Dan Berindei
Fix For: 8.0.0.Final
By default, the intermediate cache is non-transactional, so a topology change will cause write commands to be retried. Because a {{PutKeyValueCommand(K, DeltaAwareList)}} command is not idempotent, a retried command will append extra intermediate values to the list.
The M/R framework tries to guard against this by waiting for all the nodes to initialize the intermediate cache before starting the reduce phase, but it cannot guard against nodes joining or leaving during the reduce phase.
--
This message was sent by Atlassian JIRA
(v6.3.15#6346)
9 years, 7 months
[JBoss JIRA] (ISPN-5509) Memory based eviction doesn't work with store as binary
by Galder Zamarreño (JIRA)
[ https://issues.jboss.org/browse/ISPN-5509?page=com.atlassian.jira.plugin.... ]
Galder Zamarreño updated ISPN-5509:
-----------------------------------
Status: Resolved (was: Pull Request Sent)
Fix Version/s: 8.0.0.Final
Resolution: Done
> Memory based eviction doesn't work with store as binary
> -------------------------------------------------------
>
> Key: ISPN-5509
> URL: https://issues.jboss.org/browse/ISPN-5509
> Project: Infinispan
> Issue Type: Enhancement
> Components: Eviction
> Affects Versions: 8.0.0.Alpha1
> Reporter: William Burns
> Assignee: William Burns
> Priority: Critical
> Fix For: 8.0.0.Alpha2, 8.0.0.Final
>
>
> The memory based eviction only currently works with byte[] for key and values.
> We would like this also to support any key/value pair when store as binary is enabled. We would need to add a EntrySizeCalculator that would support MarshalledValue, the various MarshalledValueByteStream classes as well as all the types that MarshalledValue doesn't wrap (ie. Integer, Long etc.)
--
This message was sent by Atlassian JIRA
(v6.3.15#6346)
9 years, 7 months
[JBoss JIRA] (ISPN-4312) Distributed(Intermediate)SharedCacheFourNodesMapReduceTest randomly fails on Windows
by Dan Berindei (JIRA)
[ https://issues.jboss.org/browse/ISPN-4312?page=com.atlassian.jira.plugin.... ]
Dan Berindei commented on ISPN-4312:
------------------------------------
The recent timeout failures are probably caused by ISPN-5524.
> Distributed(Intermediate)SharedCacheFourNodesMapReduceTest randomly fails on Windows
> ------------------------------------------------------------------------------------
>
> Key: ISPN-4312
> URL: https://issues.jboss.org/browse/ISPN-4312
> Project: Infinispan
> Issue Type: Bug
> Components: Core, Distributed Execution and Map/Reduce, Test Suite - Core
> Affects Versions: 7.0.0.Alpha4
> Environment: Windows machines, 2k8r2, 2k12r2
> Reporter: Tomas Sykora
> Priority: Blocker
> Labels: testsuite_stability
>
> It looks like tests have problems with fitting into 15 sec timeout on Windows.
> org.infinispan.commons.CacheException: java.util.concurrent.ExecutionException: Reduce phase executing at DistributedSharedCacheFourNodesMapReduceTest-NodeA-36515 did not complete within 15 sec timeout
> at org.infinispan.distexec.mapreduce.MapReduceTask.executeHelper(MapReduceTask.java:498)
> at org.infinispan.distexec.mapreduce.MapReduceTask.execute(MapReduceTask.java:447)
> at org.infinispan.distexec.mapreduce.BaseWordCountMapReduceTest.testInvokeMapReduceOnAllKeysWithResultCache(BaseWordCountMapReduceTest.java:173)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> at java.lang.reflect.Method.invoke(Method.java:597)
> at org.testng.internal.MethodInvocationHelper.invokeMethod(MethodInvocationHelper.java:80)
> at org.testng.internal.Invoker.invokeMethod(Invoker.java:714)
> at org.testng.internal.Invoker.invokeTestMethod(Invoker.java:901)
> at org.testng.internal.Invoker.invokeTestMethods(Invoker.java:1231)
> at org.testng.internal.TestMethodWorker.invokeTestMethods(TestMethodWorker.java:127)
> at org.testng.internal.TestMethodWorker.run(TestMethodWorker.java:111)
> at org.testng.TestRunner.privateRun(TestRunner.java:767)
> at org.testng.TestRunner.run(TestRunner.java:617)
> at org.testng.SuiteRunner.runTest(SuiteRunner.java:334)
> at org.testng.SuiteRunner.access$000(SuiteRunner.java:37)
> at org.testng.SuiteRunner$SuiteWorker.run(SuiteRunner.java:368)
> at org.testng.internal.thread.ThreadUtil$2.call(ThreadUtil.java:64)
> at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
> at java.util.concurrent.FutureTask.run(FutureTask.java:138)
> at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:895)
> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:918)
> at java.lang.Thread.run(Thread.java:662)
> Caused by: java.util.concurrent.ExecutionException: Reduce phase executing at DistributedSharedCacheFourNodesMapReduceTest-NodeA-36515 did not complete within 15 sec timeout
> at org.infinispan.distexec.mapreduce.MapReduceTask.executeReducePhase(MapReduceTask.java:748)
> at org.infinispan.distexec.mapreduce.MapReduceTask.executeHelper(MapReduceTask.java:495)
> ... 23 more
> Caused by: org.infinispan.util.concurrent.TimeoutException: Node DistributedSharedCacheFourNodesMapReduceTest-NodeD-54406 timed out
> at org.infinispan.remoting.transport.jgroups.CommandAwareRpcDispatcher.invokeRemoteCommand(CommandAwareRpcDispatcher.java:174)
> at org.infinispan.remoting.transport.jgroups.JGroupsTransport.invokeRemotely(JGroupsTransport.java:521)
> at org.infinispan.remoting.rpc.RpcManagerImpl.invokeRemotely(RpcManagerImpl.java:280)
> at org.infinispan.remoting.rpc.RpcManagerImpl$2.call(RpcManagerImpl.java:312)
> ... 5 more
> Caused by: org.jgroups.TimeoutException: timeout sending message to DistributedSharedCacheFourNodesMapReduceTest-NodeD-54406
> at org.jgroups.blocks.MessageDispatcher.sendMessage(MessageDispatcher.java:419)
> at org.infinispan.remoting.transport.jgroups.CommandAwareRpcDispatcher.processSingleCall(CommandAwareRpcDispatcher.java:349)
> at org.infinispan.remoting.transport.jgroups.CommandAwareRpcDispatcher.invokeRemoteCommand(CommandAwareRpcDispatcher.java:167)
> ... 8 more
--
This message was sent by Atlassian JIRA
(v6.3.15#6346)
9 years, 7 months
[JBoss JIRA] (ISPN-5532) HotRod client retries on the same server after receiving an IllegalLifecycleStateException
by Dan Berindei (JIRA)
Dan Berindei created ISPN-5532:
----------------------------------
Summary: HotRod client retries on the same server after receiving an IllegalLifecycleStateException
Key: ISPN-5532
URL: https://issues.jboss.org/browse/ISPN-5532
Project: Infinispan
Issue Type: Bug
Components: Remote Protocols, Test Suite - Server
Affects Versions: 8.0.0.Alpha1, 7.2.2.Final
Reporter: Dan Berindei
Priority: Blocker
Fix For: 8.0.0.Alpha2
When it receives an IllegalLifecycleStateException, the HotRod client retries the operation, but it may do so on the same server and it may have to retry again. Since the ISPN-5477 fix, {{DistTopologyChangeUnderLoadTest}} uses only one retry, and the test will fail if that retry happens to be on the same server. It would be better to exclude the server that returned the error when retrying.
{noformat}
java.util.concurrent.ExecutionException: org.infinispan.client.hotrod.exceptions.RemoteIllegalLifecycleStateException:Request for message id[9390] returned server error (status=0x88): org.infinispan.IllegalLifecycleStateException: ISPN000323: Default cache is in 'TERMINATED' state and so it does not accept new invocations. Either restart it or recreate the cache container.
at java.util.concurrent.FutureTask.report(FutureTask.java:122)
at java.util.concurrent.FutureTask.get(FutureTask.java:192)
at org.infinispan.client.hotrod.DistTopologyChangeUnderLoadTest.testPutsSucceedWhileTopologyChanges(DistTopologyChangeUnderLoadTest.java:67)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:483)
at org.testng.internal.MethodInvocationHelper.invokeMethod(MethodInvocationHelper.java:84)
at org.testng.internal.Invoker.invokeMethod(Invoker.java:714)
at org.testng.internal.Invoker.invokeTestMethod(Invoker.java:901)
at org.testng.internal.Invoker.invokeTestMethods(Invoker.java:1231)
at org.testng.internal.TestMethodWorker.invokeTestMethods(TestMethodWorker.java:127)
at org.testng.internal.TestMethodWorker.run(TestMethodWorker.java:111)
at org.testng.TestRunner.privateRun(TestRunner.java:767)
at org.testng.TestRunner.run(TestRunner.java:617)
at org.testng.SuiteRunner.runTest(SuiteRunner.java:348)
at org.testng.SuiteRunner.access$000(SuiteRunner.java:38)
at org.testng.SuiteRunner$SuiteWorker.run(SuiteRunner.java:382)
at org.testng.internal.thread.ThreadUtil$2.call(ThreadUtil.java:64)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
Caused by: org.infinispan.client.hotrod.exceptions.RemoteIllegalLifecycleStateException:Request for message id[9390] returned server error (status=0x88): org.infinispan.IllegalLifecycleStateException: ISPN000323: Default cache is in 'TERMINATED' state and so it does not accept new invocations. Either restart it or recreate the cache container.
at org.infinispan.client.hotrod.impl.protocol.Codec20.checkForErrorsInResponseStatus(Codec20.java:340)
at org.infinispan.client.hotrod.impl.protocol.Codec20.readPartialHeader(Codec20.java:126)
at org.infinispan.client.hotrod.impl.protocol.Codec20.readHeader(Codec20.java:112)
at org.infinispan.client.hotrod.impl.operations.HotRodOperation.readHeaderAndValidate(HotRodOperation.java:56)
at org.infinispan.client.hotrod.impl.operations.AbstractKeyValueOperation.sendPutOperation(AbstractKeyValueOperation.java:57)
at org.infinispan.client.hotrod.impl.operations.PutOperation.executeOperation(PutOperation.java:31)
at org.infinispan.client.hotrod.impl.operations.PutOperation.executeOperation(PutOperation.java:20)
at org.infinispan.client.hotrod.impl.operations.RetryOnFailureOperation.execute(RetryOnFailureOperation.java:52)
at org.infinispan.client.hotrod.impl.RemoteCacheImpl.put(RemoteCacheImpl.java:247)
at org.infinispan.client.hotrod.impl.RemoteCacheSupport.put(RemoteCacheSupport.java:79)
at org.infinispan.client.hotrod.DistTopologyChangeUnderLoadTest$PutHammer.call(DistTopologyChangeUnderLoadTest.java:80)
at org.infinispan.client.hotrod.DistTopologyChangeUnderLoadTest$PutHammer.call(DistTopologyChangeUnderLoadTest.java:72)
at org.infinispan.test.AbstractInfinispanTest$LoggingCallable.call(AbstractInfinispanTest.java:435)
{noformat}
--
This message was sent by Atlassian JIRA
(v6.3.15#6346)
9 years, 7 months
[JBoss JIRA] (ISPN-5531) java.lang.UnsupportedOperationException during remove (using RemoteCacheManager)
by Enrico Olivelli (JIRA)
[ https://issues.jboss.org/browse/ISPN-5531?page=com.atlassian.jira.plugin.... ]
Enrico Olivelli updated ISPN-5531:
----------------------------------
Forum Reference: https://developer.jboss.org/message/933152
> java.lang.UnsupportedOperationException during remove (using RemoteCacheManager)
> --------------------------------------------------------------------------------
>
> Key: ISPN-5531
> URL: https://issues.jboss.org/browse/ISPN-5531
> Project: Infinispan
> Issue Type: Bug
> Components: Core
> Affects Versions: 7.2.2.Final
> Reporter: Enrico Olivelli
>
> during a "remove" from a java HotRod client we got this error.
> This error happens only when there are more than one hotrod server in the cluster.
> We are using hotrod server in library mode (embedded)
> {code}
> 15/06/09 08:05:34 ERROR interceptors.InvocationContextInterceptor: ISPN000136: Execution error
> org.infinispan.commons.CacheListenerException: ISPN000280: Caught exception [java.lang.UnsupportedOperationException] while invoking method [public void org.infinispan.n
> otifications.cachelistener.cluster.RemoteClusterListener.handleClusterEvents(org.infinispan.notifications.cachelistener.event.CacheEntryEvent) throws java.lang.Exception] on listener instance: org.infinispan.not
> ifications.cachelistener.cluster.RemoteClusterListener@28d94aee
> at org.infinispan.notifications.impl.AbstractListenerImpl$ListenerInvocationImpl$1.run(AbstractListenerImpl.java:291)
> at org.infinispan.util.concurrent.WithinThreadExecutor.execute(WithinThreadExecutor.java:22)
> at org.infinispan.notifications.impl.AbstractListenerImpl$ListenerInvocationImpl.invoke(AbstractListenerImpl.java:309)
> at org.infinispan.notifications.cachelistener.CacheNotifierImpl$BaseCacheEntryListenerInvocation.doRealInvocation(CacheNotifierImpl.java:1180)
> at org.infinispan.notifications.cachelistener.CacheNotifierImpl$BaseCacheEntryListenerInvocation.invokeNoChecks(CacheNotifierImpl.java:1175)
> at org.infinispan.notifications.cachelistener.CacheNotifierImpl$BaseCacheEntryListenerInvocation.invoke(CacheNotifierImpl.java:1152)
> at org.infinispan.notifications.cachelistener.CacheNotifierImpl.notifyCacheEntryRemoved(CacheNotifierImpl.java:333)
> at org.infinispan.commands.write.RemoveCommand.notify(RemoveCommand.java:97)
> at org.infinispan.interceptors.locking.ClusteringDependentLogic$AbstractClusteringDependentLogic.notifyCommitEntry(ClusteringDependentLogic.java:130)
> at org.infinispan.interceptors.locking.ClusteringDependentLogic$DistributionLogic.commitSingleEntry(ClusteringDependentLogic.java:493)
> at org.infinispan.interceptors.locking.ClusteringDependentLogic$AbstractClusteringDependentLogic.commitEntry(ClusteringDependentLogic.java:108)
> at org.infinispan.interceptors.EntryWrappingInterceptor.commitContextEntry(EntryWrappingInterceptor.java:371)
> at org.infinispan.interceptors.EntryWrappingInterceptor.commitEntryIfNeeded(EntryWrappingInterceptor.java:549)
> at org.infinispan.interceptors.EntryWrappingInterceptor.commitContextEntries(EntryWrappingInterceptor.java:348)
> at org.infinispan.interceptors.EntryWrappingInterceptor.invokeNextAndApplyChanges(EntryWrappingInterceptor.java:422)
> at org.infinispan.interceptors.EntryWrappingInterceptor.setSkipRemoteGetsAndInvokeNextForDataCommand(EntryWrappingInterceptor.java:453)
> at org.infinispan.interceptors.EntryWrappingInterceptor.visitRemoveCommand(EntryWrappingInterceptor.java:252)
> at org.infinispan.commands.write.RemoveCommand.acceptVisitor(RemoveCommand.java:58)
> at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:97)
> at org.infinispan.interceptors.distribution.L1LastChanceInterceptor.visitDataWriteCommand(L1LastChanceInterceptor.java:79)
> at org.infinispan.interceptors.distribution.L1LastChanceInterceptor.visitRemoveCommand(L1LastChanceInterceptor.java:75)
> at org.infinispan.commands.write.RemoveCommand.acceptVisitor(RemoveCommand.java:58)
> at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:97)
> at org.infinispan.interceptors.locking.AbstractLockingInterceptor.visitNonTxDataWriteCommand(AbstractLockingInterceptor.java:88)
> at org.infinispan.interceptors.locking.NonTransactionalLockingInterceptor.visitDataWriteCommand(NonTransactionalLockingInterceptor.java:40)
> at org.infinispan.interceptors.locking.AbstractLockingInterceptor.visitRemoveCommand(AbstractLockingInterceptor.java:65)
> at org.infinispan.commands.write.RemoveCommand.acceptVisitor(RemoveCommand.java:58)
> at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:97)
> at org.infinispan.interceptors.base.CommandInterceptor.handleDefault(CommandInterceptor.java:111)
> at org.infinispan.commands.AbstractVisitor.visitRemoveCommand(AbstractVisitor.java:49)
> at org.infinispan.commands.write.RemoveCommand.acceptVisitor(RemoveCommand.java:58)
> at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:97)
> at org.infinispan.statetransfer.StateTransferInterceptor.handleNonTxWriteCommand(StateTransferInterceptor.java:324)
> at org.infinispan.statetransfer.StateTransferInterceptor.handleWriteCommand(StateTransferInterceptor.java:256)
> at org.infinispan.statetransfer.StateTransferInterceptor.visitRemoveCommand(StateTransferInterceptor.java:130)
> at org.infinispan.commands.write.RemoveCommand.acceptVisitor(RemoveCommand.java:58)
> at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:97)
> at org.infinispan.interceptors.CacheMgmtInterceptor.visitRemoveCommand(CacheMgmtInterceptor.java:209)
> at org.infinispan.commands.write.RemoveCommand.acceptVisitor(RemoveCommand.java:58)
> at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:97)
> at org.infinispan.interceptors.compat.BaseTypeConverterInterceptor.visitRemoveCommand(BaseTypeConverterInterceptor.java:228
> at org.infinispan.commands.write.RemoveCommand.acceptVisitor(RemoveCommand.java:58)
> at org.infinispan.interceptors.base.CommandInterceptor.invokeNextInterceptor(CommandInterceptor.java:97)
> at org.infinispan.interceptors.InvocationContextInterceptor.handleAll(InvocationContextInterceptor.java:102)
> at org.infinispan.interceptors.InvocationContextInterceptor.handleDefault(InvocationContextInterceptor.java:71)
> at org.infinispan.commands.AbstractVisitor.visitRemoveCommand(AbstractVisitor.java:49)
> at org.infinispan.commands.write.RemoveCommand.acceptVisitor(RemoveCommand.java:58)
> at org.infinispan.interceptors.InterceptorChain.invoke(InterceptorChain.java:336)
> at org.infinispan.cache.impl.CacheImpl.executeCommandAndCommitIfNeeded(CacheImpl.java:1617)
> at org.infinispan.cache.impl.CacheImpl.removeInternal(CacheImpl.java:579)
> at org.infinispan.cache.impl.CacheImpl.remove(CacheImpl.java:572)
> at org.infinispan.cache.impl.DecoratedCache.remove(DecoratedCache.java:442)
> at org.infinispan.server.hotrod.CacheDecodeContext.remove(CacheDecodeContext.scala:235)
> at org.infinispan.server.hotrod.HotRodDecoder.handleModification(HotRodDecoder.scala:216)
> at org.infinispan.server.hotrod.HotRodDecoder.org$infinispan$server$hotrod$HotRodDecoder$$decodeKey(HotRodDecoder.scala:102)
> at org.infinispan.server.hotrod.HotRodDecoder$$anonfun$decode$1.apply$mcV$sp(HotRodDecoder.scala:48)
> at org.infinispan.server.hotrod.HotRodDecoder$$anon$1.run(HotRodDecoder.scala:206)
> at org.infinispan.server.hotrod.HotRodDecoder$$anon$1.run(HotRodDecoder.scala:205)
> at org.infinispan.security.Security.doAs(Security.java:143)
> at org.infinispan.server.hotrod.HotRodDecoder.wrapSecurity(HotRodDecoder.scala:205)
> at org.infinispan.server.hotrod.HotRodDecoder.decode(HotRodDecoder.scala:45)
> at io.netty.handler.codec.ReplayingDecoder.callDecode(ReplayingDecoder.java:370)
> at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:168)
> at org.infinispan.server.hotrod.HotRodDecoder.org$infinispan$server$core$transport$StatsChannelHandler$$super$channelRead(HotRodDecoder.scala:31)
> at org.infinispan.server.core.transport.StatsChannelHandler$class.channelRead(StatsChannelHandler.scala:32)
> at org.infinispan.server.hotrod.HotRodDecoder.channelRead(HotRodDecoder.scala:31)
> at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308)
> at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294)
> at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:846)
> at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:130)
> at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:511)
> at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:468)
> at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:382)
> at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:354)
> at io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:116)
> at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:137)
> at java.lang.Thread.run(Thread.java:745)
> Caused by: java.lang.UnsupportedOperationException
> at java.util.AbstractCollection.add(AbstractCollection.java:262)
> at java.util.AbstractCollection.addAll(AbstractCollection.java:344)
> at org.infinispan.notifications.cachelistener.cluster.impl.BatchingClusterEventManagerImpl$UnicastEventContext.addTargets(BatchingClusterEventManagerImpl.java:82)
> at org.infinispan.notifications.cachelistener.cluster.impl.BatchingClusterEventManagerImpl.addEvents(BatchingClusterEventManagerImpl.java:43)
> at org.infinispan.notifications.cachelistener.cluster.RemoteClusterListener.handleClusterEvents(RemoteClusterListener.java:108)
> at sun.reflect.GeneratedMethodAccessor24.invoke(Unknown Source)
> at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:497)
> at org.infinispan.notifications.impl.AbstractListenerImpl$ListenerInvocationImpl$1.run(AbstractListenerImpl.java:286)
> ... 76 more
> {code}
--
This message was sent by Atlassian JIRA
(v6.3.15#6346)
9 years, 7 months