[JBoss JIRA] (ISPN-9116) Server marshallers/transcoders don't support whitelist when deserializing
by Diego Lovison (Jira)
[ https://issues.jboss.org/browse/ISPN-9116?page=com.atlassian.jira.plugin.... ]
Diego Lovison closed ISPN-9116.
-------------------------------
> Server marshallers/transcoders don't support whitelist when deserializing
> -------------------------------------------------------------------------
>
> Key: ISPN-9116
> URL: https://issues.jboss.org/browse/ISPN-9116
> Project: Infinispan
> Issue Type: Bug
> Components: Server
> Affects Versions: 9.3.0.Final, 9.2.5.Final
> Reporter: Gustavo Fernandes
> Assignee: Gustavo Fernandes
> Priority: Major
> Fix For: 9.4.0.Alpha1, 9.4.0.CR3, 9.3.1.Final
>
>
> The server deserializes binary payloads and json/xml payload without any checks. This happens when:
> * Compatibility mode is on
> * Remote listeners with filters
> * Remote iteration with filters
> * Remote tasks with parameters
> * Server is configured with MediaType.APPLICATION_OBJECT
> * Potentially with JSON and XML contents sent via REST
> The remote endpoints affected are REST, Hot Rod and Memcached.
--
This message was sent by Atlassian Jira
(v7.12.1#712002)
7 years, 6 months
[JBoss JIRA] (ISPN-3791) Silence "Received invalid rebalance confirmation from NodeX" exceptions
by Dan Berindei (Jira)
[ https://issues.jboss.org/browse/ISPN-3791?page=com.atlassian.jira.plugin.... ]
Dan Berindei reassigned ISPN-3791:
----------------------------------
Assignee: Dan Berindei
> Silence "Received invalid rebalance confirmation from NodeX" exceptions
> -----------------------------------------------------------------------
>
> Key: ISPN-3791
> URL: https://issues.jboss.org/browse/ISPN-3791
> Project: Infinispan
> Issue Type: Bug
> Components: State Transfer
> Affects Versions: 6.0.0.Final
> Reporter: Dan Berindei
> Assignee: Dan Berindei
> Priority: Minor
>
> When the coordinator shuts down, it tries to shut down each of its caches first. This triggers a rebalance for the rest of the members, but the rebalance usually finishes only after the coordinator's channel also shuts down.
> The nodes who finish their state transfer will then send a REBALANCE_CONFIRM command to the new coordinator, but the new coordinator doesn't know about that rebalance (it will start the rebalance process from scratch). This results in exceptions like this in the new coordinator's log:
> {noformat}
> 12:36:04,977 WARN [org.infinispan.topology.CacheTopologyControlCommand] (remote-thread-2,ISPN-Node-1) ISPN000071: Caught exception when handling command CacheTopologyControlCommand{cache=MyCoolCache, type=REBALANCE_CONFIRM, sender=ISPN-Node-3-54019, joinInfo=null, topologyId=8, currentCH=null, pendingCH=null, throwable=null, viewId=4}: org.infinispan.commons.CacheException: Received invalid rebalance confirmation from ISPN-Node-3-54019 for cache MyCoolCache, we don't have a rebalance in progress
> at org.infinispan.topology.ClusterTopologyManagerImpl.handleRebalanceCompleted(ClusterTopologyManagerImpl.java:190) [infinispan-core-6.0.0.Final.jar:6.0.0.Final]
> at org.infinispan.topology.CacheTopologyControlCommand.doPerform(CacheTopologyControlCommand.java:147) [infinispan-core-6.0.0.Final.jar:6.0.0.Final]
> at org.infinispan.topology.CacheTopologyControlCommand.perform(CacheTopologyControlCommand.java:124) [infinispan-core-6.0.0.Final.jar:6.0.0.Final]
> at org.infinispan.remoting.transport.jgroups.CommandAwareRpcDispatcher$4.run(CommandAwareRpcDispatcher.java:270) [infinispan-core-6.0.0.Final.jar:6.0.0.Final]
> at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) [rt.jar:1.7.0_45]
> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) [rt.jar:1.7.0_45]
> at java.lang.Thread.run(Thread.java:744) [rt.jar:1.7.0_45]
> {noformat}
> A simple way to avoid these warnings would be to keep track of the coordinator that initiated a particular rebalance on each node, and only send the confirmation message to that coordinator. The same warnings seem to appear on the old coordinator, when it receives a confirmation after its ClusterTopologyManager started shutting down, so we may need another check there.
> A more ambitious approach would be to keep the old rebalance when the new coordinator takes over, and have another round in the cluster state recovery asking if any members have already sent REBALANCE_CONFIRMATION commands (after the new coordinator is ready to process those commands). This should eliminate the duplicate state transfer that happens now.
--
This message was sent by Atlassian Jira
(v7.12.1#712002)
7 years, 6 months
[JBoss JIRA] (ISPN-8813) org.infinispan.server.test.query.RemoteQueryStringIT.testFullTextTermRightOperandAnalyzed fails randomly
by Adrian Nistor (Jira)
[ https://issues.jboss.org/browse/ISPN-8813?page=com.atlassian.jira.plugin.... ]
Adrian Nistor updated ISPN-8813:
--------------------------------
Status: Pull Request Sent (was: Open)
Git Pull Request: https://github.com/infinispan/infinispan/pull/6331 (was: https://github.com/infinispan/infinispan/pull/5753)
> org.infinispan.server.test.query.RemoteQueryStringIT.testFullTextTermRightOperandAnalyzed fails randomly
> --------------------------------------------------------------------------------------------------------
>
> Key: ISPN-8813
> URL: https://issues.jboss.org/browse/ISPN-8813
> Project: Infinispan
> Issue Type: Bug
> Components: Remote Querying
> Reporter: Adrian Nistor
> Assignee: Adrian Nistor
> Priority: Major
> Fix For: 9.4.1.Final
>
>
> To be fair, this failure happens almost all the time. But I've seen some rare builds where it did not fail, so I'm marking it as a random failure.
> {code}
> [0m[33m10:10:13,714 WARN [org.infinispan.factories.ComponentRegistry] (MSC service thread 1-2) ISPN000189: While stopping a cache or cache manager, one of its components failed to stop: org.infinispan.commons.CacheException: Unable to invoke method public void org.infinispan.notifications.cachelistener.CacheNotifierImpl.stop() on object of type CacheNotifierImpl
> at org.infinispan.commons.util.ReflectionUtil.invokeAccessibly(ReflectionUtil.java:172)
> at org.infinispan.factories.AbstractComponentRegistry$PrioritizedMethod.invoke(AbstractComponentRegistry.java:878)
> at org.infinispan.factories.AbstractComponentRegistry.internalStop(AbstractComponentRegistry.java:679)
> at org.infinispan.factories.AbstractComponentRegistry.stop(AbstractComponentRegistry.java:581)
> at org.infinispan.factories.ComponentRegistry.stop(ComponentRegistry.java:256)
> at org.infinispan.cache.impl.CacheImpl.performImmediateShutdown(CacheImpl.java:937)
> at org.infinispan.cache.impl.CacheImpl.stop(CacheImpl.java:901)
> at org.infinispan.cache.impl.AbstractDelegatingCache.stop(AbstractDelegatingCache.java:420)
> at org.infinispan.server.infinispan.SecurityActions$6.run(SecurityActions.java:148)
> at org.infinispan.server.infinispan.SecurityActions$6.run(SecurityActions.java:145)
> at org.infinispan.security.Security.doPrivileged(Security.java:76)
> at org.infinispan.server.infinispan.SecurityActions.doPrivileged(SecurityActions.java:69)
> at org.infinispan.server.infinispan.SecurityActions.stopCache(SecurityActions.java:152)
> at org.jboss.as.clustering.infinispan.subsystem.CacheService.stop(CacheService.java:103)
> at org.jboss.msc.service.ServiceControllerImpl$StopTask.stopService(ServiceControllerImpl.java:2150)
> at org.jboss.msc.service.ServiceControllerImpl$StopTask.run(ServiceControllerImpl.java:2101)
> at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> at java.lang.Thread.run(Thread.java:745)
> Caused by: java.lang.NullPointerException
> at org.infinispan.query.dsl.embedded.impl.BaseJPAFilterIndexingServiceProvider.stop(BaseJPAFilterIndexingServiceProvider.java:68)
> at org.infinispan.notifications.cachelistener.CacheNotifierImpl.stop(CacheNotifierImpl.java:275)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:497)
> at org.infinispan.commons.util.ReflectionUtil.invokeAccessibly(ReflectionUtil.java:168)
> ... 18 more
> [0m[0m10:10:13,728 INFO [org.jboss.as.clustering.infinispan] (MSC service thread 1-2) DGISPN0002: Stopped default cache from local container
> [0m[33m10:10:13,748 WARN [org.infinispan.factories.ComponentRegistry] (MSC service thread 1-2) ISPN000189: While stopping a cache or cache manager, one of its components failed to stop: org.infinispan.commons.CacheException: Unable to invoke method public void org.infinispan.notifications.cachelistener.CacheNotifierImpl.stop() on object of type CacheNotifierImpl
> at org.infinispan.commons.util.ReflectionUtil.invokeAccessibly(ReflectionUtil.java:172)
> at org.infinispan.factories.AbstractComponentRegistry$PrioritizedMethod.invoke(AbstractComponentRegistry.java:878)
> at org.infinispan.factories.AbstractComponentRegistry.internalStop(AbstractComponentRegistry.java:679)
> at org.infinispan.factories.AbstractComponentRegistry.stop(AbstractComponentRegistry.java:581)
> at org.infinispan.factories.ComponentRegistry.stop(ComponentRegistry.java:256)
> at org.infinispan.cache.impl.CacheImpl.performImmediateShutdown(CacheImpl.java:937)
> at org.infinispan.cache.impl.CacheImpl.stop(CacheImpl.java:901)
> at org.infinispan.cache.impl.AbstractDelegatingCache.stop(AbstractDelegatingCache.java:420)
> at org.infinispan.manager.DefaultCacheManager.terminate(DefaultCacheManager.java:679)
> at org.infinispan.manager.DefaultCacheManager.stopCaches(DefaultCacheManager.java:719)
> at org.infinispan.manager.DefaultCacheManager.stop(DefaultCacheManager.java:696)
> at org.infinispan.manager.impl.AbstractDelegatingEmbeddedCacheManager.stop(AbstractDelegatingEmbeddedCacheManager.java:190)
> at org.infinispan.server.infinispan.SecurityActions$2.run(SecurityActions.java:98)
> at org.infinispan.server.infinispan.SecurityActions$2.run(SecurityActions.java:94)
> at org.infinispan.security.Security.doPrivileged(Security.java:76)
> at org.infinispan.server.infinispan.SecurityActions.doPrivileged(SecurityActions.java:69)
> at org.infinispan.server.infinispan.SecurityActions.stopAndUnregisterContainer(SecurityActions.java:106)
> at org.jboss.as.clustering.infinispan.subsystem.CacheContainerBuilder.stop(CacheContainerBuilder.java:106)
> at org.jboss.msc.service.ServiceControllerImpl$StopTask.stopService(ServiceControllerImpl.java:2150)
> at org.jboss.msc.service.ServiceControllerImpl$StopTask.run(ServiceControllerImpl.java:2101)
> at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> at java.lang.Thread.run(Thread.java:745)
> Caused by: java.lang.NullPointerException
> at org.infinispan.query.dsl.embedded.impl.BaseJPAFilterIndexingServiceProvider.stop(BaseJPAFilterIndexingServiceProvider.java:68)
> at org.infinispan.notifications.cachelistener.CacheNotifierImpl.stop(CacheNotifierImpl.java:275)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:497)
> at org.infinispan.commons.util.ReflectionUtil.invokeAccessibly(ReflectionUtil.java:168)
> ... 22 more
> {code}
--
This message was sent by Atlassian Jira
(v7.12.1#712002)
7 years, 6 months
[JBoss JIRA] (ISPN-3791) Silence "Received invalid rebalance confirmation from NodeX" exceptions
by Dan Berindei (Jira)
[ https://issues.jboss.org/browse/ISPN-3791?page=com.atlassian.jira.plugin.... ]
Dan Berindei commented on ISPN-3791:
------------------------------------
The 4-phase rebalance introduced with ISPN-5021 made this exception much more common, and also harmless: if the coordinator sends a topology update, all the nodes that already confirmed the rebalance will resend the confirmation with the new topology id.
> Silence "Received invalid rebalance confirmation from NodeX" exceptions
> -----------------------------------------------------------------------
>
> Key: ISPN-3791
> URL: https://issues.jboss.org/browse/ISPN-3791
> Project: Infinispan
> Issue Type: Bug
> Components: State Transfer
> Affects Versions: 6.0.0.Final
> Reporter: Dan Berindei
> Priority: Minor
>
> When the coordinator shuts down, it tries to shut down each of its caches first. This triggers a rebalance for the rest of the members, but the rebalance usually finishes only after the coordinator's channel also shuts down.
> The nodes who finish their state transfer will then send a REBALANCE_CONFIRM command to the new coordinator, but the new coordinator doesn't know about that rebalance (it will start the rebalance process from scratch). This results in exceptions like this in the new coordinator's log:
> {noformat}
> 12:36:04,977 WARN [org.infinispan.topology.CacheTopologyControlCommand] (remote-thread-2,ISPN-Node-1) ISPN000071: Caught exception when handling command CacheTopologyControlCommand{cache=MyCoolCache, type=REBALANCE_CONFIRM, sender=ISPN-Node-3-54019, joinInfo=null, topologyId=8, currentCH=null, pendingCH=null, throwable=null, viewId=4}: org.infinispan.commons.CacheException: Received invalid rebalance confirmation from ISPN-Node-3-54019 for cache MyCoolCache, we don't have a rebalance in progress
> at org.infinispan.topology.ClusterTopologyManagerImpl.handleRebalanceCompleted(ClusterTopologyManagerImpl.java:190) [infinispan-core-6.0.0.Final.jar:6.0.0.Final]
> at org.infinispan.topology.CacheTopologyControlCommand.doPerform(CacheTopologyControlCommand.java:147) [infinispan-core-6.0.0.Final.jar:6.0.0.Final]
> at org.infinispan.topology.CacheTopologyControlCommand.perform(CacheTopologyControlCommand.java:124) [infinispan-core-6.0.0.Final.jar:6.0.0.Final]
> at org.infinispan.remoting.transport.jgroups.CommandAwareRpcDispatcher$4.run(CommandAwareRpcDispatcher.java:270) [infinispan-core-6.0.0.Final.jar:6.0.0.Final]
> at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) [rt.jar:1.7.0_45]
> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) [rt.jar:1.7.0_45]
> at java.lang.Thread.run(Thread.java:744) [rt.jar:1.7.0_45]
> {noformat}
> A simple way to avoid these warnings would be to keep track of the coordinator that initiated a particular rebalance on each node, and only send the confirmation message to that coordinator. The same warnings seem to appear on the old coordinator, when it receives a confirmation after its ClusterTopologyManager started shutting down, so we may need another check there.
> A more ambitious approach would be to keep the old rebalance when the new coordinator takes over, and have another round in the cluster state recovery asking if any members have already sent REBALANCE_CONFIRMATION commands (after the new coordinator is ready to process those commands). This should eliminate the duplicate state transfer that happens now.
--
This message was sent by Atlassian Jira
(v7.12.1#712002)
7 years, 6 months