[JBoss JIRA] (ISPN-7470) Distributed executor does not fail over unless future.get() is called
by Galder Zamarreño (JIRA)
[ https://issues.jboss.org/browse/ISPN-7470?page=com.atlassian.jira.plugin.... ]
Galder Zamarreño updated ISPN-7470:
-----------------------------------
Fix Version/s: 9.4.0.Final
(was: 9.3.0.Final)
> Distributed executor does not fail over unless future.get() is called
> ---------------------------------------------------------------------
>
> Key: ISPN-7470
> URL: https://issues.jboss.org/browse/ISPN-7470
> Project: Infinispan
> Issue Type: Bug
> Components: Core
> Affects Versions: 8.2.6.Final, 9.0.0.CR1
> Reporter: Dan Berindei
> Assignee: William Burns
> Fix For: 9.4.0.Final
>
>
> After ISPN-6392, {{DistributedExecutorService.submit(...)}} nominally returns a {{CompletableFuture}}. However, it doesn't behave like a regular {{CompletableFuture}}, because it doesn't run the failure policy until the user calls {{future.get()}}.
> {{future.isComplete()}} will return {{true}} before running the failure policy, and {{future.whenComplete(callback)}} will also execute the callback before running the failure policy.
--
This message was sent by Atlassian JIRA
(v7.5.0#75005)
6 years, 6 months
[JBoss JIRA] (ISPN-7711) Improve server-side protostream marshaller registration
by Galder Zamarreño (JIRA)
[ https://issues.jboss.org/browse/ISPN-7711?page=com.atlassian.jira.plugin.... ]
Galder Zamarreño updated ISPN-7711:
-----------------------------------
Fix Version/s: 9.4.0.Final
(was: 9.3.0.Final)
> Improve server-side protostream marshaller registration
> -------------------------------------------------------
>
> Key: ISPN-7711
> URL: https://issues.jboss.org/browse/ISPN-7711
> Project: Infinispan
> Issue Type: Enhancement
> Reporter: Galder Zamarreño
> Assignee: Adrian Nistor
> Fix For: 9.4.0.Final
>
>
> A more user friendly way to register protostream marshallers remotely is needed when the protostream marshaller is used for compatibility mode (or transcoding) in the future. It's currently possible to do it with a remote server task but that's a bit hacky.
> This server task shows how it can be done:
> {code}
> package delays.java.stream;
> import org.infinispan.manager.EmbeddedCacheManager;
> import org.infinispan.query.remote.ProtobufMetadataManager;
> import org.infinispan.tasks.ServerTask;
> import org.infinispan.tasks.TaskContext;
> import org.infinispan.tasks.TaskExecutionMode;
> import delays.java.stream.pojos.Station;
> import delays.java.stream.pojos.Stop;
> import delays.java.stream.pojos.Train;
> public class AddProtobufTask implements ServerTask {
> private TaskContext ctx;
> @Override
> public void setTaskContext(TaskContext ctx) {
> this.ctx = ctx;
> }
> @Override
> public String getName() {
> return "add-protobuf";
> }
> @Override
> public Object call() throws Exception {
> EmbeddedCacheManager cm = ctx.getCache().get().getCacheManager();
> ProtobufMetadataManager protobufMetadataManager = cm.getGlobalComponentRegistry().getComponent(ProtobufMetadataManager.class);
> protobufMetadataManager.registerMarshaller(new Stop.Marshaller());
> protobufMetadataManager.registerMarshaller(new Station.Marshaller());
> protobufMetadataManager.registerMarshaller(new Train.Marshaller());
> return null;
> }
> @Override
> public TaskExecutionMode getExecutionMode() {
> // Registering marshallers should be done in all nodes
> return TaskExecutionMode.ALL_NODES;
> }
> }
> {code}
--
This message was sent by Atlassian JIRA
(v7.5.0#75005)
6 years, 6 months
[JBoss JIRA] (ISPN-7678) Server error when updating default cache configuration
by Galder Zamarreño (JIRA)
[ https://issues.jboss.org/browse/ISPN-7678?page=com.atlassian.jira.plugin.... ]
Galder Zamarreño updated ISPN-7678:
-----------------------------------
Fix Version/s: 9.4.0.Final
(was: 9.3.0.Final)
> Server error when updating default cache configuration
> ------------------------------------------------------
>
> Key: ISPN-7678
> URL: https://issues.jboss.org/browse/ISPN-7678
> Project: Infinispan
> Issue Type: Bug
> Components: Configuration, Server
> Affects Versions: 9.0.0.CR4
> Environment: Run the server with the default configuration in domain mode.
> Reporter: Ryan Emerson
> Assignee: Tristan Tarrant
> Fix For: 9.4.0.Final
>
>
> {code:java}
> [Server:server-one] 13:34:26,304 INFO [org.jboss.as.clustering.infinispan] (MSC service thread 1-3) DGISPN0002: Stopped default cache from clustered container
> [Server:server-two] 13:34:26,311 INFO [org.infinispan.CLUSTER] (remote-thread--p2-t2) ISPN000310: Starting cluster-wide rebalance for cache default, topology CacheTopology{id=5, rebalanceId=3, currentCH=DefaultConsistentHash{ns=20, owners = (1)[master:server-two: 20+0]}, pendingCH=DefaultConsistentHash{ns=20, owners = (2)[master:server-two: 10+10, master:server-one: 10+10]}, unionCH=null, actualMembers=[master:server-two, master:server-one], persistentUUIDs=[ca1b2fd7-2f90-4439-88e6-ca202f1bded4, 069b35d1-2d9d-4ca4-bf14-c01fa80bf86a]}
> [Server:server-two] 13:34:26,311 INFO [org.infinispan.CLUSTER] (remote-thread--p2-t2) [Context=default][Context=master:server-two]ISPN100002: Started local rebalance
> [Server:server-two] 13:34:26,312 INFO [org.infinispan.CLUSTER] (transport-thread--p4-t9) [Context=default][Context=master:server-two]ISPN100003: Finished local rebalance
> [Server:server-two] 13:34:26,318 INFO [org.infinispan.CLUSTER] (remote-thread--p2-t6) [Context=default][Context=master:server-one]ISPN100003: Finished local rebalance
> [Server:server-two] 13:34:26,320 INFO [org.infinispan.CLUSTER] (remote-thread--p2-t6) ISPN000336: Finished cluster-wide rebalance for cache default, topology id = 5
> [Server:server-one] 13:34:26,321 INFO [org.jboss.as.clustering.infinispan] (MSC service thread 1-2) DGISPN0001: Started default cache from clustered container
> [Server:server-one] 13:34:26,321 INFO [org.infinispan.server.endpoint] (MSC service thread 1-2) DGENDPT10000: MemcachedServer starting
> [Server:server-one] 13:34:26,321 INFO [org.infinispan.server.endpoint] (MSC service thread 1-2) DGENDPT10001: MemcachedServer listening on 127.0.0.1:11211
> [Server:server-one] 13:34:26,321 INFO [org.infinispan.server.endpoint] (MSC service thread 1-3) DGENDPT10000: REST starting
> [Server:server-one] 13:34:26,321 INFO [org.infinispan.server.endpoint] (MSC service thread 1-4) DGENDPT10000: WebSocketServer starting
> [Server:server-one] 13:34:26,322 INFO [org.infinispan.server.endpoint] (MSC service thread 1-4) DGENDPT10001: WebSocketServer listening on 127.0.0.1:8181
> [Server:server-one] 13:34:26,321 INFO [org.infinispan.server.endpoint] (MSC service thread 1-5) DGENDPT10000: HotRodServer starting
> [Server:server-one] 13:34:26,322 INFO [org.infinispan.server.endpoint] (MSC service thread 1-5) DGENDPT10001: HotRodServer listening on 127.0.0.1:11222
> [Server:server-two] 13:34:26,341 INFO [org.jboss.as.clustering.infinispan] (MSC service thread 1-8) DGISPN0002: Stopped default cache from clustered container
> [Server:server-one] 13:34:26,349 INFO [org.infinispan.rest.embedded.netty4.NettyRestServer] (MSC service thread 1-3) ISPN012003: REST server starting, listening on 127.0.0.1:8080
> [Server:server-one] 13:34:26,350 INFO [org.infinispan.server.endpoint] (MSC service thread 1-3) DGENDPT10002: REST mapped to rest/rest
> [Server:server-two] 13:34:26,351 INFO [org.infinispan.CLUSTER] (MSC service thread 1-8) ISPN000310: Starting cluster-wide rebalance for cache default, topology CacheTopology{id=8, rebalanceId=4, currentCH=DefaultConsistentHash{ns=20, owners = (1)[master:server-one: 20+0]}, pendingCH=DefaultConsistentHash{ns=20, owners = (2)[master:server-one: 10+10, master:server-two: 10+10]}, unionCH=null, actualMembers=[master:server-one, master:server-two], persistentUUIDs=[069b35d1-2d9d-4ca4-bf14-c01fa80bf86a, ca1b2fd7-2f90-4439-88e6-ca202f1bded4]}
> [Server:server-two] 13:34:26,351 INFO [org.infinispan.CLUSTER] (MSC service thread 1-8) [Context=default][Context=master:server-two]ISPN100002: Started local rebalance
> [Server:server-two] 13:34:26,353 INFO [org.infinispan.CLUSTER] (remote-thread--p2-t7) [Context=default][Context=master:server-one]ISPN100003: Finished local rebalance
> [Server:server-two] 13:34:26,366 INFO [org.infinispan.CLUSTER] (transport-thread--p4-t19) [Context=default][Context=master:server-two]ISPN100003: Finished local rebalance
> [Server:server-two] 13:34:26,367 INFO [org.infinispan.CLUSTER] (transport-thread--p4-t19) ISPN000336: Finished cluster-wide rebalance for cache default, topology id = 8
> [Server:server-two] 13:34:26,368 INFO [org.jboss.as.clustering.infinispan] (MSC service thread 1-8) DGISPN0001: Started default cache from clustered container
> [Server:server-two] 13:34:26,368 INFO [org.infinispan.server.endpoint] (MSC service thread 1-8) DGENDPT10000: WebSocketServer starting
> [Server:server-two] 13:34:26,368 INFO [org.infinispan.server.endpoint] (MSC service thread 1-6) DGENDPT10000: HotRodServer starting
> [Server:server-two] 13:34:26,368 INFO [org.infinispan.server.endpoint] (MSC service thread 1-1) DGENDPT10000: MemcachedServer starting
> [Server:server-two] 13:34:26,369 INFO [org.infinispan.server.endpoint] (MSC service thread 1-8) DGENDPT10001: WebSocketServer listening on 127.0.0.1:8331
> [Server:server-two] 13:34:26,369 INFO [org.infinispan.server.endpoint] (MSC service thread 1-6) DGENDPT10001: HotRodServer listening on 127.0.0.1:11372
> [Server:server-two] 13:34:26,369 INFO [org.infinispan.server.endpoint] (MSC service thread 1-1) DGENDPT10001: MemcachedServer listening on 127.0.0.1:11361
> [Server:server-two] 13:34:26,369 INFO [org.infinispan.server.endpoint] (MSC service thread 1-5) DGENDPT10000: REST starting
> [Server:server-two] 13:34:26,407 INFO [org.infinispan.rest.embedded.netty4.NettyRestServer] (MSC service thread 1-5) ISPN012003: REST server starting, listening on 127.0.0.1:8230
> [Server:server-two] 13:34:26,408 INFO [org.infinispan.server.endpoint] (MSC service thread 1-5) DGENDPT10002: REST mapped to rest/rest
> [Server:server-one] 13:34:27,437 ERROR [org.jboss.msc.service.fail] (MSC service thread 1-5) MSC000001: Failed to start service jboss.datagrid-infinispan-endpoint.hotrod.hotrod-connector: org.jboss.msc.service.StartException in service jboss.datagrid-infinispan-endpoint.hotrod.hotrod-connector: DGENDPT10004: Failed to start HotRodServer
> [Server:server-one] at org.infinispan.server.endpoint.subsystem.ProtocolServerService.start(ProtocolServerService.java:156)
> [Server:server-one] at org.jboss.msc.service.ServiceControllerImpl$StartTask.startService(ServiceControllerImpl.java:1948)
> [Server:server-one] at org.jboss.msc.service.ServiceControllerImpl$StartTask.run(ServiceControllerImpl.java:1881)
> [Server:server-one] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> [Server:server-one] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> [Server:server-one] at java.lang.Thread.run(Thread.java:745)
> [Server:server-one] Caused by: org.infinispan.commons.CacheConfigurationException: ISPN000358: A cache configuration named ___hotRodTopologyCache already exists. This cannot be configured externally by the user.
> [Server:server-one] at org.infinispan.registry.impl.InternalCacheRegistryImpl.registerInternalCache(InternalCacheRegistryImpl.java:58)
> [Server:server-one] at org.infinispan.server.hotrod.HotRodServer.defineTopologyCacheConfig(HotRodServer.java:265)
> [Server:server-one] at org.infinispan.server.hotrod.HotRodServer.startInternal(HotRodServer.java:163)
> [Server:server-one] at org.infinispan.server.hotrod.HotRodServer.startInternal(HotRodServer.java:82)
> [Server:server-one] at org.infinispan.server.core.AbstractProtocolServer.start(AbstractProtocolServer.java:80)
> [Server:server-one] at org.infinispan.server.endpoint.subsystem.SecurityActions$6.run(SecurityActions.java:136)
> [Server:server-one] at org.infinispan.server.endpoint.subsystem.SecurityActions$6.run(SecurityActions.java:133)
> [Server:server-one] at org.infinispan.security.Security.doPrivileged(Security.java:76)
> [Server:server-one] at org.infinispan.server.endpoint.subsystem.SecurityActions.doPrivileged(SecurityActions.java:42)
> [Server:server-one] at org.infinispan.server.endpoint.subsystem.SecurityActions.startProtocolServer(SecurityActions.java:140)
> [Server:server-one] at org.infinispan.server.endpoint.subsystem.ProtocolServerService.startProtocolServer(ProtocolServerService.java:190)
> [Server:server-one] at org.infinispan.server.endpoint.subsystem.ProtocolServerService.start(ProtocolServerService.java:148)
> [Server:server-one] ... 5 more
> [Server:server-one]
> [Server:server-two] 13:34:27,493 ERROR [org.jboss.msc.service.fail] (MSC service thread 1-6) MSC000001: Failed to start service jboss.datagrid-infinispan-endpoint.hotrod.hotrod-connector: org.jboss.msc.service.StartException in service jboss.datagrid-infinispan-endpoint.hotrod.hotrod-connector: DGENDPT10004: Failed to start HotRodServer
> [Server:server-two] at org.infinispan.server.endpoint.subsystem.ProtocolServerService.start(ProtocolServerService.java:156)
> [Server:server-two] at org.jboss.msc.service.ServiceControllerImpl$StartTask.startService(ServiceControllerImpl.java:1948)
> [Server:server-two] at org.jboss.msc.service.ServiceControllerImpl$StartTask.run(ServiceControllerImpl.java:1881)
> [Server:server-two] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> [Server:server-two] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> [Server:server-two] at java.lang.Thread.run(Thread.java:745)
> [Server:server-two] Caused by: org.infinispan.commons.CacheConfigurationException: ISPN000358: A cache configuration named ___hotRodTopologyCache already exists. This cannot be configured externally by the user.
> [Server:server-two] at org.infinispan.registry.impl.InternalCacheRegistryImpl.registerInternalCache(InternalCacheRegistryImpl.java:58)
> [Server:server-two] at org.infinispan.server.hotrod.HotRodServer.defineTopologyCacheConfig(HotRodServer.java:265)
> [Server:server-two] at org.infinispan.server.hotrod.HotRodServer.startInternal(HotRodServer.java:163)
> [Server:server-two] at org.infinispan.server.hotrod.HotRodServer.startInternal(HotRodServer.java:82)
> [Server:server-two] at org.infinispan.server.core.AbstractProtocolServer.start(AbstractProtocolServer.java:80)
> [Server:server-two] at org.infinispan.server.endpoint.subsystem.SecurityActions$6.run(SecurityActions.java:136)
> [Server:server-two] at org.infinispan.server.endpoint.subsystem.SecurityActions$6.run(SecurityActions.java:133)
> [Server:server-two] at org.infinispan.security.Security.doPrivileged(Security.java:76)
> [Server:server-two] at org.infinispan.server.endpoint.subsystem.SecurityActions.doPrivileged(SecurityActions.java:42)
> [Server:server-two] at org.infinispan.server.endpoint.subsystem.SecurityActions.startProtocolServer(SecurityActions.java:140)
> [Server:server-two] at org.infinispan.server.endpoint.subsystem.ProtocolServerService.startProtocolServer(ProtocolServerService.java:190)
> [Server:server-two] at org.infinispan.server.endpoint.subsystem.ProtocolServerService.start(ProtocolServerService.java:148)
> [Server:server-two] ... 5 more
> [Server:server-two]
> [Server:server-two] 13:34:27,528 INFO [org.jboss.as.controller] (ServerService Thread Pool -- 15) WFLYCTL0183: Service status report
> [Server:server-two] WFLYCTL0186: Services which failed to start: service jboss.datagrid-infinispan-endpoint.hotrod.hotrod-connector: org.jboss.msc.service.StartException in service jboss.datagrid-infinispan-endpoint.hotrod.hotrod-connector: DGENDPT10004: Failed to start HotRodServer
> [Server:server-two]
> [Server:server-one] 13:34:27,528 INFO [org.jboss.as.controller] (ServerService Thread Pool -- 30) WFLYCTL0183: Service status report
> [Server:server-one] WFLYCTL0186: Services which failed to start: service jboss.datagrid-infinispan-endpoint.hotrod.hotrod-connector: org.jboss.msc.service.StartException in service jboss.datagrid-infinispan-endpoint.hotrod.hotrod-connector: DGENDPT10004: Failed to start HotRodServer
> [Server:server-one]
> {code}
--
This message was sent by Atlassian JIRA
(v7.5.0#75005)
6 years, 6 months
[JBoss JIRA] (ISPN-7672) NonTotalOrderTxPerCacheInboundInvocationHandler throws warning when adding cache entry using Spring Session
by Galder Zamarreño (JIRA)
[ https://issues.jboss.org/browse/ISPN-7672?page=com.atlassian.jira.plugin.... ]
Galder Zamarreño updated ISPN-7672:
-----------------------------------
Fix Version/s: 9.4.0.Final
(was: 9.3.0.Final)
> NonTotalOrderTxPerCacheInboundInvocationHandler throws warning when adding cache entry using Spring Session
> -----------------------------------------------------------------------------------------------------------
>
> Key: ISPN-7672
> URL: https://issues.jboss.org/browse/ISPN-7672
> Project: Infinispan
> Issue Type: Bug
> Components: Cloud, Spring Integration
> Affects Versions: 9.0.0.CR4
> Reporter: Sebastian Łaskawiec
> Assignee: Galder Zamarreño
> Priority: Minor
> Fix For: 9.4.0.Final
>
>
> When I'm trying to add an entry using Spring Session integration with a [transactional cache|https://github.com/slaskawi/presentations/blob/master/2017_spring_s...], the server throws a warning:
> {code}
> [transactions-repository-1-2cbrv] 06:53:40,773 WARN [org.infinispan.remoting.inboundhandler.NonTotalOrderTxPerCacheInboundInvocationHandler] (remote-thread--p2-t18) ISPN000071: Caught exception when handling command DistributedExecuteCommand [cache=Cache 'sessions'@transactions-repository-1-2cbrv, keys=[], callable=ClusterEventCallable{identifier=b345211e-fbd7-4305-b3a6-6979301e0360, events=[ClusterEvent {type=CACHE_ENTRY_CREATED, cache=Cache 'sessions'@transactions-repository-1-2cbrv, key=[B@8c75820, value=[B@76856353, oldValue=null, transaction=RecoveryAwareGlobalTransaction{xid=< 131077, 29, 36, 0000000000-1-1-84170374-96-629488-44-62370001349, 0000000000-1-1-84170374-96-629488-44-62370001400000000 >, internalId=562954248388609} GlobalTx:transactions-repository-1-cwk6f:1, retryCommand=false, origin=transactions-repository-1-cwk6f}]}]: org.infinispan.commons.CacheListenerException: ISPN000280: Caught exception [java.lang.ClassCastException] while invoking method [public void org.infinispan.server.hotrod.ClientListenerRegistry$BaseClientEventSender.onCacheEvent(org.infinispan.notifications.cachelistener.event.CacheEntryEvent)] on listener instance: org.infinispan.server.hotrod.ClientListenerRegistry$StatelessClientEventSender@7b97a57
> [transactions-repository-1-2cbrv] at org.infinispan.notifications.impl.AbstractListenerImpl$ListenerInvocationImpl$1.run(AbstractListenerImpl.java:401)
> [transactions-repository-1-2cbrv] at org.infinispan.util.concurrent.WithinThreadExecutor.execute(WithinThreadExecutor.java:20)
> [transactions-repository-1-2cbrv] at org.infinispan.notifications.impl.AbstractListenerImpl$ListenerInvocationImpl.invoke(AbstractListenerImpl.java:419)
> [transactions-repository-1-2cbrv] at org.infinispan.notifications.cachelistener.CacheNotifierImpl$BaseCacheEntryListenerInvocation.doRealInvocation(CacheNotifierImpl.java:1512)
> [transactions-repository-1-2cbrv] at org.infinispan.notifications.cachelistener.CacheNotifierImpl$BaseCacheEntryListenerInvocation.doRealInvocation(CacheNotifierImpl.java:1508)
> [transactions-repository-1-2cbrv] at org.infinispan.notifications.cachelistener.CacheNotifierImpl$BaseCacheEntryListenerInvocation.invokeNoChecks(CacheNotifierImpl.java:1503)
> [transactions-repository-1-2cbrv] at org.infinispan.notifications.cachelistener.CacheNotifierImpl.notifyClusterListeners(CacheNotifierImpl.java:711)
> [transactions-repository-1-2cbrv] at org.infinispan.notifications.cachelistener.cluster.ClusterEventCallable.call(ClusterEventCallable.java:49)
> [transactions-repository-1-2cbrv] at org.infinispan.notifications.cachelistener.cluster.ClusterEventCallable.call(ClusterEventCallable.java:25)
> [transactions-repository-1-2cbrv] at org.infinispan.commands.read.DistributedExecuteCommand.invokeAsync(DistributedExecuteCommand.java:99)
> [transactions-repository-1-2cbrv] at org.infinispan.remoting.inboundhandler.BasePerCacheInboundInvocationHandler.invokeCommand(BasePerCacheInboundInvocationHandler.java:90)
> [transactions-repository-1-2cbrv] at org.infinispan.remoting.inboundhandler.BaseBlockingRunnable.invoke(BaseBlockingRunnable.java:90)
> [transactions-repository-1-2cbrv] at org.infinispan.remoting.inboundhandler.BaseBlockingRunnable.runAsync(BaseBlockingRunnable.java:68)
> [transactions-repository-1-2cbrv] at org.infinispan.remoting.inboundhandler.BaseBlockingRunnable.run(BaseBlockingRunnable.java:40)
> [transactions-repository-1-2cbrv] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> [transactions-repository-1-2cbrv] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> [transactions-repository-1-2cbrv] at java.lang.Thread.run(Thread.java:745)
> [transactions-repository-1-2cbrv] Caused by: java.lang.ClassCastException: org.infinispan.container.versioning.SimpleClusteredVersion cannot be cast to org.infinispan.container.versioning.NumericVersion
> [transactions-repository-1-2cbrv] at org.infinispan.server.hotrod.ClientListenerRegistry$BaseClientEventSender.onCacheEvent(ClientListenerRegistry.java:363)
> [transactions-repository-1-2cbrv] at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> [transactions-repository-1-2cbrv] at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> [transactions-repository-1-2cbrv] at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> [transactions-repository-1-2cbrv] at java.lang.reflect.Method.invoke(Method.java:498)
> [transactions-repository-1-2cbrv] at org.infinispan.notifications.impl.AbstractListenerImpl$ListenerInvocationImpl$1.run(AbstractListenerImpl.java:396)
> [transactions-repository-1-2cbrv] ... 16 more
> [transactions-repository-1-2cbrv]
> {code}
> This didn't happen in {{CR2}} release, so there must be something that changed since then. I also noticed that this sometimes leads to exceptions in the Hot Rod client.
--
This message was sent by Atlassian JIRA
(v7.5.0#75005)
6 years, 6 months
[JBoss JIRA] (ISPN-7811) Improve out-of-the-box server security in cloud
by Galder Zamarreño (JIRA)
[ https://issues.jboss.org/browse/ISPN-7811?page=com.atlassian.jira.plugin.... ]
Galder Zamarreño updated ISPN-7811:
-----------------------------------
Fix Version/s: 9.4.0.Final
(was: 9.3.0.Final)
> Improve out-of-the-box server security in cloud
> -----------------------------------------------
>
> Key: ISPN-7811
> URL: https://issues.jboss.org/browse/ISPN-7811
> Project: Infinispan
> Issue Type: Enhancement
> Components: Security, Server
> Affects Versions: 9.0.0.Final
> Reporter: Galder Zamarreño
> Assignee: Galder Zamarreño
> Fix For: 9.4.0.Final
>
>
> When running Infinispan 9.0.0.Final in a cloud env, the default security code enforcements are causing issues when trying to register a proto file.
> The "___protobuf_metadata" cache cannot be written remotely any more. Accessing this cache to add protofile descriptors to server. The default configuration throws this error:
> {code}
> [datagrid-1-akxoi]
> [datagrid-1-akxoi] 12:15:56,602 ERROR [org.infinispan.server.hotrod.CacheDecodeContext] (HotRod-ServerWorker-4-2) ISPN005003: Exception reported: org.infinispan.server.hotrod.RequestParsingException: Remote requests are allowed to protected caches only over loopback or if authorization is enabled. Do no send remote requests to cache '___protobuf_metadata'
> [datagrid-1-akxoi] at org.infinispan.server.hotrod.CacheDecodeContext.obtainCache(CacheDecodeContext.java:116)
> [datagrid-1-akxoi] at org.infinispan.server.hotrod.HotRodDecoder.decodeHeader(HotRodDecoder.java:162)
> [datagrid-1-akxoi] at org.infinispan.server.hotrod.HotRodDecoder.decode(HotRodDecoder.java:93)
> {code}
> The code in CacheDecodeContext that enables this check does the following:
> {code}
> if (!cacheManager.getCacheManagerConfiguration().security().authorization().enabled()...
> {code}
> In order to have better out-of-the-box experience in cloud but still be secured, the following should be done:
> * Remove the code check for authorization in CacheDecodeContext.
> * Server's default configuration should require authentication.
> * Docker image allows passing in APP_USER and APP_PASS as env variables easily, but it provides default usernames and passwords for both APP and MGMT. These defaults should be removed since they're a security risk.
> * Docker image should have the possibility to set APP_GROUPS so that we can pass in optionally the role groups associated with a user. This is handy for making it easier in the future for users to add authorization on top of authentication.
> I will create JIRA subtasks for these so that the work can be divided.
--
This message was sent by Atlassian JIRA
(v7.5.0#75005)
6 years, 6 months
[JBoss JIRA] (ISPN-7801) RehashWithL1Test.testPutWithRehashAndCacheClear random failures
by Galder Zamarreño (JIRA)
[ https://issues.jboss.org/browse/ISPN-7801?page=com.atlassian.jira.plugin.... ]
Galder Zamarreño updated ISPN-7801:
-----------------------------------
Fix Version/s: 9.4.0.Final
(was: 9.3.0.Final)
> RehashWithL1Test.testPutWithRehashAndCacheClear random failures
> ---------------------------------------------------------------
>
> Key: ISPN-7801
> URL: https://issues.jboss.org/browse/ISPN-7801
> Project: Infinispan
> Issue Type: Bug
> Components: Test Suite - Core
> Affects Versions: 9.0.0.Final
> Reporter: Dan Berindei
> Assignee: Dan Berindei
> Labels: testsuite_stability
> Fix For: 9.4.0.Final
>
>
> The test kills the only owner of a key and checks that when a node starts owning an L1 entry, it doesn't send it to other nodes during state transfer. Then it adds a new node (owning the key) and checks that the key isn't transferred to the new node, and it's deleted from L1 on the old nodes. The problem is that it doesn't wait, it assumes all the nodes have already removed it by the time {{getCache()}} returns on the joiner.
> {noformat}
> 03:24:27,606 TRACE (jgroups-5,Test-NodeB-54331:[]) [L1WriteSynchronizer] Caching remotely retrieved entry for key k0 in L1
> 03:24:27,607 TRACE (jgroups-5,Test-NodeB-54331:[]) [DefaultDataContainer] Store MortalCacheEntry{key=k0, value=some data} in container
> 03:24:26,754 DEBUG (testng-Test:[]) [Test] Populating L1 on Test-NodeA-2588
> 03:24:27,514 DEBUG (testng-Test:[]) [Test] Populating L1 on Test-NodeB-54331
> 03:24:27,777 DEBUG (testng-Test:[]) [Test] Populating L1 on Test-NodeC-65326
> 03:24:27,777 DEBUG (testng-Test:[]) [Test] Killing node Test-NodeC-65326
> 03:24:27,781 TRACE (transport-thread-Test-NodeA-p51-t2:[Topology-___defaultcache]) [DefaultDataContainer] Removed MortalCacheEntry{key=k0, value=some data} from container
> *** The entry is not removed from NodeB at this point
> 03:24:27,936 DEBUG (testng-Test:[]) [Test] Checking values on Test-NodeA-2588
> 03:24:27,998 TRACE (jgroups-5,Test-NodeB-54331:[]) [CommandAwareRpcDispatcher] About to send back response SuccessfulResponse{responseValue=MortalCacheValue{value=some data, lifespan=600000, created=1493943867607}} for command ClusteredGetCommand{key=k0, flags=[]}
> 03:24:28,034 TRACE (jgroups-7,Test-NodeA-2588:[]) [L1WriteSynchronizer] Caching remotely retrieved entry for key k0 in L1
> 03:24:28,044 TRACE (jgroups-7,Test-NodeA-2588:[]) [DefaultDataContainer] Store MortalCacheEntry{key=k0, value=some data} in container
> 03:24:28,519 DEBUG (testng-Test:[]) [Test] Checking values on Test-NodeB-54331
> 03:24:28,595 DEBUG (testng-Test:[]) [Test] Starting a new joiner
> 03:24:30,261 TRACE (transport-thread-Test-NodeA-p51-t6:[Topology-___defaultcache]) [InvocationContextInterceptor] Invoked with command InvalidateCommand{keys=[k0, k1, k2, k3, k4, k5, k6, k7, k8, k9]} and InvocationContext [org.infinispan.context.impl.NonTxInvocationContext@54c5cc1d]
> 03:24:30,292 DEBUG (testng-Test:[]) [Test] Checking values on Test-NodeA-2588
> 03:24:30,355 ERROR (testng-Test:[]) [TestSuiteProgress] Test failed: org.infinispan.distribution.rehash.RehashWithL1Test.testPutWithRehashAndCacheClear
> java.lang.AssertionError: wrong value for k0
> at org.testng.AssertJUnit.fail(AssertJUnit.java:59) ~[testng-6.8.8.jar:?]
> at org.testng.AssertJUnit.assertTrue(AssertJUnit.java:24) ~[testng-6.8.8.jar:?]
> at org.testng.AssertJUnit.assertNull(AssertJUnit.java:282) ~[testng-6.8.8.jar:?]
> at org.infinispan.distribution.rehash.RehashWithL1Test.testPutWithRehashAndCacheClear(RehashWithL1Test.java:78) ~[test-classes/:?]
> *** Too late
> 03:24:30,360 TRACE (transport-thread-Test-NodeA-p51-t6:[Topology-___defaultcache]) [DefaultDataContainer] Removed MortalCacheEntry{key=k0, value=some data} from container
> {noformat}
--
This message was sent by Atlassian JIRA
(v7.5.0#75005)
6 years, 6 months