[JBoss JIRA] (ISPN-3160) RemoteCacheManager javadocs list wrong default pool size
by Adrian Nistor (JIRA)
[ https://issues.jboss.org/browse/ISPN-3160?page=com.atlassian.jira.plugin.... ]
Adrian Nistor updated ISPN-3160:
--------------------------------
Fix Version/s: 7.0.0.CR1
(was: 7.0.0.Beta2)
> RemoteCacheManager javadocs list wrong default pool size
> --------------------------------------------------------
>
> Key: ISPN-3160
> URL: https://issues.jboss.org/browse/ISPN-3160
> Project: Infinispan
> Issue Type: Bug
> Components: Remote Protocols
> Affects Versions: 5.1.8.Final
> Reporter: Dennis Reed
> Assignee: Tristan Tarrant
> Priority: Minor
> Fix For: 7.0.0.CR1
>
>
> RemoteCacheManager javadocs list the default pool size as 10:
> "infinispan.client.hotrod.default_executor_factory.pool_size, default = 10."
> However, the default is actually 99:
> org/infinispan/client/hotrodimpl/ConfigurationProperties.java: return props.getIntProperty(DEFAULT_EXECUTOR_FACTORY_POOL_SIZE, 99);
--
This message was sent by Atlassian JIRA
(v6.3.1#6329)
11 years, 6 months
[JBoss JIRA] (ISPN-3223) HotRod 2.0 Client recieves stale toplogy view on instance leave
by Adrian Nistor (JIRA)
[ https://issues.jboss.org/browse/ISPN-3223?page=com.atlassian.jira.plugin.... ]
Adrian Nistor updated ISPN-3223:
--------------------------------
Fix Version/s: 7.0.0.CR1
(was: 7.0.0.Beta2)
> HotRod 2.0 Client recieves stale toplogy view on instance leave
> ---------------------------------------------------------------
>
> Key: ISPN-3223
> URL: https://issues.jboss.org/browse/ISPN-3223
> Project: Infinispan
> Issue Type: Bug
> Components: Remote Protocols
> Affects Versions: 5.2.4.Final, 5.3.0.CR1, 7.0.0.Beta1
> Reporter: Takayoshi Kimura
> Assignee: Galder Zamarreño
> Fix For: 7.0.0.CR1, 7.0.0.Final
>
> Attachments: hr-protocol-13.log, hr-protocol-20.log
>
>
> When killed a HotRod server node, HotRod Clinet sometimes recieves a stale toplogy view which includes the dead node and uses it as a latest view. In this case the client keeps trying to connect that node and keeps failing.
> Looks like the AbstractEncoder1x.generateTopologyResponse() takes care of node join but doesn't handle node leave:
> {noformat}
> if (!serverEndpointsMap.keySet.containsAll(cacheMembers)) {
> {noformat}
> For example, serverEndpointsMap.keySet is [A, B, C] and the actual cacheMembers is [A, B].
--
This message was sent by Atlassian JIRA
(v6.3.1#6329)
11 years, 6 months
[JBoss JIRA] (ISPN-3273) Dist L1 owners that aren't primary don't respect assumeOriginKeptEntryInL1
by Adrian Nistor (JIRA)
[ https://issues.jboss.org/browse/ISPN-3273?page=com.atlassian.jira.plugin.... ]
Adrian Nistor updated ISPN-3273:
--------------------------------
Fix Version/s: 7.0.0.CR1
(was: 7.0.0.Beta2)
> Dist L1 owners that aren't primary don't respect assumeOriginKeptEntryInL1
> --------------------------------------------------------------------------
>
> Key: ISPN-3273
> URL: https://issues.jboss.org/browse/ISPN-3273
> Project: Infinispan
> Issue Type: Bug
> Components: Core
> Reporter: William Burns
> Assignee: William Burns
> Fix For: 7.0.0.CR1
>
> Attachments: DistSyncFuncTest.java
>
>
> When a write operation occurs causing a L1 invalidation, there is a boolean to say assumeOriginKeptEntryInL1 which means the owner won't send an invalidation to the originating node that caused this update. This works fine for the primary owner, however any additional backups think the origin is the primary owner and such send invalidations to possibly the real origin.
> -This affects both tx and non tx caches. Tx caches that are sync don't see the problem since locking prevents the invalidation, however it causes an unneeded network roundtrip which can cause delay.-
> Actually this only affects non-tx caches, as tx caches send the prepare/commit directly to the owner(s) instead of having it relayed.
--
This message was sent by Atlassian JIRA
(v6.3.1#6329)
11 years, 6 months
[JBoss JIRA] (ISPN-3336) Extended Statistics check list
by Adrian Nistor (JIRA)
[ https://issues.jboss.org/browse/ISPN-3336?page=com.atlassian.jira.plugin.... ]
Adrian Nistor updated ISPN-3336:
--------------------------------
Fix Version/s: 7.0.0.CR1
(was: 7.0.0.Beta2)
> Extended Statistics check list
> ------------------------------
>
> Key: ISPN-3336
> URL: https://issues.jboss.org/browse/ISPN-3336
> Project: Infinispan
> Issue Type: Task
> Components: Core
> Reporter: Pedro Ruivo
> Assignee: Pedro Ruivo
> Priority: Critical
> Labels: core
> Fix For: 7.0.0.CR1
>
>
> * add documentation
> * check for duplicate stats;
> * check for not exposed statistics
> * try to improve the code (reduce the number of access to the CHM)
> * add tests to check if they are exported via JMX correctly
> * failing test: OptimisticLockingTxClusterExtendedStatisticLogicTest.testWriteSkewOnNonOwner
> * failing test: PessimisticLockingTxClusterExtendedStatisticLogicTest.testDeadlockOnNonOwnerWithRemoteTx
--
This message was sent by Atlassian JIRA
(v6.3.1#6329)
11 years, 6 months
[JBoss JIRA] (ISPN-3244) TopologyAwareSyncConsistentHashFactory should limit the number of segments per node
by Adrian Nistor (JIRA)
[ https://issues.jboss.org/browse/ISPN-3244?page=com.atlassian.jira.plugin.... ]
Adrian Nistor updated ISPN-3244:
--------------------------------
Fix Version/s: 7.0.0.CR1
(was: 7.0.0.Beta2)
> TopologyAwareSyncConsistentHashFactory should limit the number of segments per node
> -----------------------------------------------------------------------------------
>
> Key: ISPN-3244
> URL: https://issues.jboss.org/browse/ISPN-3244
> Project: Infinispan
> Issue Type: Bug
> Components: State Transfer
> Affects Versions: 5.2.6.Final, 5.3.0.CR2
> Reporter: Dan Berindei
> Assignee: Dan Berindei
> Fix For: 7.0.0.CR1
>
>
> Let's say we have a cluster with 5 nodes: A(r1), B(r2), C(r2), D(r3), E(r3)
> TopologyAwareConsistentSyncHashFactory will spread the segments equally on each rack, meaning A will own 2x segments compared to the other nodes.
> TopologyAwareConsistentHashFactory limits the maximum number per node, so that A owns just as many segments as the other nodes. With a slight limitation: the number of racks must be greater than numOwners, otherwise each rack must hold (at least) one copy of all the data.
> TopologyAwareConsistentSyncHashFactory is a little random, so we can't distribute the data perfectly, but we can limit the number of segments on each node to something like 1.5x the average number of segments.
--
This message was sent by Atlassian JIRA
(v6.3.1#6329)
11 years, 6 months
[JBoss JIRA] (ISPN-3337) Define an approach to allow the injection of Configuration and GlobalConfiguration beans for overriding
by Adrian Nistor (JIRA)
[ https://issues.jboss.org/browse/ISPN-3337?page=com.atlassian.jira.plugin.... ]
Adrian Nistor updated ISPN-3337:
--------------------------------
Fix Version/s: 7.0.0.CR1
(was: 7.0.0.Beta2)
> Define an approach to allow the injection of Configuration and GlobalConfiguration beans for overriding
> -------------------------------------------------------------------------------------------------------
>
> Key: ISPN-3337
> URL: https://issues.jboss.org/browse/ISPN-3337
> Project: Infinispan
> Issue Type: Feature Request
> Components: Spring Integration
> Reporter: Navin Surtani
> Assignee: Gustavo Fernandes
> Fix For: 7.0.0.CR1
>
>
> ISPN-3279 removed the ConfigurationOverrides and GlobalConfigurationOverrides classes in the Spring module. These classes contained older set() calls in order to override Cache configurations and given the way that programmatic Configuration currently works in Infinispan, that approach is not maintainable.
> We hence need a clean way to provide (Global) Configuration beans as a newer, cleaner approach.
--
This message was sent by Atlassian JIRA
(v6.3.1#6329)
11 years, 6 months