[JBoss JIRA] (ISPN-5368) Out of order events produced when using the MassIndexer with async backend
by RH Bugzilla Integration (JIRA)
[ https://issues.jboss.org/browse/ISPN-5368?page=com.atlassian.jira.plugin.... ]
RH Bugzilla Integration commented on ISPN-5368:
-----------------------------------------------
Dave Stahl <dstahl(a)redhat.com> changed the Status of [bug 1208162|https://bugzilla.redhat.com/show_bug.cgi?id=1208162] from MODIFIED to ON_QA
> Out of order events produced when using the MassIndexer with async backend
> --------------------------------------------------------------------------
>
> Key: ISPN-5368
> URL: https://issues.jboss.org/browse/ISPN-5368
> Project: Infinispan
> Issue Type: Bug
> Affects Versions: 7.2.0.Beta2, 7.1.1.Final
> Reporter: Gustavo Fernandes
> Assignee: Gustavo Fernandes
> Fix For: 7.2.0.CR1, 7.2.0.Final
>
>
> When using async indexing backend on DIST caches with shared index (InfinispanIndexManager), the MassIndexer fails to re-index all the entries, if it is run from a node that is not
> the indexing master.
> Normally the operation sequence of the MassIndexer in the above configuration, for a two node cluster is:
> * Purge the index
> * Send index job to node A and to node B
> * Flush
> Given the backend is async, all index commands are sent to the master RPC-wise asynchronously, and so a reorder can occur and produce like:
> * Send index job to node A
> * Purge
> * Send index job to node B
> * Flush
> Causing previously re-indexed entries to be wiped
--
This message was sent by Atlassian JIRA
(v6.3.11#6341)
10 years, 11 months
[JBoss JIRA] (ISPN-5019) After coordinator change, cache topologies should be installed in parallel
by RH Bugzilla Integration (JIRA)
[ https://issues.jboss.org/browse/ISPN-5019?page=com.atlassian.jira.plugin.... ]
RH Bugzilla Integration commented on ISPN-5019:
-----------------------------------------------
Dave Stahl <dstahl(a)redhat.com> changed the Status of [bug 1195565|https://bugzilla.redhat.com/show_bug.cgi?id=1195565] from MODIFIED to ON_QA
> After coordinator change, cache topologies should be installed in parallel
> --------------------------------------------------------------------------
>
> Key: ISPN-5019
> URL: https://issues.jboss.org/browse/ISPN-5019
> Project: Infinispan
> Issue Type: Bug
> Components: Core
> Affects Versions: 7.0.2.Final
> Reporter: Dan Berindei
> Assignee: Dan Berindei
> Priority: Critical
> Fix For: 7.2.0.CR1, 7.2.0.Final
>
>
> When the coordinator crashes, the new coordinator has to recover the cache topologies from all the nodes in the cluster and install updated topologies for all the caches. This is done on a single thread, and it can take a long time when there are a lot of caches.
> We should be accelerate this by doing the topology installation on separate threads. However, we have to be careful with the async transport pool, because {{executeOnClusterAsync}} actually needs to spawn a new thread in the same pool.
--
This message was sent by Atlassian JIRA
(v6.3.11#6341)
10 years, 11 months
[JBoss JIRA] (ISPN-4491) Cluster Listener Event Batching
by RH Bugzilla Integration (JIRA)
[ https://issues.jboss.org/browse/ISPN-4491?page=com.atlassian.jira.plugin.... ]
RH Bugzilla Integration commented on ISPN-4491:
-----------------------------------------------
Dave Stahl <dstahl(a)redhat.com> changed the Status of [bug 1212012|https://bugzilla.redhat.com/show_bug.cgi?id=1212012] from MODIFIED to ON_QA
> Cluster Listener Event Batching
> -------------------------------
>
> Key: ISPN-4491
> URL: https://issues.jboss.org/browse/ISPN-4491
> Project: Infinispan
> Issue Type: Enhancement
> Components: Listeners
> Affects Versions: 7.0.2.Final
> Reporter: William Burns
> Assignee: William Burns
> Fix For: 7.1.0.Beta1, 7.1.0.Final
>
>
> Currently when a local listener which was installed for a cluster listener finds an event to send back to the parent it does this 1 message per listener. It might be more beneficial if we had batching so that it wouldn't send 1 message per.
> There are 2 cases I can think of which this would benefit.
> # When the underlying transport is UDP. In this case we can send just 1 message to all the nodes for the event instead of N Unicasts
> # When a node has more than 1 cluster listener installed we could send a single message to notifiy more than 1 listener.
--
This message was sent by Atlassian JIRA
(v6.3.11#6341)
10 years, 11 months
[JBoss JIRA] (ISPN-4888) ExampleConfigsIT.testSSLHotRodConfig fails on JDK8
by RH Bugzilla Integration (JIRA)
[ https://issues.jboss.org/browse/ISPN-4888?page=com.atlassian.jira.plugin.... ]
RH Bugzilla Integration commented on ISPN-4888:
-----------------------------------------------
Dave Stahl <dstahl(a)redhat.com> changed the Status of [bug 1197655|https://bugzilla.redhat.com/show_bug.cgi?id=1197655] from MODIFIED to ON_QA
> ExampleConfigsIT.testSSLHotRodConfig fails on JDK8
> --------------------------------------------------
>
> Key: ISPN-4888
> URL: https://issues.jboss.org/browse/ISPN-4888
> Project: Infinispan
> Issue Type: Bug
> Components: Integration , Test Suite - Server
> Affects Versions: 7.0.0.CR2
> Environment: java version "1.8.0_20"
> Reporter: Dan Berindei
> Assignee: Tristan Tarrant
> Priority: Blocker
> Labels: testsuite_stability
> Fix For: 7.0.0.Final
>
>
> Java 8 doesn't support DSA keys longer than 1024 bits, so we need to change the test SSL key:
> {noformat}
> org.infinispan.client.hotrod.exceptions.TransportException: javax.net.ssl.SSLException: Server key
> at sun.security.provider.DSA$LegacyDSA.checkKey(DSA.java:487)
> at sun.security.provider.DSA.engineInitVerify(DSA.java:152)
> at java.security.Signature$Delegate.init(Signature.java:1104)
> at java.security.Signature$Delegate.chooseProvider(Signature.java:1067)
> at java.security.Signature$Delegate.engineInitVerify(Signature.java:1122)
> at java.security.Signature.initVerify(Signature.java:452)
> at sun.security.ssl.HandshakeMessage$DH_ServerKeyExchange.<init>(HandshakeMessage.java:848)
> at sun.security.ssl.ClientHandshaker.processMessage(ClientHandshaker.java:208)
> at sun.security.ssl.Handshaker.processLoop(Handshaker.java:925)
> at sun.security.ssl.Handshaker.process_record(Handshaker.java:860)
> at sun.security.ssl.SSLSocketImpl.readRecord(SSLSocketImpl.java:1043)
> at sun.security.ssl.SSLSocketImpl.performInitialHandshake(SSLSocketImpl.java:1343)
> at sun.security.ssl.SSLSocketImpl.writeRecord(SSLSocketImpl.java:728)
> at sun.security.ssl.AppOutputStream.write(AppOutputStream.java:123)
> at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82)
> at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140)
> at org.infinispan.client.hotrod.impl.transport.tcp.TcpTransport.flush(TcpTransport.java:164)
> at org.infinispan.client.hotrod.impl.operations.PingOperation.execute(PingOperation.java:42)
> at org.infinispan.client.hotrod.impl.operations.FaultTolerantPingOperation.executeOperation(FaultTolerantPingOperation.java:32)
> at org.infinispan.client.hotrod.impl.operations.FaultTolerantPingOperation.executeOperation(FaultTolerantPingOperation.java:18)
> at org.infinispan.client.hotrod.impl.operations.RetryOnFailureOperation.execute(RetryOnFailureOperation.java:50)
> at org.infinispan.client.hotrod.impl.RemoteCacheImpl.ping(RemoteCacheImpl.java:573)
> at org.infinispan.client.hotrod.RemoteCacheManager.ping(RemoteCacheManager.java:650)
> at org.infinispan.client.hotrod.RemoteCacheManager.createRemoteCache(RemoteCacheManager.java:632)
> at org.infinispan.client.hotrod.RemoteCacheManager.getCache(RemoteCacheManager.java:533)
> at org.infinispan.client.hotrod.RemoteCacheManager.getCache(RemoteCacheManager.java:529)
> at org.infinispan.server.test.util.RemoteCacheManagerFactory.createCache(RemoteCacheManagerFactory.java:26)
> at org.infinispan.server.test.configs.ExampleConfigsIT.createCache(ExampleConfigsIT.java:722)
> at org.infinispan.server.test.configs.ExampleConfigsIT.testSSLHotRodConfig(ExampleConfigsIT.java:371)
> {noformat}
--
This message was sent by Atlassian JIRA
(v6.3.11#6341)
10 years, 11 months