[JBoss JIRA] (ISPN-7428) PutAll does not scale (triangle)
by Pedro Ruivo (JIRA)
[ https://issues.jboss.org/browse/ISPN-7428?page=com.atlassian.jira.plugin.... ]
Pedro Ruivo updated ISPN-7428:
------------------------------
Description:
PutAll command for some reason now performs worse as the batch size grows. There must be some exponential computation done now that needs to be cleaned up.
The new algorithm used for the triangle splits the command per segments. This slows down everything:
From
Before: 2150 iterations (430 per thread)
After: 104 iterations (20.8 per thread)
Note:
* The per-segments are only need for primary->backup communication. It is possible to split by primary owner in the originator and the primary owner would order and split by segment.
was:PutAll command for some reason now performs worse as the batch size grows. There must be some exponential computation done now that needs to be cleaned up.
> PutAll does not scale (triangle)
> --------------------------------
>
> Key: ISPN-7428
> URL: https://issues.jboss.org/browse/ISPN-7428
> Project: Infinispan
> Issue Type: Bug
> Components: Core
> Reporter: William Burns
> Assignee: Pedro Ruivo
>
> PutAll command for some reason now performs worse as the batch size grows. There must be some exponential computation done now that needs to be cleaned up.
> The new algorithm used for the triangle splits the command per segments. This slows down everything:
> From
> Before: 2150 iterations (430 per thread)
> After: 104 iterations (20.8 per thread)
> Note:
> * The per-segments are only need for primary->backup communication. It is possible to split by primary owner in the originator and the primary owner would order and split by segment.
--
This message was sent by Atlassian JIRA
(v7.2.3#72005)
7 years, 11 months
[JBoss JIRA] (ISPN-7429) ExampleConfigsIT#testCompatibilityModeConfig and ExampleConfigsIT#testRestRollingUpgrades always fail on Solaris Sparc
by Anna Manukyan (JIRA)
Anna Manukyan created ISPN-7429:
-----------------------------------
Summary: ExampleConfigsIT#testCompatibilityModeConfig and ExampleConfigsIT#testRestRollingUpgrades always fail on Solaris Sparc
Key: ISPN-7429
URL: https://issues.jboss.org/browse/ISPN-7429
Project: Infinispan
Issue Type: Bug
Components: Test Suite - Core
Reporter: Anna Manukyan
ExampleConfigsIT#testCompatibilityModeConfig and ExampleConfigsIT#testRestRollingUpgrades are always fail on Solaris Sparc with the following exception:
{code}
java.lang.AssertionError: expected:<200> but was:<400>
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotEquals(Assert.java:743)
at org.junit.Assert.assertEquals(Assert.java:118)
at org.junit.Assert.assertEquals(Assert.java:555)
at org.junit.Assert.assertEquals(Assert.java:542)
at org.infinispan.server.test.configs.ExampleConfigsIT.testCompatibilityModeConfig(ExampleConfigsIT.java:267)
{code}
The test execution log is:
{code}
&#27;[0m02:54:28,986 INFO [org.jboss.modules] (main) JBoss Modules version 1.5.1.Final-redhat-1
&#27;[0m&#27;[0m02:54:30,270 INFO [org.jboss.msc] (main) JBoss MSC version 1.2.6.Final-redhat-1
&#27;[0m&#27;[0m02:54:30,846 INFO [org.jboss.as] (MSC service thread 1-7) WFLYSRV0049: Data Grid 7.1.0 (WildFly Core 2.1.10.Final-redhat-1) starting
&#27;[0m&#27;[0m02:54:37,098 INFO [org.jboss.as.server] (Controller Boot Thread) WFLYSRV0039: Creating http management service using socket-binding (management-http)
&#27;[0m&#27;[0m02:54:37,235 INFO [org.xnio] (MSC service thread 1-2) XNIO version 3.4.1.Final-redhat-1
&#27;[0m&#27;[0m02:54:37,287 INFO [org.xnio.nio] (MSC service thread 1-2) XNIO NIO Implementation Version 3.4.1.Final-redhat-1
&#27;[0m&#27;[33m02:54:37,447 WARN [org.jboss.as.txn] (ServerService Thread Pool -- 29) WFLYTX0013: Node identifier property is set to the default value. Please make sure it is unique.
&#27;[0m&#27;[0m02:54:37,463 INFO [org.jboss.as.clustering.infinispan] (ServerService Thread Pool -- 19) Activating Infinispan subsystem.
&#27;[0m&#27;[0m02:54:37,512 INFO [org.jboss.as.naming] (ServerService Thread Pool -- 25) WFLYNAM0001: Activating Naming Subsystem
&#27;[0m&#27;[0m02:54:37,560 INFO [org.jboss.as.security] (ServerService Thread Pool -- 27) WFLYSEC0002: Activating Security Subsystem
&#27;[0m&#27;[0m02:54:37,606 INFO [org.wildfly.extension.io] (ServerService Thread Pool -- 21) WFLYIO001: Worker 'default' has auto-configured to 256 core threads with 2048 task threads based on your 128 available processors
&#27;[0m&#27;[0m02:54:37,652 INFO [org.jboss.as.security] (MSC service thread 1-4) WFLYSEC0001: Current PicketBox version=4.9.7.Final-redhat-1
&#27;[0m&#27;[0m02:54:37,818 INFO [org.jboss.as.connector] (MSC service thread 1-2) WFLYJCA0009: Starting JCA Subsystem (WildFly/IronJacamar 1.3.5.Final-redhat-1)
&#27;[0m&#27;[0m02:54:37,997 INFO [org.jboss.remoting] (MSC service thread 1-3) JBoss Remoting version 4.0.21.Final-redhat-1
&#27;[0m&#27;[0m02:54:38,000 INFO [org.jboss.as.naming] (MSC service thread 1-6) WFLYNAM0003: Starting Naming Service
&#27;[0m&#27;[0m02:54:38,038 INFO [org.jboss.as.connector.subsystems.datasources] (ServerService Thread Pool -- 18) WFLYJCA0004: Deploying JDBC-compliant driver class org.h2.Driver (version 1.3)
&#27;[0m&#27;[0m02:54:38,074 INFO [org.jboss.as.connector.deployers.jdbc] (MSC service thread 1-7) WFLYJCA0018: Started Driver service with driver-name = h2
&#27;[0m&#27;[0m02:54:40,189 INFO [org.jboss.as.server.deployment.scanner] (MSC service thread 1-6) WFLYDS0013: Started FileSystemDeploymentService for directory /mnt/hudson_workspace/workspace/jdg-func-ispn-testsuite-solaris-os/e69eb0d4/infinispan/server/integration/testsuite/target/server/node1/standalone/deployments
&#27;[0m&#27;[0m02:54:40,294 INFO [org.jboss.as.connector.subsystems.datasources] (MSC service thread 1-8) WFLYJCA0001: Bound data source [java:jboss/datasources/ExampleDS]
&#27;[0m&#27;[0m02:54:42,112 INFO [org.infinispan.factories.GlobalComponentRegistry] (MSC service thread 1-4) ISPN000128: Infinispan version: Infinispan 'Chakra' 8.4.0.ER4-redhat-1
&#27;[0m&#27;[0m02:54:44,645 INFO [org.infinispan.expiration.impl.ExpirationManagerImpl] (MSC service thread 1-6) ISPN000025: wakeUpInterval is <= 0, not starting expired purge thread
&#27;[0m&#27;[0m02:54:44,654 INFO [org.jboss.as.clustering.infinispan] (MSC service thread 1-6) DGISPN0001: Started default cache from local container
&#27;[0m&#27;[0m02:54:44,917 INFO [org.infinispan.server.endpoint] (MSC service thread 1-6) DGENDPT10000: HotRodServer starting
&#27;[0m&#27;[0m02:54:44,917 INFO [org.infinispan.server.endpoint] (MSC service thread 1-1) DGENDPT10000: REST starting
&#27;[0m&#27;[0m02:54:44,917 INFO [org.infinispan.server.endpoint] (MSC service thread 1-8) DGENDPT10000: MemcachedServer starting
&#27;[0m&#27;[0m02:54:44,924 INFO [org.infinispan.server.endpoint] (MSC service thread 1-8) DGENDPT10001: MemcachedServer listening on 127.0.0.1:11211
&#27;[0m&#27;[0m02:54:44,924 INFO [org.infinispan.server.endpoint] (MSC service thread 1-6) DGENDPT10001: HotRodServer listening on 127.0.0.1:11222
&#27;[0m&#27;[33m02:54:45,257 WARN [org.infinispan.server.core.transport.NettyTransport] (MSC service thread 1-8) ISPN005028: Native Epoll transport not available, using NIO instead: java.lang.ExceptionInInitializerError
&#27;[0m&#27;[0m02:54:45,402 INFO [org.jboss.as.clustering.infinispan] (MSC service thread 1-1) DGISPN0001: Started ___protobuf_metadata cache from local container
&#27;[0m&#27;[0m02:54:45,539 INFO [org.jboss.as.clustering.infinispan] (MSC service thread 1-6) DGISPN0001: Started ___script_cache cache from local container
&#27;[0m&#27;[33m02:54:46,831 WARNING [io.netty.channel.DefaultChannelId] (MSC service thread 1-6) Failed to find a usable hardware address from the network interfaces; using random bytes: a9:be:8b:b4:0a:5b:e4:21
&#27;[0m&#27;[0m02:54:47,443 INFO [org.infinispan.rest.NettyRestServer] (MSC service thread 1-1) ISPN012003: REST server starting, listening on 127.0.0.1:8080
&#27;[0m&#27;[0m02:54:47,445 INFO [org.infinispan.server.endpoint] (MSC service thread 1-1) DGENDPT10002: REST mapped to /rest
&#27;[0m&#27;[0m02:54:49,415 INFO [org.jboss.as] (Controller Boot Thread) WFLYSRV0060: Http management interface listening on http://127.0.0.1:9990/management
&#27;[0m&#27;[0m02:54:49,425 INFO [org.jboss.as] (Controller Boot Thread) WFLYSRV0051: Admin console listening on http://127.0.0.1:9990
&#27;[0m&#27;[0m02:54:49,426 INFO [org.jboss.as] (Controller Boot Thread) WFLYSRV0025: Data Grid 7.1.0 (WildFly Core 2.1.10.Final-redhat-1) started in 22504ms - Started 150 of 156 services (42 services are lazy, passive or on-demand)
&#27;[0m&#27;[33m02:54:51,459 WARN [org.infinispan.server.hotrod.Decoder2x] (HotRodServerWorker-6-1) ISPN006011: Operation 'PUT' forced to return previous value should be used on transactional caches, otherwise data inconsistency issues could arise under failure situations
&#27;[0m&#27;[31m02:54:52,532 ERROR [org.jboss.resteasy.resteasy_jaxrs.i18n] (nioEventLoopGroup-7-1) RESTEASY002005: Failed executing PUT /rest/default/2: org.jboss.resteasy.spi.ReaderException: java.lang.NullPointerException
at org.jboss.resteasy.core.MessageBodyParameterInjector.inject(MessageBodyParameterInjector.java:184)
at org.jboss.resteasy.core.MethodInjectorImpl.injectArguments(MethodInjectorImpl.java:91)
at org.jboss.resteasy.core.MethodInjectorImpl.invoke(MethodInjectorImpl.java:114)
at org.jboss.resteasy.core.ResourceMethodInvoker.invokeOnTarget(ResourceMethodInvoker.java:295)
at org.jboss.resteasy.core.ResourceMethodInvoker.invoke(ResourceMethodInvoker.java:249)
at org.jboss.resteasy.core.ResourceMethodInvoker.invoke(ResourceMethodInvoker.java:236)
at org.jboss.resteasy.core.SynchronousDispatcher.invoke(SynchronousDispatcher.java:402)
at org.jboss.resteasy.core.SynchronousDispatcher.invoke(SynchronousDispatcher.java:209)
at org.jboss.resteasy.plugins.server.netty.RequestDispatcher.service(RequestDispatcher.java:83)
at org.jboss.resteasy.plugins.server.netty.RequestHandler.channelRead0(RequestHandler.java:55)
at io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:372)
at io.netty.channel.AbstractChannelHandlerContext.access$600(AbstractChannelHandlerContext.java:38)
at io.netty.channel.AbstractChannelHandlerContext$7.run(AbstractChannelHandlerContext.java:363)
at io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java:163)
at io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:418)
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:440)
at io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:873)
at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:144)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
at org.jboss.resteasy.util.ReadFromStream.readFromStream(ReadFromStream.java:30)
at org.jboss.resteasy.plugins.providers.ByteArrayProvider.readFrom(ByteArrayProvider.java:35)
at org.jboss.resteasy.plugins.providers.ByteArrayProvider.readFrom(ByteArrayProvider.java:23)
at org.jboss.resteasy.core.interception.AbstractReaderInterceptorContext.readFrom(AbstractReaderInterceptorContext.java:61)
at org.jboss.resteasy.core.interception.ServerReaderInterceptorContext.readFrom(ServerReaderInterceptorContext.java:60)
at org.jboss.resteasy.core.interception.AbstractReaderInterceptorContext.proceed(AbstractReaderInterceptorContext.java:53)
at org.jboss.resteasy.plugins.interceptors.encoding.GZIPDecodingInterceptor.aroundReadFrom(GZIPDecodingInterceptor.java:59)
at org.jboss.resteasy.core.interception.AbstractReaderInterceptorContext.proceed(AbstractReaderInterceptorContext.java:55)
at org.jboss.resteasy.core.MessageBodyParameterInjector.inject(MessageBodyParameterInjector.java:151)
... 19 more
&#27;[0m&#27;[0m02:54:52,586 INFO [org.jboss.as.server] (management-handler-thread - 4) WFLYSRV0236: Suspending server with no timeout.
&#27;[0m&#27;[0m02:54:52,594 INFO [org.jboss.as.server] (Management Triggered Shutdown) WFLYSRV0241: Shutting down in response to management operation 'shutdown'
&#27;[0m&#27;[0m02:54:52,717 INFO [org.jboss.as.connector.subsystems.datasources] (MSC service thread 1-2) WFLYJCA0010: Unbound data source [java:jboss/datasources/ExampleDS]
&#27;[0m&#27;[0m02:54:52,813 INFO [org.jboss.as.connector.deployers.jdbc] (MSC service thread 1-7) WFLYJCA0019: Stopped Driver service with driver-name = h2
&#27;[0m&#27;[0m02:54:53,188 INFO [org.jboss.as.clustering.infinispan] (MSC service thread 1-8) DGISPN0002: Stopped default cache from local container
&#27;[0m&#27;[0m02:54:53,256 INFO [org.jboss.as.clustering.infinispan] (MSC service thread 1-6) DGISPN0002: Stopped ___protobuf_metadata cache from local container
&#27;[0m&#27;[0m02:54:53,271 INFO [org.jboss.as.clustering.infinispan] (MSC service thread 1-6) DGISPN0002: Stopped ___script_cache cache from local container
&#27;[0m&#27;[0m02:54:53,605 INFO [org.jboss.as] (MSC service thread 1-5) WFLYSRV0050: Data Grid 7.1.0 (WildFly Core 2.1.10.Final-redhat-1) stopped in 923ms
&#27;[0m
{code}
--
This message was sent by Atlassian JIRA
(v7.2.3#72005)
7 years, 11 months
[JBoss JIRA] (ISPN-7428) PutAll does not scale
by William Burns (JIRA)
William Burns created ISPN-7428:
-----------------------------------
Summary: PutAll does not scale
Key: ISPN-7428
URL: https://issues.jboss.org/browse/ISPN-7428
Project: Infinispan
Issue Type: Bug
Components: Core
Reporter: William Burns
PutAll command for some reason now performs worse as the batch size grows. There must be some exponential computation done now that needs to be cleaned up.
--
This message was sent by Atlassian JIRA
(v7.2.3#72005)
7 years, 11 months
[JBoss JIRA] (ISPN-7130) Failed clearing MariaDB JDBC cache store
by Ryan Emerson (JIRA)
[ https://issues.jboss.org/browse/ISPN-7130?page=com.atlassian.jira.plugin.... ]
Ryan Emerson closed ISPN-7130.
------------------------------
Resolution: Out of Date
[~dcihak][~pferraro] I'm closing this issue as I believe the problem is with MariaDB JDBC driver versions < 1.3.7, not Infinispan.
> Failed clearing MariaDB JDBC cache store
> ----------------------------------------
>
> Key: ISPN-7130
> URL: https://issues.jboss.org/browse/ISPN-7130
> Project: Infinispan
> Issue Type: Bug
> Components: Loaders and Stores
> Affects Versions: 8.2.4.Final
> Reporter: Paul Ferraro
> Assignee: Ryan Emerson
> Priority: Minor
>
> During scenarion {code}eap-7x-failover-db-session-shutdown-repl-sync-mariadb-10{code} (failover test when session are stored in MariaDB database) we saw errors on server when the server was shutting down. According to stacktrace it could be related to Infinispan.
> {code}
> 10:25:43,315 INFO [org.jboss.test.clusterbench.common.session.CommonHttpSessionServlet] (default task-1) New session created: OcWgFKcNH9Qyn1lQO2wJBdIkrf3d-_wr2gVoUeBp
> [JBossINF] 10:25:49,725 ERROR [org.infinispan.persistence.jdbc.binary.JdbcBinaryStore] (expiration-thread--p18-t1) ISPN008001: Failed clearing cache store: java.sql.SQLException: No Parameters set. The command addBatch() must have been set
> [JBossINF] at org.mariadb.jdbc.internal.util.ExceptionMapper.getSqlException(ExceptionMapper.java:149)
> [JBossINF] at org.mariadb.jdbc.MariaDbServerPreparedStatement.executeBatch(MariaDbServerPreparedStatement.java:211)
> [JBossINF] at org.jboss.jca.adapters.jdbc.CachedPreparedStatement.executeBatch(CachedPreparedStatement.java:714)
> [JBossINF] at org.jboss.jca.adapters.jdbc.WrappedStatement.executeBatch(WrappedStatement.java:1190)
> [JBossINF] at org.infinispan.persistence.jdbc.binary.JdbcBinaryStore.purge(JdbcBinaryStore.java:329)
> [JBossINF] at org.infinispan.persistence.manager.PersistenceManagerImpl.purgeExpired(PersistenceManagerImpl.java:373)
> [JBossINF] at org.infinispan.expiration.impl.ClusterExpirationManager.processExpiration(ClusterExpirationManager.java:90)
> [JBossINF] at org.infinispan.expiration.impl.ExpirationManagerImpl$ScheduledTask.run(ExpirationManagerImpl.java:231)
> [JBossINF] at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> [JBossINF] at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
> [JBossINF] at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
> [JBossINF] at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
> [JBossINF] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> [JBossINF] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> [JBossINF] at org.jboss.as.clustering.infinispan.ClassLoaderThreadFactory.lambda$newThread$0(ClassLoaderThreadFactory.java:48)
> [JBossINF] at java.lang.Thread.run(Thread.java:745)
> {code}
> It was shortly after cluster-wide rebalance finished
> {code}
> 10:25:07,403 INFO [org.infinispan.CLUSTER] (remote-thread--p8-t1) ISPN000310: Starting cluster-wide rebalance for cache clusterbench-ee7.ear/clusterbench-ee7-ejb.jar, topology CacheTopology{id=6, rebalanceId=4, currentCH=DefaultConsistentHash{ns=256, owners = (3)[dev212: 90+92, dev215: 85+81, dev213: 81+83]}, pendingCH=DefaultConsistentHash{ns=256, owners = (4)[dev212: 67+72, dev215: 65+63, dev213: 59+55, dev214: 65+66]}, unionCH=null, actualMembers=[dev212, dev215, dev213, dev214]}
> [JBossINF] 10:25:07,403 INFO [org.infinispan.CLUSTER] (remote-thread--p8-t1) [Context=clusterbench-ee7.ear/clusterbench-ee7-ejb.jar][Scope=dev212]ISPN100002: Started local rebalance
> [JBossINF] 10:25:07,404 INFO [org.infinispan.CLUSTER] (transport-thread--p14-t21) [Context=clusterbench-ee7.ear/clusterbench-ee7-ejb.jar][Scope=dev212]ISPN100003: Finished local rebalance
> [JBossINF] 10:25:07,408 INFO [org.infinispan.CLUSTER] (remote-thread--p8-t1) [Context=clusterbench-ee7.ear/clusterbench-ee7-ejb.jar][Scope=dev213]ISPN100003: Finished local rebalance
> [JBossINF] 10:25:07,413 INFO [org.infinispan.CLUSTER] (remote-thread--p8-t1) [Context=clusterbench-ee7.ear/clusterbench-ee7-ejb.jar][Scope=dev215]ISPN100003: Finished local rebalance
> [JBossINF] 10:25:07,436 INFO [org.infinispan.CLUSTER] (remote-thread--p8-t1) [Context=clusterbench-ee7.ear/clusterbench-ee7-ejb.jar][Scope=dev214]ISPN100003: Finished local rebalance
> [JBossINF] 10:25:07,436 INFO [org.infinispan.CLUSTER] (remote-thread--p8-t1) ISPN000336: Finished cluster-wide rebalance for cache clusterbench-ee7.ear/clusterbench-ee7-ejb.jar, topology id = 6
> [JBossINF] 10:25:08,270 INFO [org.infinispan.CLUSTER] (remote-thread--p6-t5) [Context=clusterbench-ee7.ear.clusterbench-ee7-web-granular.war][Scope=dev214]ISPN100003: Finished local rebalance
> [JBossINF] 10:25:08,270 INFO [org.infinispan.CLUSTER] (remote-thread--p6-t5) ISPN000336: Finished cluster-wide rebalance for cache clusterbench-ee7.ear.clusterbench-ee7-web-granular.war, topology id = 6
> {code}
> or after new cluster view was received
> {code}
> 10:29:53,838 INFO [org.infinispan.remoting.transport.jgroups.JGroupsTransport] (thread-4,ee,dev212) ISPN000094: Received new cluster view for channel ejb: [dev215|6] (3) [dev215, dev214, dev212]
> {code}
> No errors or warning occured on client at that time.
> Link to server log:
> http://jenkins.mw.lab.eng.bos.redhat.com/hudson/job/eap-7x-failover-db-se...
--
This message was sent by Atlassian JIRA
(v7.2.3#72005)
7 years, 11 months
[JBoss JIRA] (ISPN-7130) Failed clearing MariaDB JDBC cache store
by Daniel Čihák (JIRA)
[ https://issues.jboss.org/browse/ISPN-7130?page=com.atlassian.jira.plugin.... ]
Daniel Čihák edited comment on ISPN-7130 at 2/1/17 9:22 AM:
------------------------------------------------------------
[~rhusar] MariaDB JDBC driver in version 1.3.4 was used.
was (Author: dcihak):
MariaDB JDBC driver in version 1.3.4 was used.
> Failed clearing MariaDB JDBC cache store
> ----------------------------------------
>
> Key: ISPN-7130
> URL: https://issues.jboss.org/browse/ISPN-7130
> Project: Infinispan
> Issue Type: Bug
> Components: Loaders and Stores
> Affects Versions: 8.2.4.Final
> Reporter: Paul Ferraro
> Assignee: Ryan Emerson
> Priority: Minor
>
> During scenarion {code}eap-7x-failover-db-session-shutdown-repl-sync-mariadb-10{code} (failover test when session are stored in MariaDB database) we saw errors on server when the server was shutting down. According to stacktrace it could be related to Infinispan.
> {code}
> 10:25:43,315 INFO [org.jboss.test.clusterbench.common.session.CommonHttpSessionServlet] (default task-1) New session created: OcWgFKcNH9Qyn1lQO2wJBdIkrf3d-_wr2gVoUeBp
> [JBossINF] 10:25:49,725 ERROR [org.infinispan.persistence.jdbc.binary.JdbcBinaryStore] (expiration-thread--p18-t1) ISPN008001: Failed clearing cache store: java.sql.SQLException: No Parameters set. The command addBatch() must have been set
> [JBossINF] at org.mariadb.jdbc.internal.util.ExceptionMapper.getSqlException(ExceptionMapper.java:149)
> [JBossINF] at org.mariadb.jdbc.MariaDbServerPreparedStatement.executeBatch(MariaDbServerPreparedStatement.java:211)
> [JBossINF] at org.jboss.jca.adapters.jdbc.CachedPreparedStatement.executeBatch(CachedPreparedStatement.java:714)
> [JBossINF] at org.jboss.jca.adapters.jdbc.WrappedStatement.executeBatch(WrappedStatement.java:1190)
> [JBossINF] at org.infinispan.persistence.jdbc.binary.JdbcBinaryStore.purge(JdbcBinaryStore.java:329)
> [JBossINF] at org.infinispan.persistence.manager.PersistenceManagerImpl.purgeExpired(PersistenceManagerImpl.java:373)
> [JBossINF] at org.infinispan.expiration.impl.ClusterExpirationManager.processExpiration(ClusterExpirationManager.java:90)
> [JBossINF] at org.infinispan.expiration.impl.ExpirationManagerImpl$ScheduledTask.run(ExpirationManagerImpl.java:231)
> [JBossINF] at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> [JBossINF] at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
> [JBossINF] at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
> [JBossINF] at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
> [JBossINF] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> [JBossINF] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> [JBossINF] at org.jboss.as.clustering.infinispan.ClassLoaderThreadFactory.lambda$newThread$0(ClassLoaderThreadFactory.java:48)
> [JBossINF] at java.lang.Thread.run(Thread.java:745)
> {code}
> It was shortly after cluster-wide rebalance finished
> {code}
> 10:25:07,403 INFO [org.infinispan.CLUSTER] (remote-thread--p8-t1) ISPN000310: Starting cluster-wide rebalance for cache clusterbench-ee7.ear/clusterbench-ee7-ejb.jar, topology CacheTopology{id=6, rebalanceId=4, currentCH=DefaultConsistentHash{ns=256, owners = (3)[dev212: 90+92, dev215: 85+81, dev213: 81+83]}, pendingCH=DefaultConsistentHash{ns=256, owners = (4)[dev212: 67+72, dev215: 65+63, dev213: 59+55, dev214: 65+66]}, unionCH=null, actualMembers=[dev212, dev215, dev213, dev214]}
> [JBossINF] 10:25:07,403 INFO [org.infinispan.CLUSTER] (remote-thread--p8-t1) [Context=clusterbench-ee7.ear/clusterbench-ee7-ejb.jar][Scope=dev212]ISPN100002: Started local rebalance
> [JBossINF] 10:25:07,404 INFO [org.infinispan.CLUSTER] (transport-thread--p14-t21) [Context=clusterbench-ee7.ear/clusterbench-ee7-ejb.jar][Scope=dev212]ISPN100003: Finished local rebalance
> [JBossINF] 10:25:07,408 INFO [org.infinispan.CLUSTER] (remote-thread--p8-t1) [Context=clusterbench-ee7.ear/clusterbench-ee7-ejb.jar][Scope=dev213]ISPN100003: Finished local rebalance
> [JBossINF] 10:25:07,413 INFO [org.infinispan.CLUSTER] (remote-thread--p8-t1) [Context=clusterbench-ee7.ear/clusterbench-ee7-ejb.jar][Scope=dev215]ISPN100003: Finished local rebalance
> [JBossINF] 10:25:07,436 INFO [org.infinispan.CLUSTER] (remote-thread--p8-t1) [Context=clusterbench-ee7.ear/clusterbench-ee7-ejb.jar][Scope=dev214]ISPN100003: Finished local rebalance
> [JBossINF] 10:25:07,436 INFO [org.infinispan.CLUSTER] (remote-thread--p8-t1) ISPN000336: Finished cluster-wide rebalance for cache clusterbench-ee7.ear/clusterbench-ee7-ejb.jar, topology id = 6
> [JBossINF] 10:25:08,270 INFO [org.infinispan.CLUSTER] (remote-thread--p6-t5) [Context=clusterbench-ee7.ear.clusterbench-ee7-web-granular.war][Scope=dev214]ISPN100003: Finished local rebalance
> [JBossINF] 10:25:08,270 INFO [org.infinispan.CLUSTER] (remote-thread--p6-t5) ISPN000336: Finished cluster-wide rebalance for cache clusterbench-ee7.ear.clusterbench-ee7-web-granular.war, topology id = 6
> {code}
> or after new cluster view was received
> {code}
> 10:29:53,838 INFO [org.infinispan.remoting.transport.jgroups.JGroupsTransport] (thread-4,ee,dev212) ISPN000094: Received new cluster view for channel ejb: [dev215|6] (3) [dev215, dev214, dev212]
> {code}
> No errors or warning occured on client at that time.
> Link to server log:
> http://jenkins.mw.lab.eng.bos.redhat.com/hudson/job/eap-7x-failover-db-se...
--
This message was sent by Atlassian JIRA
(v7.2.3#72005)
7 years, 11 months
[JBoss JIRA] (ISPN-7130) Failed clearing MariaDB JDBC cache store
by Daniel Čihák (JIRA)
[ https://issues.jboss.org/browse/ISPN-7130?page=com.atlassian.jira.plugin.... ]
Daniel Čihák commented on ISPN-7130:
------------------------------------
MariDB JDBC driver in version 1.3.4 was used.
> Failed clearing MariaDB JDBC cache store
> ----------------------------------------
>
> Key: ISPN-7130
> URL: https://issues.jboss.org/browse/ISPN-7130
> Project: Infinispan
> Issue Type: Bug
> Components: Loaders and Stores
> Affects Versions: 8.2.4.Final
> Reporter: Paul Ferraro
> Assignee: Ryan Emerson
> Priority: Minor
>
> During scenarion {code}eap-7x-failover-db-session-shutdown-repl-sync-mariadb-10{code} (failover test when session are stored in MariaDB database) we saw errors on server when the server was shutting down. According to stacktrace it could be related to Infinispan.
> {code}
> 10:25:43,315 INFO [org.jboss.test.clusterbench.common.session.CommonHttpSessionServlet] (default task-1) New session created: OcWgFKcNH9Qyn1lQO2wJBdIkrf3d-_wr2gVoUeBp
> [JBossINF] 10:25:49,725 ERROR [org.infinispan.persistence.jdbc.binary.JdbcBinaryStore] (expiration-thread--p18-t1) ISPN008001: Failed clearing cache store: java.sql.SQLException: No Parameters set. The command addBatch() must have been set
> [JBossINF] at org.mariadb.jdbc.internal.util.ExceptionMapper.getSqlException(ExceptionMapper.java:149)
> [JBossINF] at org.mariadb.jdbc.MariaDbServerPreparedStatement.executeBatch(MariaDbServerPreparedStatement.java:211)
> [JBossINF] at org.jboss.jca.adapters.jdbc.CachedPreparedStatement.executeBatch(CachedPreparedStatement.java:714)
> [JBossINF] at org.jboss.jca.adapters.jdbc.WrappedStatement.executeBatch(WrappedStatement.java:1190)
> [JBossINF] at org.infinispan.persistence.jdbc.binary.JdbcBinaryStore.purge(JdbcBinaryStore.java:329)
> [JBossINF] at org.infinispan.persistence.manager.PersistenceManagerImpl.purgeExpired(PersistenceManagerImpl.java:373)
> [JBossINF] at org.infinispan.expiration.impl.ClusterExpirationManager.processExpiration(ClusterExpirationManager.java:90)
> [JBossINF] at org.infinispan.expiration.impl.ExpirationManagerImpl$ScheduledTask.run(ExpirationManagerImpl.java:231)
> [JBossINF] at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> [JBossINF] at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
> [JBossINF] at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
> [JBossINF] at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
> [JBossINF] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> [JBossINF] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> [JBossINF] at org.jboss.as.clustering.infinispan.ClassLoaderThreadFactory.lambda$newThread$0(ClassLoaderThreadFactory.java:48)
> [JBossINF] at java.lang.Thread.run(Thread.java:745)
> {code}
> It was shortly after cluster-wide rebalance finished
> {code}
> 10:25:07,403 INFO [org.infinispan.CLUSTER] (remote-thread--p8-t1) ISPN000310: Starting cluster-wide rebalance for cache clusterbench-ee7.ear/clusterbench-ee7-ejb.jar, topology CacheTopology{id=6, rebalanceId=4, currentCH=DefaultConsistentHash{ns=256, owners = (3)[dev212: 90+92, dev215: 85+81, dev213: 81+83]}, pendingCH=DefaultConsistentHash{ns=256, owners = (4)[dev212: 67+72, dev215: 65+63, dev213: 59+55, dev214: 65+66]}, unionCH=null, actualMembers=[dev212, dev215, dev213, dev214]}
> [JBossINF] 10:25:07,403 INFO [org.infinispan.CLUSTER] (remote-thread--p8-t1) [Context=clusterbench-ee7.ear/clusterbench-ee7-ejb.jar][Scope=dev212]ISPN100002: Started local rebalance
> [JBossINF] 10:25:07,404 INFO [org.infinispan.CLUSTER] (transport-thread--p14-t21) [Context=clusterbench-ee7.ear/clusterbench-ee7-ejb.jar][Scope=dev212]ISPN100003: Finished local rebalance
> [JBossINF] 10:25:07,408 INFO [org.infinispan.CLUSTER] (remote-thread--p8-t1) [Context=clusterbench-ee7.ear/clusterbench-ee7-ejb.jar][Scope=dev213]ISPN100003: Finished local rebalance
> [JBossINF] 10:25:07,413 INFO [org.infinispan.CLUSTER] (remote-thread--p8-t1) [Context=clusterbench-ee7.ear/clusterbench-ee7-ejb.jar][Scope=dev215]ISPN100003: Finished local rebalance
> [JBossINF] 10:25:07,436 INFO [org.infinispan.CLUSTER] (remote-thread--p8-t1) [Context=clusterbench-ee7.ear/clusterbench-ee7-ejb.jar][Scope=dev214]ISPN100003: Finished local rebalance
> [JBossINF] 10:25:07,436 INFO [org.infinispan.CLUSTER] (remote-thread--p8-t1) ISPN000336: Finished cluster-wide rebalance for cache clusterbench-ee7.ear/clusterbench-ee7-ejb.jar, topology id = 6
> [JBossINF] 10:25:08,270 INFO [org.infinispan.CLUSTER] (remote-thread--p6-t5) [Context=clusterbench-ee7.ear.clusterbench-ee7-web-granular.war][Scope=dev214]ISPN100003: Finished local rebalance
> [JBossINF] 10:25:08,270 INFO [org.infinispan.CLUSTER] (remote-thread--p6-t5) ISPN000336: Finished cluster-wide rebalance for cache clusterbench-ee7.ear.clusterbench-ee7-web-granular.war, topology id = 6
> {code}
> or after new cluster view was received
> {code}
> 10:29:53,838 INFO [org.infinispan.remoting.transport.jgroups.JGroupsTransport] (thread-4,ee,dev212) ISPN000094: Received new cluster view for channel ejb: [dev215|6] (3) [dev215, dev214, dev212]
> {code}
> No errors or warning occured on client at that time.
> Link to server log:
> http://jenkins.mw.lab.eng.bos.redhat.com/hudson/job/eap-7x-failover-db-se...
--
This message was sent by Atlassian JIRA
(v7.2.3#72005)
7 years, 11 months