[JBoss JIRA] (ISPN-8646) PessimisticTxPartitionAndMergeDuringRollbackTest.testDegradedPartitionWithDiscard failures
by Pedro Ruivo (JIRA)
[ https://issues.jboss.org/browse/ISPN-8646?page=com.atlassian.jira.plugin.... ]
Pedro Ruivo updated ISPN-8646:
------------------------------
Status: Open (was: New)
> PessimisticTxPartitionAndMergeDuringRollbackTest.testDegradedPartitionWithDiscard failures
> ------------------------------------------------------------------------------------------
>
> Key: ISPN-8646
> URL: https://issues.jboss.org/browse/ISPN-8646
> Project: Infinispan
> Issue Type: Bug
> Components: Test Suite - Core
> Reporter: Pedro Ruivo
> Assignee: Pedro Ruivo
>
> The discard filter isn't working properly. the discard is only expects (and discard) a single command but if the retries happens to fast, the discard let it go through.
> this may be the cause of the other *discard* related failures in this package
> log relevant parts
> {noformat}
> #Node C discards the rollback (as expected)
> 15:34:04,611 TRACE (jgroups-11,PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeC-27886) [org.infinispan.partitionhandling.BaseTxPartitionAndMergeTest] [Discard] Checking command RollbackCommand {gtx=GlobalTx:PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeA-2726:6, cacheName='pes-cache', topologyId=13}.
> 15:34:04,611 TRACE (jgroups-11,PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeC-27886) [org.infinispan.partitionhandling.BaseTxPartitionAndMergeTest] [Discard] Discarding command RollbackCommand {gtx=GlobalTx:PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeA-2726:6, cacheName='pes-cache', topologyId=13}.
> #we split the cluster now
> 15:34:04,612 DEBUG (testng-PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]) [org.infinispan.partitionhandling.PessimisticTxPartitionAndMergeDuringRollbackTest] Splitting cluster in equal partition
> #topology changed on originator and it retries the rollback
> 15:34:04,623 TRACE (remote-thread-PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeA-p34-t6) [org.infinispan.remoting.transport.impl.MultiTargetRequest] Target PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeC-27886 of request 25 left the cluster view
> 15:34:04,623 TRACE (remote-thread-PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeA-p34-t6) [org.infinispan.statetransfer.StateTransferInterceptor] Retrying command RollbackCommand {gtx=GlobalTx:PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeA-2726:6, cacheName='pes-cache', topologyId=14} for topology 14
> #command isn't discarded and it is enqueued. NodeC is still in topology 13.
> #this is wrong. the test expects the rollback to be lost.
> 15:34:04,629 TRACE (jgroups-11,PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeC-27886) [org.infinispan.partitionhandling.BaseTxPartitionAndMergeTest] [Discard] Checking command RollbackCommand {gtx=GlobalTx:PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeA-2726:6, cacheName='pes-cache', topologyId=14}.
> 15:34:04,629 TRACE (jgroups-11,PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeC-27886) [org.infinispan.statetransfer.StateTransferLockImpl] Checking if transaction data was received for topology 14, current topology is 13
> 15:34:04,629 TRACE (jgroups-11,PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeC-27886) [org.infinispan.util.concurrent.BlockingTaskAwareExecutorServiceImpl]
> Added a new task to the queue: 1 task(s) are waiting
> {noformat}
--
This message was sent by Atlassian JIRA
(v7.5.0#75005)
7 years, 4 months
[JBoss JIRA] (ISPN-8646) PessimisticTxPartitionAndMergeDuringRollbackTest.testDegradedPartitionWithDiscard failures
by Pedro Ruivo (JIRA)
[ https://issues.jboss.org/browse/ISPN-8646?page=com.atlassian.jira.plugin.... ]
Pedro Ruivo updated ISPN-8646:
------------------------------
Summary: PessimisticTxPartitionAndMergeDuringRollbackTest.testDegradedPartitionWithDiscard failures (was: PessimisticTxPartitionAndMergeDuringRollbackTest.testDegradedPartitionWithDiscard)
> PessimisticTxPartitionAndMergeDuringRollbackTest.testDegradedPartitionWithDiscard failures
> ------------------------------------------------------------------------------------------
>
> Key: ISPN-8646
> URL: https://issues.jboss.org/browse/ISPN-8646
> Project: Infinispan
> Issue Type: Bug
> Components: Test Suite - Core
> Reporter: Pedro Ruivo
> Assignee: Pedro Ruivo
>
> The discard filter isn't working properly. the discard is only expects (and discard) a single command but if the retries happens to fast, the discard let it go through.
> this may be the cause of the other *discard* related failures in this package
> log relevant parts
> {noformat}
> #Node C discards the rollback (as expected)
> 15:34:04,611 TRACE (jgroups-11,PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeC-27886) [org.infinispan.partitionhandling.BaseTxPartitionAndMergeTest] [Discard] Checking command RollbackCommand {gtx=GlobalTx:PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeA-2726:6, cacheName='pes-cache', topologyId=13}.
> 15:34:04,611 TRACE (jgroups-11,PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeC-27886) [org.infinispan.partitionhandling.BaseTxPartitionAndMergeTest] [Discard] Discarding command RollbackCommand {gtx=GlobalTx:PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeA-2726:6, cacheName='pes-cache', topologyId=13}.
> #we split the cluster now
> 15:34:04,612 DEBUG (testng-PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]) [org.infinispan.partitionhandling.PessimisticTxPartitionAndMergeDuringRollbackTest] Splitting cluster in equal partition
> #topology changed on originator and it retries the rollback
> 15:34:04,623 TRACE (remote-thread-PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeA-p34-t6) [org.infinispan.remoting.transport.impl.MultiTargetRequest] Target PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeC-27886 of request 25 left the cluster view
> 15:34:04,623 TRACE (remote-thread-PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeA-p34-t6) [org.infinispan.statetransfer.StateTransferInterceptor] Retrying command RollbackCommand {gtx=GlobalTx:PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeA-2726:6, cacheName='pes-cache', topologyId=14} for topology 14
> #command isn't discarded and it is enqueued. NodeC is still in topology 13.
> #this is wrong. the test expects the rollback to be lost.
> 15:34:04,629 TRACE (jgroups-11,PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeC-27886) [org.infinispan.partitionhandling.BaseTxPartitionAndMergeTest] [Discard] Checking command RollbackCommand {gtx=GlobalTx:PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeA-2726:6, cacheName='pes-cache', topologyId=14}.
> 15:34:04,629 TRACE (jgroups-11,PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeC-27886) [org.infinispan.statetransfer.StateTransferLockImpl] Checking if transaction data was received for topology 14, current topology is 13
> 15:34:04,629 TRACE (jgroups-11,PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeC-27886) [org.infinispan.util.concurrent.BlockingTaskAwareExecutorServiceImpl]
> Added a new task to the queue: 1 task(s) are waiting
> {noformat}
--
This message was sent by Atlassian JIRA
(v7.5.0#75005)
7 years, 4 months
[JBoss JIRA] (ISPN-8646) PessimisticTxPartitionAndMergeDuringRollbackTest.testDegradedPartitionWithDiscard
by Pedro Ruivo (JIRA)
Pedro Ruivo created ISPN-8646:
---------------------------------
Summary: PessimisticTxPartitionAndMergeDuringRollbackTest.testDegradedPartitionWithDiscard
Key: ISPN-8646
URL: https://issues.jboss.org/browse/ISPN-8646
Project: Infinispan
Issue Type: Bug
Components: Test Suite - Core
Reporter: Pedro Ruivo
Assignee: Pedro Ruivo
The discard filter isn't working properly. the discard is only expects (and discard) a single command but if the retries happens to fast, the discard let it go through.
this may be the cause of the other *discard* related failures in this package
log relevant parts
{noformat}
#Node C discards the rollback (as expected)
15:34:04,611 TRACE (jgroups-11,PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeC-27886) [org.infinispan.partitionhandling.BaseTxPartitionAndMergeTest] [Discard] Checking command RollbackCommand {gtx=GlobalTx:PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeA-2726:6, cacheName='pes-cache', topologyId=13}.
15:34:04,611 TRACE (jgroups-11,PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeC-27886) [org.infinispan.partitionhandling.BaseTxPartitionAndMergeTest] [Discard] Discarding command RollbackCommand {gtx=GlobalTx:PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeA-2726:6, cacheName='pes-cache', topologyId=13}.
#we split the cluster now
15:34:04,612 DEBUG (testng-PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]) [org.infinispan.partitionhandling.PessimisticTxPartitionAndMergeDuringRollbackTest] Splitting cluster in equal partition
#topology changed on originator and it retries the rollback
15:34:04,623 TRACE (remote-thread-PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeA-p34-t6) [org.infinispan.remoting.transport.impl.MultiTargetRequest] Target PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeC-27886 of request 25 left the cluster view
15:34:04,623 TRACE (remote-thread-PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeA-p34-t6) [org.infinispan.statetransfer.StateTransferInterceptor] Retrying command RollbackCommand {gtx=GlobalTx:PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeA-2726:6, cacheName='pes-cache', topologyId=14} for topology 14
#command isn't discarded and it is enqueued. NodeC is still in topology 13.
#this is wrong. the test expects the rollback to be lost.
15:34:04,629 TRACE (jgroups-11,PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeC-27886) [org.infinispan.partitionhandling.BaseTxPartitionAndMergeTest] [Discard] Checking command RollbackCommand {gtx=GlobalTx:PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeA-2726:6, cacheName='pes-cache', topologyId=14}.
15:34:04,629 TRACE (jgroups-11,PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeC-27886) [org.infinispan.statetransfer.StateTransferLockImpl] Checking if transaction data was received for topology 14, current topology is 13
15:34:04,629 TRACE (jgroups-11,PessimisticTxPartitionAndMergeDuringRollbackTest[DIST_SYNC]-NodeC-27886) [org.infinispan.util.concurrent.BlockingTaskAwareExecutorServiceImpl]
Added a new task to the queue: 1 task(s) are waiting
{noformat}
--
This message was sent by Atlassian JIRA
(v7.5.0#75005)
7 years, 4 months
[JBoss JIRA] (ISPN-8644) RemoteGetDuringStateTransferTest.testScenario_2322 random test failures
by Dan Berindei (JIRA)
Dan Berindei created ISPN-8644:
----------------------------------
Summary: RemoteGetDuringStateTransferTest.testScenario_2322 random test failures
Key: ISPN-8644
URL: https://issues.jboss.org/browse/ISPN-8644
Project: Infinispan
Issue Type: Bug
Components: Test Suite - Core
Affects Versions: 9.2.0.Beta2
Reporter: Dan Berindei
Assignee: Dan Berindei
Fix For: 9.2.0.CR1
{noformat}
java.util.concurrent.ExecutionException: org.infinispan.commons.CacheException: java.lang.AssertionError: expected:<RemoteGetDuringStateTransferTest-NodeC-51737> but was:<RemoteGetDuringStateTransferTest-NodeB-41634>
at java.util.concurrent.FutureTask.report(FutureTask.java:122)
at java.util.concurrent.FutureTask.get(FutureTask.java:192)
at org.infinispan.statetransfer.RemoteGetDuringStateTransferTest.testScenario_2xy2(RemoteGetDuringStateTransferTest.java:526)
at org.infinispan.statetransfer.RemoteGetDuringStateTransferTest.testScenario_2322(RemoteGetDuringStateTransferTest.java:452)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Caused by: org.infinispan.commons.CacheException: java.lang.AssertionError: expected:<RemoteGetDuringStateTransferTest-NodeC-51737> but was:<RemoteGetDuringStateTransferTest-NodeB-41634>
at org.infinispan.interceptors.impl.InvocationContextInterceptor.rethrowException(InvocationContextInterceptor.java:141)
at org.infinispan.interceptors.impl.InvocationContextInterceptor.lambda$new$0(InvocationContextInterceptor.java:62)
at org.infinispan.interceptors.InvocationExceptionFunction.apply(InvocationExceptionFunction.java:21)
at org.infinispan.interceptors.impl.QueueAsyncInvocationStage.invokeQueuedHandlers(QueueAsyncInvocationStage.java:118)
at org.infinispan.interceptors.impl.QueueAsyncInvocationStage.accept(QueueAsyncInvocationStage.java:81)
at org.infinispan.interceptors.impl.QueueAsyncInvocationStage.accept(QueueAsyncInvocationStage.java:30)
at java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:760)
at java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:736)
at java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
at java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1962)
at org.infinispan.remoting.transport.AbstractRequest.complete(AbstractRequest.java:66)
at org.infinispan.remoting.transport.impl.SingleTargetRequest.receiveResponse(SingleTargetRequest.java:56)
at org.infinispan.remoting.transport.impl.SingleTargetRequest.onResponse(SingleTargetRequest.java:35)
at org.infinispan.remoting.transport.impl.RequestRepository.addResponse(RequestRepository.java:53)
at org.infinispan.remoting.transport.jgroups.JGroupsTransport.processResponse(JGroupsTransport.java:1329)
at org.infinispan.remoting.transport.jgroups.JGroupsTransport.processMessage(JGroupsTransport.java:1239)
at org.infinispan.remoting.transport.jgroups.JGroupsTransport.access$200(JGroupsTransport.java:121)
at org.infinispan.remoting.transport.jgroups.JGroupsTransport$ChannelCallbacks.receive(JGroupsTransport.java:1367)
at org.jgroups.JChannel.up(JChannel.java:819)
at org.jgroups.stack.ProtocolStack.up(ProtocolStack.java:893)
at org.jgroups.protocols.RSVP.up(RSVP.java:163)
at org.jgroups.protocols.FlowControl.up(FlowControl.java:343)
at org.jgroups.protocols.tom.TOA.up(TOA.java:112)
at org.jgroups.protocols.pbcast.GMS.up(GMS.java:864)
at org.jgroups.protocols.pbcast.STABLE.up(STABLE.java:240)
at org.jgroups.protocols.UNICAST3.deliverMessage(UNICAST3.java:1002)
at org.jgroups.protocols.UNICAST3.handleDataReceived(UNICAST3.java:728)
at org.jgroups.protocols.UNICAST3.up(UNICAST3.java:383)
at org.jgroups.protocols.pbcast.NAKACK2.up(NAKACK2.java:600)
at org.jgroups.protocols.Discovery.up(Discovery.java:267)
at org.jgroups.protocols.TP.passMessageUp(TP.java:1229)
at org.jgroups.util.SubmitToThreadPool$SingleMessageHandler.run(SubmitToThreadPool.java:87)
... 3 more
Caused by: java.lang.AssertionError: expected:<RemoteGetDuringStateTransferTest-NodeC-51737> but was:<RemoteGetDuringStateTransferTest-NodeB-41634>
at org.infinispan.statetransfer.RemoteGetDuringStateTransferTest.lambda$testScenario_2xy2$6(RemoteGetDuringStateTransferTest.java:500)
at org.infinispan.tx.dld.ControlledRpcManager.lambda$checkResponses$0(ControlledRpcManager.java:86)
at org.infinispan.tx.dld.ControlledRpcManager.lambda$afterInvokeRemotely$1(ControlledRpcManager.java:132)
at java.util.ArrayList.forEach(ArrayList.java:1255)
at org.infinispan.tx.dld.ControlledRpcManager.afterInvokeRemotely(ControlledRpcManager.java:132)
at org.infinispan.util.AbstractControlledRpcManager.lambda$invokeRemotelyAsync$0(AbstractControlledRpcManager.java:43)
at java.util.concurrent.CompletableFuture.uniApply(CompletableFuture.java:602)
at java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:577)
... 27 more
{noformat}
This cannot happen anymore after the ISPN-7919 fix landed, because {{ControlledRpcManager.checkResponses()}} doesn't see who sent the response, but we should go back and check that the response is from the node we expected.
--
This message was sent by Atlassian JIRA
(v7.5.0#75005)
7 years, 4 months
[JBoss JIRA] (ISPN-8643) ClientConnectionPoolingTest.testMaxActiveReached random failures
by Dan Berindei (JIRA)
Dan Berindei created ISPN-8643:
----------------------------------
Summary: ClientConnectionPoolingTest.testMaxActiveReached random failures
Key: ISPN-8643
URL: https://issues.jboss.org/browse/ISPN-8643
Project: Infinispan
Issue Type: Bug
Components: Test Suite - Server
Affects Versions: 9.2.0.Beta2
Reporter: Dan Berindei
Fix For: 9.2.0.CR1
Looks like waiting for 1s for all the connections to be releases is not enough:
{noformat}
java.lang.AssertionError:
at org.infinispan.test.AbstractInfinispanTest.eventually(AbstractInfinispanTest.java:229)
at org.infinispan.test.AbstractInfinispanTest.eventually(AbstractInfinispanTest.java:211)
at org.infinispan.test.AbstractInfinispanTest.eventually(AbstractInfinispanTest.java:187)
at org.infinispan.client.hotrod.ClientConnectionPoolingTest.testMaxActiveReached(ClientConnectionPoolingTest.java:208)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
{noformat}
http://ci.infinispan.org/job/Infinispan/job/master/336
--
This message was sent by Atlassian JIRA
(v7.5.0#75005)
7 years, 4 months
[JBoss JIRA] (ISPN-8305) PessimisticTxPartitionAndMergeDuringPrepareTest.testPrimaryOwnerIsolatedPartitionWithDiscard[DIST_SYNC] randomly failing
by Dan Berindei (JIRA)
[ https://issues.jboss.org/browse/ISPN-8305?page=com.atlassian.jira.plugin.... ]
Dan Berindei reopened ISPN-8305:
--------------------------------
Assignee: Dan Berindei (was: Pedro Ruivo)
> PessimisticTxPartitionAndMergeDuringPrepareTest.testPrimaryOwnerIsolatedPartitionWithDiscard[DIST_SYNC] randomly failing
> ------------------------------------------------------------------------------------------------------------------------
>
> Key: ISPN-8305
> URL: https://issues.jboss.org/browse/ISPN-8305
> Project: Infinispan
> Issue Type: Bug
> Affects Versions: 9.1.0.Final
> Reporter: Galder Zamarreño
> Assignee: Dan Berindei
> Labels: testsuite_stability
> Fix For: 9.1.1.Final
>
>
> Dan said it could be related to ISPN-8232 (Dan had not seen failures since ISPN-6997)
> {code}
> Error Message
> Expected key 'MagicKey#k1{151C/F56895B4/127@PessimisticTxPartitionAndMergeDuringPrepareTest[DIST_SYNC]-NodeB-47561}' to be locked on cache 'Cache 'pes-cache'@PessimisticTxPartitionAndMergeDuringPrepareTest[DIST_SYNC]-NodeB-47561'
> Stacktrace
> java.lang.AssertionError: Expected key 'MagicKey#k1{151C/F56895B4/127@PessimisticTxPartitionAndMergeDuringPrepareTest[DIST_SYNC]-NodeB-47561}' to be locked on cache 'Cache 'pes-cache'@PessimisticTxPartitionAndMergeDuringPrepareTest[DIST_SYNC]-NodeB-47561'
> at org.infinispan.test.AbstractCacheTest.assertLocked(AbstractCacheTest.java:77)
> at org.infinispan.partitionhandling.PessimisticTxPartitionAndMergeDuringPrepareTest.checkLocksDuringPartition(PessimisticTxPartitionAndMergeDuringPrepareTest.java:75)
> at org.infinispan.partitionhandling.BasePessimisticTxPartitionAndMergeTest.doTest(BasePessimisticTxPartitionAndMergeTest.java:79)
> at org.infinispan.partitionhandling.PessimisticTxPartitionAndMergeDuringPrepareTest.testPrimaryOwnerIsolatedPartitionWithDiscard(PessimisticTxPartitionAndMergeDuringPrepareTest.java:41)
> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> at java.lang.Thread.run(Thread.java:748)
> ... Removed 18 stack frames{code}
--
This message was sent by Atlassian JIRA
(v7.5.0#75005)
7 years, 4 months
[JBoss JIRA] (ISPN-8642) Administration console - updating scheme doesn't close the update dialog
by Roman Macor (JIRA)
Roman Macor created ISPN-8642:
---------------------------------
Summary: Administration console - updating scheme doesn't close the update dialog
Key: ISPN-8642
URL: https://issues.jboss.org/browse/ISPN-8642
Project: Infinispan
Issue Type: Bug
Components: JMX, reporting and management
Affects Versions: 9.1.4.Final
Reporter: Roman Macor
Priority: Minor
Click on cache container -> configuration -> Schema -> Edit -> update schema
Result: the schema is updated, but the update dialog remains open
Expected result: the schema should be updated and update dialog should be closed
Please close this issue if this is the intended behaviour.
--
This message was sent by Atlassian JIRA
(v7.5.0#75005)
7 years, 4 months