[JBoss JIRA] (WFLY-1066) Automatic configuration of 'Initial_hosts' for a cluster using JGroups TCP-stack in domain mode (aka DOMAIN_PING)
by Sebastian Łaskawiec (JIRA)
[ https://issues.jboss.org/browse/WFLY-1066?page=com.atlassian.jira.plugin.... ]
Sebastian Łaskawiec updated WFLY-1066:
--------------------------------------
Description:
It is complicated to keep the subsystem JGroups in sync if the tcp-stack is used in domain mode.
All new servers that join/leave a clustered server group (configuration) must be added or removed by hand for the jgroup configuration.
The domain server will receive the information if a host-controller enrol and register server to a clustered server-group.
So the configuration of the initial_hosts can be done automatically to avoid old entries which cause unnecessary checks and ensure that all active servers are known.
was:
It is complicated to keep the subsystem JGroups in sync if the tcp-stack is used in domain mode.
All new servers that join/leave a clustered server group (configuration) must be added or removed by hand for the jgroup configuration.
The domain server will receive the informations if a host-controller enrol and register server to a clustered server-group.
So the configuration of the initial_hosts can be done automaticaly to avoid old entries which cause unnecessary checks and ensure that all active servers are known.
> Automatic configuration of 'Initial_hosts' for a cluster using JGroups TCP-stack in domain mode (aka DOMAIN_PING)
> -----------------------------------------------------------------------------------------------------------------
>
> Key: WFLY-1066
> URL: https://issues.jboss.org/browse/WFLY-1066
> Project: WildFly
> Issue Type: Feature Request
> Components: Domain Management
> Environment: Server running in Domain mode
> Reporter: Wolf-Dieter Fink
> Assignee: Sebastian Łaskawiec
> Labels: clustering, domain, jgroups
>
> It is complicated to keep the subsystem JGroups in sync if the tcp-stack is used in domain mode.
> All new servers that join/leave a clustered server group (configuration) must be added or removed by hand for the jgroup configuration.
> The domain server will receive the information if a host-controller enrol and register server to a clustered server-group.
> So the configuration of the initial_hosts can be done automatically to avoid old entries which cause unnecessary checks and ensure that all active servers are known.
--
This message was sent by Atlassian JIRA
(v6.4.11#64026)
10 years, 1 month
[JBoss JIRA] (DROOLS-465) Cannot find KieModule with LATEST
by RH Bugzilla Integration (JIRA)
[ https://issues.jboss.org/browse/DROOLS-465?page=com.atlassian.jira.plugin... ]
RH Bugzilla Integration commented on DROOLS-465:
------------------------------------------------
Tomas David <tdavid(a)redhat.com> changed the Status of [bug 1303018|https://bugzilla.redhat.com/show_bug.cgi?id=1303018] from ON_QA to VERIFIED
> Cannot find KieModule with LATEST
> ---------------------------------
>
> Key: DROOLS-465
> URL: https://issues.jboss.org/browse/DROOLS-465
> Project: Drools
> Issue Type: Bug
> Affects Versions: 6.1.0.Beta1
> Reporter: Takami Hirata
> Assignee: Mario Fusco
> Fix For: 6.4.0.Beta2
>
> Attachments: reproducer.zip
>
>
> I install a kmodule into local repository with command {{mvn clean install}}.
> Loading this kmodule with 'LATEST' works on the day, but will fail on next day.
> Maven will find update on next day because of default updatePolicy 'daily', then maven check local repository ({{org.kie.scanner.Aether}} treats as a remote repository) but local artifact cache does not have 'maven-metadata.xml' which remote artifact should have, so finding update fails and {{org.sonatype.aether.connector.file.FileRepositoryWorker}} removes {{maven-metadata-local.xml}} for cleanup.
> {{org.kie.scanner.Aether}} adds local repository to remote repositories, but this may be incorrect usage.
--
This message was sent by Atlassian JIRA
(v6.4.11#64026)
10 years, 1 month
[JBoss JIRA] (JBJCA-1318) list driven ExceptionSorter, StaleConnectionChecker
by Chao Wang (JIRA)
[ https://issues.jboss.org/browse/JBJCA-1318?page=com.atlassian.jira.plugin... ]
Chao Wang reopened JBJCA-1318:
------------------------------
> list driven ExceptionSorter, StaleConnectionChecker
> ---------------------------------------------------
>
> Key: JBJCA-1318
> URL: https://issues.jboss.org/browse/JBJCA-1318
> Project: IronJacamar
> Issue Type: Task
> Components: JDBC
> Reporter: Johnathon Lee
> Assignee: Johnathon Lee
> Priority: Minor
> Fix For: 1.2.8.Final
>
>
> Configuration defined ExceptionSorter, StaleConnectionChecker.
> Using an externally (from the implementation) defined "," separated list.
> Defined via '<config-property name=prop_name >value1,value2,value3</>' in the standalone/domain configuration files.
> ie:
> set via <config-property name="FatalExceptions">10099,10100,10101</>
> set via <config-property name="StaleExceptions">10099,10100,10101</>
> Documentation updated to reflect suggested error codes for vendor based upon the current implementations.
--
This message was sent by Atlassian JIRA
(v6.4.11#64026)
10 years, 1 month
[JBoss JIRA] (JBJCA-1318) list driven ExceptionSorter, StaleConnectionChecker
by RH Bugzilla Integration (JIRA)
[ https://issues.jboss.org/browse/JBJCA-1318?page=com.atlassian.jira.plugin... ]
RH Bugzilla Integration commented on JBJCA-1318:
------------------------------------------------
Chao Wang <chaowan(a)redhat.com> changed the Status of [bug 1320720|https://bugzilla.redhat.com/show_bug.cgi?id=1320720] from NEW to ASSIGNED
> list driven ExceptionSorter, StaleConnectionChecker
> ---------------------------------------------------
>
> Key: JBJCA-1318
> URL: https://issues.jboss.org/browse/JBJCA-1318
> Project: IronJacamar
> Issue Type: Task
> Components: JDBC
> Reporter: Johnathon Lee
> Assignee: Johnathon Lee
> Priority: Minor
> Fix For: 1.2.8.Final
>
>
> Configuration defined ExceptionSorter, StaleConnectionChecker.
> Using an externally (from the implementation) defined "," separated list.
> Defined via '<config-property name=prop_name >value1,value2,value3</>' in the standalone/domain configuration files.
> ie:
> set via <config-property name="FatalExceptions">10099,10100,10101</>
> set via <config-property name="StaleExceptions">10099,10100,10101</>
> Documentation updated to reflect suggested error codes for vendor based upon the current implementations.
--
This message was sent by Atlassian JIRA
(v6.4.11#64026)
10 years, 1 month
[JBoss JIRA] (WFLY-6294) Session draining always takes maximum configured timeout
by Radoslav Husar (JIRA)
[ https://issues.jboss.org/browse/WFLY-6294?page=com.atlassian.jira.plugin.... ]
Radoslav Husar edited comment on WFLY-6294 at 3/23/16 6:08 PM:
---------------------------------------------------------------
Awesome investigation [~aogburn], thanks. There is bunch of problems in the current implementation that I am fixing:
# *all* sessions that are in the node will be subject to draining while we want to only attempt to drain sessions for which given node is the primary owner (Jira TBD)
# passivated sessions are not accounted for session draining -- this will NOT be fixed as the sessions are likely to be not used any more, but mainly this would mean we would bring all the sessions out of passivation, trigger session activated events and other undesirable side effects
# the notification schema does not work, since the notification will always happen BEFORE the session is actually removed from the manager -- we will poll and wait with a smaller timeouts -- submitted as fix for MODCLUSTER-478
# the server might look hang while draining, so we should log what is it going to attempt to drain -- submitted as part of PR for MODCLUSTER-478
was (Author: rhusar):
Awesome investigation [~aogburn], thanks. There is bunch of problems in the current implementation that will be fixed:
1/ *all* sessions that are in the node will be subject to draining while we want to only attempt to drain sessions for which given node is the primary owner (Jira TBD)
2/ passivated sessions are not accounted for session draining -- this will NOT be fixed as the sessions are likely to be not used any more, but mainly this would mean we would bring all the sessions out of passivation, trigger session activated events and other undesirable side effects
3/ the notification schema does not work, since the notification will always happen BEFORE the session is actually removed from the manager -- we will poll and wait with a smaller timeouts -- submitted as fix for MODCLUSTER-478
4/ the server might look hang while draining, so we should log what is it going to attempt to drain -- submitted as part of PR for MODCLUSTER-478
> Session draining always takes maximum configured timeout
> --------------------------------------------------------
>
> Key: WFLY-6294
> URL: https://issues.jboss.org/browse/WFLY-6294
> Project: WildFly
> Issue Type: Bug
> Components: Clustering
> Affects Versions: 10.0.0.Final
> Reporter: Aaron Ogburn
> Assignee: Radoslav Husar
> Priority: Minor
>
> The mod_cluster session drain wait is not ending as expected. mod_cluster adds a session listener to be notified of session destruction. That is fired appropriately, but when the listener is invoked, the infinispan session manager still reports the session as active. Thus, this drain loop doesn't end after the notify because it still sees the active session:
> {code}
> while ((remainingSessions > 0) && (noTimeout || (timeout > 0))) {
> ModClusterLogger.LOGGER.drainSessions(remainingSessions, context.getHost(), context);
> listener.wait(noTimeout ? 0 : timeout);
> current = System.currentTimeMillis();
> timeout = end - current;
> remainingSessions = context.getActiveSessionCount();
> }
> {code}
> Can the listeners be invoked when the session is fully removed and no longer considered active?
--
This message was sent by Atlassian JIRA
(v6.4.11#64026)
10 years, 1 month
[JBoss JIRA] (WFLY-6294) Session draining always takes maximum configured timeout
by Radoslav Husar (JIRA)
[ https://issues.jboss.org/browse/WFLY-6294?page=com.atlassian.jira.plugin.... ]
Radoslav Husar commented on WFLY-6294:
--------------------------------------
Awesome investigation [~aogburn], thanks. There is bunch of problems in the current implementation that will be fixed:
1/ *all* sessions that are in the node will be subject to draining while we want to only attempt to drain sessions for which given node is the primary owner (Jira TBD)
2/ passivated sessions are not accounted for session draining -- this will NOT be fixed as the sessions are likely to be not used any more, but mainly this would mean we would bring all the sessions out of passivation, trigger session activated events and other undesirable side effects
3/ the notification schema does not work, since the notification will always happen BEFORE the session is actually removed from the manager -- we will poll and wait with a smaller timeouts -- submitted as fix for MODCLUSTER-478
4/ the server might look hang while draining, so we should log what is it going to attempt to drain -- submitted as part of PR for MODCLUSTER-478
> Session draining always takes maximum configured timeout
> --------------------------------------------------------
>
> Key: WFLY-6294
> URL: https://issues.jboss.org/browse/WFLY-6294
> Project: WildFly
> Issue Type: Bug
> Components: Clustering
> Affects Versions: 10.0.0.Final
> Reporter: Aaron Ogburn
> Assignee: Radoslav Husar
> Priority: Minor
>
> The mod_cluster session drain wait is not ending as expected. mod_cluster adds a session listener to be notified of session destruction. That is fired appropriately, but when the listener is invoked, the infinispan session manager still reports the session as active. Thus, this drain loop doesn't end after the notify because it still sees the active session:
> {code}
> while ((remainingSessions > 0) && (noTimeout || (timeout > 0))) {
> ModClusterLogger.LOGGER.drainSessions(remainingSessions, context.getHost(), context);
> listener.wait(noTimeout ? 0 : timeout);
> current = System.currentTimeMillis();
> timeout = end - current;
> remainingSessions = context.getActiveSessionCount();
> }
> {code}
> Can the listeners be invoked when the session is fully removed and no longer considered active?
--
This message was sent by Atlassian JIRA
(v6.4.11#64026)
10 years, 1 month
[JBoss JIRA] (WFLY-6294) Session draining always takes maximum configured timeout
by Radoslav Husar (JIRA)
[ https://issues.jboss.org/browse/WFLY-6294?page=com.atlassian.jira.plugin.... ]
Radoslav Husar commented on WFLY-6294:
--------------------------------------
I would consider this as minor as it is not the default configuration and neither session draining ALWAYS makes much sense for distributable session manager. By default only requests are drained which is not even necessary for stop since we now support graceful shutdown.
> Session draining always takes maximum configured timeout
> --------------------------------------------------------
>
> Key: WFLY-6294
> URL: https://issues.jboss.org/browse/WFLY-6294
> Project: WildFly
> Issue Type: Bug
> Components: Clustering
> Affects Versions: 10.0.0.Final
> Reporter: Aaron Ogburn
> Assignee: Radoslav Husar
> Priority: Minor
>
> The mod_cluster session drain wait is not ending as expected. mod_cluster adds a session listener to be notified of session destruction. That is fired appropriately, but when the listener is invoked, the infinispan session manager still reports the session as active. Thus, this drain loop doesn't end after the notify because it still sees the active session:
> {code}
> while ((remainingSessions > 0) && (noTimeout || (timeout > 0))) {
> ModClusterLogger.LOGGER.drainSessions(remainingSessions, context.getHost(), context);
> listener.wait(noTimeout ? 0 : timeout);
> current = System.currentTimeMillis();
> timeout = end - current;
> remainingSessions = context.getActiveSessionCount();
> }
> {code}
> Can the listeners be invoked when the session is fully removed and no longer considered active?
--
This message was sent by Atlassian JIRA
(v6.4.11#64026)
10 years, 1 month
[JBoss JIRA] (WFLY-6294) Session draining always takes maximum configured timeout
by Radoslav Husar (JIRA)
[ https://issues.jboss.org/browse/WFLY-6294?page=com.atlassian.jira.plugin.... ]
Radoslav Husar updated WFLY-6294:
---------------------------------
Summary: Session draining always takes maximum configured timeout (was: mod_cluster session drain wait does not end)
> Session draining always takes maximum configured timeout
> --------------------------------------------------------
>
> Key: WFLY-6294
> URL: https://issues.jboss.org/browse/WFLY-6294
> Project: WildFly
> Issue Type: Bug
> Components: Clustering
> Affects Versions: 10.0.0.Final
> Reporter: Aaron Ogburn
> Assignee: Radoslav Husar
>
> The mod_cluster session drain wait is not ending as expected. mod_cluster adds a session listener to be notified of session destruction. That is fired appropriately, but when the listener is invoked, the infinispan session manager still reports the session as active. Thus, this drain loop doesn't end after the notify because it still sees the active session:
> {code}
> while ((remainingSessions > 0) && (noTimeout || (timeout > 0))) {
> ModClusterLogger.LOGGER.drainSessions(remainingSessions, context.getHost(), context);
> listener.wait(noTimeout ? 0 : timeout);
> current = System.currentTimeMillis();
> timeout = end - current;
> remainingSessions = context.getActiveSessionCount();
> }
> {code}
> Can the listeners be invoked when the session is fully removed and no longer considered active?
--
This message was sent by Atlassian JIRA
(v6.4.11#64026)
10 years, 1 month
[JBoss JIRA] (WFLY-6294) Session draining always takes maximum configured timeout
by Radoslav Husar (JIRA)
[ https://issues.jboss.org/browse/WFLY-6294?page=com.atlassian.jira.plugin.... ]
Radoslav Husar updated WFLY-6294:
---------------------------------
Priority: Minor (was: Major)
> Session draining always takes maximum configured timeout
> --------------------------------------------------------
>
> Key: WFLY-6294
> URL: https://issues.jboss.org/browse/WFLY-6294
> Project: WildFly
> Issue Type: Bug
> Components: Clustering
> Affects Versions: 10.0.0.Final
> Reporter: Aaron Ogburn
> Assignee: Radoslav Husar
> Priority: Minor
>
> The mod_cluster session drain wait is not ending as expected. mod_cluster adds a session listener to be notified of session destruction. That is fired appropriately, but when the listener is invoked, the infinispan session manager still reports the session as active. Thus, this drain loop doesn't end after the notify because it still sees the active session:
> {code}
> while ((remainingSessions > 0) && (noTimeout || (timeout > 0))) {
> ModClusterLogger.LOGGER.drainSessions(remainingSessions, context.getHost(), context);
> listener.wait(noTimeout ? 0 : timeout);
> current = System.currentTimeMillis();
> timeout = end - current;
> remainingSessions = context.getActiveSessionCount();
> }
> {code}
> Can the listeners be invoked when the session is fully removed and no longer considered active?
--
This message was sent by Atlassian JIRA
(v6.4.11#64026)
10 years, 1 month