[JBoss JIRA] (MODCLUSTER-540) ManagerBalancerName variable is lowercased
by Jean-Frederic Clere (JIRA)
[ https://issues.jboss.org/browse/MODCLUSTER-540?page=com.atlassian.jira.pl... ]
Jean-Frederic Clere closed MODCLUSTER-540.
------------------------------------------
Resolution: Rejected
> ManagerBalancerName variable is lowercased
> ------------------------------------------
>
> Key: MODCLUSTER-540
> URL: https://issues.jboss.org/browse/MODCLUSTER-540
> Project: mod_cluster
> Issue Type: Bug
> Affects Versions: 1.3.3.Final
> Reporter: Bogdan Sikora
> Assignee: Michal Karm Babacek
>
> Documentation
> {noformat}
> 3.5.8. ManagerBalancerName
> ManagerBalancerName: That is the name of balancer to use when the JBoss AS/JBossWeb/Tomcat doesn't provide a balancer name.
> Default: mycluster
> {noformat}
> Issue:
> Apache Httpd (2.4.23-ER1) did not recognize ManagerBalancerName uppercase letters and is turning whole name to lowercase. If worker passes balancer name then is correctly used as documentation suggest and even with uppercase letters.
> Reproduce:
> 1. Set up balancer (httpd) with worker (for example EAP-7, do not set Balancer variable)
> 2. Set ManagerBalancerName to QA-bAlAnCeR in mod_cluster.conf
> 3. Start everything and access mod_cluster status page
> 4. Look for Balancer variable under your worker, should be cAmeLcAse but isn't
> Workaround:
> Set balancer name in each worker, as documentation says it will override variable set in mod_cluster.conf (ManagerBalancerName)
> Verbose:
> Httpd debug node join part
> {noformat}
> 08:13:22.058 [INFO] RESPONSE: <!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 3.2 Final//EN">
> <html><head>
> <title>Mod_cluster Status</title>
> </head><body>
> <h1>mod_cluster/1.3.3.Final</h1><a href="/mcm?nonce=2cfb1542-913e-11e6-9142-9fb646370036&refresh=10">Auto Refresh</a> <a href="/mcm?nonce=2cfb1542-913e-11e6-9142-9fb646370036&Cmd=DUMP&Range=ALL">show DUMP output</a> <a href="/mcm?nonce=2cfb1542-913e-11e6-9142-9fb646370036&Cmd=INFO&Range=ALL">show INFO output</a>
> <h1> Node jboss-eap-7.1 (ajp://192.168.122.88:8009): </h1>
> <a href="/mcm?nonce=2cfb1542-913e-11e6-9142-9fb646370036&Cmd=ENABLE-APP&Range=NODE&JVMRoute=jboss-eap-7.1">Enable Contexts</a> <a href="/mcm?nonce=2cfb1542-913e-11e6-9142-9fb646370036&Cmd=DISABLE-APP&Range=NODE&JVMRoute=jboss-eap-7.1">Disable Contexts</a> <a href="/mcm?nonce=2cfb1542-913e-11e6-9142-9fb646370036&Cmd=STOP-APP&Range=NODE&JVMRoute=jboss-eap-7.1">Stop Contexts</a><br/>
> Balancer:qa-balancer,LBGroup: ,Flushpackets: Off,Flushwait: 10000,Ping: 10000000,Smax: 2,Ttl: 60000000,Status: OK,Elected: 0,Read: 0,Transferred: 0,Connected: 0,Load: 1
> <h2> Virtual Host 1:</h2><h3>Contexts:</h3><pre>/clusterbench, Status: ENABLED Request: 0 <a href="/mcm?nonce=2cfb1542-913e-11e6-9142-9fb646370036&Cmd=DISABLE-APP&Range=CONTEXT&JVMRoute=jboss-eap-7.1&Alias=default-host&Context=/clusterbench">Disable</a> <a href="/mcm?nonce=2cfb1542-913e-11e6-9142-9fb646370036&Cmd=STOP-APP&Range=CONTEXT&JVMRoute=jboss-eap-7.1&Alias=default-host&Context=/clusterbench">Stop</a>
> </pre><h3>Aliases:</h3><pre>default-host
> localhost
> </pre></body></html>
> {noformat}
> {noformat}
> [Thu Oct 13 08:12:04.337611 2016] [:debug] [pid 12216] mod_manager.c(3018): manager_handler CONFIG (/) processing: "JVMRoute=jboss-eap-7.1&Host=192.168.122.88&Maxattempts=1&Port=8009&StickySessionForce=No&Type=ajp&ping=10"
> [Thu Oct 13 08:12:04.340202 2016] [:debug] [pid 12216] mod_manager.c(3068): manager_handler CONFIG OK
> [Thu Oct 13 08:12:04.342416 2016] [:debug] [pid 12217] mod_manager.c(2302): manager_trans ENABLE-APP (/)
> [Thu Oct 13 08:12:04.342499 2016] [authz_core:debug] [pid 12217] mod_authz_core.c(809): [client 192.168.122.88:54652] AH01626: authorization result of Require all granted: granted
> [Thu Oct 13 08:12:04.342509 2016] [authz_core:debug] [pid 12217] mod_authz_core.c(809): [client 192.168.122.88:54652] AH01626: authorization result of <RequireAny>: granted
> [Thu Oct 13 08:12:04.342576 2016] [:debug] [pid 12217] mod_proxy_cluster.c(1054): update_workers_node starting
> [Thu Oct 13 08:12:04.343186 2016] [:debug] [pid 12217] mod_proxy_cluster.c(695): add_balancer_node: Create balancer balancer://qa-balancer
> [Thu Oct 13 08:12:04.343243 2016] [:debug] [pid 12217] mod_proxy_cluster.c(293): Created: worker for ajp://192.168.122.88:8009
> [Thu Oct 13 08:12:04.343259 2016] [proxy:debug] [pid 12217] proxy_util.c(1779): AH00925: initializing worker ajp://192.168.122.88 shared
> [Thu Oct 13 08:12:04.343262 2016] [proxy:debug] [pid 12217] proxy_util.c(1821): AH00927: initializing worker ajp://192.168.122.88 local
> [Thu Oct 13 08:12:04.343279 2016] [proxy:debug] [pid 12217] proxy_util.c(1872): AH00931: initialized single connection worker in child 12217 for (192.168.122.88)
> [Thu Oct 13 08:12:04.343288 2016] [:debug] [pid 12217] mod_proxy_cluster.c(293): Created: worker for ajp://192.168.122.88:8009
> [Thu Oct 13 08:12:04.343290 2016] [proxy:debug] [pid 12217] proxy_util.c(1774): AH00924: worker ajp://192.168.122.88 shared already initialized
> [Thu Oct 13 08:12:04.343292 2016] [proxy:debug] [pid 12217] proxy_util.c(1821): AH00927: initializing worker ajp://192.168.122.88 local
> [Thu Oct 13 08:12:04.343312 2016] [proxy:debug] [pid 12217] proxy_util.c(1872): AH00931: initialized single connection worker in child 12217 for (192.168.122.88)
> [Thu Oct 13 08:12:04.343318 2016] [:debug] [pid 12217] mod_proxy_cluster.c(1066): update_workers_node done
> {noformat}
> Configuration (mod_cluster.conf)
> {noformat}
> <IfModule manager_module>
> Listen 192.168.122.88:8747
> LogLevel debug
> <VirtualHost 192.168.122.88:8747>
> ServerName localhost.localdomain:8747
> <Directory />
> Require all granted
> </Directory>
> KeepAliveTimeout 60
> MaxKeepAliveRequests 0
> ServerAdvertise on
> AdvertiseFrequency 5
> ManagerBalancerName QA-bAlAnCeR
> AdvertiseGroup 224.0.5.88:23364
> AdvertiseBindAddress 192.168.122.88:23364
> EnableMCPMReceive
> <Location /mcm>
> SetHandler mod_cluster-manager
> Require all granted
> </Location>
> </VirtualHost>
> </IfModule>
> {noformat}
--
This message was sent by Atlassian JIRA
(v7.2.2#72004)