Apache Load Balancer Agent
The integration dynamically creates and updates the mod_proxy_balancer configuration, based on the state of the deployed web applications. Once changes occur (relocation / failover / changes to the number of web application instances), the utility identifies the change, updates the balancer configuration, and sends a soft restart to Apache to take the new configuration into account.
General information on how to enable the mod_proxy load balancer and configuration can be found in the Jetty mod_proxy docs. Note, the load balancer actual configuration of specific balancer members are generated automatically as explained below.
Similar agents can be implemented for other HTTP load-balancers such as F5, Radware, CISCO ACE, Barracuda etc., using the Administration and Monitoring API. See the Web Load Balancer Agent PU for details.
How it Works?
Here is a description of the flow that eventually scales the web application dynamically:
- Monitoring Service getting statistics about the behavior of the web application. It could be statistics for a specific web application instance or multiple ones.
- The Monitoring Service identifies that some SLA has been breached. It waits some time to make sure it is not a sudden behavior that should be ignored. If the value monitored still breach it threshold, The Monitor Service calling the relevant API to scale the web application by adding web application instances or removing existing web application instances.
- The GSM receives the command and tries to deploy additional instance(s) of the web application. Based on the SLA the GSM will provision new web application instances to GSC that already hosting web application instances or to an empty ones. It can also start new GSCs via the GSA. If there are no available GSCs, a provisioning failure event will be thrown.
- Once a new instance of the web application has been fully deployed and started the Apache HTTPD Load-Balancer Agent will be notified. It will reconfigure the HTTPD Load-Balancer mapping to have the location of the new web application added and will perform soft re-start of the Apache HTTPD Load-Balancer to accept the new configuration.
- New HTTP requests will be routed (based on the HTTPD load-balancing policy) into the new web application to serve the users requests.
The agent resides within
XAP Root folder/tools/apache
and includes theapache-lb-agent` script. The script should run within the same box where the Apache server is running.
The agent uses two main configurations. The first is the directory where the load-balancer configuration files are created (there is one per web application). The second is the command the script executes in order to update Apache with the changes made to the load-balancer configuration files.
By default, the script assumes that Apache is installed under
PROGRAM_FILES/Apache Software Foundation/Apache2.2 for Windows, and
/opt/local/apache2 for Unix. The location of Apache can be controlled using the
-apache command line parameter. Here is an example for executing the script with a custom Apache location:
apache-lb-agent.sh -apache /opt/apache
apache-lb-agent.bat -apache c:\Apache2.2
When using just the Apache location, the configuration files are created under [apache]/conf/gigaspaces. The command executed is
[apache]/bin/apachectl graceful on Unix, and
[apache]\bin\httpd.exe -k restart on Windows.
The specific location of the configuration directory where the load-balancer configuration files are created, can be passed using the
-conf-dir command line parameter. For example:
apache-lb-agent.sh -apache /opt/apache -conf-dir /var/apache/conf/gigaspaces
apache-lb-agent.bat -apache c:\Apache2.2 -conf-dir c:\Apache2.2\conf\gigaspaces
The Apache configuration should be updated to include the load-balancer configuration directory, the relevant modules required and optionally to enable the load-balancer console. Here is an example of the configuration sections that should be added to Apache:
Include /opt/local/apache2/conf/gigaspaces/*.conf LoadModule proxy_module modules/mod_proxy.so LoadModule proxy_balancer_module modules/mod_proxy_balancer.so LoadModule proxy_http_module modules/mod_proxy_http.so LoadModule status_module modules/mod_status.so ProxyPass /balancer ! <Location /balancer-manager> SetHandler balancer-manager Order deny,allow Deny from all Allow from 127.0.0.1 </Location>
|Command Line parameter||Description||Default Value|
||The installation location of apache (
||windows/unix common locations|
||The directory where the load balancer config files will be created.||
||The update interval command. The interval (in milliseconds) when the load balancer conf files will be updated). The agent updates the configuration files and sends the restart command periodically. All changes happening to the deployment are accumulated and then flushed.||10000 (10 seconds)|
||The full apache restart command.||windows: httpd -k restart
unix: apachectl graceful
||Controls the The
||Lookup group name (or coma separated group names) to locate the lookup service via multicast discovery protocol. This option is relevant only when the load-balancer machine has multicast connectivity with the GSM machine.||The XAP default group name specified via the
||Host machine name(or coma separated host names) or IP(s) running the lookup service.||The XAP default locators specified via the
The agent listens for changes happening in the web application deployment topology running within the Service Grid. If custom lookup groups or lookup locators are used, the agent scripts use the built-in options within the product to change them (for example, in setenv.(bat/sh)).
Load Balancer Configuration
A load-balancer configuration is created for each web application deployed into the service grid. The template used to generate the load-balancer configuration can be found under
XAP Root folder/tools/apache/balancer-template.vm. The file is a velocity template, used to generate the actual load-balancer configuration. It is passed with an instance of the
org.openspaces.pu.container.jee.lb.apache.LoadBalancerInfo object, which holds all the information regarding the currently deployed web application instances locations.
Processing unit specific templates can be created by placing a processing unit name.vm file in the same location. Note, the processing unit name is the name of the processing unit which defaults to the war file name, but can be overridden using the