Tuning Rundeck
Tuning Rundeck
File descriptors
The Rundeck server process opens a number of files during normal operation. These
include system and java libraries, logs, and sockets.
Your system restricts the number of open file handles per process
but these limitations can be adjusted.
If your installation attempts to exceed the limit, you will see an error
like the one shown below in your service.log file.
Too many open files
On Linux nodes
List the current limit with the ulimit command:
ulimit -n
If the limit is low (eg 1024
) it should be raised.
You can get the current number of open file descriptors used by the
Rundeck server process with lsof:
lsof -p <rundeck pid> | wc -l
Increase the limit for a wide margin.
Edit /etc/security/limits.conf file
to raise the hard and soft limits. Here they are raised to 65535
for
the "rundeck" system account:
rundeck hard nofile 65535
rundeck soft nofile 65535
The system file descriptor limit is set in /proc/sys/fs/file-max.
The following command will increase the limit to 65535:
echo 65535 > /proc/sys/fs/file-max
In a new shell, run the ulimit command to set the new level:
ulimit -n 65535
The ulimit setting can be set in the rundeckd
startup script, or profile.
Restart Rundeck.
Java heap size
The rundeckd
startup script sets initial and maximum heap sizes
for the server process. For many installations it will be sufficient.
If the Rundeck JVM runs out of memory, the following error occurs:
Exception in thread "main" java.lang.OutOfMemoryError: Java heap space
Heap size is governed by the following startup parameters:-Xms<initial heap size>
and -Xmx<maximum heap size>
You can increase these by updating the Rundeck profile.
To see the current values, grep the profile
for
the Xmx and Xms patterns:
Launcher installs:
egrep '(Xmx|Xms)' $RDECK_BASE/etc/profile
RPM and DEB installs:
egrep '(Xmx|Xms)' /etc/rundeck/profile
The default settings initialized by the installer sets these to 1024 megabytes maximum and 256 megabytes initial.
Sizing advice
Several factors drive memory usage in Rundeck:
- User sessions
- Concurrent threads
- Concurrent jobs
- Number of managed nodes
For example, if your installation has dozens of active users that manage a large environment (1000+ nodes), and has sufficient system memory, the following sizings might be more suitable:
In the Launcher Install you can edit the $RDECK_BASE/etc/profile
file.
In RPM create/edit /etc/sysconfig/rundeckd
and add below line.
RDECK_JVM_SETTINGS="$RDECK_JVM_SETTINGS -Xmx4096m -Xms1024m"
In DEB create/edit /etc/default/rundeckd
and add below line.
RDECK_JVM_SETTINGS="$RDECK_JVM_SETTINGS -Xmx4096m -Xms1024m"
Quartz job threadCount
The maximum number of threads used by Rundeck for concurrent jobs
by default is set to 10
.
You can change this value, by updating therundeck-config.properties
file.
Please refer to the Quartz site for detailed information:
Quartz - Configure ThreadPool Settings.
Update rundeck-config
Use the properties mentioned in the Quartz documentation, but replace the org.quartz
prefix, with the prefix quartz
.
e.g. in rundeck-config.properties
:
quartz.threadPool.threadCount = 20
Set the threadCount value to the max number of threads you want to run concurrently.
Memory and Threads Example
As an example, an environment which has 200 scheduled jobs that will trigger all at the same time, while using a threadPool size of 50.
When these jobs are triggered, only 50 will run (sometimes less due to Rundeck using some threads for a temporary background process at that specific time). The other 150 jobs are queued, waiting for more threads to be released so more jobs can be triggered. So jobs will trigger at a delayed time but will trigger. This delay always depends on how much time the active jobs take to finish and free up the next thread.
The relation between the Java heap size allocated and threadPool size is, the more threads you set, the more memory size will be needed.
It is recommended to start from low numbers like 50, 100, 200, etc. (50 is a good number to start at) These settings will depend on how much workload/jobs/memory the server will need to handle within it's environment.
This threads size does not affect only to scheduled jobs but ad-hoc commands, manually triggered jobs, and health-checks.
JMX instrumentation
You may wish to monitor the internal operation of your Rundeck server via JMX.
JMX provides introspection on the JVM, the application server,
and classes all through a consistent interface.
These various components are exposed to the management console
via JMX managed beans - MBeans for short.
Note: For more background information on JMX, see
"Java theory and practice: Instrumenting applications with JMX.".
Enable local JMX monitoring by adding the com.sun.management.jmxremote
flag to the startup parameters in the profile for Launcher Install.
export RDECK_JVM="$RDECK_JVM -Dcom.sun.management.jmxremote"
For RPM create/edit /etc/sysconfig/rundeckd
and add below line.
RDECK_JVM_SETTINGS="$RDECK_JVM_SETTINGS -Dcom.sun.management.jmxremote"
For DEB create/edit /etc/default/rundeckd
and add below line.
RDECK_JVM_SETTINGS="$RDECK_JVM_SETTINGS -Dcom.sun.management.jmxremote"
You use a JMX client to monitor JMX agents.
This can be a desktop GUI like JConsole run locally.
jconsole <rundeck pid>
For instructions on remote JMX monitoring for Grails, Spring and log4j see:
Grails in the enterprise.
Node execution
If you are executing commands across many hundreds or thousands of hosts, the bundled SSH node executor may not meet your performance requirements. Each SSH connection uses multiple threads and the encryption/decryption of messages uses CPU cycles and memory. Depending on your environment, you might choose another Node executor like MCollective, Salt or something similar. This essentially delegates remote execution to another tool designed for asynchronous fan out and thus relieving Rundeck of managing remote task execution.
Built in SSH plugins
If you are interested in using the built in SSH plugins, here are some details about how it performs when executing commands across very large numbers of nodes. For these tests, Rundeck was running on an 8 core, 32GB RAM m2.4xlarge AWS EC2 instance.
We chose the rpm -q
command which checks against the rpm database to see if a particular package was installed. For 1000 nodes we saw an average execution of 52 seconds. A 4000 node cluster took roughly 3.5 minutes, and 8000 node cluster about 7 minutes.
The main limitation appears to be memory of the JVM instance relative to the number of concurrent requests. We tuned the max memory to be 12GB with a 1000 Concurrent Dispatch Threads to 1GB of Memory. GC appears to behave well during the runs given the "bursty" nature of them.
SSL and HTTPS performance
It is possible to offload SSL connection processing by using an SSL termination proxy. This can be accomplished by setting up Apache httpd or Nginx as a frontend to your Rundeck instances.
Resource provider
Rundeck projects obtain information about nodes via a
resource provider. If your resource provider is a long blocking process (due to slow responses from a backend service), it can slow down or even hang up Rundeck. Be sure to make your resource provider work asynchronously.
Also, consider using caching when possible.