Magnolia 4.5 reached end of life on June 30, 2016. This branch is no longer supported, see End-of-life policy.
The Magnolia Scheduler module allows you to schedule commands to run at regularly scheduled times and is powered by the Quartz Engine.
The Scheduler module is bundled with Magnolia and typically already installed. You can download it from our Magnolia Store or Nexus repository.
Scheduler is a community module bundled with both editions and typically already installed. Go to Magnolia Store > Installed modules in AdminCentral to check. To install the module individually, see the general module installation instructions.
See the general module uninstalling instructions and advice.
The Scheduler module can be used to execute any configured command at regular intervals. For example, it could be used to:
The Scheduler module is used to execute commands that are typically configured in other modules. See Commands for more information on configuring commands.
The scheduled tasks are configured in
modules/scheduler/config/jobs. The example
demo configuration can be adapted to set up your own scheduled jobs:
params: Parameters passed to the command. Depends on the command. For example, the
activatecommand expects to receive a repository name and a content path, and the
generatePlanetDatacommand used by the RSS Aggregator module, expects only a
path: Content path to the item that the command should use.
repository: Workspace where the content item resides
active: Enables (
true) and disables (
false) the job.
catalog: Name of the catalog where the command resides
command: Name of the command
cron: CRON expression that sets the scheduled execution time. For example
0 0 1 5 11 ? 2010means "run on November 5th, 2010 at 01:00 am" (as opposed to
0 0 1 5 11 ? *which will run annually on "Nov 5th at 01:00 am") . Cronmaker is a useful tool for building expressions
description: Description of the job
The Synchronization, Backup and RSS Aggregator modules use the Scheduler module for scheduling their execution.
In a clustered configuration one or more workspaces is stored in a shared, clustered storage. See Clustering for more information. Cluster nodes (Magnolia instances) access the clustered workspace rather than their own workspaces. This can lead to a situation where multiple scheduled jobs attempt to access the same content simultaneously and a lockup occurs. To avoid this situation, identify the cluster nodes and run the job on only one node.
magnolia.clusteridproperty in the
magnolia.propertiesfile of the cluster node. The file is in the
/<CATALINA_HOME>/webapps/<contextPath>/WEB-INF/config/defaultfolder. The property value can be a literal cluster name such as
magnolia.clusterid=public123) or a variable such as
jobsand edit the job configuration.
params node, add a
clusterId property and set the value to match the
magnolia.clusterId of the cluster node where you want to run the job.
Job configurations are stored in the
config workspace. If you want to run a particular job on all cluster nodes you would have to cluster the
config workspace so that all instances can read the configuration or create the same job configuration on all cluster nodes. This can be time consuming. As a workaround, configure the job once on a clustered instance without the
clusterId property. This has the effect that the job will run on all cluster nodes.
Scheduled tasks on cluster nodes:
Clustering Nodes are usually clustered due to load balancing or high availability issues. With the above solution to bundle the scheduled job to a fix node you'll loose high availability and will have to reconfigure all scheduled jobs to a surfiving node due to failure of the prefered node.
In a clustered environment should be a automatic handshake between all involved (and Online - heartbeated) clustered nodes and one should determine that it will do the job. To prevent to long communication tries if a node is down there should also be a heartbeat connection between all involved clustered nodes to get stati of all nodes - and if one (or more) aren't accessible it should get marked as offline until it will get up again and will be unmarked. Marked (unavailable) nodes shouldn't be involved in handshake mechanism.