KumoMTA uses Lua in place of a more traditional formatted configuration file.
Lua is a surprisingly powerful configuration language, allowing you to either statically define configuration or dynamically build it by pulling from multiple data sources.
While using a scripting language for configuration may sound complicated, in practice many configurations end up being very succinct and readable. Take a look at the example policy to see how a Lua configuration can appear quite similar to a traditional configuration file.
For more information on implementing policies in KumoMTA, refer to the policy chapter.
By default, the server will load from
/opt/kumomta/etc/policy/init.lua on startup. We recommend using this location, but information on starting the server with an alternate path can be found in the chapter on starting KumoMTA.
To better understand how a KumoMTA configuration is built, it helps to review the general flow of a message through the KumoMTA server. This is covered in more detail in the Queues chapter of the reference manual but will be summarized here.
graph TD SQ["Scheduled Queue: campaign:email@example.com"] SMTPL["ESMTP Listener"] HTTPI["Injection API"] RQ1["Ready Queue: 10.0.0.1->MX(domain.com)"] RQ2["Ready Queue: 10.0.0.2->MX(domain.com)"] POOL["egress pool"] IP1["Source: 10.0.0.1"] IP2["Source: 10.0.0.2"] MAINT["Queue Maintainer"] DESTSITE["domain.com"] SMTPL --> SQ HTTPI --> SQ SQ --> MAINT IP1 --> POOL IP2 --> POOL POOL -- per tenant:domain config --> MAINT MAINT -- throttle per tenant:domain config --> RQ1 MAINT -- throttle per tenant:domain config --> RQ2 RQ1 -- throttle per source:domain config\nconnect via 10.0.0.1 --> DESTSITE RQ2 -- throttle per source:domain config\nconnect via 10.0.0.2 --> DESTSITE
2) The message is assigned into a Scheduled Queue based on the combination of its campaign, tenant, and destination domain. If there is no defined campaign or tenant, the message is placed in a queue based on the elements that are present.
At this point, the behavior of the queue can be configured to control things such as the age of a message, the retry intervals, and the routing of a message. These options are described in the Configuring Queue Management chapter.
3) The KumoMTA server moves the message from the Scheduled Queue into the Ready Queue based on retry intervals configured for the Standby Queue. If a message is on its first attempt, it will be moved to the Ready Queue immediately.
4) Messages move from the Ready Queue to their destination via an egress path that was configured for the Ready Queue. This egress path is defined as a combination of an egress source and a site name. Traffic shaping and other similar options are configured based on this combination, see the Configuring Traffic Shaping chapter for more information.
The egress source is a configured structure that defines the name, source IP, and ehlo domain of a given pathway, and it is added to an egress pool, which the message is assigned to as part of the queue config. Note: While routing is assigned at the egress pool level, traffic shaping happens at the egress source level.
The site name is an identifier string created by merging the combined MX hostnames for a given destination domain. This approach allows the server to queue and throttle based not on the destination domain for a given message, but on the aggregate of all domains that share the same set of MXes.
5) All delivery attempts are logged, and any messages that receive a 4xx tempfail response from the remote host are returned to the Scheduled Queue to await a retry attempt. See the Configuring Logging chapter for more information on logging.
There is a lot of flexibility in how a KumoMTA policy file is laid out, but a few things are generally consistent:
The majority of the base server configuration will reside within an init event handler. The init event is fired when the server first starts up, making it the appropriate time for basic server configuration.
Because these attributes are only loaded on init, an explicit reload must be triggered when anything in the init handler is changed, whether the change is in the policy script itself or a change in a data source or file accessed by the policy script.
Attributes that are needed at the time of queueing and sending are handled in events that are called repeatedly as messages pass through the server. Any such events are constantly firing, and as such any file or data source access involved in those events will update immediately unless caching is configured.
That said, any modification to the policy script itself is subject to caching of the Lua policy, which is refreshed every 300 seconds or 1024 executions by default.
Because the configuration is implemented through policy, the traditional practice of breaking things up into discrete files and assembling them using include statements does not apply.
Since includes were often used for dynamic information such as relay domains or relay hosts, the recommended practice is to store that specific data in a data file or data source and load it using Lua data access functions.
For example, DKIM signing information by domain and selector could be stored in a JSON file like this:
The data file could then be read and used to control signing:
That said, Lua does support Include and Require directives, but they operate similarly to the Include and Require directives commonly used in programming languages.