Configuration
Scaler comes with a number of settings that can be tuned for performance. Reasonable defaults are chosen for these that will yield decent performance, but users can tune these settings to get the best performance for their use case. A full list of the available settings can be found by calling the CLI with the -h
flag.
Object storage server settings
For the list of available settings, use the CLI command:
scaler_object_storage_server -h
Scheduler Settings
For the list of available settings, use the CLI command:
scaler_scheduler -h
Protected Mode
The Scheduler is started in protected mode by default, which means that it can not be shut down by the Client. This is because multiple Clients can connect to a long-running Scheduler, and calling shutdown()
may inadverdently kill work that another user is executing. This is also because Scaler encourages strong decoupling between Client, Scheduler, and Workers. To turn off protected mode, start the scheduler with:
scaler_scheduler tcp://127.0.0.1:8516 -p False
Or if using the programmatic API, pass protected=False
:
from scaler import SchedulerClusterCombo
cluster = SchedulerClusterCombo(
address=f"tcp://127.0.0.1:{port}",
n_workers=2,
protected=False, # this will turn off protected mode
)
Event Loop
Scaler uses Python’s built-in asyncio
event loop by default, however users can choose to use uvloop
for the event loop. uvloop
is a faster implementation of the event loop and may improve performance of Scaler.
uvloop
needs to be installed separately:
pip install uvloop
uvloop
can be set when starting the scheduler through the CLI:
scaler_scheduler tcp://127.0.0.1:8516 -e uvloop
from scaler import SchedulerClusterCombo
cluster = SchedulerClusterCombo(
address=f"tcp://127.0.0.1:{port}",
n_workers=2,
event_loop="uvloop"
)
Worker Settings
For the list of available settings, use the CLI command:
scaler_cluster -h
Preload Hook
Workers can execute an optional initialization function before processing tasks using the --preload
option. This enables workers to:
Set up environments on demand
Preload data, libraries, or models
Initialize connections or state
The preload specification follows the format module.path:function(args, kwargs)
where:
module.path
is the Python module to importfunction
is the callable to executeargs
andkwargs
are literal values (strings, numbers, booleans, lists, dicts)
# Simple function call with no arguments
scaler_cluster tcp://127.0.0.1:8516 --preload "mypackage.init:setup"
# Function call with arguments
scaler_cluster tcp://127.0.0.1:8516 --preload "mypackage.init:configure('production', debug=False)"
The preload function is executed once per processor during initialization, before any tasks are processed. If the preload function fails, the error is logged and the processor will terminate.
Example preload module (mypackage/init.py
):
import logging
def setup():
"""Basic setup with no arguments"""
logging.info("Worker initialized")
def configure(environment, debug=True):
"""Setup with configuration parameters"""
logging.info(f"Configuring for {environment}, debug={debug}")
# Initialize connections, load models, etc.
Death Timeout
Workers are spun up with a death_timeout_seconds
, which indicates how long the worker will stay alive without being connected to a Scheduler. The default setting is 300 seconds. This is intended for the workers to clean up if the Scheduler crashes.
This can be set using the CLI:
scaler_cluster -n 10 tcp://127.0.0.1:8516 -ds 300
Or through the programmatic API:
Configuring with TOML Files
While all Scaler components can be configured using command-line flags, using TOML files is the recommended approach for production or shareable setups. Configuration files make your setup explicit, easier to manage, and allow you to check your infrastructure’s configuration into version control.
For convenience, you can define the settings for all components in a single, sectioned TOML file. Each component automatically loads its configuration from its corresponding section.
Core Concepts
Usage: To use a configuration file, pass its path via the
--config
or-c
flag.scaler_scheduler --config /path/to/your/example_config.toml
Precedence: Settings are loaded in a specific order, with later sources overriding earlier ones. The hierarchy is:
Command-Line Flags > TOML File Settings > Built-in Default Values
Naming Convention: The keys in the TOML file must match the long-form command-line arguments. The rule is to replace any hyphens (
-
) with underscores (_
).For example, the flag
--num-of-workers
becomes the TOML keynum_of_workers
.One can discover all available keys by running any command with the
-h
or--help
flag.
Supported Components and Section Names
The following table maps each Scaler command to its corresponding section name in the TOML file.
Command |
TOML Section Name |
---|---|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Practical Scenarios & Examples
Scenario 1: Unified Configuration File
Here is an example of a single example_config.toml
file that configures multiple components using sections.
example_config.toml
# This is a unified configuration file for all Scaler components.
[scheduler]
scheduler_address = "tcp://127.0.0.1:6378"
object_storage_address = "tcp://127.0.0.1:6379"
monitor_address = "tcp://127.0.0.1:6380"
allocate_policy = "even"
logging_level = "INFO"
logging_paths = ["/dev/stdout", "/var/log/scaler/scheduler.log"]
[cluster]
scheduler_address = "tcp://127.0.0.1:6378"
num_of_workers = 8
per_worker_capabilities = "linux,cpu=8"
task_timeout_seconds = 600
[object_storage_server]
object_storage_address = "tcp://127.0.0.1:6379"
[webui]
monitor_address = "tcp://127.0.0.1:6380"
web_port = 8081
With this single file, starting your entire stack is simple and consistent:
scaler_object_storage_server --config example_config.toml &
scaler_scheduler --config example_config.toml &
scaler_cluster --config example_config.toml &
scaler_ui --config example_config.toml &
Scenario 2: Overriding a Section’s Setting
You can override any value from the TOML file by providing it as a command-line flag. For example, to use the example_config.toml
file but test the cluster with 12 workers instead of 8:
# The --num-of-workers flag will take precedence over the [cluster] section
scaler_cluster --config example_config.toml --num-of-workers 12
The cluster will start with 12 workers, but all other settings (like scheduler_address
) will still be loaded from the [cluster]
section of example_config.toml
.