Signals allows decoupled applications to receive notifications when certain actions occur elsewhere in the application.
Celery ships with many signals that you application can hook into to augment behavior of certain actions.
Several kinds of events trigger signals, you can connect to these signals to perform actions as they trigger.
Example connecting to the :signal:`task_sent` signal:
from celery.signals import task_sent
@task_sent.connect
def task_sent_handler(sender=None, task_id=None, task=None, args=None,
kwargs=None, \*\*kwds):
print('Got signal task_sent for task id {0}'.format(task_id))
Some signals also have a sender which you can filter by. For example the :signal:`task_sent` signal uses the task name as a sender, so you can connect your handler to be called only when tasks with name "tasks.add" has been sent by providing the sender argument to :class:`~celery.utils.dispatch.signal.Signal.connect`:
@task_sent.connect(task_sent_handler, sender='tasks.add')
def task_sent_handler(sender=None, task_id=None, task=None, args=None,
kwargs=None, \*\*kwds):
print('Got signal task_sent for task id {0}'.format(task_id)
.. signal:: task_sent
Dispatched when a task has been sent to the broker. Note that this is executed in the client process, the one sending the task, not in the worker.
Sender is the name of the task being sent.
Provides arguments:
- task_id
- Id of the task to be executed.
- task
- The task being executed.
- args
- the tasks positional arguments.
- kwargs
- The tasks keyword arguments.
- eta
- The time to execute the task.
- taskset
- Id of the taskset this task is part of (if any).
.. signal:: task_prerun
Dispatched before a task is executed.
Sender is the task class being executed.
Provides arguments:
- task_id
- Id of the task to be executed.
- task
- The task being executed.
- args
- the tasks positional arguments.
- kwargs
- The tasks keyword arguments.
.. signal:: task_postrun
Dispatched after a task has been executed.
Sender is the task class executed.
Provides arguments:
- task_id
Id of the task to be executed.
- task
The task being executed.
- args
The tasks positional arguments.
- kwargs
The tasks keyword arguments.
- retval
The return value of the task.
state
Name of the resulting state.
.. signal:: task_success
Dispatched when a task succeeds.
Sender is the task class executed.
Provides arguments
- result
- Return value of the task.
.. signal:: task_failure
Dispatched when a task fails.
Sender is the task class executed.
Provides arguments:
- task_id
- Id of the task.
- exception
- Exception instance raised.
- args
- Positional arguments the task was called with.
- kwargs
- Keyword arguments the task was called with.
- traceback
- Stack trace object.
- einfo
- The :class:`celery.datastructures.ExceptionInfo` instance.
.. signal:: task_revoked
Dispatched when a task is revoked/terminated by the worker.
Sender is the task class revoked/terminated.
Provides arguments:
- terminated
- Set to :const:`True` if the task was terminated.
- signum
- Signal number used to terminate the task. If this is :const:`None` and terminated is :const:`True` then :sig:`TERM` should be assumed.
- expired Set to :const:`True` if the task expired.
.. signal:: celeryd_after_setup
This signal is sent after the worker instance is set up, but before it calls run. This means that any queues from the :option:`-Q` option is enabled, logging has been set up and so on.
It can be used to e.g. add custom queues that should always be consumed from, disregarding the :option:`-Q` option. Here's an example that sets up a direct queue for each worker, these queues can then be used to route a task to any specific worker:
from celery.signals import celeryd_after_setup
@celeryd_after_setup.connect
def setup_direct_queue(sender, instance, **kwargs):
queue_name = '{0}.dq'.format(sender) # sender is the hostname of the worker
instance.app.queues.select_add(queue_name)
Provides arguments:
- sender Hostname of the worker.
- instance
- This is the :class:`celery.apps.worker.Worker` instance to be initialized.
Note that only the :attr:`app` and :attr:`hostname` attributes have been
set so far, and the rest of
__init__
has not been executed.
- conf
- The configuration of the current app.
.. signal:: celeryd_init
This is the first signal sent when :program:`celeryd` starts up.
The sender
is the host name of the worker, so this signal can be used
to setup worker specific configuration:
from celery.signals import celeryd_init
@celeryd_init.connect(sender='worker12.example.com')
def configure_worker12(conf=None, **kwargs):
conf.CELERY_DEFAULT_RATE_LIMIT = '10/m'
or to set up configuration for multiple workers you can omit specifying a sender when you connect:
from celery.signals import celeryd_init
@celeryd_init.connect
def configure_workers(sender=None, conf=None, **kwargs):
if sender in ('worker1.example.com', 'worker2.example.com'):
conf.CELERY_DEFAULT_RATE_LIMIT = '10/m'
if sender == 'worker3.example.com':
conf.CELERYD_PREFETCH_MULTIPLIER = 0
Provides arguments:
- sender Hostname of the worker.
- instance
- This is the :class:`celery.apps.worker.Worker` instance to be initialized.
Note that only the :attr:`app` and :attr:`hostname` attributes have been
set so far, and the rest of
__init__
has not been executed.
- conf
- The configuration of the current app.
.. signal:: worker_init
Dispatched before the worker is started.
.. signal:: worker_ready
Dispatched when the worker is ready to accept work.
.. signal:: worker_process_init
Dispatched by each new pool worker process when it starts.
.. signal:: worker_shutdown
Dispatched when the worker is about to shut down.
.. signal:: beat_init
Dispatched when celerybeat starts (either standalone or embedded). Sender is the :class:`celery.beat.Service` instance.
.. signal:: beat_embedded_init
Dispatched in addition to the :signal:`beat_init` signal when celerybeat is started as an embedded process. Sender is the :class:`celery.beat.Service` instance.
.. signal:: eventlet_pool_started
Sent when the eventlet pool has been started.
Sender is the :class:`celery.concurrency.eventlet.TaskPool` instance.
.. signal:: eventlet_pool_preshutdown
Sent when the worker shutdown, just before the eventlet pool is requested to wait for remaining workers.
Sender is the :class:`celery.concurrency.eventlet.TaskPool` instance.
.. signal:: eventlet_pool_postshutdown
Sent when the pool has been joined and the worker is ready to shutdown.
Sender is the :class:`celery.concurrency.eventlet.TaskPool` instance.
.. signal:: eventlet_pool_apply
Sent whenever a task is applied to the pool.
Sender is the :class:`celery.concurrency.eventlet.TaskPool` instance.
Provides arguments:
target
The target function.
args
Positional arguments.
kwargs
Keyword arguments.
.. signal:: setup_logging
Celery won't configure the loggers if this signal is connected, so you can use this to completely override the logging configuration with your own.
If you would like to augment the logging configuration setup by Celery then you can use the :signal:`after_setup_logger` and :signal:`after_setup_task_logger` signals.
Provides arguments:
- loglevel
- The level of the logging object.
- logfile
- The name of the logfile.
- format
- The log format string.
- colorize
- Specify if log messages are colored or not.
.. signal:: after_setup_logger
Sent after the setup of every global logger (not task loggers). Used to augment logging configuration.
Provides arguments:
- logger
- The logger object.
- loglevel
- The level of the logging object.
- logfile
- The name of the logfile.
- format
- The log format string.
- colorize
- Specify if log messages are colored or not.
.. signal:: after_setup_task_logger
Sent after the setup of every single task logger. Used to augment logging configuration.
Provides arguments:
- logger
- The logger object.
- loglevel
- The level of the logging object.
- logfile
- The name of the logfile.
- format
- The log format string.
- colorize
- Specify if log messages are colored or not.