Author: | Ask Solem (ask at celeryproject.org ) |
---|
Change history
What's new documents describe the changes in major versions, we also have a :ref:`changelog` that lists the changes in bugfix releases (0.0.x), while older series are archived under the :ref:`history` section.
Celery is a simple, flexible, and reliable distributed system to process vast amounts of messages, while providing operations with the tools required to maintain such a system.
It's a task queue with focus on real-time processing, while also supporting task scheduling.
Celery has a large and diverse community of users and contributors, you should come join us :ref:`on IRC <irc-channel>` or :ref:`our mailing-list <mailing-list>`.
To read more about Celery you should go read the :ref:`introduction <intro>`.
While this version is backward compatible with previous versions it's important that you read the following section.
This version is officially supported on CPython 2.7, 3.4, and 3.5. and also supported on PyPy.
Table of Contents
Make sure you read the important notes before upgrading to this version.
- Preface
- Important Notes
- Dropped support for Python 2.6
- Last major version to support Python 2
- Django support
- Removed features
- New Task Message Protocol
- Lowercase setting names
- JSON is now the default serializer
- The Task base class no longer automatically register tasks
- Task argument checking
- Redis Events not backward compatible
- Django: Auto-discover now supports Django app configurations
- Worker direct queues no longer use auto-delete
- Old command-line programs removed
- News
- New protocol highlights
- Prefork: Tasks now log from the child process
- Prefork: One log-file per child process
- Configure broker URL for read/write separately
- Canvas Refactor
- Amazon SQS transport now officially supported
- Apache QPid transport now officially supported
- Schedule tasks based on sunrise, sunset, dawn and dusk
- New API for configuring periodic tasks
- Handling task connection errors
- RabbitMQ priority queue support
- Prefork: Limit child process resident memory size
- Redis: Result backend optimizations
- New Riak result backend Introduced
- New CouchDB result backend introduced
- New Consul result backend introduced
- Brand new Cassandra result backend
- New Elasticsearch result backend introduced
- New File-system result backend introduced
- Event Batching
Task.replace
- Optimized Beat implementation
- Task Auto-retry Decorator
- Remote Task Tracebacks
- Async Result API
- RPC Result Backend matured
- New Task Router API
- In Other News
- Reorganization, Deprecations, and Removals
- Deprecation Time-line Changes
XXX To be written
XXX Needs update
Aaron McMillin, Adam Renberg, Adrien Guinet, Ahmet Demir, Aitor Gómez-Goiri, Albert Wang, Alex Koshelev, Alex Rattray, Alex Williams, Alexander Koshelev, Alexander Lebedev, Alexander Oblovatniy, Alexey Kotlyarov, Ali Bozorgkhan, Alice Zoë Bevan–McGregor, Allard Hoeve, Alman One, Andrea Rabbaglietti, Andrea Rosa, Andrei Fokau, Andrew Rodionoff, Andriy Yurchuk, Aneil Mallavarapu, Areski Belaid, Artyom Koval, Ask Solem, Balthazar Rouberol, Berker Peksag, Bert Vanderbauwhede, Brian Bouterse, Chris Duryee, Chris Erway, Chris Harris, Chris Martin, Corey Farwell, Craig Jellick, Cullen Rhodes, Dallas Marlow, Daniel Wallace, Danilo Bargen, Davanum Srinivas, Dave Smith, David Baumgold, David Harrigan, David Pravec, Dennis Brakhane, Derek Anderson, Dmitry Malinovsky, Dudás Ádám, Dustin J. Mitchell, Ed Morley, Fatih Sucu, Feanil Patel, Felix Schwarz, Fernando Rocha, Flavio Grossi, Frantisek Holop, Gao Jiangmiao, Gerald Manipon, Gilles Dartiguelongue, Gino Ledesma, Hank John, Hogni Gylfason, Ilya Georgievsky, Ionel Cristian Mărieș, James Pulec, Jared Lewis, Jason Veatch, Jasper Bryant-Greene, Jeremy Tillman, Jocelyn Delalande, Joe Jevnik, John Anderson, John Kirkham, John Whitlock, Joshua Harlow, Juan Rossi, Justin Patrin, Kai Groner, Kevin Harvey, Konstantinos Koukopoulos, Kouhei Maeda, Kracekumar Ramaraju, Krzysztof Bujniewicz, Latitia M. Haskins, Len Buckens, Lorenzo Mancini, Lucas Wiman, Luke Pomfrey, Marcio Ribeiro, Marin Atanasov Nikolov, Mark Parncutt, Maxime Vdb, Mher Movsisyan, Michael (:github_user:`michael-k`), Michael Duane Mooring, Michael Permana, Mickaël Penhard, Mike Attwood, Morton Fox, Môshe van der Sterre, Nat Williams, Nathan Van Gheem, Nik Nyby, Omer Katz, Omer Korner, Ori Hoch, Paul Pearce, Paulo Bu, Philip Garnero, Piotr Maślanka, Radek Czajka, Raghuram Srinivasan, Randy Barlow, Rodolfo Carvalho, Roger Hu, Rongze Zhu, Ross Deane, Ryan Luckie, Rémy Greinhofer, Samuel Jaillet, Sergey Azovskov, Sergey Tikhonov, Seungha Kim, Steve Peak, Sukrit Khera, Tadej Janež, Tewfik Sadaoui, Thomas French, Thomas Grainger, Tobias Schottdorf, Tocho Tochev, Valentyn Klindukh, Vic Kumar, Vladimir Bolshakov, Vladimir Gorbunov, Wayne Chang, Wil Langford, Will Thompson, William King, Yury Selivanov, Zoran Pavlovic, 許邱翔, :github_user:`allenling`, :github_user:`bee-keeper`, :github_user:`ffeast`, :github_user:`flyingfoxlee`, :github_user:`gdw2`, :github_user:`gitaarik`, :github_user:`hankjin`, :github_user:`m-vdb`, :github_user:`mdk`, :github_user:`nokrik`, :github_user:`ocean1`, :github_user:`orlo666`, :github_user:`raducc`, :github_user:`wanglei`, :github_user:`worldexception`.
Celery now requires Python 2.7 or later, and also drops support for Python 3.3 so supported versions are:
- CPython 2.7
- CPython 3.4
- CPython 3.5
- PyPy 5.3 (
pypy2
) - PyPy 2.4 (
pypy3
)
Starting from Celery 5.0 only Python 3.5+ will be supported.
To make sure you're not affected by this change you should pin
the Celery version in your requirements file, either to a specific
version: celery==4.0.0
, or a range: celery>=4.0,<5.0
.
Dropping support for Python 2 will enable us to remove massive amounts of compatibility code, and going with Python 3.5 allows us to take advantage of typing, async/await, asyncio, and similar concepts there's no alternative for in older versions.
Celery 4.x will continue to work on Python 2.7, 3.4, 3.5; just as Celery 3.x still works on Python 2.6.
Celery now supports Django 1.8 and newer versions.
- Microsoft Windows is no longer supported.
- Jython is no longer supported.
Webhook task machinery (
celery.task.http
) has been removed.Nowadays it's easy to use the :pypi:`requests` module to write webhook tasks manually. We would love to use requests but we are simply unable to as there's a very vocal 'anti-dependency' mob in the Python community
If you need backwards compatibility you can simply copy + paste the 3.1 version of the module and make sure it's imported by the worker: https://github.com/celery/celery/blob/3.1/celery/task/http.py
Task no longer sends error emails.
This also removes support for
app.mail_admins
, and any functionality related to sending emails.celery.contrib.batches
has been removed.This was an experimental feature, so not covered by our deprecation timeline guarantee.
We announced with the 3.1 release that some transports were moved to experimental status, and that there'd be no official support for the transports, citing a lack of resources.
As this subtle hint for the need of funding failed we've removed them completely, breaking backwards compatibility.
- Using MongoDB as a broker is no longer supported.
- Using the Django ORM as a broker is no longer supported.
- Using SQLAlchemy as a broker is no longer supported.
- Using CouchDB as a broker is no longer supported.
- Using IronMQ as a broker is no longer supported.
- Using Beanstalk as a broker is no longer supported.
In addition some features have been removed completely so that attempting to use them will raise an exception:
The
--autoreload
feature has been removed.This was an experimental feature, and not covered by our deprecation timeline guarantee. The flag is removed completely so the worker will crash at startup when present. Luckily this flag isn't used in production systems.
The
--autoscale
feature has been removed.This flag is only used by companies to save money, but had bugs either nobody cared to work on, or sponsor a few hours of work to get it fixed.
The flag has been removed completely, so you must remove this command-line argument or your workers will crash.
The experimental
threads
pool is no longer supported and has been removed.The force_execv feature is no longer supported.
The
celery worker
command now ignores the--no-execv
,--force-execv
, and theCELERYD_FORCE_EXECV
setting.This flag will be removed completely in 5.0 and the worker will raise an error.
The old legacy "amqp" result backend has been deprecated, and will be removed in Celery 5.0.
Please use the
rpc
result backend for RPC-style calls, and a persistent result backend for multi-consumer results.
Now to the good news...
This version introduces a brand new task message protocol, the first major change to the protocol since the beginning of the project.
The new protocol is enabled by default in this version and since the new version isn't backwards compatible you have to be careful when upgrading.
The 3.1.24 version was released to add compatibility with the new protocol so the easiest way to upgrade is to upgrade to that version first, then upgrade to 4.0 in a second deployment.
If you wish to keep using the old protocol you may also configure the protocol version number used:
app = Celery()
app.conf.task_protocol = 1
Read more about the features available in the new protocol in the news section found later in this document.
In the pursuit of beauty all settings are now renamed to be in all lowercase and some setting names have been renamed for consistency.
This change is fully backwards compatible so you can still use the uppercase setting names, but we would like you to upgrade as soon as possible and you can this automatically using the :program:`celery upgrade settings` command:
$ celery upgrade settings proj/settings.py
This command will modify your module in-place to use the new lower-case
names (if you want uppercase with a "CELERY
" prefix see block below),
and save a backup in :file:`proj/settings.py.orig`.
For Django users and others who want to keep uppercase names
If you're loading Celery configuration from the Django settings module then you'll want to keep using the uppercase names.
You also want to use a CELERY_
prefix so that no Celery settings
collide with Django settings used by other apps.
To do this, you'll first need to convert your settings file to use the new consistent naming scheme, and add the prefix to all Celery related settings:
$ celery upgrade settings --django proj/settings.py
After upgrading the settings file, you need to set the prefix explicitly
in your proj/celery.py
module:
app.config_from_object('django.conf:settings', namespace='CELERY')
You can find the most up to date Django Celery integration example here: :ref:`django-first-steps`.
Note that this will also add a prefix to settings that didn't previously
have one, like BROKER_URL
.
Luckily you don't have to manually change the files, as the :program:`celery upgrade settings --django` program should do the right thing.
The loader will try to detect if your configuration is using the new format, and act accordingly, but this also means you're not allowed to mix and match new and old setting names, that's unless you provide a value for both alternatives.
The major difference between previous versions, apart from the lower case
names, are the renaming of some prefixes, like celerybeat_
to beat_
,
celeryd_
to worker_
.
The celery_
prefix has also been removed, and task related settings
from this name-space is now prefixed by task_
, worker related settings
with worker_
.
Apart from this most of the settings will be the same in lowercase, apart from a few special ones:
You can see a full table of the changes in :ref:`conf-old-settings-map`.
The time has finally come to end the reign of :mod:`pickle` as the default serialization mechanism, and json is the default serializer starting from this version.
This change was :ref:`announced with the release of Celery 3.1 <last-version-to-enable-pickle>`.
If you're still depending on :mod:`pickle` being the default serializer, then you have to configure your app before upgrading to 4.0:
task_serializer = 'pickle'
result_serializer = 'pickle'
accept_content = {'pickle'}
The :class:`~@Task` class is no longer using a special meta-class that automatically registers the task in the task registry.
Instead this is now handled by the :class:`@task` decorators.
If you're still using class based tasks, then you need to register these manually:
class CustomTask(Task):
def run(self):
print('running')
app.tasks.register(CustomTask())
The best practice is to use custom task classes only for overriding general behavior, and then using the task decorator to realize the task:
@app.task(bind=True, base=CustomTask)
def custom(self):
print('running')
This change also means the abstract
attribute of the task
no longer has any effect.
The arguments of the task are now verified when calling the task, even asynchronously:
>>> @app.task
... def add(x, y):
... return x + y
>>> add.delay(8, 8)
<AsyncResult: f59d71ca-1549-43e0-be41-4e8821a83c0c>
>>> add.delay(8)
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "celery/app/task.py", line 376, in delay
return self.apply_async(args, kwargs)
File "celery/app/task.py", line 485, in apply_async
check_arguments(*(args or ()), **(kwargs or {}))
TypeError: add() takes exactly 2 arguments (1 given)
The Redis fanout_patterns
and fanout_prefix
transport
options are now enabled by default.
Workers/monitors without these flags enabled won't be able to see workers with this flag disabled. They can still execute tasks, but they cannot receive each others monitoring messages.
You can upgrade in a backward compatible manner by first configuring your 3.1 workers and monitors to enable the settings, before the final upgrade to 4.0:
BROKER_TRANSPORT_OPTIONS = {
'fanout_patterns': True,
'fanout_prefix': True,
}
The autodiscover_tasks()
function can now be called without arguments,
and the Django handler will automatically find your installed apps:
app.autodiscover_tasks()
The Django integration :ref:`example in the documentation <django-first-steps>` has been updated to use the argument-less call.
This also ensures compatibility with the new, ehm, AppConfig
stuff
introduced in recent Django versions.
Workers/clients running 4.0 will no longer be able to send worker direct messages to workers running older versions, and vice versa.
If you're relying on worker direct messages you should upgrade your 3.x workers and clients to use the new routing settings first, by replacing :func:`celery.utils.worker_direct` with this implementation:
from kombu import Exchange, Queue
worker_direct_exchange = Exchange('C.dq2')
def worker_direct(hostname):
return Queue(
'{hostname}.dq2'.format(hostname),
exchange=worker_direct_exchange,
routing_key=hostname,
)
This feature closed Issue #2492.
Installing Celery will no longer install the celeryd
,
celerybeat
and celeryd-multi
programs.
This was announced with the release of Celery 3.1, but you may still have scripts pointing to the old names so make sure you update these to use the new umbrella command:
Program | New Status | Replacement |
---|---|---|
celeryd |
REMOVED | :program:`celery worker` |
celerybeat |
REMOVED | :program:`celery beat` |
celeryd-multi |
REMOVED | :program:`celery multi` |
The new protocol fixes many problems with the old one, and enables some long-requested features:
Most of the data are now sent as message headers, instead of being serialized with the message body.
In version 1 of the protocol the worker always had to deserialize the message to be able to read task meta-data like the task id, name, etc. This also meant that the worker was forced to double-decode the data, first deserializing the message on receipt, serializing the message again to send to child process, then finally the child process deserializes the message again.
Keeping the meta-data fields in the message headers means the worker doesn't actually have to decode the payload before delivering the task to the child process, and also that it's now possible for the worker to reroute a task written in a language different from Python to a different worker.
A new
lang
message header can be used to specify the programming language the task is written in.Worker stores results for internal errors like
ContentDisallowed
, and other deserialization errors.Worker stores results and sends monitoring events for unregistered task errors.
Worker calls callbacks/errbacks even when the result is sent by the parent process (e.g., :exc:`WorkerLostError` when a child process terminates, deserialization errors, unregistered tasks).
A new
origin
header contains information about the process sending the task (worker node-name, or PID and host-name information).A new
shadow
header allows you to modify the task name used in logs.This is useful for dispatch like patterns, like a task that calls any function using pickle (don't do this at home):
from celery import Task from celery.utils.imports import qualname class call_as_task(Task): def shadow_name(self, args, kwargs, options): return 'call_as_task:{0}'.format(qualname(args[0])) def run(self, fun, *args, **kwargs): return fun(*args, **kwargs) call_as_task = app.tasks.register(call_as_task())
New
argsrepr
andkwargsrepr
fields contain textual representations of the task arguments (possibly truncated) for use in logs, monitors, etc.This means the worker doesn't have to deserialize the message payload to display the task arguments for informational purposes.
Chains now use a dedicated
chain
field enabling support for chains of thousands and more tasks.New
parent_id
androot_id
headers adds information about a tasks relationship with other tasks.parent_id
is the task id of the task that called this taskroot_id
is the first task in the work-flow.
These fields can be used to improve monitors like flower to group related messages together (like chains, groups, chords, complete work-flows, etc).
app.TaskProducer
replaced by :meth:`@amqp.create_task_message` and :meth:`@amqp.send_task_message`.Dividing the responsibilities into creating and sending means that people who want to send messages using a Python AMQP client directly, doesn't have to implement the protocol.
The :meth:`@amqp.create_task_message` method calls either :meth:`@amqp.as_task_v2`, or :meth:`@amqp.as_task_v1` depending on the configured task protocol, and returns a special :class:`~celery.app.amqp.task_message` tuple containing the headers, properties and body of the task message.
.. seealso:: The new task protocol is documented in full here: :ref:`message-protocol-task-v2`.
Logging of task success/failure now happens from the child process executing the task. As a result logging utilities, like Sentry can get full information about tasks, including variables in the traceback stack.
Init-scrips and :program:`celery multi` now uses the %I log file format option (e.g., :file:`/var/log/celery/%n%I.log`).
This change was necessary to ensure each child process has a separate log file after moving task logging to the child process, as multiple processes writing to the same log file can cause corruption.
You're encouraged to upgrade your init-scripts and :program:`celery multi` arguments to use this new option.
New :setting:`broker_read_url` and :setting:`broker_write_url` settings have been added so that separate broker URLs can be provided for connections used for consuming/publishing.
In addition to the configuration options, two new methods have been added the app API:
app.connection_for_read()
app.connection_for_write()
These should now be used in place of app.connection()
to specify
the intent of the required connection.
Note
Two connection pools are available: app.pool
(read), and
app.producer_pool
(write). The latter doesn't actually give connections
but full :class:`kombu.Producer` instances.
def publish_some_message(app, producer=None):
with app.producer_or_acquire(producer) as producer:
...
def consume_messages(app, connection=None):
with app.connection_or_acquire(connection) as connection:
...
The canvas/work-flow implementation have been heavily refactored to fix some long outstanding issues.
Error callbacks can now take real exception and traceback instances (Issue #2538).
>>> add.s(2, 2).on_error(log_error.s()).delay()
Where
log_error
could be defined as:@app.task def log_error(request, exc, traceback): with open(os.path.join('/var/errors', request.id), 'a') as fh: print('--\n\n{0} {1} {2}'.format( task_id, exc, traceback), file=fh)
See :ref:`guide-canvas` for more examples.
Now unrolls groups within groups into a single group (Issue #1509).
chunks/map/starmap tasks now routes based on the target task
chords and chains can now be immutable.
Fixed bug where serialized signatures weren't converted back into signatures (Issue #2078)
Fix contributed by Ross Deane.
Fixed problem where chains and groups didn't work when using JSON serialization (Issue #2076).
Fix contributed by Ross Deane.
Creating a chord no longer results in multiple values for keyword argument 'task_id' (Issue #2225).
Fix contributed by Aneil Mallavarapu.
Fixed issue where the wrong result is returned when a chain contains a chord as the penultimate task.
Fix contributed by Aneil Mallavarapu.
Special case of
group(A.s() | group(B.s() | C.s()))
now works.Chain: Fixed bug with incorrect id set when a subtask is also a chain.
group | group
is now flattened into a single group (Issue #2573).Fixed issue where
group | task
wasn't upgrading correctly to chord (Issue #2922).
The SQS broker transport has been rewritten to use async I/O and as such joins RabbitMQ and Redis as officially supported transports.
The new implementation also takes advantage of long polling, and closes several issues related to using SQS as a broker.
This work was sponsored by Nextdoor.
Contributed by Brian Bouterse.
See :ref:`beat-solar` for more information.
Contributed by Mark Parncutt.
This new API enables you to use signatures when defining periodic tasks, removing the chance of mistyping task names.
An example of the new API is :ref:`here <beat-entries>`.
Connection related errors occuring while sending a task is now re-raised as a :exc:`kombu.exceptions.OperationalError` error:
>>> try:
... add.delay(2, 2)
... except add.OperationalError as exc:
... print('Could not send task %r: %r' % (add, exc))
See :ref:`calling-connection-errors` for more information.
See :ref:`routing-options-rabbitmq-priorities` for more information.
Contributed by Gerald Manipon.
You can now limit the maximum amount of memory allocated per prefork pool child process by setting the worker :option:`--max-memory-per-child <celery worker --max-memory-per-child>` option, or the :setting:`worker_max_memory_per_child` setting.
The limit is for RSS/resident memory size and is specified in kilobytes.
A child process having exceeded the limit will be terminated and replaced with a new process after the currently executing task returns.
See :ref:`worker-max-memory-per-child` for more information.
Contributed by Dave Smith.
Calling result.get()
when using the Redis result backend
used to be extremely expensive as it was using polling to wait
for the result to become available. A default polling
interval of 0.5 seconds didn't help performance, but was
necessary to avoid a spin loop.
The new implementation is using Redis Pub/Sub mechanisms to publish and retrieve results immediately, greatly improving task round-trip times.
Contributed by Yaroslav Zhavoronkov and Ask Solem.
This was an experimental feature introduced in Celery 3.1,
that could only be enabled by adding ?new_join=1
to the
result backend URL configuration.
We feel that the implementation has been tested thoroughly enough to be considered stable and enabled by default.
The new implementation greatly reduces the overhead of chords, and especially with larger chords the performance benefit can be massive.
See :ref:`conf-riak-result-backend` for more information.
Contributed by Gilles Dartiguelongue, Alman One and NoKriK.
See :ref:`conf-couchdb-result-backend` for more information.
Contributed by Nathan Van Gheem.
Add support for Consul as a backend using the Key/Value store of Consul.
Consul has an HTTP API where through you can store keys with their values.
The backend extends KeyValueStoreBackend and implements most of the methods.
Mainly to set, get and remove objects.
This allows Celery to store Task results in the K/V store of Consul.
Consul also allows to set a TTL on keys using the Sessions from Consul. This way the backend supports auto expiry of Task results.
For more information on Consul visit http://consul.io/
The backend uses :pypi:`python-consul` for talking to the HTTP API. This package is fully Python 3 compliant just as this backend is:
$ pip install python-consul
That installs the required package to talk to Consul's HTTP API from Python.
Contributed by Wido den Hollander.
A brand new Cassandra backend utilizing the new :pypi:`cassandra-driver` library is replacing the old result backend using the older :pypi:`pycassa` library.
See :ref:`conf-cassandra-result-backend` for more information.
See :ref:`conf-elasticsearch-result-backend` for more information.
Contributed by Ahmet Demir.
See :ref:`conf-filesystem-result-backend` for more information.
Contributed by Môshe van der Sterre.
Events are now buffered in the worker and sent as a list, reducing the overhead required to send monitoring events.
For authors of custom event monitors there will be no action required as long as you're using the Python Celery helpers (:class:`[email protected]`) to implement your monitor.
However, if you're parsing raw event messages you must now account for batched event messages, as they differ from normal event messages in the following way:
- The routing key for a batch of event messages will be set to
<event-group>.multi
where the only batched event group is currentlytask
(giving a routing key oftask.multi
). - The message body will be a serialized list-of-dictionaries instead of a dictionary. Each item in the list can be regarded as a normal event message body.
Task.replace changed, removes Task.replace_in_chord.
The two methods had almost the same functionality, but the old
Task.replace
would force the new task to inherit the
callbacks/errbacks of the existing task.
If you replace a node in a tree, then you wouldn't expect the new node to inherit the children of the old node, so this seems like unexpected behavior.
So self.replace(sig)
now works for any task, in addition sig
can now
be a group.
Groups are automatically converted to a chord, where the callback will "accumulate" the results of the group tasks.
A new built-in task (celery.accumulate was added for this purpose)
Closes #817
The :program:`celery beat` implementation has been optimized for millions of periodic tasks by using a heap to schedule entries.
Contributed by Ask Solem and Alexander Koshelev.
Writing custom retry handling for exception events is so common that we now have built-in support for it.
For this a new autoretry_for
argument is now supported by
the task decorators, where you can specify a tuple of exceptions
to automatically retry for.
See :ref:`task-autoretry` for more information.
Contributed by Dmitry Malinovsky.
The new :setting:`task_remote_tracebacks` will make task tracebacks more useful by injecting the stack of the remote worker.
This feature requires the additional :pypi:`tblib` library.
Contributed by Ionel Cristian Mărieș.
eventlet/gevent drainers, promises, BLA BLA
Closed issue #2529.
Lots of bugs in the previously experimental RPC result backend have been fixed and we now consider it production ready.
Contributed by Ask Solem, Morris Tweed.
The :setting:`task_routes` setting can now hold functions, and map routes now support glob patterns and regexes.
Instead of using router classes you can now simply define a function:
def route_for_task(name, args, kwargs, options, task=None, **kwargs):
from proj import tasks
if name == tasks.add.name:
return {'queue': 'hipri'}
If you don't need the arguments you can use start arguments, just make sure you always also accept star arguments so that we have the ability to add more features in the future:
def route_for_task(name, *args, **kwargs):
from proj import tasks
if name == tasks.add.name:
return {'queue': 'hipri', 'priority': 9}
Both the options
argument and the new task
keyword argument
are new to the function-style routers, and will make it easier to write
routers based on execution options, or properties of the task.
The optional task
keyword argument won't be set if a task is called
by name using :meth:`@send_task`.
For more examples, including using glob/regexes in routers please see :setting:`task_routes` and :ref:`routing-automatic`.
- Now depends on :ref:`Kombu 4.0 <kombu:version-4.0>`.
- Now depends on :pypi:`billiard` version 3.5.
- No longer depends on :pypi:`anyjson`. Good-bye old friend :(
The "anon-exchange" is now used for simple name-name direct routing.
This increases performance as it completely bypasses the routing table, in addition it also improves reliability for the Redis broker transport.
An empty ResultSet now evaluates to True.
Fix contributed by Colin McIntosh.
New :setting:`task_reject_on_worker_lost` setting, and :attr:`[email protected]_on_worker_lost` task attribute decides what happens when the child worker process executing a late ack task is terminated.
Contributed by Michael Permana.
Task.subtask
renamed toTask.signature
with alias.Task.subtask_from_request
renamed toTask.signature_from_request
with alias.The
delivery_mode
attribute for :class:`kombu.Queue` is now respected (Issue #1953).Routes in :setting:`task-routes` can now specify a :class:`~kombu.Queue` instance directly.
Example:
task_routes = {'proj.tasks.add': {'queue': Queue('add')}}
AsyncResult
now raises :exc:`ValueError` if task_id is None. (Issue #1996).Retried tasks didn't forward expires setting (Issue #3297).
result.get()
now supports anon_message
argument to set a callback to be called for every message received.New abstract classes added:
:class:`~celery.utils.abstract.CallableTask`
Looks like a task.
:class:`~celery.utils.abstract.CallableSignature`
Looks like a task signature.
Task.replace
now properly forwards callbacks (Issue #2722).Fix contributed by Nicolas Unravel.
Task.replace
: Append to chain/chord (Closes #3232)Fixed issue #3232, adding the signature to the chain (if there's any). Fixed the chord suppress if the given signature contains one.
Fix contributed by :github_user:`honux`.
Task retry now also throws in eager mode.
Fix contributed by Feanil Patel.
Fixed crontab infinite loop with invalid date.
When occurrence can never be reached (example, April, 31th), trying to reach the next occurrence would trigger an infinite loop.
Try fixing that by raising a RuntimeError after 2,000 iterations
(Also added a test for crontab leap years in the process)
Fix contributed by Romuald Brunet.
Now ensures the program exits with a non-zero exit code when an exception terminates the service.
Fix contributed by Simon Peeters.
Dates are now always timezone aware even if :setting:`enable_utc` is disabled (Issue #943).
Fix contributed by Omer Katz.
Config: App preconfiguration is now also pickled with the configuration.
Fix contributed by Jeremy Zafran.
- The application can now change how task names are generated using
the :meth:`~@gen_task_name` method.
Contributed by Dmitry Malinovsky.
App has new
app.current_worker_task
property that returns the task that's currently being worked on (or :const:`None`). (Issue #2100).
Eventlet/Gevent: Fixed race condition leading to "simultaneous read" errors (Issue #2812).
Prefork: Prefork pool now uses
poll
instead ofselect
where available (Issue #2373).Prefork: Fixed bug where the pool would refuse to shut down the worker (Issue #2606).
Eventlet: Now returns pool size in :program:`celery inspect stats` command.
Contributed by Alexander Oblovatniy.
- Redis Transport: The Redis transport now supports the :setting:`broker_use_ssl` option.
:program:`celery multi`:
%n
format for is now synonym with%N
to be consistent with :program:`celery worker`.:program:`celery inspect`/:program:`celery control`: now supports a new :option:`--json <celery inspect --json>` option to give output in json format.
:program:`celery inspect registered`: now ignores built-in tasks.
:program:`celery purge` now takes
-Q
and-X
options used to specify what queues to include and exclude from the purge.New :program:`celery logtool`: Utility for filtering and parsing celery worker log-files
:program:`celery multi`: now passes through %i and %I log file formats.
General:
%p
can now be used to expand to the full worker node-name in log-file/pid-file arguments.- A new command line option
:option:`--executable <celery worker --executable>` is now available for daemonizing programs (:program:`celery worker` and :program:`celery beat`).
Contributed by Bert Vanderbauwhede.
:program:`celery worker`: supports new :option:`--prefetch-multiplier <celery worker --prefetch-multiplier>` option.
Contributed by Mickaël Penhard.
Improvements and fixes for :class:`~celery.utils.collections.LimitedSet`.
Getting rid of leaking memory + adding
minlen
size of the set: the minimal residual size of the set after operating for some time.minlen
items are kept, even if they should've been expired.Problems with older and even more old code:
- Heap would tend to grow in some scenarios (like adding an item multiple times).
- Adding many items fast wouldn't clean them soon enough (if ever).
- When talking to other workers, revoked._data was sent, but it was processed on the other side as iterable. That means giving those keys new (current) time-stamp. By doing this workers could recycle items forever. Combined with 1) and 2), this means that in large set of workers, you're getting out of memory soon.
All those problems should be fixed now.
This should fix issues #3095, #3086.
Contributed by David Pravec.
Worker now only starts the remote control command consumer if the broker transport used actually supports them.
Gossip now sets
x-message-ttl
for event queue to heartbeat_interval s. (Issue #2005).Now preserves exit code (Issue #2024).
Fixed crash when the
-purge
argument was used.Log--level for unrecoverable errors changed from
error
tocritical
.Improved rate limiting accuracy.
Account for missing timezone information in task expires field.
Fix contributed by Albert Wang.
- The worker no longer has a
Queues
bootsteps, as it is now superfluous.
- The worker no longer has a
Now emits the "Received task" line even for revoked tasks. (Issue #3155).
Now respects :setting:`broker_connection_retry` setting.
Fix contributed by Nat Williams.
New :data:`celery.worker.state.requests` enables O(1) loookup of active/reserved tasks by id.
Auto-scale didn't always update keep-alive when scaling down.
Fix contributed by Philip Garnero.
Fixed typo
options_list
->option_list
.Fix contributed by Greg Wilbur.
:mod:`celery.contrib.rdb`: Changed remote debugger banner so that you can copy and paste the address easily (no longer has a period in the address).
Contributed by Jonathan Vanasco.
Fixed compatibility with recent :pypi:`psutil` versions (Issue #3262).
App: New signals for app configuration/finalization:
Task: New task signals for rejected task messages:
Worker: New signal for when a heartbeat event is sent.
:data:`celery.signals.heartbeat_sent`
Contributed by Kevin Richardson.
Event messages now uses the RabbitMQ
x-message-ttl
option to ensure older event messages are discarded.The default is 5 seconds, but can be changed using the :setting:`event_queue_ttl` setting.
Event monitors now sets the :setting:`event_queue_expires` setting by default.
The queues will now expire after 60 seconds after the monitor stops consuming from it.
Fixed a bug where a None value wasn't handled properly.
Fix contributed by Dongweiming.
New :setting:`event_queue_prefix` setting can now be used to change the default
celeryev
queue prefix for event receiver queues.Contributed by Takeshi Kanemoto.
State.tasks_by_type
andState.tasks_by_worker
can now be used as a mapping for fast access to this information.
chunks
/map
/starmap
are now routed based on the target task.Signature.link
now works when argument is scalar (not a list)- (Issue #2019).
Generic init-scripts now support :envvar:`CELERY_SU` and :envvar:`CELERYD_SU_ARGS` environment variables to set the path and arguments for :command:`su` (:manpage:`su(1)`).
Generic init-scripts now better support FreBSD and other BSD systems by searching :file:`/usr/local/etc/` for the configuration file.
Contributed by Taha Jahangir.
Generic init-script: Fixed strange bug for
celerybeat
where restart didn't always work (Issue #3018).The systemd init script now uses a shell when executing services.
Contributed by Tomas Machalek.
Redis: Now has a default socket timeout of 5 seconds.
The default can be changed using the new :setting:`redis_socket_timeout` setting.
Contributed by Raghuram Srinivasan.
RPC Backend result queues are now auto delete by default (Issue #2001).
RPC Backend: Fixed problem where exception wasn't deserialized properly with the json serializer (Issue #2518).
Fix contributed by Allard Hoeve.
CouchDB: Fixed typo causing the backend to not be found (Issue #3287).
Fix contributed by Andrew Stewart.
MongoDB: Now supports setting the :setting:`result_serialzier` setting to
bson
to use the MongoDB libraries own serializer.Contributed by Davide Quarta.
- MongoDB: URI handling has been improved to use
database name, user and password from the URI if provided.
Contributed by Samuel Jaillet.
SQLAlchemy result backend: Now ignores all result engine options when using NullPool (Issue #1930).
SQLAlchemy result backend: Now sets max char size to 155 to deal with brain damaged MySQL unicode implementation (Issue #1748).
General: All Celery exceptions/warnings now inherit from common :class:`~celery.exceptions.CeleryError`/:class:`~celery.exceptions.CeleryWarning`. (Issue #2643).
Contributed by:
- Adam Chainz
- Arthur Vuillard
- Batiste Bieler
- Daniel Devine
- Edward Betts
- Jason Veatch
- Jeff Widman
- Manuel Kaufmann
- Maxime Beauchemin
- Mitchel Humpherys
- Rik
- Tayfun Sen
Prefork: Calling
result.get()
or joining any result from within a task now raises :exc:`RuntimeError`.In previous versions this would emit a warning.
:mod:`celery.worker.consumer` is now a package, not a module.
Module
celery.worker.job
renamed to :mod:`celery.worker.request`.Beat:
Scheduler.Publisher
/.publisher
renamed to.Producer
/.producer
.Result: The task_name argument/attribute of :class:`@AsyncResult` was removed.
This was historically a field used for :mod:`pickle` compatibility, but is no longer needed.
Backends: Arguments named
status
renamed tostate
.Backends:
backend.get_status()
renamed tobackend.get_state()
.
The experimental :mod:`celery.contrib.methods` feature has been removed, as there were far many bugs in the implementation to be useful.
The CentOS init-scripts have been removed.
These didn't really add any features over the generic init-scripts, so you're encouraged to use them instead, or something like :pypi:`supervisor`.
These symbols have been renamed, and while there's an alias available in this version for backward compatibility, they will be removed in Celery 5.0, so make sure you rename these ASAP to make sure it won't break for that release.
Chances are that you'll only use the first in this list, but you never know:
celery.utils.worker_direct
-> :meth:`celery.utils.nodenames.worker_direct`.celery.utils.nodename
-> :meth:`celery.utils.nodenames.nodename`.celery.utils.anon_nodename
-> :meth:`celery.utils.nodenames.anon_nodename`.celery.utils.nodesplit
-> :meth:`celery.utils.nodenames.nodesplit`.celery.utils.default_nodename
-> :meth:`celery.utils.nodenames.default_nodename`.celery.utils.node_format
-> :meth:`celery.utils.nodenames.node_format`.celery.utils.host_format
-> :meth:`celery.utils.nodenames.host_format`.
Module
celery.worker.job
has been renamed to :mod:`celery.worker.request`.This was an internal module so shouldn't have any effect. It's now part of the public API so must not change again.
Module
celery.task.trace
has been renamed tocelery.app.trace
as thecelery.task
package is being phased out. The module will be removed in version 5.0 so please change any import from:from celery.task.trace import X
to:
from celery.app.trace import X
Old compatibility aliases in the :mod:`celery.loaders` module has been removed.
- Removed
celery.loaders.current_loader()
, use:current_app.loader
- Removed
celery.loaders.load_settings()
, use:current_app.conf
- Removed
AsyncResult.serializable()
andcelery.result.from_serializable
has been removed:
Use instead:
>>> tup = result.as_tuple() >>> from celery.result import result_from_tuple >>> result = result_from_tuple(tup)
Removed
BaseAsyncResult
, useAsyncResult
for instance checks instead.Removed
TaskSetResult
, useGroupResult
instead.TaskSetResult.total
->len(GroupResult)
TaskSetResult.taskset_id
->GroupResult.id
Removed
ResultSet.subtasks
, useResultSet.results
instead.
TaskSet has been renamed to group and TaskSet will be removed in version 4.0.
Old:
>>> from celery.task import TaskSet >>> TaskSet(add.subtask((i, i)) for i in xrange(10)).apply_async()
New:
>>> from celery import group >>> group(add.s(i, i) for i in xrange(10))()
Removals for class :class:`celery.events.state.Worker`:
Worker._defaults
attribute.Use
{k: getattr(worker, k) for k in worker._fields}
.Worker.update_heartbeat
Use
Worker.event(None, timestamp, received)
Worker.on_online
Use
Worker.event('online', timestamp, received, fields)
Worker.on_offline
Use
Worker.event('offline', timestamp, received, fields)
Worker.on_heartbeat
Use
Worker.event('heartbeat', timestamp, received, fields)
Removals for class :class:`celery.events.state.Task`:
Task._defaults
attribute.Use
{k: getattr(task, k) for k in task._fields}
.Task.on_sent
Use
Worker.event('sent', timestamp, received, fields)
Task.on_received
Use
Task.event('received', timestamp, received, fields)
Task.on_started
Use
Task.event('started', timestamp, received, fields)
Task.on_failed
Use
Task.event('failed', timestamp, received, fields)
Task.on_retried
Use
Task.event('retried', timestamp, received, fields)
Task.on_succeeded
Use
Task.event('succeeded', timestamp, received, fields)
Task.on_revoked
Use
Task.event('revoked', timestamp, received, fields)
Task.on_unknown_event
Use
Task.event(short_type, timestamp, received, fields)
Task.update
Use
Task.event(short_type, timestamp, received, fields)
Task.merge
Contact us if you need this.
Support for the very old magic keyword arguments accepted by tasks is finally removed in this version.
If you're still using these you have to rewrite any task still
using the old celery.decorators
module and depending
on keyword arguments being passed to the task,
for example:
from celery.decorators import task @task() def add(x, y, task_id=None): print('My task id is %r' % (task_id,))
should be rewritten into:
from celery import task @task(bind=True) def add(self, x, y): print('My task id is {0.request.id}'.format(self))
The following settings have been removed, and is no longer supported:
Setting name | Replace with |
---|---|
CELERYD_LOG_LEVEL |
:option:`celery worker --loglevel` |
CELERYD_LOG_FILE |
:option:`celery worker --logfile` |
CELERYBEAT_LOG_LEVEL |
:option:`celery beat --loglevel` |
CELERYBEAT_LOG_FILE |
:option:`celery beat --loglevel` |
CELERYMON_LOG_LEVEL |
celerymon is deprecated, use flower |
CELERYMON_LOG_FILE |
celerymon is deprecated, use flower |
CELERYMON_LOG_FORMAT |
celerymon is deprecated, use flower |
Setting name | Replace with |
---|---|
CELERY_CHORD_PROPAGATES |
N/A |
Module
celery.datastructures
renamed to :mod:`celery.utils.collections`.Module
celery.utils.timeutils
renamed to :mod:`celery.utils.time`.celery.utils.datastructures.DependencyGraph
moved to :mod:`celery.utils.graph`.celery.utils.jsonify
is now :func:`celery.utils.serialization.jsonify`.celery.utils.strtobool
is now :func:`celery.utils.serialization.strtobool`.celery.utils.is_iterable
has been removed.Instead use:
isinstance(x, collections.Iterable)
celery.utils.lpmerge
is now :func:`celery.utils.collections.lpmerge`.celery.utils.cry
is now :func:`celery.utils.debug.cry`.celery.utils.isatty
is now :func:`celery.platforms.isatty`.celery.utils.gen_task_name
is now :func:`celery.utils.imports.gen_task_name`.celery.utils.deprecated
is now :func:`celery.utils.deprecated.Callable`celery.utils.deprecated_property
is now :func:`celery.utils.deprecated.Property`.celery.utils.warn_deprecated
is now :func:`celery.utils.deprecated.warn`
See the :ref:`deprecation-timeline`.