This is an alpha library in active development, so the API may change.
Automatic retry functionality for ActiveJob. Just include ActiveJob::Retry
in your job
class and call one of constant_retry
, variable_retry
, exponential_retry
or retry_with
to define your
retry strategy:
class ProcessWebhook < ActiveJob::Base
include ActiveJob::Retry
queue_as :webhooks
# Constant delay between attempts:
constant_retry limit: 3, delay: 5.minutes, retryable_exceptions: [TimeoutError, NetworkError]
# Or, variable delay between attempts:
variable_retry delays: [1.minute, 5.minutes, 10.minutes, 30.minutes]
# Or, exponential delay between attempts:
exponential_retry limit: 25
# You can also use a custom backoff strategy by passing an object which responds to
# `should_retry?(attempt, exception)`, and `retry_delay(attempt, exception)`
# to `retry_with`:
module ChaoticBackoffStrategy
def self.should_retry?(retry_attempt, exception)
[true, true, true, true, false].sample
end
def self.retry_delay(retry_attempt, exception)
(0..10).to_a.sample
end
end
retry_with ChaoticBackoffStrategy
def perform(webhook)
webhook.process!
end
end
The retry will get executed before any rescue_from
blocks, which will only get executed
if the exception is not going to be retried, or has failed the final retry.
Option | Default | Description |
---|---|---|
limit |
1 |
Maximum number of times to attempt the job (default: 1). |
unlimited_retries |
false |
If set to true , this job will be repeated indefinitely until in succeeds. Use with extreme caution. |
delay |
0 |
Time between attempts in seconds (default: 0). |
retryable_exceptions |
nil |
A whitelist of exceptions to retry. When nil , all exceptions will result in a retry. |
fatal_exceptions |
[] |
A blacklist of exceptions to not retry (default: []). |
Option | Default | Description |
---|---|---|
limit |
1 |
Maximum number of times to attempt the job (default: 1). |
unlimited_retries |
false |
If set to true , this job will be repeated indefinitely until in succeeds. Use with extreme caution. |
retryable_exceptions |
nil |
Same as for constant_retry. |
fatal_exceptions |
[] |
Same as for constant_retry. |
Option | Default | Description |
---|---|---|
delays |
required An array of delays between attempts in seconds. The first attempt will occur whenever you originally enqueued the job to happen. | |
min_delay_multiplier |
If supplied, each delay will be multiplied by a random number between this and max_delay_multiplier . |
|
max_delay_multiplier |
The other end of the range for min_delay_multiplier . If one is supplied, both must be. |
|
retryable_exceptions |
nil |
Same as for constant_retry. |
fatal_exceptions |
[] |
Same as for constant_retry. |
Any queue adapter which supports delayed enqueuing (i.e. the enqueue_at
method) will work with ActiveJob::Retry, however some queue backends have
automatic retry logic, which should be disabled. The cleanest way to do this is
to use a rescue_from
in the jobs for which you're using ActiveJob::Retry, so
the queue backend never perceives the job as having failed. E.g.:
class MyJob < ActiveJob::Base
include ActiveJob::Retry
queue_as :some_job
constant_retry limit: 3, delay: 5, retryable_exceptions: [TimeoutError, NetworkError]
rescue_from(StandardError) { |error| MyErrorService.record(error) }
def perform
raise "Weird!"
end
end
Since rescue_from
s are only executed once all retries have been attempted,
this will send the recurring error to your error service (e.g. Airbrake,
Sentry), but will make it appear to the queue backend (e.g. Que, Sidekiq) as if
the job has succeeded.
An alternative is to alter the appropriate JobWrapper
to alter the
configuration of the backend to disable retries globally. For Sidekiq this
would be:
# config/initializers/disable_sidekiq_retries.rb
ActiveJob::QueueAdapters::SidekiqAdapter::JobWrapper.sidekiq_options(retry: false)
This has the advantages of moving failed jobs to the Dead Job Queue instead of
just executing the logic in the rescue_from
, which makes manual re-enqueueing
easier. On the other hand it does disable Sidekiq's automatic retrying for all
ActiveJob jobs.
Contributions are very welcome! Please open a PR or issue on this repo.