Benthos is a generic and high performance streaming service, able to connect various sources and sinks and perform arbitrary actions, transformations and filters on payloads. It is ready to drop into your pipeline either as a static binary or a docker image.
A Benthos instance (stream) consists of four components; inputs,
optional buffer, processor workers and
outputs. Inputs and outputs can be combined in a range of broker
patterns. It is possible to run multiple isolated streams within a single
Benthos instance using --streams
mode.
Benthos is crash resilient by default. When connecting to at-least-once sources and sinks without a buffer it guarantees at-least-once delivery without needing to persist messages during transit.
When running a Benthos stream with a buffer there are various options for choosing a level of resiliency that meets your needs.
- Amazon (S3, SQS)
- Elasticsearch
- File
- HTTP(S)
- Kafka
- MQTT
- Nanomsg
- NATS
- NATS Streaming
- NSQ
- RabbitMQ (AMQP 0.91)
- Redis
- Stdin/Stdout
- Websocket
- ZMQ4
Documentation for Benthos components, concepts and recommendations can be found in the docs directory.
For some applied examples of Benthos such as streaming and deduplicating the Twitter firehose to Kafka check out the cookbook section.
benthos -c ./config.yaml
Or, with docker:
# Send HTTP /POST data to Kafka:
docker run --rm \
-e "BENTHOS_INPUT=http_server" \
-e "BENTHOS_OUTPUT=kafka" \
-e "KAFKA_OUTPUT_BROKER_ADDRESSES=kafka-server:9092" \
-e "KAFKA_OUTPUT_TOPIC=benthos_topic" \
-p 4195:4195 \
jeffail/benthos
# Using your own config file:
docker run --rm -v /path/to/your/config.yaml:/benthos.yaml jeffail/benthos
The configuration file for a Benthos stream is made up of four main sections; input, buffer, pipeline, output. If we were to pipe stdin directly to Kafka it would look like this:
input:
type: stdin
buffer:
type: none
pipeline:
threads: 1
processors: []
output:
type: kafka
kafka:
addresses:
- localhost:9092
topic: benthos_stream
There are example configs demonstrating each input, output, buffer and processor option which can be found here.
You can print a configuration file containing fields for all types with the following command:
benthos --print-yaml --all > config.yaml
benthos --print-json --all | jq '.' > config.json
There are also sections for setting logging, metrics and HTTP server options.
It is possible to select fields inside a configuration file to be set via environment variables. The docker image, for example, is built with a config file where all common fields can be set this way.
Build with Go:
go get github.com/Jeffail/benthos/cmd/benthos
Or, pull the docker image:
docker pull jeffail/benthos
Or, grab a binary for your OS from here.
There's a multi-stage Dockerfile
for creating a Benthos docker image which
results in a minimal image from scratch. You can build it with:
make docker
Then use the image:
docker run --rm \
-v /path/to/your/benthos.yaml:/config.yaml \
-v /tmp/data:/data \
-p 4195:4195 \
benthos -c /config.yaml
There are a few examples here that show you some ways of
setting up Benthos containers using docker-compose
.
Benthos supports ZMQ4 for both data input and output. To add this you need to install libzmq4 and use the compile time flag when building Benthos:
go install -tags "ZMQ4" ./cmd/...