Plumber Relay
Relaying data using plumber is the most reliable and performant way to get data into Streamdal.
You can launch plumber
relays in multiple ways:
-
Running
plumber
in single-relay mode via CLI- Best for quick, one-offs
-
Running
plumber
as a docker container- Best for ephemeral workloads
-
Running
plumber
in server mode- Best for production
The following examples show how to run plumber
in single relay mode.
For production deployments, we suggest to deploy plumber
running in server mode.
plumber relay kafka \
--address "your-kafka-address.com:9092" \
--token YOUR-COLLECTION-TOKEN-HERE \
--topics orders \
--tls-skip-verify
In this example, all messages from kafka topic new_orders will be automatically sent to the collection with the specified relay token.
docker run --name plumber-rabbit -p 8080:8080 \
-e PLUMBER_RELAY_TYPE=rabbit \
-e PLUMBER_RELAY_TOKEN=$YOUR-BATCHSH-TOKEN-HERE \
-e PLUMBER_RELAY_RABBIT_EXCHANGE=my_exchange \
-e PLUMBER_RELAY_RABBIT_QUEUE=my_queue \
-e PLUMBER_RELAY_RABBIT_ROUTING_KEY=some.routing.key \
-e PLUMBER_RELAY_RABBIT_QUEUE_EXCLUSIVE=false \
-e PLUMBER_RELAY_RABBIT_QUEUE_DURABLE=true \
streamdal/plumber \
rabbit
In this example, all messages sent to my_exchange
that match the routing key some.routing.key
will be sent to my_queue
.
At that point, plumber will pick up the messages and send them to Streamdal using the specified token.
A full suite of environment variables are provided in ENV.md for configuring plumber’s relay mode.
Example of running plumber via kubernetes
apiVersion: apps/v1
kind: Deployment
metadata:
name: plumber-deployment
spec:
selector:
matchLabels:
app: plumber
replicas: 1
template:
metadata:
labels:
app: plumber
spec:
containers:
- name: plumber
image: streamdal/plumber:latest
command: ["/plumber-linux", "relay", "kafka"]
args: ["--stats-enable"]
ports:
- containerPort: 9191
env:
- name: PLUMBER_RELAY_TOKEN
value: "--- COLLECTION TOKEN HERE ---"
- name: PLUMBER_RELAY_KAFKA_ADDRESS
value: "kafka.server.com:9092"
- name: PLUMBER_RELAY_KAFKA_TOPIC
value: "new-orders"
- name: PLUMBER_RELAY_KAFKA_GROUP_ID
value: "plumber"
resources:
requests:
memory: "256Mi"
cpu: "250m"
limits:
memory: "512Mi"
cpu: "500m"
More examples of relaying from various systems can be found in EXAMPLES.md.
When should you use this API?
Plumber is the easiest way to relay throughput heavy workloads and should be used by anyone wanting to get up and running quickly.
Throughput
plumber
uses gRPC under the hood to communicate with Streamdal’s collectors.
You should be able to comfortably reach 25K-50K messages/sec on a single plumber
instance. To reach higher levels, you should run plumber
in cluster server mode and launch 2+ replicas of plumber
.
Make sure to use the same consumer group if relaying for backends such as Kafka or NATS.