Before reading about queuing webhook processing using Redis, you may wish to read the documentation regarding horizontal scaling.

When combining queuing and horizontal scalability, it's highly recommended that you use a third-party driver like Redis. Redis helps ensure that once a webhook is triggered it will be completely processed because the message to send the webhook will remain in-memory within Redis. Therefore, even if the soketi server goes down, the webhook will still be sent.

Each webhook message is processed by a worker. In addition, each worker can spawn multiple queue listeners. In soketi's case, each worker represents one of the listed events within the app webhooks documentation. This way, soketi ensures that webhooks are being processed quickly and efficiently. This behavior may be subject to change in the future. For example, queues for each app might eventually be needed to ensure high-performance message processing in all situations.

To decouple the queue processors from the active WS/HTTP server, consider setting MODE=worker and run a separate fleet for your workers.

In case you want to scale your queue workers with Prometheus, the best solution is to use **** bull_exporter

Redis Cluster mode may be broken in some cases. Read more about BullMQ Redis Cluster configurations.

Environment Variables

NameDefaultPossible valuesDescription



Any integer

The number of webhook messages that can be processed in parallel for each event.



false, true

Whether the client should be initialized for Redis Cluster. You have to specify the DB_REDIS_CLUSTER_NODES value.

Last updated