prwnr / laravel-streamer
Events streaming package for Laravel that uses Redis 5 streams
Installs: 118 976
Dependents: 1
Suggesters: 0
Security: 0
Stars: 97
Watchers: 2
Forks: 12
Open Issues: 2
Requires
- php: ^8.1|^8.2
- ext-json: *
- ext-redis: *
- illuminate/console: ^10.0|^11.0
- illuminate/container: ^10.0|^11.0
- illuminate/redis: ^10.0|^11.0
- illuminate/support: ^10.0|^11.0
Requires (Dev)
- laravel/pint: ^1.2
- orchestra/testbench: ^8.0|^9.0
- phpunit/phpunit: ^10.0
- predis/predis: ^1.1
- rector/rector: ^1.0
- dev-master
- 4.3.0
- 4.2.0
- 4.1.0
- 4.0.0
- 3.5.1
- 3.5.0
- 3.4.1
- 3.4.0
- 3.4.0-beta
- 3.3.0
- 3.2.0
- 3.1.0
- 3.0.0
- 2.x-dev
- 2.7.1
- 2.7.0
- 2.6.1
- 2.6.0
- 2.5.1
- 2.5.0
- 2.4.0
- 2.3.1
- 2.3.0
- 2.2.0
- 2.1.0
- 2.0.1
- 2.0.0
- 1.6.3
- 1.6.2
- 1.6.1
- 1.6.0
- 1.5.1
- 1.5.0
- 1.4.0
- 1.3.1
- 1.3.0
- 1.2.2
- 1.2.1
- 1.2.0
- 1.1.0
- 1.0.x-dev
- 1.0.1
- 1.0.0
- dev-predis-back
- dev-update-php8.1
- dev-xread-bool-return-fix
- dev-minor-fixes
- dev-reports
- dev-php74
This package is auto-updated.
Last update: 2024-12-09 13:31:51 UTC
README
Streamer is a Laravel package for events functionality between different applications, powered by Redis Streams. This package utilizes all main commands of Redis 5.0 Streams providing a simple usage of Streams as Events.
Main concept of this package is to provide easy way of emitting new events from your application and to allow listening to them in your other applications that are using same Redis server.
Installation
- Install package via composer command
composer require prwnr/laravel-streamer
or by adding it to your composer.json file with version. - Discover the package
- Publish configuration with
vendor:publish
command. - Make sure that you have running Redis 5.0 instance and that Laravel is configured to use it
- Make sure that you have PHPRedis extension installed.
Usage
There are two main ends of this package usage - emiting new event and listening to events. Whereas emiting requires a bit more work to get it used, such as creating own Event classes, then listening to events is available with artisan command and is working without much work needed.
Version Compatibility
Emiting new events
In order to emit new event few things needs to be done.
First of all, you will need to have a valid class that implements Prwnr\Streamer\Contracts\Event
like this:
class ExampleStreamerEvent implements Prwnr\Streamer\Contracts\Event { /** * Require name method, must return a string. * Event name can be anything, but remember that it will be used for listening */ public function name(): string { return 'example.streamer.event'; } /** * Required type method, must return a string. * Type can be any string or one of predefined types from Event */ public function type(): string { return Event::TYPE_EVENT; } /** * Required payload method, must return array * This array will be your message data content */ public function payload(): array { return ['message' => 'content']; } }
Then, at any point in your application all you need to do is to emit that event by using either Streamer
instance or Streamer
facade.
$event = new ExampleStreamerEvent(); $id = \Prwnr\Streamer\Facades\Streamer::emit($event);
This will create a message on a stream named (if such does not exist): example.streamer.event
. Emit method will return
an ID if emitting your event ended up with success.
Listening for new messages on events
In order to listen to event you will have to properly configure config/streamer.php
(or use ListenerStack::add
method) and run php artisan streamer:listen
command.
At config file you will find Application listeners configuration with default values for it, that should be changed if you want to start listening with streamer listen command.
Other way to add listeners for events is to use ListenersStack
static class. This class is being booted with listeners from configuration file and is then used by
command to get all of them. So, the addition of this class is that it allows adding listeners not only via configuration file, but also programmatically.
Remember that local listener should implement MessageReceiver
contract to ensure that it has handle
method
which accepts ReceivedMessage
as an argument.
/* |-------------------------------------------------------------------------- | Application listeners |-------------------------------------------------------------------------- | | Listeners classes that should be invoked with Streamer listen command | based on streamer.event.name => [local_handlers] pairs | | Local listeners should implement MessageReceiver contract | */ 'listen_and_fire' => [ 'example.streamer.event' => [ //List of local listeners that should be invoked //\App\Listeners\ExampleListener::class ], ],
Above configuration is an array of Streamer events with an array of local listeners related to Streamer event. When
listening to example.streamer.event
all local listeners from its config definition will be created and their handle
method fired with message received from Stream. For listener instance creation this package uses Laravel Container,
therefore you can type hint anything into your listener constructor to use Laravel dependency injection.
To start listening for an event, use listen command.
Commands
Listen
streamer:listen example.streamer.event
This command will start listening on a given stream (or streams separated by comma) starting from "now". It will be listening in a blocking way, meaning that it will run until Redis will time out or crash. All listener related errors are being caught and logged into console as well as stored in Failed Messages list for later debugging and/or retrying.
That's a basic usage of this command, where event name is a required argument (unless --all
option is provided).
So in this case it simply starts listening for only new events.
This command however has few options that are extending its usage, those are:
--all= : Will trigger listener mode to start listening on all events that are registered with local listeners classes (from the ListenersStack). Event name argument is no longer required in this case.
--group= : Name of your streaming group. Only when group is provided listener will listen on group as consumer
--consumer= : Name of your group consumer. If not provided a name will be created as groupname-timestamp
--reclaim= : Milliseconds of pending messages idle time, that should be reclaimed for current consumer in this group. Can be only used with group listening
--last_id= : ID from which listener should start reading messages (using 0-0 will process all old messages)
--keep-alive : Will keep listener alive when any unexpected non-listener related error will occur by simply restarting listening
--max-attempts= : Number of maximum attempts to restart a listener on an unexpected non-listener related error (requires --keep-alive to be used)
--purge : Will remove message from the stream if it will be processed successfully by all local listeners in the current stack.
--archive : Will remove message from the stream and store it in database if it will be processed successfully by all local listeners in the current stack.
When consumer
and group
options are being in use, every message on a stream will be marked as acknowledged for the
given consumer, thus it will not be processed by consequent
streamer:listen
command call with the same options. Note that listening from a specific ID without consumer and group
being set will ignore acknowledgments.
The purge
and archive
options (available since v2.6) are designed to be used to release memory or storage of the
Redis instance
(in a cases when there are tons of streamed messages or the payloads are big and Redis runs out of memory/storage). When
using those options, keep in mind, that they are not going to take into account listeners running in other instances or
other servers - meaning, that when first listener hooked to specific event will process its messages, the purge
and
archive
options will delete the message not waiting for other listeners to finish. To fully use archive
option
see [Stream Archive][#stream-archive] for more details and instructions.
Using multiple events in argument or the --all
option with any other option (like group, consumer, last_id)
will apply those options to every stream event that is being in use.
Failed List
streamer:failed:list
This command will show list of stream messages that failed to be handled by their listeners. It will yield all the important information about them like: ID, stream name, listener class, error message that cause it to fail, and a date when that happened.
Table example:
+-----+-----------+---------------------------+-------+---------------------+
| ID | Stream | Receiver | Error | Date |
+-----+-----------+---------------------------+-------+---------------------+
| 123 | foo.bar | Tests\Stubs\LocalListener | error | 2021-12-12 12:12:12 |
| 321 | other.bar | Tests\Stubs\LocalListener | error | 2021-12-12 12:15:12 |
+-----+-----------+---------------------------+-------+---------------------+
There's one addition option for this command, called --compact
which will limit the table output to only ID, Stream
and Error columns.
Failed Retry
streamer:failed:retry
This command is meant to try again failed listening. It simply reads the message from a stream and attempts to handle it again by the listener that it was originally processed.
When the listener fails to process the message again, the message failed information will be re-stored (with a newer date and updated error message) and will be available to be retried again. There's no limit to how many times message can be processed. It will remain available after each fail unless flush command will be used.
This command has few options that are available:
--all : retries all existing failed messages
--id= : retries only those messages that are matching given ID
--stream= : retries only those messages that are matching given stream name
--receiver= : retries only those messages that are matching given listener full class name (may require to be in quotation)
At least one of those options is required to be used with the command to process failed messages. The all
option can
be only used solely, while the other three options can be used together or not. This means, that any combination of id
, stream
and receiver
can be used to match any number of failed messages and retry them. So, for example a stream
can be used together with id
or in other case id
can be used with receiver
, or only one of them can be used, or
all three at once, its all up to the use case.
Failed Flush
streamer:failed:flush
This command will remove all existing failed messages from the messages' repository. Can be used to prune entries that cannot be processed at all by listeners.
This command WILL NOT remove the message from the Stream itself - the message will remain there untouched, but acknowledged by its original consumer (if used).
List
streamer:list
This command will list all registered events, and their associated listeners. The option --compact
will yield only a
list of the events, skipping listeners column.
This command may be useful to see what events are being actually handled by a listener, what can help to find out what's missing. This list can be also used to start listening to available events by 3rd party app.
Table example:
+------------------------+------------------------------------+
| Event | Listeners |
+------------------------+------------------------------------+
| example.streamer.event | none |
| foo.bar | Tests\Stubs\LocalListener |
| other.foo.bar | Tests\Stubs\LocalListener |
| | Tests\Stubs\AnotherLocalListener |
+------------------------+------------------------------------+
Archive
streamer:archive
This command will archive messages from a selected streams older than days/weeks or so. It will process all stream
messages, verifying their created
timestamp and will attempt to archive (deleting them from redis and attempting to
store them in associated archive storage)
each one of them.
This command has two required options:
--streams : list of streams separated by comma to archive messages from
--older_than= : information how old messages should be to archive them. The suggested format is: 60 min, 1 day, 1 week, 5 days, 2 weeks etc.
Be aware of using this command, as it will not take into account whether listeners processed messages it tries to archive or not. This should be used with caution and only for older messages, so that it will be more certain, that all listeners processed their messages.
Purge
streamer:purge
This command will purge messages from a selected streams older than days/weeks or so. It will process all stream
messages, verifying their created
timestamp and will attempt to purge them (deleting them from the redis entirely).
This command has two required options:
--streams : list of streams separated by comma to purge messages from
--older_than= : information how old messages should be to purge them. The suggested format is: 60 min, 1 day, 1 week, 5 days, 2 weeks etc.
Be aware of using this command, as it will not take into account whether listeners processed messages it tries to purge or not. This should be used with caution and only for older messages, so that it will be more certain, that all listeners processed their messages.
Archive Restore
streamer:archive:restore
This command will restore archived stream messages from the associated archive storage. It will essentially fetch messages (all or selection) and will try to put them back onto the stream, while also deleting them from the archive. This action will trigger listeners that are hooked to the restored streams!
This command has few options that are available:
--all : restores all archived messages back to the stream.
--id= : restores archived message back to the stream by ID. Requires --stream option to be used as well.
--stream= : restores all archived messages from a selected stream.
At least one of those options is required to attempt restoration of the messages. If any error occurs while restoring a message, it will be reported for that particular attempt not preventing other message from being processed.
Restoring message puts it back onto a stream with NEW ID - this is Redis requirement and limitation, that any message
added to stream, needs to have ID higher than the last generated one. The original ID of the message that is being
restored will be stored in meta information in original_id
field.
Stream Archive
Stream Archive allows storing processed message in any kind of storage, to free up Redis memory and/or space since 2.6 version. Archive allows restoring those messages, releasing them back onto the stream.
To fully use archive storage, a new storage driver needs to be written and added to manager. To do so, few quick steps needs to be finished:
- define your storage driver class and make it implement
\Prwnr\Streamer\Contracts\ArchiveStorage
contract. This interface is mandatory. - extend the manager with your driver like this:
$manager = $this->app->make(StorageManager::class); $manager->extend('your_driver_name', static function () { return new YourDriver(); });
- define your driver as default in streamer config file
'archive' => [ 'storage_driver' => 'your_driver_name' ]
Replaying Events
Since 2.2 version, Stream events can be "replayed". This means, that the specific message (with a unique identifier) can be "reconstructed" until "now" (or until a selected date).
What "replaying" messages really means? It means, that all the messages that are in the stream, will be read from the very beginning, and payload of each single entry will be "combined" into a final version of the message - each filed will be replaced with its "newer" value, if such exists in the history.
This is going to be useful with events that don't hold all the information about the resource they may represent, but have only data about fields that changed.
So, for example having a resource with fields name
and surname
,
we will emit 3 different events:
- first for its creation populating both fields with values (
name: foo; surname: bar
) - second event that will change only
name
intofoo bar
- third event that changes name again to
bar foo
.
While replaying this set of messages (remember that each one has the same unique identifier)
our final replayed resource will be: name: bar foo; surname: bar
. If we replay the event until the time before third
change, we would have name: foo bar; surname: bar
Usage
To make Event replayable, it needs to implement the Prwnr\Streamer\Contracts\Replayable
contract.
This will enforce adding a getIdentifier
method, that should return unique identifier for the resource
(like UUID of the resource that this event represents). With this contract being fulfilled, all events that will go
through Streamer
emit method, will be also "marked" as available to be replayed.
To actually replay messages, Hsitory
interface implementation needs to be used.
Method that should be used is: replay(string $event, string $identifier, Carbon $until = null): array
.
This method will return the "current state" of the event, rebuilding it from its history. As seen in method definition,
it asks for event string name and resource identifier (that was applied by Replayable
contract). Third parameter is
optional and if used, it will stop replaying messages when first message with matching date will be encountered.
Eloquent Model Events
With use of a EmitsStreamerEvents
trait you can easily make your Eloquent Models emit basic events. This trait will
integrate your model with Streamer and will emit events on actions like: save
, create
and delete
. It will emit an
event of your model name with suffix of the action and a payload of what happened. In case of a create
and save
actions the payload will have a list of changed fields and a before/after for each of those fields (with
create action fields before will basically have all values set to null), in case of a delete
action, payload will
simply state that the model has been deleted. Each payload includes a [key_name => key_value]
pair of your model ID.
By default, events will take names from their models with a suffix of the action, but the name can be changed by
assigning it to a baseEventName
attribute. This name will replace the model name but will keep suffix of what action
has been taken.
Check example's directory in this package to see how can you exactly use each command with package Stream and Consumer instances.