gdbots / pbjx
Messaging tools for pbj.
Installs: 25 550
Dependents: 4
Suggesters: 0
Security: 0
Stars: 0
Watchers: 6
Forks: 0
Open Issues: 0
Requires
- php: >=8.1
- ext-json: *
- gdbots/query-parser: ^3.0
- gdbots/schemas: ^3.0
- psr/log: ^1|^2|^3
- symfony/event-dispatcher: ^6.4 || ^7.0
Requires (Dev)
- aws/aws-sdk-php: ^3.138
- gdbots/acme-schemas: ^3.0
- phpunit/phpunit: ^10.5
- ruflin/elastica: ^7.1
- dev-master
- v4.2.0
- v4.1.1
- v4.1.0
- v4.0.0
- v3.1.6
- v3.1.5
- v3.1.4
- v3.1.3
- v3.1.2
- v3.1.1
- v3.1.0
- v3.0.1
- v3.0.0
- v2.3.14
- v2.3.13
- v2.3.12
- v2.3.11
- v2.3.10
- v2.3.9
- v2.3.8
- v2.3.7
- v2.3.6
- v2.3.5
- v2.3.4
- v2.3.3
- v2.3.2
- v2.3.1
- v2.3.0
- v2.2.5
- v2.2.4
- v2.2.3
- v2.2.2
- v2.2.1
- v2.2.0
- v2.1.1
- v2.1.0
- v2.0.1
- v2.0.0
- v1.0.1
- v1.0.0
- v0.3.0
- v0.2.2
- v0.2.1
- v0.2.0
- v0.1.1
- v0.1.0
This package is auto-updated.
Last update: 2024-10-24 06:34:34 UTC
README
This library provides the messaging tools for Pbj.
Pbj stands for "Private Business Json". Pbjc stands for "Private Business Json Compiler", a tool that creates php classes from a schema configuration. Pbjx stands for "Private Business Json Exchanger", a tool that exchanges pbj through various transports and storage engines.
Using this library assumes that you've already created and compiled your own pbj classes using the Pbjc.
Pbjx Primary Methods
- send: asynchronous message delivery to a single recipient with no return payload.
- publish: asynchronous message broadcast which can be subscribed to.
- request: synchronous message delivery to a single recipient with an expected return payload.
If you have configured the scheduler then these methods will also work:
- sendAt: schedules a command to send at a later time.
- cancelJobs: cancels previously scheduled commands by their job ids.
The strategy behind this library is similar to GRPC and CQRS.
If your project is using Symfony use the gdbots/pbjx-bundle-php to simplify the integration .
Transports
When pbj (aka messages) are exchanged a transport is used to perform that action. Your application/domain logic should never deal directly with the transports.
Available transports:
- AWS Firehose
- AWS Kinesis
- In Memory
Routers
Some transports require a router to determine the delivery channel (stream name, gearman channel, etc.) to route the message through. The router implementation can be fixed per type of message (command, event, request) or content specific as the pbj message itself is provided to the router.
For example:
interface Router { public function forCommand(Message $command): string; ... }
Pbjx::send
Processes a command (asynchronously if transport supports it).
When using the send method it implies that there is a single handler for that command, stated another way... if a "PublishArticle" command exists, there MUST be a service that handles that command.
In the gdbots/pbjx-bundle-php the
SchemaCurie
is used to derive the service id.
All command handlers MUST implement Gdbots\Pbjx\CommandHandler
.
Example handler for a "PublishArticle" command:
<?php declare(strict_types = 1); final class PublishArticleHandler implements CommandHandler { protected function handleCommand(Message $command, Pbjx $pbjx): void { // handle the command here } }
Invoking the command handler is never done directly (except in unit tests). In this made up example, you might have a controller that creates and sends the command.
<?php declare(strict_types = 1); final class ArticleController extends Controller { /** * @Route("/articles/{article_id}/publish", requirements={"article_id": "^[0-9A-Fa-f]+$"}) * @Method("POST") * @Security("is_granted('acme:blog:command:publish-article')") * * @param Request $request * * @return Response */ public function publishAction(Request $request): Response { $command = PublishArticleV1::create()->set('article_id', $request->attributes->get('article_id')); $this->getPbjx()->send($command); $this->addFlash('success', 'Article was published'); return $this->redirectToRoute('app_article_index'); } }
Pbjx::publish
Publishes events to all subscribers (asynchronously if transport supports it). All subscribers will receive the event unless a fatal error occurs.
Publisher/Subscriber is the pattern used. This is important because it may look like the Symfony EventDispatcher but a Pbjx subscriber cannot stop the propagation of events like they can in a Symfony subscriber/listener.
Pbjx published events are distinct from application or "lifecycle" events. These are your "domain events". The names you'll use here would make sense to most people, including the non-developer folks. MoneyDeposited, ArticlePublished, AccountClosed, UserUpgraded, etc.
Subscribing to a Pbjx published event requires that you know the SchemaCurie
of the event or its mixins.
Continuing the example above, let's imagine that PublishArticleHandler
created and published an event called ArticlePublished
and its curie was "acme:blog:event:article-published". In your subscriber you could listen to any of:
- acme:blog:event:article-published:v1
- acme:blog:event:article-published
- acme:blog:* all events in "acme:blog" namespace
- * all events
And any of its mixins:
- vendor:package:mixin:some-event:v1
- vendor:package:mixin:some-event
The method signature of all pbjx event subscribers should be the interface of the event and then the Pbjx service itself.
<?php declare(strict_types = 1); namespace Acme\Blog; use Gdbots\Pbjx\EventSubscriber; final class MyEventSubscriber implements EventSubscriber { public function onArticlePublished(Message $event, Pbjx $pbjx): void { // do something with this event. } public static function getSubscribedEvents() { return [ 'acme:blog:event:article-published' => 'onArticlePublished', ]; } }
When subscribing to multiple events you can use the convenient EventSubscriberTrait
which will automatically call methods matching any events it receives, e.g. "onUserRegistered", "onUserUpdated", "onUserDeleted".
Pbjx::request
Processes a request synchronously and returns the response. If the transport supports it, it may not be running in the current process (gearman for example). This is similar to "send" above but in this case, a response MUST be returned.
All request handlers MUST implement Gdbots\Pbjx\RequestHandler
.
Example handler for a "GetArticleRequest":
<?php declare(strict_types = 1); final class GetArticleRequestHandler implements RequestHandler { protected function handleRequest(Message $request, Pbjx $pbjx): Message { $response = GetArticleResponseV1::create(); // imaginary repository $article = $this->repository->getArticle($request->get('article_id')); return $response->set('article', $article); } }
Invoking the request handler is never done directly (except in unit tests).
<?php declare(strict_types = 1); final class ArticleController extends Controller { /** * @Route("/articles/{article_id}", requirements={"article_id": "^[0-9A-Fa-f]+$"}) * @Method("GET") * @Security("is_granted('acme:blog:request:get-article-request')") * * @param Request $request * * @return Response */ public function getAction(Request $request): Response { $getArticleRequest = GetArticleRequestV1::create()->set('article_id', $request->attributes->get('article_id')); $getArticleResponse = $this->getPbjx()->request($getArticleRequest); return $this->render('article.html.twig', ['article' => $getArticleResponse->get('article')]); } }
Pbjx Lifecycle Events
When a message is processed (send, publish, request) it goes through a lifecycle which allows for "in process" modification and validation. The method of subscribing to these events is similar to how a Symfony event subscriber/listener works and can stop propagation.
The lifecycle event names (what your subscriber/listener must be bound to) all have a standard format, e.g: "gdbots:pbjx:mixin:command.bind". These are named in the same way that the SimpleEventBus
names them. See SimplePbjx::trigger
method for insight into how this is done.
The lifecycle events, in order of occurrence are:
bind
Data that must be bound to the message by the environment its running in is done in the "bind" event. This is data that generally comes from the environment variables, http request itself, request context, etc. It's a cheap operation (in most cases).
Binding user agent, ip address, input from a form are good examples of things to do in the "bind" event.
validate
Before a message is allowed to be processed it should be validated. This is where business rules are generally implemented. This is more than just schema validation (which is done for you by pbj).
Checking permissions is generally done in this event. In a Symfony app this would be where you run the
AuthorizationChecker
and/or security voters. Additional examples:
- Checking inventory level on "AddProductToCart"
- Optimistic concurrency control on "UpdateArticle"
- Check for available username on "RegisterUser"
- Validate catchpa on "SubmitContactForm"
- Limit uploaded file size on "UploadVideo"
enrich
Once you've decided that a message is going to be processed you can perform additional enrichment that would have been expensive or not worth doing up until now. The enrichment is the final phase so once this is done the message will be frozen and then transported.
Geo2Ip enrichment, sentiment analysis, adding related data to events, etc. is a good use of the "enrich" event.
EventStore
Publishing events and storing/retrieving them is such a common need that we added the Pbjx::getEventStore
method to make the service available without any extra wiring of services to handlers, subscribers, etc.
The service will not be instantiated until you call the method so there's no performance penalty for having this capability built into Pbjx. At this time the only available implementation is DynamoDb.
Key concepts of the EventStore
Streams
Events are appended to a stream which is identified by a StreamId
(see below) and are ordered by the occurred_at
field of the event within that stream. This is not (forgive the ten dollar word) a "monotonically increasing gapless sequence number" (v1, v2, ..., v10). You could certainly achieve that design with your own implementation but for the original design/implemenation, it was more important to be able to collect data from many different servers (including different regions) without having to coordinate with a centralized id/sequence service.
The
occurred_at
field is 10 digits (unix timestamp) concatenated with 6 microsecond digits.
In order to mitigate collision possibilities we use more granular stream ids when the volume of data is expected to be high. Users editing articles in a cms vs users voting on polls for example, you know the write volume of poll votes would be MUCH higher.
EventStores can have billions of streams and have no performance issues so long as:
- The StreamIds you append events to are distributed enough to avoid hot keys.
- The total volume of events on a single stream doesn't exceed 10GB (This limit comes from DynamoDb). You can use date based storage to deal with long-lived streams, e.g. events-YYYYMM.
- The underlying storage is setup to handle the data size and write/read rates of your application.
IMPORTANT: An event may exist in one or more streams. The StreamId is NOT a property of the event itself.
StreamId
A stream id represents a stream of events. The parts of the id are delimited by a colon for our purposes but can easily be converted to acceptable formats for SNS, Kafka, etc.
It may also be desirable to only use parts of the stream id (e.g. topic) for broadcast.
Using a partition and optionally a sub-partition makes it possible to group all of those records together in storage and also guarantee their sequence is exactly in the order that they were added to the stream.
StreamId Format: vendor:topic:partition:sub-partition
Examples:
"twitter" (vendor), "user.timeline" (topic), "homer-simpson" (partition), "yyyymm" (sub-partition)
twitter:user.timeline:homer-simpson:201501 twitter:user.timeline:homer-simpson:201502 twitter:user.timeline:homer-simpson:201503
"acme" (vendor), "bank-account" (topic), "homer-simpson" (partition)
acme:bank-account:homer-simpson
"acme" (vendor), "poll.votes" (topic), "batman-vs-superman" (partition), "yyyymm.[0-9a-f][0-9a-f]" (sub-partition)
Note the sub-partition here is two hexidecimal digits allowing for 256 separate stream ids. Useful when you need to avoid hot keys and ordering in the overall partition isn't important.
acme:poll.votes:batman-vs-superman:20160301.0a acme:poll.votes:batman-vs-superman:20160301.1b acme:poll.votes:batman-vs-superman:20160301.c2
StreamSlice
Getting data out of the EventStore is done via piping from a single stream or ALL streams or by getting a slice of a stream. Think of the StreamSlice as the php function array_slice. You can get slices forward and backward from a stream and the events are ordered by the occurred_at
field.
EventStore::putEvents
In most cases, you'd be writing to the EventStore in a command handler. Continuing the publish article example above, let's actually make the event.
<?php declare(strict_types = 1); final class PublishArticleHandler implements CommandHandler { protected function handleCommand(Message $command, Pbjx $pbjx): void { // in this example it's ultra basic, create the event and push it to a stream $event = ArticlePublishedV1::create()->set('article_id', $command->get('article_id')); // copies contextual data from the previous message (ctx_* fields) $pbjx->copyContext($command, $event); $streamId = StreamId::fromString(sprintf('acme:article:%s', $command->get('article_id'))); $pbjx->getEventStore()->putEvents($streamId, [$event]); // after the event is persisted it will be published either via a // two phase commit or a publisher reading the EventStore streams // (DynamoDb streams for example) } }
One thing to note is that the state of the article was not modified here, we just put an event on a stream. An event subscriber would listen for those events and then update the article. That is just one way to handle state updates.
This design implies eventual consistency.
EventSearch
Storing and retrieving events are all handled by the EventStore but often times you need to be able to search events too. Comments, product reviews, post reactions, survey responses, etc. Lots of use cases can benefit from indexing.
Similar to the EventStore the EventSearch service is only instantiated if requested. The only implementation we have right now is ElasticSearch. To use this feature, the events you want to index must be using the "gdbots:pbjx:mixin:indexed" mixin. When using the gdbots/pbjx-bundle-php you can enable the indexing with a simple configuration option.
Searching events is generally done in a request handler. Here is an example of searching events:
protected function handleRequest(Message $request, Pbjx $pbjx): Message { $parsedQuery = ParsedQuery::fromArray(json_decode($request->get('parsed_query_json', '{}'), true)); $response = SearchCommentsResponseV1::create(); $pbjx->getEventSearch()->searchEvents( $request, $parsedQuery, $response, [SchemaCurie::fromString('acme:blog:event:comment-posted')] ); return $response; }
Scheduler
Commands can be scheduled to send at a later time using the sendAt
method. You construct a command in the usual way, decide when it should send and optionally provide a jobId of your own.
At this time the only available implementation is DynamoDb which also depends on AWS Step Functions. An example implementation of the state machine and lambda task will be open sourced in the future. For now, here is a simple state machine defined in CloudFormation that can deal with the deferred sending.
# Expects you to define your own SchedulerRole and SchedulerFunction resource SchedulerStateMachine: Type: AWS::StepFunctions::StateMachine DependsOn: SchedulerRole Properties: StateMachineName: !Sub 'acme-${AppEnv}-pbjx-scheduler' RoleArn: !GetAtt 'SchedulerRole.Arn' DefinitionString: !Sub |- { "StartAt": "WaitUntil", "States": { "WaitUntil" : { "Type": "Wait", "TimestampPath": "$.send_at", "Next": "Send" }, "Send": { "Type": "Task", "Resource": "${SchedulerFunction.Arn}", "TimeoutSeconds": 60, "End": true } } }
In the above state machine the SchedulerFunction.Arn
would be a Lambda function that receives the input to the state machine and then sends the payload (stored in DynamoDb) to your pbjx endpoint.
{ "send_at": "2016-08-18T17:33:00Z", "job_id": "my-job-id-or-an-autogenerated-uuid" }
If the send_at
is more than a year in the future your lambda will have to deal with restarting the execution. This is a limitation of Step Functions at this time. To make this simple the payload includes the future dates to use.
{ "send_at": "2017-12-25T12:00:00Z", "job_id": "my-job-id-or-an-autogenerated-uuid", "resend_at": [ "2018-12-25T12:00:00Z", "2019-12-25T12:00:00Z", "2020-12-25T12:00:00Z" ] }
When the resend_at
array is present and not empty:
- shift the first item off the
resend_at
array - use it as the new
send_at
value for the input - start the execution again (providing the new input)
- update the dynamodb record with key of the
job_id
with the new executionArn.
If there is no
resend_at
array or it's empty, process as normal.
JobId
If you provide your own jobId to the sendAt method it will automatically stop an existing job using that id. This makes it possible to ensure only job exist for a given process. Example use case is publishing or expiring content. When a user changes the publish/expiry date you want to cancel the old scheduled command.
Canceling Jobs
The sendAt method returns a job id (or simply returns the one you provided). This id is required to cancel a command (using the cancelJobs
method) and there is no secondary method at this time to query for jobs and cancel the results.
PbjxDebugger
When developing applications using Pbjx you need to be able to see what's being exchanged. The PbjxDebugger
will push "debug" messages with all of the pbj data that is being handled. A Psr\Log\LoggerInterface
logger must be provided when using this service.
If you're using monolog you can route all of these entries to their own file in json line delimited format. This is the recommended use because it makes it possible to use other great tools like jq.
Example yaml config for use in Symfony.
services: monolog_json_formatter: class: Monolog\Formatter\JsonFormatter arguments: [!php/const:Monolog\Formatter\JsonFormatter::BATCH_MODE_NEWLINES] monolog: handlers: pbjx_debugger: type: stream path: '%kernel.logs_dir%/pbjx-debugger.log' level: debug formatter: monolog_json_formatter channels: ['pbjx.debugger']