Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/eugene-khyst/postgresql-event-sourcing
A reference implementation of an event-sourced system that uses PostgreSQL as an event store built with Spring Boot. Fork the repository and use it as a template for your projects. Or clone the repository and run end-to-end tests to see how everything works together.
https://github.com/eugene-khyst/postgresql-event-sourcing
cqrs ddd event-driven event-driven-architecture event-sourcing event-store eventsourcing postgres postgres-notifications postgresql solution-architecture
Last synced: 22 days ago
JSON representation
A reference implementation of an event-sourced system that uses PostgreSQL as an event store built with Spring Boot. Fork the repository and use it as a template for your projects. Or clone the repository and run end-to-end tests to see how everything works together.
- Host: GitHub
- URL: https://github.com/eugene-khyst/postgresql-event-sourcing
- Owner: eugene-khyst
- License: apache-2.0
- Created: 2021-04-25T16:59:29.000Z (over 3 years ago)
- Default Branch: main
- Last Pushed: 2024-06-10T09:26:10.000Z (5 months ago)
- Last Synced: 2024-10-01T20:42:33.636Z (about 1 month ago)
- Topics: cqrs, ddd, event-driven, event-driven-architecture, event-sourcing, event-store, eventsourcing, postgres, postgres-notifications, postgresql, solution-architecture
- Language: Java
- Homepage:
- Size: 1.61 MB
- Stars: 1,077
- Watchers: 14
- Forks: 98
- Open Issues: 3
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
README
# Event Sourcing with PostgreSQL
- [Introduction](#1)
- [Example domain](#2)
- [Event sourcing and CQRS basics](#3)
- [State-oriented persistence](#3-1)
- [Event sourcing](#3-2)
- [Snapshotting](#3-3)
- [Querying the data](#3-4)
- [CQRS](#3-5)
- [Event handlers](#3-6)
- [Domain events vs Integration events](#3-7)
- [Advantages of CQRS](#3-8)
- [Advantages of event sourcing](#3-9)
- [Solution architecture](#4)
- [Component diagram](#4-1)
- [ER diagram](#4-2)
- [Optimistic concurrency control](#4-3)
- [Snapshotting](#4-4)
- [Loading any revision of the aggregate](#4-5)
- [Synchronously updating projections](#4-6)
- [Asynchronously sending integration events to a message broker](#4-7)
- [Transactional outbox using transaction ID](#4-7-1)
- [Transactional outbox using table-level lock](#4-7-2)
- [Database polling](#4-7-3)
- [Listen/Notify as an alternative to database polling](#4-7-4)
- [Adding new asynchronous event handlers](#4-8)
- [Drawbacks](#4-9)
- [Project structure](#5)
- [Gradle subprojects](#5-1)
- [Database schema migrations](#5-2)
- [Class diagrams](#5-3)
- [Class diagram of the domain model](#5-3-1)
- [Class diagram of the projections](#5-3-2)
- [Class diagram of the service layer](#5-3-3)
- [How to adapt it to your domain?](#6)
- [How to run the sample?](#7)Usually, our applications operate with the current state of a domain object.
But sometimes, we need to know the entire history of the domain object changes.
For example, we want to know how an order got into its current state.The audit trail (also called the audit log) is a chronological record of the history
and details of the actions that affected the system.
An audit trail may be a regulatory or business requirement.We can store all changes to the domain object state as a sequence of events in an append-only event stream.
Thus, event streams will contain an entire history of changes.
But how can we be sure that this history is authentic and error-free?
We can use event streams as a primary source of truth in a system.
To get the current state of an object, we have to replay all events in the order of occurrence.
This pattern is called event sourcing. The database for storing event streams is called an event store.
Event sourcing provides a complete and accurate record of all changes made to a system.
Event sourcing is an industry standard for implementing audit trail.There are specialized databases for event sourcing.
Developer Advocates working for the companies behind these specialized databases said
you shouldn't implement event sourcing with traditional relational or document-oriented databases.
Is this true or just a marketing ploy?Specialized databases for event sourcing are convenient and provide the necessary functionality out of the box.
But PostgreSQL, the world's most advanced open-source database, is also suitable for event sourcing.
You can use PostgreSQL as an event store without additional frameworks or extensions
instead of setting up and maintaining a separate specialized database for event sourcing.This repository provides a reference implementation of an event-sourced system
that uses PostgreSQL as an event store built with Spring Boot.
[Fork](https://github.com/eugene-khyst/postgresql-event-sourcing/fork) the repository and use it as a template for your projects.
Or clone the repository and run end-to-end tests to see how everything works together.![PostgreSQL Logo](img/potgresql-logo.png)
See also
* [Event Sourcing with EventStoreDB](https://github.com/eugene-khyst/eventstoredb-event-sourcing)
* [Event Sourcing with Kafka and ksqlDB](https://github.com/eugene-khyst/ksqldb-event-souring)This sample uses a simplified domain model of the ride-hailing system.
* A rider can place an order for a ride along a route specifying a price.
* A rider can edit an order price to pay more instead of waiting in cases of very high demand.
* A driver can accept an order.
* A driver can complete previously accepted order.
* An order can be canceled before completion.![Domain use case diagram](img/domain-1.svg)
![Domain state diagram](img/domain-2.svg)
## Event sourcing and CQRS basics
### State-oriented persistence
State-oriented persistence (CRUD) applications store only the latest version of an entity.
Database records present entities.
When an entity is updated, the corresponding database record gets updated too.
SQL `INSERT`, `UPDATE` and `DELETE` statements are used.![State-oriented persistence](img/state-oriented-persistence.svg)
Event sourcing applications persist the state of an entity as a sequence of immutable state-changing events.
![Event sourcing](img/event-sourcing-1.svg)
Whenever the state of an entity changes, a new event is appended to the list of events.
Only SQL `INSERT` statements are used.
Events are immutables, so SQL `UPDATE` and `DELETE` statements are not used.![Event sourcing](img/event-sourcing-2.svg)
The current state of an entity can be restored by replaying all its events.
Event sourcing is closely related to domain-driven design (DDD) and shares some terminology.
An entity in event sourcing is called an **aggregate**.
A sequence of events for the same aggregate is called a **stream**.
Event sourcing is best suited for short-living entities with a small total number of
events (e.g., orders).Restoring the state of the short-living entity by replaying all its events doesn't have any
performance impact. Thus, no optimizations for restoring state are required for short-living
entities.For endlessly stored entities (e.g., users, bank accounts) with thousands of events restoring state
by replaying all events is not optimal, and snapshotting should be considered.Snapshotting is an optimization technique where a snapshot of the aggregate's state is also saved,
so an application can restore the current state of the aggregate from the snapshot rather than from
all the events (potentially thousands).On every *nth* event, make an aggregate snapshot by storing an aggregate state and its version.
To restore an aggregate state:
1. first read the latest snapshot,
2. then read events forward from the original stream starting from the version pointed by the snapshot.![Snapshotting in event souring](img/event-sourcing-snapshotting.svg)
It's easy to find an aggregate by ID, but other queries are difficult.
Since aggregates are stored as append-only lists of immutable events,
querying the data using SQL, as we used to, is impossible.
To find an aggregate by some field, we need to first read all the events and replay them to restore all the aggregates.To bring back all the querying power a relational database has to offer,
we can create a dedicated read model derived from the event stream.The event stream is the write model and the primary source of truth.
The read model is a "denormalized" view of the write model, allowing faster and more convenient querying.
Read models are projections of the system state.
Therefore, read models are also known as **projections**.Projections provide a view of data for a single aggregate type
or perform aggregations and combine data from multiple aggregate types.That's where CQRS comes in handy.
Command-query responsibility segregation (CQRS) stands for segregating the responsibility between
commands (write requests) and queries (read requests).
The write requests and the read requests are processed by different handlers.A command generates zero or more events or results in an error.
![CQRS](img/cqrs-1.svg)
CQRS is a self-sufficient architectural pattern and doesn't require event sourcing.
But in practice, event sourcing is usually used in conjunction with CQRS.
Event store is used as a write database, and SQL or NoSQL database as a read database.![CQRS with event sourcing](img/cqrs-2.svg)
Commands generate events.
Event processing is done by **event handlers**.
As a part of event processing, we may need to update projections,
send a message to a message broker, or make an API call.There are two types of event handlers: **synchronous** and **asynchronous**.
Storing the write model and read model in the same database allows for transactional updates of the read model.
Each time we append a new event, the projection is updated **synchronously** in the same transaction.
The projection is **consistent** with the event stream.When an event handler communicates with an external system or middleware (e.g., sends a message to Kafka),
it should run **asynchronously** after the transaction updating the write model.
Asynchronous execution leads to **eventual consistency**.Communication with external systems should not occur in the same transaction updating the write model.
The external system call may succeed, but the transaction will later be rolled back, resulting in an inconsistency.Anyway, distributed systems should be designed with eventual consistency in mind.
### Domain events vs Integration events
Events in event sourcing are **domain events**.
The domain event is a part of a bounded context and should not be used "as-is" for integration with other bounded
contexts.For communication between bounded contexts **integration events** are used.
The integration event represents the current state of an aggregate, not just changes to the aggregate as a domain event.* Independent scaling of the read and write databases.
* Optimized data schema for the read database (e.g. the read databases can be denormalized).
* Simpler queries (e.g. complex `JOIN` operations can be avoided).### Advantages of event sourcing
* A true history of the system (audit and traceability).
An industry standard for implementing audit trail.
* Ability to put the system in any prior state (e.g. for debugging).
* New read-side projections can be created as needed (later) from events.
It allows responding to future needs and new requirements.PostgreSQL can be used as an event store.
It will natively support appending events, concurrency control and reading events.
Subscribing on events requires additional implementation.![PostgreSQL event store component diagram](img/postgresql-event-sourcing.svg)
![PostgreSQL event store ER diagram](img/er-diagram.svg)
Events are stored in the `ES_EVENT` table.
### Optimistic concurrency control
Latest aggregate version is stored in the `ES_AGGREGATE` table.
Version checking is used for optimistic concurrency control.
Version checking uses version numbers to detect conflicting updates (and to prevent lost updates).Appending an event operation consists of 2 SQL statements in a single transaction:
1. check the actual and expected version match and increment the version
```sql
UPDATE ES_AGGREGATE
SET VERSION = :newVersion
WHERE ID = :aggregateId
AND VERSION = :expectedVersion
```
2. insert new event
```sql
INSERT INTO ES_EVENT (TRANSACTION_ID, AGGREGATE_ID, VERSION, EVENT_TYPE, JSON_DATA)
VALUES(pg_current_xact_id(), :aggregateId, :version, :eventType, :jsonObj::json)
RETURNING ID, TRANSACTION_ID::text, EVENT_TYPE, JSON_DATA
```
`pg_current_xact_id()` returns the current transaction's ID. The need for this will be explained later.![Optimistic Concurrency Control](img/event-sourcing-concurrency.svg)
On every *nth* event insert an aggregate state (snapshot) to the `ES_AGGREGATE_SNAPSHOT` table specifying the version
```sql
INSERT INTO ES_AGGREGATE_SNAPSHOT (AGGREGATE_ID, VERSION, JSON_DATA)
VALUES (:aggregateId, :version, :jsonObj::json)
```Snapshotting for an aggregate type can be disabled and configured in the [`application.yml`](src/main/resources/application.yml)
```yaml
event-sourcing:
snapshotting:
# com.example.eventsourcing.domain.AggregateType
ORDER:
enabled: true
# Create a snapshot on every nth event
nth-event: 10
```### Loading any revision of the aggregate
To restore any revision of the aggregate:
1. first read the latest value of the snapshot
```sql
SELECT a.AGGREGATE_TYPE,
s.JSON_DATA
FROM ES_AGGREGATE_SNAPSHOT s
JOIN ES_AGGREGATE a ON a.ID = s.AGGREGATE_ID
WHERE s.AGGREGATE_ID = :aggregateId
AND (:version IS NULL OR s.VERSION <= :version)
ORDER BY s.VERSION DESC
LIMIT 1
```
2. then read forward from the event stream from the revision the snapshot points to
```sql
SELECT ID,
TRANSACTION_ID::text,
EVENT_TYPE,
JSON_DATA
FROM ES_EVENT
WHERE AGGREGATE_ID = :aggregateId
AND (:fromVersion IS NULL OR VERSION > :fromVersion)
AND (:toVersion IS NULL OR VERSION <= :toVersion)
ORDER BY VERSION ASC
```### Synchronously updating projections
Using PostgreSQL as an event store and a read database allows for transactional updates of the read model.
Each time we append a new event, the projection is updated synchronously in the same transaction.
It's a big advantage because sometimes consistency is not so easy to achieve.You can't get consistent projections when a separate database is used as an event store.
### Asynchronously sending integration events to a message broker
Integration events should be sent asynchronously after the transaction updating the write model.
PostgreSQL doesn't allow subscribing on changes, so the solution is a Transactional Outbox pattern.
A service that uses a database inserts events into an outbox table as part of the local transaction.
A separate process publishes the events inserted into database to a message broker.![Transactional outbox pattern](img/transactional-outbox-1.svg)
We may have multiple asynchronous event handlers or so-called subscriptions.
The subscription concept is required to keep track of delivered events separately for different event handlers.
The last event processed by the event handler (subscription) is stored in the separate table `ES_EVENT_SUBSCRIPTION`.
New events are processed by polling the `ES_EVENT` table.![Transactional outbox pattern with subscriptions](img/transactional-outbox-2.svg)
Since multiple instances of this application can run in parallel,
we need to ensure that any processing only affects the event once.
We don't want more than one event handler instance to handle the same event.This is achieved by acquiring locks on the rows of the `ES_EVENT_SUBSCRIPTION` table.
We lock the row (`SELECT FOR UPDATE`) of the currently processed subscription.To not hang other backend instances, we want to skip already locked rows (`SELECT FOR UPDATE SKIP LOCKED`)
and lock the "next" subscription.
It allows multiple backend instances to select different subscriptions that do not overlap.
This way, we improve availability and scalability.The event subscription processor polls `ES_EVENT_SUBSCRIPTION` table every second (interval is configurable) for new events
and processes them:1. read the last transaction ID and event ID processed by the subscription and acquire lock
```sql
SELECT LAST_TRANSACTION_ID::text,
LAST_EVENT_ID
FROM ES_EVENT_SUBSCRIPTION
WHERE SUBSCRIPTION_NAME = :subscriptionName
FOR UPDATE SKIP LOCKED
```
2. read new events
```sql
SELECT e.ID,
e.TRANSACTION_ID::text,
e.EVENT_TYPE,
e.JSON_DATA
FROM ES_EVENT e
JOIN ES_AGGREGATE a on a.ID = e.AGGREGATE_ID
WHERE a.AGGREGATE_TYPE = :aggregateType
AND (e.TRANSACTION_ID, e.ID) > (:lastProcessedTransactionId::xid8, :lastProcessedEventId)
AND e.TRANSACTION_ID < pg_snapshot_xmin(pg_current_snapshot())
ORDER BY e.TRANSACTION_ID ASC, e.ID ASC
```
A comparison like `(a, b) > (c, d)` is a row comparison and is equivalent to `a > c OR (a = c AND b > d)`.
3. update the last transaction ID and event ID processed by the subscription
```sql
UPDATE ES_EVENT_SUBSCRIPTION
SET LAST_TRANSACTION_ID = :lastProcessedTransactionId::xid8,
LAST_EVENT_ID = :lastProcessedEventId
WHERE SUBSCRIPTION_NAME = :subscriptionName
```#### Transactional outbox using transaction ID
Using only the event ID to track events processed by the subscription is unreliable
and can result in lost events.The `ID` column of the `ES_EVENT` table is of type `BIGSERIAL`.
It's a notational convenience for creating ID columns having their default values assigned from a `SEQUENCE` generator.PostgreSQL sequences can't be rolled back.
`SELECT nextval('ES_EVENT_ID_SEQ')` increments and returns the sequence value.
Even if the transaction is not yet committed, the new sequence value becomes visible to other transactions.If transaction #2 started after transaction #1 but committed first,
the event subscription processor can read the events created by transaction #2, update the last processed event ID,
and thus lose the events created by transaction #1.![PostgreSQL naive transactional outbox](img/postgresql-naive-outbox.svg)
We use transaction ID with event ID to build a reliable PostgreSQL polling mechanism that doesn't lose events.
Each event is supplemented with the current transaction ID.
`pg_current_xact_id()` returns the current transaction's ID of type `xid8`.
`xid8` values increase strictly monotonically and cannot be reused in the lifetime of a database cluster.The latest event that is "safe" to process is right before the `xmin` of the current snapshot.
`pg_current_snapshot()` returns a current snapshot, a data structure showing which transaction IDs are now in-progress.
`pg_snapshot_xmin(pg_snapshot)` returns the `xmin` of a snapshot.
`xmin` is the lowest transaction ID that was still active.
All transaction IDs less than `xmin` are either committed and visible, or rolled back.Even if transaction #2 started after transaction #1 and committed first,
the events it created won't be read by the event subscription processor until transaction #1 is committed.![PostgreSQL reliable transactional outbox using transaction ID](img/postgresql-reliable-outbox.svg)
> **NOTE**
> The transaction ID solution is used by default as it is non-blocking.#### Transactional outbox using table-level lock
With the transaction ID solution, event subscription processor doesn't wait for in-progress transactions to complete.
Events created by already committed transactions will not be available for processing
while transactions started earlier are still in-progress.
These events will be processed immediately after these earlier transactions have completed.An alternative solution is to use PostgreSQL explicit locking to make event subscription processor wait for in-progress transactions.
Before reading new events, the event subscription processor
* gets the most recently issued ID sequence number,
* very briefly locks the table for writes to wait for all pending writes to complete.The most recently issued `ES_EVENT_ID_SEQ` sequence number is obtained using the `pg_sequence_last_value` function:
`SELECT pg_sequence_last_value('ES_EVENT_ID_SEQ')`.Events are created with the command `INSERT INTO ES_EVENT...` that acquires the `ROW EXCLUSIVE` lock mode on `ES_EVENT` table.
`ROW EXCLUSIVE (RowExclusiveLock)` lock mode is acquired by any command that modifies data in a table.The command `LOCK ES_EVENT IN SHARE ROW EXCLUSIVE MODE` acquires the `SHARE ROW EXCLUSIVE` lock mode on `ES_EVENT` table.
`SHARE ROW EXCLUSIVE (ShareRowExclusiveLock)` mode protects a table against concurrent data changes,
and is self-exclusive so that only one session can hold it at a time.`SHARE ROW EXCLUSIVE` lock must be acquired in a separate transaction (`Propagation.REQUIRES_NEW`).
The transaction must contain only this command and commit quickly to release the lock, so writes can resume.When the lock is acquired and released, it means
that there are no more uncommitted writes with an ID less than or equal to the ID returned by `pg_sequence_last_value`.![PostgreSQL reliable transactional outbox using table-level lock](img/postgresql-reliable-outbox-with-lock.svg)
To get new events from the `ES_EVENT` table, the application has to poll the database.
The shorter the polling period, the shorter the delay between persisting a new event and processing it by the subscription.
But the lag is inevitable. If the polling period is 1 second, then the lag is at most 1 second.The polling mechanism implementation [ScheduledEventSubscriptionProcessor](src/main/java/com/example/eventsourcing/service/ScheduledEventSubscriptionProcessor.java)
uses a Spring annotation [@Scheduled](https://docs.spring.io/spring-framework/docs/current/javadoc-api/org/springframework/scheduling/annotation/Scheduled.html) to poll database with a fixed period.The polling event subscription processing can be enabled and configured in the [`application.yml`](src/main/resources/application.yml)
```yaml
event-sourcing:
subscriptions: polling # Enable database polling subscription processing
polling-subscriptions:
polling-initial-delay: PT1S
polling-interval: PT1S
```#### Listen/Notify as an alternative to database polling
To reduce the lag associated with database polling, the polling period can be set to a very low value,
such as 1 second.
But this means that there will be 3600 database queries per hour and 86400 per day, even if there are no new events.PostgreSQL `LISTEN` and `NOTIFY` feature can be used instead of polling.
This mechanism allows for sending asynchronous notifications across database connections.
Notifications are not sent directly from the application,
but via the database [trigger](src/main/resources/db/migration/V2__notify_trigger.sql) on a table.To use this functionality an unshared [PgConnection](https://jdbc.postgresql.org/documentation/publicapi/org/postgresql/jdbc/PgConnection.html)
which remains open is required.
The long-lived dedicated JDBC `Connection` for receiving notifications has to be created using the `DriverManager` API,
instead of getting from a pooled `DataSource`.PostgreSQL JDBC driver can't receive asynchronous notifications
and must poll the backend to check if any notifications were issued.
A timeout can be given to the poll function `getNotifications(int timeoutMillis)`,
but then the execution of statements from other threads will block.
When `timeoutMillis` = 0, blocks forever or until at least one notification has been received.
It means that notification is delivered almost immediately, without a lag.
If more than one notification is about to be received, these will be returned in one batch.This solution significantly reduces the number of issued queries
and also solves the lag problem that the polling solution suffers from.The Listen/Notify mechanism implementation [PostgresChannelEventSubscriptionProcessor](src/main/java/com/example/eventsourcing/service/PostgresChannelEventSubscriptionProcessor.java)
is inspired by the Spring Integration class [PostgresChannelMessageTableSubscriber](https://github.com/spring-projects/spring-integration/blob/v6.0.0/spring-integration-jdbc/src/main/java/org/springframework/integration/jdbc/channel/PostgresChannelMessageTableSubscriber.java).The Listen/Notify event subscription processing can be enabled in the [`application.yml`](src/main/resources/application.yml)
```yaml
event-sourcing:
subscriptions: postgres-channel # Enable Listen/Notify event subscription processing
```> **NOTE**
> The Listen/Notify mechanism is used by default as it is more efficient.### Adding new asynchronous event handlers
After restarting the backend, existing subscriptions will only process new events after the last processed event
and not everything from the first one.> **WARNING**
> Critical content demanding immediate user attention due to potential risks.
New subscriptions (event handlers) in the first poll will read and process all events.
Be careful, if there are too many events, they may take a long time to process.Using PostgreSQL as an event store has a lot of advantages, but there are also drawbacks.
1. **Asynchronous event handlers can process the same event more than once.**
It might crash after processing an event but before recording the fact that it has done so.
When it restarts, it will then process the same event again (e.g., send an integration event).
Integration events are delivered with **at-least-once** delivery guarantee.
The exactly-once delivery guarantee is hard to achieve due to a dual-write.
A dual-write describes a situation when you need to atomically update the database and publish messages
without two-phase commit (2PC).
Consumers of integration events should be idempotent and filter duplicates and unordered events.
2. The asynchronous event handling results in the **eventual consistency between the write model and sent integration events**.
The polling database table for new events with a fixed delay introduces a full consistency lag
greater than or equal to the interval between polls (1 second by default).
3. **A long-running transaction in the same database will effectively "pause" all event handlers.**
`pg_snapshot_xmin(pg_snapshot)` will return the ID of this long-running transaction
and events created by all later transactions will be read by the event subscription processor
only after this long-running transaction is committed.This reference implementation can be easily extended to comply with your domain model.
Event sourcing related code and application specific code are located in separate Gradle subprojects:
* [`postgresql-event-sourcing-core`](postgresql-event-sourcing-core): event sourcing and PostgreSQL related code, a shared library, `eventsourcing.postgresql` package,
* [`event-sourcing-app`](event-sourcing-app): application specific code, a simplified ride-hailing sample, `com.example.eventsourcing` package.`event-sourcing-app` depends on `postgresql-event-sourcing-core`:
```groovy
dependencies {
implementation project(':postgresql-event-sourcing-core')
}
```### Database schema migrations
Event sourcing related database schema migrations:
* [V1__eventsourcing_tables.sql](event-sourcing-app/src/main/resources/db/migration/V1__eventsourcing_tables.sql)
* [V2__notify_trigger.sql](event-sourcing-app/src/main/resources/db/migration/V2__notify_trigger.sql)Application specific projections database schema migration:
* [V3__projection_tables.sql](event-sourcing-app/src/main/resources/db/migration/V3__projection_tables.sql)#### Class diagram of the domain model
![Class diagram of the domain model](img/class-domain.svg)
#### Class diagram of the projections
![Class diagram of the projections](img/class-projection.svg)
#### Class diagram of the service layer
![Class diagram of the service layer](img/class-service.svg)
## How to adapt it to your domain?
To adapt this sample to your domain model, make changes to `event-sourcing-app` subproject.
No changes to `postgresql-event-sourcing-core` subproject are required.1. Download & install [SDKMAN!](https://sdkman.io/install).
2. Install JDK 21
```bash
sdk list java
sdk install java 21-tem
```3. Install [Docker](https://docs.docker.com/engine/install/)
and [Docker Compose](https://docs.docker.com/compose/install/).4. Build Java project and Docker image
```bash
./gradlew clean build jibDockerBuild -i
```5. Run PostgreSQL, Kafka and event-sourcing-app
```bash
docker compose --env-file gradle.properties up -d --scale event-sourcing-app=2
# wait a few minutes
```6. Follow the logs of the application
```bash
docker compose logs -f event-sourcing-app
```
7. Run E2E tests and see the output
```bash
E2E_TESTING=true ./gradlew clean test -i
```8. Explore the database using the Adminer database management tool at http://localhost:8181.
Find the database name, user, and password in the [docker-compose.yml](docker-compose.yml).
You can also manually call the REST API endpoints.1. Install [curl](https://curl.se/) and [jq](https://stedolan.github.io/jq/)
```bash
sudo apt install curl jq
```2. Place new order
```bash
ORDER_ID=$(curl -s -X POST http://localhost:8080/orders -d '{"riderId":"63770803-38f4-4594-aec2-4c74918f7165","price":"123.45","route":[{"address":"Kyiv, 17A Polyarna Street","lat":50.51980052414157,"lon":30.467197278948536},{"address":"Kyiv, 18V Novokostyantynivska Street","lat":50.48509161169076,"lon":30.485170724431292}]}' -H 'Content-Type: application/json' | jq -r .orderId)
```3. Get the placed order
```bash
curl -s -X GET http://localhost:8080/orders/$ORDER_ID | jq
```4. Accept the order
```bash
curl -s -X PUT http://localhost:8080/orders/$ORDER_ID -d '{"status":"ACCEPTED","driverId":"2c068a1a-9263-433f-a70b-067d51b98378"}' -H 'Content-Type: application/json'
```5. Get the accepted order
```bash
curl -s -X GET http://localhost:8080/orders/$ORDER_ID | jq
```6. Complete the order
```bash
curl -s -X PUT http://localhost:8080/orders/$ORDER_ID -d '{"status":"COMPLETED"}' -H 'Content-Type: application/json'
```7. Get the completed order
```bash
curl -s -X GET http://localhost:8080/orders/$ORDER_ID | jq
```8. Try to cancel a completed order to simulate business rule violation
```bash
curl -s -X PUT http://localhost:8080/orders/$ORDER_ID -d '{"status":"CANCELLED"}' -H 'Content-Type: application/json' | jq
```9. Print integration events
```bash
docker compose exec kafka /bin/kafka-console-consumer --bootstrap-server localhost:9092 --topic order-events --from-beginning --property print.key=true --timeout-ms 10000
```