In this post our challenge is to load a CSV file and ingest it using the EventStoreDB HTTP API to be ingested.
To be precise, we want to convert this:
Europe,Italy,Clothes,Online,M,12/17/2013,278155219,1/10/2014,1165,109.28,35.84,127311.20,41753.60,85557.60
To this (which is an example of an HTTP POST to the Event Store HTTP API):
[
{
"eventId": "fbf4b1a1-b4a3-4dfe-a01f-ec52c34e16e4",
"eventType": "InboundDataReceived",
"data": {
"message": "Europe,Italy,Clothes,Online,M,12/17/2013,278155219,1/10/2014,1165,109.28,35.84,127311.20,41753.60,85557.60"
},
"metadata": {
"host": "box-1",
"path": "/usr/data/sales.csv"
}
}
]
In this example we set several different parts of the HTTP POST, including a unique eventId
, and enrich each message with metadata (like the host
and the file path
).
This data will be invaluable later for finding the specific mapping for that particular data source and to keep track of the origin.
The API Gateway is our internal data collection endpoint. In my opinion, it must be a simple HTTP Endpoint. There is no need for a heavyweight framework that can potentially add complexity and little value.
In this case, we are using the existing Event Store HTTP API as the API Gateway, and Logstash to monitor folders and convert the CSV lines into JSON Messages.
Logstash is a powerful tool that can keep track of the last ingested position in case something goes wrong and/or the file changes.
There are other tools available for this task, such as Telegraf, but I personally have more experience using Logstash.
Once that the messages are flowing through the API Gateway, we can then implement Domain Components or Microservices that subscribe to these messages and provide validation, mapping, and Domain business logic. The results will then be saved in their own data streams as Domain Events.
To avoid repeatedly the data in the inbound stream, you can also set an optional retention period after which the messages are automatically removed from the system.
Try it yourself:
With this Logstash input configuration, when you have a new client you can just add another folder and start ingesting the data. For production environments, you can combine Logstash with Filebeat to distribute part of the ingestion workload to separate boxes.
In some scenarios - especially when the data has not yet been discovered and you don’t have a mapper - it will be easier to direct the data straight into a semi-structured data lake for further analysis.
Hope that you enjoyed my little data ingestion example.
Here’s the 15-minute webinar if you’d like to watch me go through everything above and find out a little more: https://youtu.be/2nyzvrIdnPg