Commit Graph

13 Commits

Author SHA1 Message Date
4d37578c59
feat: add storage backend
Signed-off-by: kjuulh <contact@kjuulh.io>
2024-11-09 17:19:19 +01:00
a327cdb02e
feat: remove keys and ids
Signed-off-by: kjuulh <contact@kjuulh.io>
2024-11-09 11:47:50 +01:00
b9be9a3ef1
feat: working dagger engine
Signed-off-by: kjuulh <contact@kjuulh.io>
2024-08-19 07:29:26 +02:00
95acdf4f4a
feat: with sdk
Signed-off-by: kjuulh <contact@kjuulh.io>
2024-08-18 01:31:58 +02:00
c51cd4dfc1
feat: with dagger engine
Signed-off-by: kjuulh <contact@kjuulh.io>
2024-08-18 01:13:27 +02:00
d0ea8019e1
feat: add broker setup
this is mainly to decouple the actual sending of events, from the ingest.

we now ingest the data, and update consumer groups with the new offset. Consumer groups now in the background continously send out data from the update. they tick 1 second between checks, but if something takes long than a second, the next run just continues from where we left off

Signed-off-by: kjuulh <contact@kjuulh.io>
2024-08-16 00:25:43 +02:00
f818a18d65
feat: update
Signed-off-by: kjuulh <contact@kjuulh.io>
2024-08-15 21:07:37 +02:00
3a2a65dfa7
feat: add basic streaming data
Signed-off-by: kjuulh <contact@kjuulh.io>
2024-08-15 21:07:22 +02:00
101a266ea2
fix: offset be inclusive end
Signed-off-by: kjuulh <contact@kjuulh.io>
2024-08-14 14:46:53 +02:00
3c5c5759ca
feat: add data ingest
Signed-off-by: kjuulh <contact@kjuulh.io>
2024-08-10 00:54:10 +02:00
5395e8bb1e
refactor: into files
Signed-off-by: kjuulh <contact@kjuulh.io>
2024-08-10 00:25:46 +02:00
cf321d110e
feat: add docs
Signed-off-by: kjuulh <contact@kjuulh.io>
2024-08-10 00:17:12 +02:00
a5196b005e
feat: add base
Signed-off-by: kjuulh <contact@kjuulh.io>
2024-08-10 00:09:11 +02:00