Skip to content(if available)orjump to list(if available)

Show HN: A bi-directional, persisted KV store that is faster than Redis

Show HN: A bi-directional, persisted KV store that is faster than Redis

55 comments

·March 17, 2025

we've been working on a KV store for the past year or so which is 2-6x faster than Redis (benchmark link below) yet disk persisted! so you get the speed of in-memory KV stores but with disk persistence. To achieve this we've created our custom filesystem that is optimized for our special usecase and we're doing smart batching for writes and predictive fetching for reads.

In addition to basic operations, it also provides atomic inc/dec, atomic json patch, range scans and a unique key monitoring mechanism (pub-sub) over WebSockets which essentially allows you to receive notification on registered key changes directly from the KV store. so for example in a realtime web application, you can receive notifications directly in your front-end, with no back-end implementation (no WebSocket server management, no relay etc.) and still be secure and not expose your API keys on front-end. We have REST, WebSocket and RIOC API and we can't wait to hear your feedback.

We're only providing the free tier for now but let us know and we can increase the limits for you, if have a specific case. please either send us an email to support@hpkv.io or use http://hpkv.io/contact if you prefer that way.

sign up: http://hpkv.io/signup

documentation: http://hpkv.io/docs

realtime pub-sub: http://hpkv.io/blog/2025/03/real-time-pub-sub

benchmark vs Redis: http://hpkv.io/blog/2025/02/redis-vs-hpkv-benchmark

looking forward to hear your feedback :)

dangoodmanUT

What disks give 600ns persistence _with fsync/fdatasync_? Never heard of anything under 50us p50.

mehrant

the 600ns figure represents our optimized write path and not a full fsync operation. we achieve it -among other things- through:

1- as mentioned, we are not using any traditional filesystem and we're bypassing several VFS layers.

2- free space management is a combination of two RB trees, providing O(log n) for slice and O(log n + k) - k being the number of adjacent free spaces for merge.

3- majority of the write path employs a lock free design and where needed we're using per cpu write buffers

the transactional guarantees we provide is via:

1- atomic individual operations with retries

2- various conflict resolution strategies (timestamp, etc.)

3- durability through controlled persistence cycles with configurable commit intervals

depending on the plan, we provide persistence guarantee between 30 sec to 5 minutes

dangoodmanUT

I didn't necessarily mean exactly fsync. I guess I'll ask: Is it actually flushed to persistent disk in 600ns such that if the node crashes, the data can always be read again? Or does that not fully flush?

mehrant

yes, in that case data can potentially be lost. 30 sec in a worse case scenario without HA.

dangoodmanUT

And hold on, 600ns can't possibly be right...

A memory copy plus updating what ever internal memory structures you have is definitely going to be over 1us. Even a non-fsync NVMe write is still >=1us, so this is grossy misleading.

null

[deleted]

mehrant

our p50 is indeed 600ns for write, the way I explained it. I understand that at this point, this can be read as "trust me bro" kind of statement, but I can offer you something. we can have a quick call and I provide you access to a temp server with HPKV installed on it, with access to our test suit and you'll have a chance to run your own tests.

this can be a good learning opportunity for both of us (potentially more for us) :)

if you're interested, please send us an email to support@hpkv.io and we can arrange that

bjornsing

Interesting. I did some work on a related but different product idea (https://www.haystackdb.dev/) a few years back. Gave up though as it seemed hard to get traction / find customers. What’s your thinking on that? How are you going to reach your initial customers?

Would love to have a chat about possible collaboration or if I could help out in some way. Nice to see foundational tech coming out of the EU!

mehrant

thank you :) it would be interesting to have a chat for sure. would you mind dropping an email on the email I mentioned in OP and I'll reach out to you.

alex_smart

I don’t get it. How could you be fsyncing the WAL in 600ns? What are the transactional guarantees that you are offering?

mehrant

that's a great question. the 600ns figure represents our optimized write path and not a full fsync operation. we achieve it -among other things- through:

1- as mentioned, we are not using any traditional filesystem and we're bypassing several VFS layers.

2- free space management is a combination of two RB trees, providing O(log n) for slice and O(log n + k) - k being the number of adjacent free spaces for merge.

3- majority of the write path employs a lock free design and where needed we're using per cpu write buffers

the transactional guarantees we provide is via:

1- atomic individual operations with retries

2- various conflict resolution strategies (timestamp, etc.)

3- durability through controlled persistence cycles with configurable commit intervals

depending on the plan, we provide persistence guarantee between 30 sec to 5 minutes

buenzlikoder

What storage backend are you using?

A write operation on a SSD takes 10s of uS - without any VFS layers

mehrant

sorry for not being clear again. by saying this number does not represent full fsync operation, I meant it doesn't include the SSD write time. this is the time to update KVs internal memory structure + adding to write buffers.

this is fair because we provide transactional guarantee and immediate consistency, regardless of the state of the append-only write buffer entry. during that speed, for a given key, the value might change and a new write buffer entry might be added for the said key before the write buffer had the chance to complete (as you mentioned the actual write on disk is slower) but the conflict resolution still ensures the write of the last valid entry and skips the rest. before this operation HPKV is acting like an in-memory KV store.

edoceo

How will it be faster than my Redis or KeyVal which is very close if your servers are far away? Network time matters here, right?

mehrant

of course. the speeds down to 15us can be achieved over network over our custom protocol on the same region. for sub-microsecond latency, you need to have HPKV running on the same machine as yours :)

Snawoot

> 2-6x faster than Redis (benchmark link below) yet disk persisted!

That's a false contradistinction: Redis is also disk persisted.

The benchmark you did mentions Redis benchmarking guide and this guide has following paragraph:

> Redis is, mostly, a single-threaded server from the POV of commands execution (actually modern versions of Redis use threads for different things). It is not designed to benefit from multiple CPU cores. People are supposed to launch several Redis instances to scale out on several cores if needed. It is not really fair to compare one single Redis instance to a multi-threaded data store.

Did you just benchmarked against only single Redis instance and claimed performance win? Even if so, how do benchmarks compare against source-available competitor DragonflyDB?

Finally, documentation doesn't mention how persistence exactly works and what durability guarantees should we expect?

mehrant

thanks for taking time to write a feedback :)

> That's a false contradistinction: Redis is also disk persisted.

The performance gain mentioned was vs. Redis in memory. so we weren't claiming that Redis can't be persisted (which of course it can), but we were saying that Redis without persistence (which performs faster that with persistence) was still this much slower than HPKV with persistence. But you're correct that we probably should have been more clear in explaining this :)

>Did you just benchmarked against only single Redis instance and claimed performance win?

Signle node of Redis vs. Single node of HPKV. so it's an apples to apples comparison

>Even if so, how do benchmarks compare against source-available competitor DragonflyDB?

Benchmark with DragonFly coming soon :)

sorry about lack of that information in documentation, we'll update that. for for now, the durability guarantee on Pro is 30 seconds. on Business with HA is 5 minutes.

edoceo

Benchmark to KeyVal too please

mehrant

sure! :)

ForTheKidz

> That's a false contradistinction: Redis is also disk persisted.

This feels wildly disingenuous.

quibono

Is this 2-6x faster because of multi threading/core? Or is this actually 2-6x faster on a single core machine?

mehrant

the test was done on a single node and a single thread. on multi thread and batch operations, HPKV was still faster on the same machine

kshmir

Why pay what you're asking instead of using dragonfly or something like that and just putting a beefier node?

ehsanaslani

Well that's a technical choice depending on the context, but I can list some of the advantages of HPKV:

-Persistent by default without any performance penalties

-The pub/sub feature which is unique to HPKV and allows for a bi-directional websocket connection from clients to database

-Lower cost as we need less expensive infrastructure to provide the same service

-Simple API to use

cess11

Does it have ACID guarantees?

mehrant

We provide some elements of ACID guarantees, but not full ACID compliance as traditionally defined in database systems:

Atomicity: Yes, for individual operations. Each key-value operation is atomic (it either completes fully or not at all).

Consistency: Partial. We ensure data validity through our conflict resolution strategies, but we don't support multi-key constraints or referential integrity.

Isolation: Limited. Operations on individual keys are isolated, but we don't provide transaction isolation levels across multiple keys.

Durability: Yes. Our persistence model allows for tunable durability guarantees with corresponding performance trade-offs.

So while we provide strong guarantees for individual operations, HPKV is not a full ACID-compliant database system. We've optimized for high-performance key-value operations with practical durability assurances rather than complete ACID semantics.

theonlyvasudev

Amazing!

mehrant

thank you :)

tobyhinloopen

Not open source, not interested.

It looks neat though, but I won't burn myself on anything closed source if there's open source and/or self-hosted alternatives.

Aurornis

The comparison to Redis felt misleading when I realized it was a commercial product.

I'm not sure I like this recent trend of people registering new accounts on HN to use "Show HN" to pitch commercial products.

Show HN is fun to see projects people have been working on and want to share for the good of the community. It's less fun when it's being used as a simple advertising mechanism to drive purchases.

gkbrk

It's a suitable comparison I think. Redis is also not open-source.

https://github.com/redis/redis/blob/unstable/LICENSE.txt

old_bayes

As of 12 months ago, sure, but before that Redis was open source for 15 years.

pvg

Show HN is not limited to open source projects.

Aurornis

I never said it was. Seeing green accounts (just registered) doing nothing other than promote products doesn't feel consistent with the spirit of Show HN.

h1fra

agree, especially on something as common as Redis. Being locked in is not worth the better perf on something already super-fast.

But then I wonder what the business model is here? Even without being open-source, I'm constantly asking myself who pays for DBs that are not in a major cloud

tschellenbach

Doesn't always help. Price hikes on CockroachDB for instance have been crazy.

mehrant

thanks for taking time and commenting :) we'd still be happy if you decided to use it and give us your thoughts. as I mentioned in one of the comments below, we're hoping to go open source in future :)

verdverm

> we're hoping to go open source in future :)

That's just lip service, either you intend to or you don't, it's not up to hope

We hope you will see the dev tooling space is based around open source and you will alienate many potential users by not being open source

tobyhinloopen

Let me know when it's open source and I'd be happy to give it a try! It doesn't even have to be free - I'm fine with Unreal's model where you get access to the source even though commercial use requires a paid license.

I want something running locally on my machine that doesn't rely on calling home.

mehrant

your concern is understandable. we'll be in touch :)

hdjjhhvvhga

Give me one reason to use a closed-source Redis alternative rather than one of many open ones, starting with KeyDB. If I wanted a closed clone, I'd probably go with DragonflyDB (whose license is "feel free to run it in production unless you offer it as a managed redis service").

null

[deleted]

ranger_danger

Does not appear to be open-source.

mehrant

correct. however we're actually planning to make the system open source in future; we can't set an exact date as it depends on various factors, but hopefully not too far out. :)

notpushkin

Will be looking forward to that!

However, it feels a bit weird: at this level of performance going SaaS only kinda defeats the purpose, no?

mehrant

our approach is actually hybrid. on the other side of the performance coin, we have resource efficiency. that resource efficiency let's us provide a performant and low latency managed KV store, with lower cost, so the economy of it makes sense. the idea is that not everyone requires sub-microsecond latency, and for that group the value proposition is a low latency kv store which is feature rich with a novel bi-directional ws api. for people who need sub-microsecond latency, we're planning custom setup that allows them to make a local vectored interface call to get the sub-microsecond speeds. in between, we have the business plan that provides a custom binary protocol that is the one used in the benchmark :)

huhtenberg

In that case you need to provide an SLA for your speed claims. Otherwise the claims are basically moot.

mehrant

that's a fair point and you're correct. we will have the SLAs for latency documented and provided soon. in the mean time, please try it out and give us your feedback :)

null

[deleted]