(no title)
stefanha | 4 months ago
Also, were existing network or distributed file systems not suitable? This use case sounds like Ceph might fit, for example.
stefanha | 4 months ago
Also, were existing network or distributed file systems not suitable? This use case sounds like Ceph might fit, for example.
graveland|4 months ago
Entirely programmable storage so far has allowed us to try a few different things to try and make things efficient and give us the features we want. We've been able to try different dedup methods, copy-on-write styles, different compression methods and types, different sharding strategies... All just as a start. We can easily and quickly create a new experimental storage backends and see exactly how pg performs with it side-by-side with other backends.
We're a kubernetes shop, and we have our own CSI plugin, so we can also transparently run a pg HA pair with one pg server using EBS and the other running in our new storage layer, and easily bounce between storage types with nothing but a switchover event.
yencabulator|3 months ago
But you think you have resources to maintain a distributed strongly-consistent replicating block store?
The edge cases in RDB are literally why Ceph takes expertise to manage! Things like failure while recovering from failure while trying to maintain performance are inherently tricky.
adsharma|4 months ago
You can probably hire people to maintain it.
Was it the ramp-up cost or expertise?
kjetijor|4 months ago
adsharma|4 months ago
DAOS seemed promising a couple of years ago. But in terms of popularity it seems to be stuck. No Ubuntu packages, no wide spread deployment, Optane got killed.
Yet the NVMe + metadata approach seemed promising.
Would love to see more databases fork it to do what you need from it.
Or if folks have looked at it and decided not to do it, an analysis of why would be super interesting.
adsharma|4 months ago