I need to decide which database to use for a system where I need AP from CAP theorem. Data I constantly but slowly going in. Big queries are expected. It should be reliable - no single point of failure. I can use up to 3 instances on different nodes. In-memory solutions are bad for me because of data size - it should be running for years and I expect up to terabyte data sizes. Most guys in my team prefer SQL. But I understand that traditional SQL databases are not fault tolerant in terms of hardware failure. Any ideas?
-
I need horizontal scaling as I understandPavlo– Pavlo2016-12-19 14:44:36 +00:00Commented Dec 19, 2016 at 14:44
-
2getyarn.io/yarn-clip/ac5b054e-4f2a-4e15-a810-094a2553fe44John Boker– John Boker2016-12-19 14:46:03 +00:00Commented Dec 19, 2016 at 14:46
-
postgres-xl.orguser330315– user3303152016-12-19 14:54:13 +00:00Commented Dec 19, 2016 at 14:54
-
Pretty complex in configuration and different nodes with different roles. Maybe some other solution?Pavlo– Pavlo2016-12-19 18:03:42 +00:00Commented Dec 19, 2016 at 18:03
-
With your requirements maybe your developers who prefer SQL need to learn something new.John Boker– John Boker2016-12-19 18:13:54 +00:00Commented Dec 19, 2016 at 18:13
|
Show 5 more comments
2 Answers
Since this question was asked there have been some significant changes in the Distributed SQL or NewSQL landscape...the most noteworthy being the viability of CockroachDB. That appears to be the best option in a situation like the one referenced in this question.
No single points of failure. Easy to scale. Can handle tons of volume. You can run it wherever you want. Speaks postgres. Super fault tolerant.
1 Comment
MAbraham1
Hi Daniel, thanks for that answer. This topic also has many related answers on SO. I recommend that you also add a link to one of the more popular answers, here: stackoverflow.com/q/1145726/212950