distributed-database
Here are 232 public repositories matching this topic...
Describe the solution you'd like
Ability to export to Apache Parquet format.
EXPORT INTO PARQUET
'azure://acme-co/customer-export-data?AZURE_ACCOUNT_KEY=hash&AZURE_ACCOUNT_NAME=acme-co'
FROM TABLE bank.customers;
Describe alternatives you've considered
Using CSV format
Additional context
Parquet stores the file sc
-
Updated
Aug 17, 2021 - C++
-
Updated
Mar 28, 2021
-
Updated
Aug 17, 2021 - JavaScript
Our CMake build process produces tarball packages using the cpack TGZ
target. This produces a server and client package with a file layout meant to match where the various artifacts go.
Meanwhile, the packages we include on the download page are a flat archive of various binaries produced by the build process. This include one regular package and one debug package with debug symbols for a few
-
Updated
Aug 7, 2021
-
Updated
Aug 16, 2021 - Java
Hello Philip!
I think there is an issue with this part of the code of rqlite (store/store.go).
func (s *Store) Database(leader bool) ([]byte, error) {
if leader && s.raft.State() != raft.Leader {
return nil, ErrNotLeader
}
// Ensure only one snapshot can take place at once, and block all queries.
s.mu.Lock()
defer s.mu.Unlock()
f, err := ioutil.TempFile("", "rqlilte-snap-
-
Updated
Jul 28, 2021 - Rust
Right now it is a cumulative metric that shows the number of bootstraps since the lifetime of the cluster
cc @tsmull-11
-
Updated
Aug 17, 2021 - C
If the --server
option is used without a protocol, then it should use https
when on port 443. For example, these invocations would be equivalent, with the first one having the new behavior:
trino --server example.net:443
trino --server https://example.net:443
trino --server https://example.net
This will make the CLI consistent with the JDBC driver in this regard. While it's t
Use case:
1.) A user may want to backup all tables but no metadata like users, privileges, etc. without explicitly defining each table inside the CREATE SNAPSHOT
statement.
2.) A user may want to transfer users & privileges, custom analyzers or user-defined-functions from one cluster to another without backing up a complete cluster including all data (tables).
*Feature description
-
Updated
May 10, 2021 - Ruby
Currently we don't have any mechanism to limit the maximum number of clients that could be handled simultaneously.
This feature should be designed properly. Here is some clue: https://redis.io/topics/clients#maximum-number-of-clients
-
Updated
Aug 16, 2021 - Go
-
Updated
Aug 6, 2021 - Go
-
Updated
Aug 17, 2021 - C++
-
Updated
Jul 17, 2020 - C++
-
Updated
May 6, 2017 - C++
-
Updated
Aug 12, 2021 - Java
Now that we support more than 1 data type, we should include the values data type in cache nodes. This give us a way of easily identifying what type of data a key points to e.g. string, queue (and more types in the future)
Currently a cache node object consists of a Key
, a Value
and a TTL
. We should also add a field specifying the type of the Value
wrong time display
-
Updated
Nov 5, 2019
-
Updated
Aug 17, 2021 - Java
-
Updated
Aug 24, 2020 - Go
Improve this page
Add a description, image, and links to the distributed-database topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the distributed-database topic, visit your repo's landing page and select "manage topics."
"found" by the @discordapp troops the hard way: https://status.discordapp.com/incidents/62gt9cgjwdgf
pinging @zorkian who pointed it out to me.
trivial to repro, this is against a 3.3.13 on fedora-31.