-
Updated
Sep 18, 2021 - Go
distributed-database
Here are 234 public repositories matching this topic...
Describe the solution you'd like
Ability to export to Apache Parquet format.
EXPORT INTO PARQUET
'azure://acme-co/customer-export-data?AZURE_ACCOUNT_KEY=hash&AZURE_ACCOUNT_NAME=acme-co'
FROM TABLE bank.customers;
Describe alternatives you've considered
Using CSV format
Additional context
Parquet stores the file sc
-
Updated
Sep 18, 2021 - C++
Please add unit test for this method:
public void persist(final String schemaName) {
repository.persist(SchemaMetadataNode.getSchemaNamePath(schemaName), "");
}
-
Updated
Mar 28, 2021
-
Updated
Sep 17, 2021 - JavaScript
Our CMake build process produces tarball packages using the cpack TGZ
target. This produces a server and client package with a file layout meant to match where the various artifacts go.
Meanwhile, the packages we include on the download page are a flat archive of various binaries produced by the build process. This include one regular package and one debug package with debug symbols for a few
-
Updated
Aug 25, 2021
-
Updated
Sep 17, 2021 - Java
Hello Philip!
I think there is an issue with this part of the code of rqlite (store/store.go).
func (s *Store) Database(leader bool) ([]byte, error) {
if leader && s.raft.State() != raft.Leader {
return nil, ErrNotLeader
}
// Ensure only one snapshot can take place at once, and block all queries.
s.mu.Lock()
defer s.mu.Unlock()
f, err := ioutil.TempFile("", "rqlilte-snap-
-
Updated
Sep 9, 2021 - Rust
ysqlsh history file can be specified by --variable=HISTFILE=/path/to/file
. It seems that the default amount of entries (or lines) it keeps is 500. Once it goes past that, upon history file write, it truncates the file. However, it also truncates the header line _HiStOrY_V2_
. This makes the file invalid next time it's read. This is possibly related to YB using libedit rather than libreadli
-
Updated
Sep 17, 2021 - C
With Hive connector
trino:default> CREATE TABLE one (a varchar);
-> CREATE VIEW two AS SELECT * FROM one;
CREATE TABLE
CREATE VIEW
DROP TABLE
is rejected on a view:
trino:default> DROP TABLE two;
Query 20210906_150832_00015_id3y3 failed: line 1:1: Table 'hive.default.two' does not exist, but a view with that name exists. Did you mean DROP VIEW hive.default.t
Use case:
1.) A user may want to backup all tables but no metadata like users, privileges, etc. without explicitly defining each table inside the CREATE SNAPSHOT
statement.
2.) A user may want to transfer users & privileges, custom analyzers or user-defined-functions from one cluster to another without backing up a complete cluster including all data (tables).
*Feature description
-
Updated
May 10, 2021 - Ruby
-
Updated
Sep 17, 2021 - Go
Currently we don't have any mechanism to limit the maximum number of clients that could be handled simultaneously.
This feature should be designed properly. Here is some clue: https://redis.io/topics/clients#maximum-number-of-clients
-
Updated
Aug 6, 2021 - Go
-
Updated
Sep 18, 2021 - C++
-
Updated
Jul 17, 2020 - C++
-
Updated
May 6, 2017 - C++
-
Updated
Sep 18, 2021 - Java
-
Updated
Sep 17, 2021 - C++
Now that we support more than 1 data type, we should include the values data type in cache nodes. This give us a way of easily identifying what type of data a key points to e.g. string, queue (and more types in the future)
Currently a cache node object consists of a Key
, a Value
and a TTL
. We should also add a field specifying the type of the Value
-
Updated
Sep 16, 2021 - Go
-
Updated
Nov 5, 2019
-
Updated
Sep 17, 2021 - Java
Improve this page
Add a description, image, and links to the distributed-database topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the distributed-database topic, visit your repo's landing page and select "manage topics."