Skip to content
/ buffdb Public

Embedded storage built for multiplexing. Smart machines don't need to read JSON, they only need protocol buffers. The world's first MODMS (Machine-Oriented Database Management System), built to support RocksDB, SQLite, and DuckDB as backends.

License

Apache-2.0, MIT licenses found

Licenses found

Apache-2.0
LICENSE-APACHE
MIT
LICENSE-MIT
Notifications You must be signed in to change notification settings

buffdb/buffdb

Repository files navigation

image

🦁 buffdb 🦁

buffdb is experimental software. Join buffdb’s Discord for help and have a look at things that don’t work yet. Many basic things are not yet decided.

BuffDB is a lightweight, high-performance persistence layer for gRPC written in Rust. When using the SQLite backend and dynamic linking, the binary is under 2 MiB. It is designed to be used in environments where bandwidth is limited and performance is critical. Protobuf is the messaging format of choice, but other formats are planned.

To add BuffDB to your Rust project, run cargo add buffdb.

TypeScript interaction

Using auto-generated code from the protobuf definition, we can interact with the server in many languages. The following example demonstrates how to interact with the server in TypeScript. The server is assumed to be running on port 50051. See the examples directory for demos in other languages.

import * as grpc from "@grpc/grpc-js";
import * as protoLoader from "@grpc/proto-loader";
import type { ProtoGrpcType as BlobType } from "./proto/blob";
import type { StoreResponse } from "./proto/buffdb/blob/StoreResponse";

const proto = grpc.loadPackageDefinition(
    protoLoader.loadSync("../../proto/blob.proto")
) as unknown as BlobType;

const blob_client = new proto.buffdb.blob.Blob(
    "[::1]:50051",
    grpc.credentials.createInsecure()
);
const get = blob_client.Get();
const store = blob_client.Store();

// Be sure to set up the listeners before writing data!

get.on("data", (data) => console.log("received data from GET: ", data));
get.on("end", () => console.log("stream GET ended"));

const blob_ids: number[] = [];
store.on("data", (raw_id: StoreResponse) => {
    const id = (raw_id.id as protoLoader.Long).toNumber();
    console.log("received data from STORE: ", id);
    blob_ids.push(id);
});
store.on("end", () => console.log("stream STORE ended"));

store.write({ bytes: Buffer.from("abcdef"), metadata: "{ offset: 6 }" });
store.write({ bytes: Buffer.from("ghijkl") });

store.end();

// Give the store time to finish its operations before asking for data back.
// We could also do this in the callback of other events to be certain that it's been inserted.
setTimeout(() => {
    for (const id of blob_ids) {
        console.log(`requesting ${id}`);
        get.write({ id });
    }
    get.end();
}, 100);

This example is present in examples/typescript. To run it, you need to have Node.js installed. Run npm i to install the dependencies and npm run exec to run the example.

How to use

Supported backends

Backend Support status Raw query support Feature flag (vendored) CLI flag
SQLite Full support sqlite (vendored-sqlite) -b sqlite
DuckDB Partial duckdb (vendored-duckdb) -b duckdb
RocksDB Partial (vendored-rocksdb) only -b rocksdb

Blockers for full DuckDB support include duckdb/duckdb-rs#368, but other issues are necessary to have best performance.

By default, all backends are included and vendored. To exclude a backend, use the --no-default-features flag with cargo and re-enable the desired backend with --features. If you encounter unexpected errors, consider using a vendored backend.

Server

To run the server, you need to have Rust installed. Then, with the repository cloned, you can run

cargo run --all-features -- run

This will start the server on [::1]:50051, storing the key-value pairs in kv_store.db and the blob data in blob_store.db. All three can be configured with command line flags: --addr, --kv-store, and --blob-store respectively.

To build with optimizations enabled, run cargo build --all-features --release. The resulting binary will be located at target/release/buffdb. It is statically linked (excluding the backends depending on flags), so it can be moved anywhere on your file system without issue.

Prefer to handle the gRPC server yourself? buffdb can be used as a library as well!

Command line interface

You can use buffdb help to see the commands and flags permitted. The following operations are currently supported:

  • buffdb run [ADDR], starting the server. The default address is [::1]:50051.
  • buffdb kv get <KEY>, printing the value to stdout.
  • buffdb kv set <KEY> <VALUE>, setting the value.
  • buffdb kv delete <KEY>, deleting the value.
  • buffdb kv eq [KEYS]..., exiting successfully if the values for all provided keys are equal. Exits with an error code if any two values are not equal.
  • buffdb kv not-eq [KEYS]..., exiting successfully if the values for all provided keys are unique. Exits with an error code if any two values are equal.
  • buffdb blob get <ID>, printing the data to stdout. Note that this is arbitrary bytes!
  • buffdb blob store <FILE> [METADATA], storing the file (use - for stdin) and printing the ID to stdout. Metadata is optional.
  • buffdb blob update <ID> data <FILE>, updating the data of the blob. Use - for stdin. Metadata is unchanged.
  • buffdb blob update <ID> metadata [METADATA], updating the metadata of the blob. Data is unchanged. Omitting [METADATA] will set the metadata to null.
  • buffdb blob update <ID> all <FILE> [METADATA], updating both the data and metadata of the blob. For <FILE>, use - for stdin. Omitting [METADATA] will set the metadata to null.
  • buffdb blob delete <ID>, deleting the blob.
  • buffdb blob eq-data [IDS]..., exiting successfully if the blobs for all provided IDs are equal. Exits with an error code if any two blobs are not equal.
  • buffdb blob not-eq-data [IDS]..., exiting successfully if the blobs for all provided IDs are unique. Exits with an error code if any two blobs are equal.

Commands altering a store will exit with an error code if the key/id does not exist. An exception to this is updating the metadata of a blob to be null, as it is not required to exist beforehand.

All commands for kv and blob can use -s/--store to specify which store to use. The defaults are kv_store.db and blob_store.db respectively. To select a backend, use -b/--backend. The default varies by which backends are enabled.

Library usage in Rust

Run cargo add buffdb tonic tokio futures to add the necessary dependencies. Then you can execute the following code, which is placed in src/main.rs.

use buffdb::backend::Sqlite;
use buffdb::kv::{Key, KeyValue, Value};
use tonic::{Request, IntoRequest};
use futures::{stream, StreamExt as _};

#[tokio::main]
async fn main() -> Result<(), Box<dyn std::error::Error>> {
    let mut client = buffdb::transitive::kv_client::<_, Sqlite>("kv_store.db").await?;
    client
        .set(stream::iter([KeyValue {
            key: "key_set".to_owned(),
            value: "value_set".to_owned(),
        }]))
        .await?
        .into_inner();

    let mut stream = client
        .get(stream::iter([Key {
            key: "key_get".to_owned(),
        }]))
        .await?
        .into_inner();
    let Value { value } = stream.next().await.unwrap()?;
    assert_eq!(value, "value_get");

    Ok(())
}

This project is inspired by conversations with Michael Cahill, Professor of Practice, School of Computer Science, University of Sydney

Background

This project was inspired by our many edge customers of ours dealing with the challenges associated with low-bandwidth and high performance. We hope that we can build a solution that is helpful for teams tageting edge computing environments.

Today, buffdb’s primary focus is speed: we try to ensure some level of durability for which we pay a performance penalty, but our goal is to eventually be faster than any other embedded database.

High-level Goals

  • Reducing the overhead of serialization/deserialization.
  • Ensuring consistent data formats between local storage and network communication.
  • Providing faster read/write operations compared to JSON or XML.
  • Compact Data Storage: ProtoBufs can significantly reduce the size of stored data.
  • Interoperability: Seamless integration between the app’s local storage and backend systems.

Use Cases

  • Offline Data Access: For apps that need to function offline (e.g., note-taking apps, games, fieldwork, airline, collaborative documents, etc.).
  • IoT: For managing device configurations and states locally before syncing with cloud servers.

About

Embedded storage built for multiplexing. Smart machines don't need to read JSON, they only need protocol buffers. The world's first MODMS (Machine-Oriented Database Management System), built to support RocksDB, SQLite, and DuckDB as backends.

Topics

Resources

License

Apache-2.0, MIT licenses found

Licenses found

Apache-2.0
LICENSE-APACHE
MIT
LICENSE-MIT

Stars

Watchers

Forks

Packages

No packages published

Languages