GenDB is a Generative Query Engine that uses LLM agents to generate instance-optimized query execution code, tailored to your specific data, workloads, and hardware.
Five specialized LLM agents collaborate through a structured pipeline to generate optimized storage, indexes, and standalone native executables — all tailored to the specific data, workload, and hardware.
Profiles hardware, samples data, extracts workload characteristics
Designs layouts with encoding, compression, indexes, and zone maps
Generates resource-aware execution plans adapted to data and hardware
Implements plans as optimized native code with SIMD and parallelism
Iteratively refines code using runtime profiling feedback
Today, every new use case demands either a painful extension or an entirely new system:
Option 3 — Generate
Use LLMs to generate per-query execution code. No extension wrestling, no multi-year engineering. New techniques become reachable through prompt updates.
Instance-optimized code exploits exact data distributions, join selectivities, group cardinalities, and hardware characteristics. No general-purpose engine can match this.
Integrating new techniques requires prompting, not re-engineering. Semantic queries, GPU-native code — all reachable through prompt updates.
80% of queries repeat in 50% of clusters. Generation cost is amortized over many executions, making it cost-effective for recurring analytical workloads.
Total query execution time across all queries. GenDB variants use different LLM backbone models. All systems run on identical hardware with full parallelism enabled.
| # | System | Total Time | vs. Best GenDB | Relative |
|---|
| # | System | Total Time | vs. Best GenDB | Relative |
|---|
Different LLM backbone models offer different trade-offs between generated code quality, generation time, and cost. Ranked by average query execution time.
GenDB is under active development. Every step follows three principles:
Multi-agent pipeline for analytical queries. Evaluated on TPC-H and SEC-EDGAR, outperforming DuckDB, Umbra, ClickHouse, MonetDB, and PostgreSQL.
Agents learn from past runs, accumulate optimization experience, and improve generation quality over time — without retraining the underlying LLMs.
Generate CUDA and GPU-accelerated code targeting libcudf for cost-efficient GPU analytics, not just CPU.
Generate code for multimodal data — images, audio, text — with AI-powered operators, moving beyond SQL’s relational model.
Reusable operators across queries, query template generation, hybrid execution with traditional DBMS, and further cost reduction as LLMs become faster and cheaper.