18.6M records/sec ingestion | 6M+ rows/sec queries | Portable Parquet storage

High-Performance Columnar
Analytical Database

Fast ingestion. Fast queries. Built on DuckDB and Parquet.

Use for product analytics, observability, AI agents, IoT, or data warehousing. Single binary. No vendor lock-in.

Get started in seconds
docker run -d -p 8000:8000 \
  -e STORAGE_BACKEND=local \
  -v arc-data:/app/data \
  ghcr.io/basekick-labs/arc:latest

Single binary. No dependencies. Production-ready in minutes.

240+ deployments
Open source (AGPL-3.0)

Your Data Deserves Better Than Legacy Databases

Billions of events per day. Millions of records per second. Sub-second queries across all of it.

Traditional databases weren't built for this. Arc was.

Real Production Scale:

Product analytics: Billions of user events per day

Observability: Millions of logs, metrics, and traces per second

AI agents: Millions of conversation and context records

IoT & telemetry: Billions of sensor readings across thousands of devices

Built for billion-record workloads.

Why Arc

DuckDB SQL. No Proprietary Query Language.

Not a custom DSL. Not a query language that changes every major version.

Not vendor lock-in through proprietary syntax.

Just DuckDB-powered SQL with window functions, CTEs, and joins.

Track user behavior and conversion funnels
SELECT
  user_id,
  event_type,
  COUNT(*) OVER (
    PARTITION BY user_id
    ORDER BY timestamp
    ROWS BETWEEN UNBOUNDED PRECEDING AND CURRENT ROW
  ) as event_sequence,
  LEAD(event_type) OVER (
    PARTITION BY user_id
    ORDER BY timestamp
  ) as next_event,
  DATEDIFF('second', timestamp, LEAD(timestamp) OVER (
    PARTITION BY user_id ORDER BY timestamp
  )) as time_to_next_event
FROM analytics.events
WHERE timestamp > NOW() - INTERVAL '7 days'
  AND event_type IN ('page_view', 'add_to_cart', 'checkout', 'purchase')
ORDER BY user_id, timestamp;

If you know SQL, you know Arc. Powered by DuckDB.

Window Functions

Moving averages, ranking, and complex aggregations built-in

CTEs & Subqueries

Break down complex analysis into readable, composable parts

JOINs Across Tables

Correlate events, logs, metrics, and any data across tables

Parquet files you actually own

Your data lives in standard Parquet files on S3, MinIO, or local disk.

Arc disappears tomorrow? You still own your data.

Query it with DuckDB, ClickHouse, Snowflake, or any tool that reads Parquet.

Zero proprietary formats. Zero lock-in.

This is what "portable data" actually means.

S3 / MinIO / Local

Store anywhere you want

Standard Parquet

Industry standard format

You Own It

No vendor lock-in, ever

Performance

18.6M records/sec sustained

High-throughput ingestion and analytical queries via MessagePack columnar format.

No degradation. No memory leaks. Just stable performance.

18.6M
records/sec
sustained ingestion
6.29M
rows/sec
query throughput
0.47ms
p50 latency
p99: 3.51ms

Hardware: M3 Max 14-core, 36GB RAM

Built for mission-critical workloads

From product analytics to IoT telemetry—Arc handles the scale and performance requirements that matter.

Product Analytics

Track user events, clickstreams, and A/B tests. Query billions of events in seconds with DuckDB SQL. Window functions, CTEs, and complex aggregations out of the box.

User EventsClickstreamsA/B Testing

Observability

Unified storage for logs, metrics, and traces. Fast queries across all telemetry types. 18.6M writes/sec, 6M+ reads/sec. One database instead of three.

LogsMetricsTraces

AI Agent Memory

Store conversation history, context, and embeddings. Columnar architecture for agent memory with fast temporal queries. RAG-ready.

ConversationsContextEmbeddings

IoT & Telemetry

Sensor telemetry, manufacturing data, equipment monitoring. Native MQTT support. Multi-year retention with automatic compaction and 90% storage reduction.

SensorsMQTTTelemetry
Migration

Migrate from Your Existing Database

We help with migration at no cost.

We'll help you migrate at no cost.

Arc speaks InfluxDB Line Protocol natively.

Point Telegraf at Arc. Dual-write during migration. Cut over when ready.

No agent changes. No downtime. No data loss.

Multiple InfluxDB 1.x Enterprise customers already testing migration.

We'll help you migrate at no cost.

Keep your SQL queries. Arc uses standard DuckDB SQL with window functions, CTEs, and joins.

Own your data in Parquet. No vendor lock-in. Query with any tool. 10-50x faster queries after compaction.

We'll help you migrate at no cost.

Broader SQL support. Full window functions, CTEs, complex joins — features QuestDB doesn't support.

Better ecosystem integration. Native Grafana datasource, VSCode extension, Apache Superset dialect.

We'll help you migrate at no cost.

Simpler operations. Single binary vs complex cluster management. No ZooKeeper, no replica coordination.

Faster ingestion. 18M+ rec/s vs ClickHouse's 1-2M rec/s. Better for high-throughput workloads.

Same SQL. Keep your analytical queries. DuckDB SQL with window functions, CTEs, joins.

Portable data. Parquet format. No vendor lock-in. Query with any tool.

We'll help you migrate at no cost.

10-100x cheaper. Self-hosted on your infrastructure. No cloud warehouse bills.

Portable data. Parquet format. No vendor lock-in. Move your data anytime.

Same SQL. DuckDB SQL compatible with most Snowflake/BigQuery queries.

Single binary. No cluster complexity. Deploy in minutes.

We'll help you migrate at no cost.

Better SQL. DuckDB vs Elasticsearch DSL. Window functions, CTEs, complex joins.

Faster ingestion. 18M+ rec/s vs Elasticsearch's ~100K-500K rec/s.

Cheaper storage. Parquet on S3 vs Elasticsearch indices. 90% reduction.

Keep your tools. Native Grafana datasource, Superset dialect, API compatibility.

We'll help you migrate at no cost.

Simpler operations. Single binary vs complex cluster (coordinator, broker, historical, real-time nodes).

Better SQL. Full DuckDB SQL vs limited Druid SQL. Window functions, CTEs, subqueries.

Portable data. Parquet format vs proprietary segments. No vendor lock-in.

Faster ingestion. 18M+ rec/s sustained throughput.

Features That Matter

Grafana Integration

Official Grafana datasource plugin. Build dashboards for analytical data, real-time metrics, and operational monitoring.

Setup guide ->

VSCode Extension

Full-featured database manager. Query editor with autocomplete. Notebooks for analysis. CSV import.

Install from marketplace ->

Automatic Compaction

Small files merge into optimized 512MB Parquet files. 10-50x faster queries with zero configuration.

Retention Policies

Time-based lifecycle management. Keep 7 days of raw data, 90 days of rollups, 2 years of aggregates.

GDPR-Compliant Deletion

Precise record deletion with file rewrites. No tombstones. No query overhead.

Write-Ahead Log (WAL)

Optional durability for zero data loss. Disabled by default for maximum throughput.

Multi-Database Architecture

Organize by project, team, or environment. Isolated namespaces for multi-tenant deployments.

Apache Superset Integration

Native dialect for BI dashboards. Connect your existing visualization tools.

Learn more ->

Frequently Asked Questions

Is this production ready?

Yes. Version 25.12.1 is stable and ready for production use.

Self-hosted Arc is production-ready now.

Enterprise features available for large deployments.

What about data retention and storage costs?

Parquet compression gives 3-5x reduction vs raw data.

S3 storage costs ~$0.023/GB/month. 1TB = ~$23/month.

Built-in retention policies automatically delete old data. Or keep everything — storage is cheap, queries are fast.

Can I run this in high availability?

Single-node architecture in open source version.

Arc Enterprise includes clustering and high availability.

For OSS: Run primary + standby with object storage replication. WAL enabled for zero data loss during failover.

What about authentication and access control?

API token authentication included in all versions.

Grafana integration supports Azure AD OAuth.

Arc Enterprise includes RBAC, SSO, and multi-tenancy.

Pricing

Start free with open source. Scale with enterprise features when you need them.

Open Source

Freeforever
AGPL-3.0 licensed
  • 18M records/sec ingestion
  • Full SQL query engine (DuckDB)
  • Parquet storage (S3, GCS, local)
  • Docker and Kubernetes ready
  • Community support (Discord)
Coming Q2 2026

Enterprise

$5,000/year

Starting price for up to 8 cores. Clustering, RBAC, and dedicated support.

  • Everything in Open Source
  • Horizontal clustering and HA
  • Role-based access control (RBAC)
  • Tiered storage and auto-aggregation
  • Dedicated support and SLAs
View all plans ->

Enterprise Features

Clustering

Horizontal scaling with automatic data distribution. Query routing and load balancing across nodes.

Security

Fine-grained RBAC with database and table-level permissions. LDAP/SAML integration available.

Data Management

Automated retention policies, continuous queries for aggregation, and tiered storage for cost optimization.

Get Started

Open Source Installation

# Docker (easiest)
docker run -d -p 8000:8000 \
  -e STORAGE_BACKEND=local \
  -v arc-data:/app/data \
  ghcr.io/basekick-labs/arc:latest
# Send data (InfluxDB Line Protocol)
curl -X POST http://localhost:8000/api/v1/write/line-protocol \
  -H "Authorization: Bearer YOUR_TOKEN" \
  -H "Content-Type: text/plain" \
  -H "x-arc-database: default" \
  --data-binary "cpu,host=server01 value=0.64"
# Query with SQL
curl -X POST http://localhost:8000/api/v1/query \
  -H "Authorization: Bearer YOUR_TOKEN" \
  -H "Content-Type: application/json" \
  -d '{"sql":"SELECT * FROM cpu LIMIT 10"}'

Arc Enterprise

For organizations requiring clustering, RBAC, and dedicated support.

Multi-node clustering and high availability

Role-based access control (RBAC) and SSO

Dedicated support and SLAs

Commercial licensing available

enterprise@basekick.net

Ready to handle billion-record workloads?

Deploy Arc in minutes. Own your data in Parquet. Use for analytics, observability, AI, IoT, or data warehousing.

Get Started ->