High-Performance Columnar
Analytical Database
Fast ingestion. Fast queries. Built on DuckDB and Parquet.
Use for product analytics, observability, AI agents, IoT, or data warehousing. Single binary. No vendor lock-in.
docker run -d -p 8000:8000 \
-e STORAGE_BACKEND=local \
-v arc-data:/app/data \
ghcr.io/basekick-labs/arc:latestSingle binary. No dependencies. Production-ready in minutes.
Your Data Deserves Better Than Legacy Databases
Billions of events per day. Millions of records per second. Sub-second queries across all of it.
Traditional databases weren't built for this. Arc was.
Real Production Scale:
Product analytics: Billions of user events per day
Observability: Millions of logs, metrics, and traces per second
AI agents: Millions of conversation and context records
IoT & telemetry: Billions of sensor readings across thousands of devices
Built for billion-record workloads.
DuckDB SQL. No Proprietary Query Language.
Not a custom DSL. Not a query language that changes every major version.
Not vendor lock-in through proprietary syntax.
Just DuckDB-powered SQL with window functions, CTEs, and joins.
SELECT
user_id,
event_type,
COUNT(*) OVER (
PARTITION BY user_id
ORDER BY timestamp
ROWS BETWEEN UNBOUNDED PRECEDING AND CURRENT ROW
) as event_sequence,
LEAD(event_type) OVER (
PARTITION BY user_id
ORDER BY timestamp
) as next_event,
DATEDIFF('second', timestamp, LEAD(timestamp) OVER (
PARTITION BY user_id ORDER BY timestamp
)) as time_to_next_event
FROM analytics.events
WHERE timestamp > NOW() - INTERVAL '7 days'
AND event_type IN ('page_view', 'add_to_cart', 'checkout', 'purchase')
ORDER BY user_id, timestamp;If you know SQL, you know Arc. Powered by DuckDB.
Window Functions
Moving averages, ranking, and complex aggregations built-in
CTEs & Subqueries
Break down complex analysis into readable, composable parts
JOINs Across Tables
Correlate events, logs, metrics, and any data across tables
Parquet files you actually own
Your data lives in standard Parquet files on S3, MinIO, or local disk.
Arc disappears tomorrow? You still own your data.
Query it with DuckDB, ClickHouse, Snowflake, or any tool that reads Parquet.
This is what "portable data" actually means.
S3 / MinIO / Local
Store anywhere you want
Standard Parquet
Industry standard format
You Own It
No vendor lock-in, ever
18.6M records/sec sustained
High-throughput ingestion and analytical queries via MessagePack columnar format.
No degradation. No memory leaks. Just stable performance.
Hardware: M3 Max 14-core, 36GB RAM
See Arc in Action
Live demos running on Arc. Real data, real-time ingestion, sub-second queries.
Satellite Tracking
Loading satellite data...
Explore more demos
NYC Citibike
13 years of bike share data. 153M+ rides across New York City with interactive filtering.
Vessel Tracking
Real-time AIS data from ships in Singapore's Strait of Malacca.
Flight Tracking
Live ADS-B aircraft positions over New York City airspace.
Weather
Multi-city weather monitoring — Buenos Aires, London, and Tokyo.
System Metrics
Real-time Docker container CPU, memory, network, and disk I/O.
Built for mission-critical workloads
From product analytics to IoT telemetry—Arc handles the scale and performance requirements that matter.
Product Analytics
Track user events, clickstreams, and A/B tests. Query billions of events in seconds with DuckDB SQL. Window functions, CTEs, and complex aggregations out of the box.
Observability
Unified storage for logs, metrics, and traces. Fast queries across all telemetry types. 18.6M writes/sec, 6M+ reads/sec. One database instead of three.
AI Agent Memory
Store conversation history, context, and embeddings. Columnar architecture for agent memory with fast temporal queries. RAG-ready.
IoT & Telemetry
Sensor telemetry, manufacturing data, equipment monitoring. Native MQTT support. Multi-year retention with automatic compaction and 90% storage reduction.
Migrate from Your Existing Database
We help with migration at no cost.
We'll help you migrate at no cost.
Arc speaks InfluxDB Line Protocol natively.
Point Telegraf at Arc. Dual-write during migration. Cut over when ready.
No agent changes. No downtime. No data loss.
Multiple InfluxDB 1.x Enterprise customers already testing migration.
We'll help you migrate at no cost.
Keep your SQL queries. Arc uses standard DuckDB SQL with window functions, CTEs, and joins.
Own your data in Parquet. No vendor lock-in. Query with any tool. 10-50x faster queries after compaction.
We'll help you migrate at no cost.
Broader SQL support. Full window functions, CTEs, complex joins — features QuestDB doesn't support.
Better ecosystem integration. Native Grafana datasource, VSCode extension, Apache Superset dialect.
We'll help you migrate at no cost.
Simpler operations. Single binary vs complex cluster management. No ZooKeeper, no replica coordination.
Faster ingestion. 18M+ rec/s vs ClickHouse's 1-2M rec/s. Better for high-throughput workloads.
Same SQL. Keep your analytical queries. DuckDB SQL with window functions, CTEs, joins.
Portable data. Parquet format. No vendor lock-in. Query with any tool.
We'll help you migrate at no cost.
10-100x cheaper. Self-hosted on your infrastructure. No cloud warehouse bills.
Portable data. Parquet format. No vendor lock-in. Move your data anytime.
Same SQL. DuckDB SQL compatible with most Snowflake/BigQuery queries.
Single binary. No cluster complexity. Deploy in minutes.
We'll help you migrate at no cost.
Better SQL. DuckDB vs Elasticsearch DSL. Window functions, CTEs, complex joins.
Faster ingestion. 18M+ rec/s vs Elasticsearch's ~100K-500K rec/s.
Cheaper storage. Parquet on S3 vs Elasticsearch indices. 90% reduction.
Keep your tools. Native Grafana datasource, Superset dialect, API compatibility.
We'll help you migrate at no cost.
Simpler operations. Single binary vs complex cluster (coordinator, broker, historical, real-time nodes).
Better SQL. Full DuckDB SQL vs limited Druid SQL. Window functions, CTEs, subqueries.
Portable data. Parquet format vs proprietary segments. No vendor lock-in.
Faster ingestion. 18M+ rec/s sustained throughput.
Features That Matter
Grafana Integration
Official Grafana datasource plugin. Build dashboards for analytical data, real-time metrics, and operational monitoring.
Setup guide ->VSCode Extension
Full-featured database manager. Query editor with autocomplete. Notebooks for analysis. CSV import.
Install from marketplace ->Automatic Compaction
Small files merge into optimized 512MB Parquet files. 10-50x faster queries with zero configuration.
Retention Policies
Time-based lifecycle management. Keep 7 days of raw data, 90 days of rollups, 2 years of aggregates.
GDPR-Compliant Deletion
Precise record deletion with file rewrites. No tombstones. No query overhead.
Write-Ahead Log (WAL)
Optional durability for zero data loss. Disabled by default for maximum throughput.
Multi-Database Architecture
Organize by project, team, or environment. Isolated namespaces for multi-tenant deployments.
Apache Superset Integration
Native dialect for BI dashboards. Connect your existing visualization tools.
Learn more ->Frequently Asked Questions
Is this production ready?
Yes. Version 25.12.1 is stable and ready for production use.
Self-hosted Arc is production-ready now.
Enterprise features available for large deployments.
What about data retention and storage costs?
Parquet compression gives 3-5x reduction vs raw data.
S3 storage costs ~$0.023/GB/month. 1TB = ~$23/month.
Built-in retention policies automatically delete old data. Or keep everything — storage is cheap, queries are fast.
Can I run this in high availability?
Single-node architecture in open source version.
Arc Enterprise includes clustering and high availability.
For OSS: Run primary + standby with object storage replication. WAL enabled for zero data loss during failover.
What about authentication and access control?
API token authentication included in all versions.
Grafana integration supports Azure AD OAuth.
Arc Enterprise includes RBAC, SSO, and multi-tenancy.
Pricing
Start free with open source. Scale with enterprise features when you need them.
Open Source
- 18M records/sec ingestion
- Full SQL query engine (DuckDB)
- Parquet storage (S3, GCS, local)
- Docker and Kubernetes ready
- Community support (Discord)
Enterprise
Starting price for up to 8 cores. Clustering, RBAC, and dedicated support.
- Everything in Open Source
- Horizontal clustering and HA
- Role-based access control (RBAC)
- Tiered storage and auto-aggregation
- Dedicated support and SLAs
Enterprise Features
Clustering
Horizontal scaling with automatic data distribution. Query routing and load balancing across nodes.
Security
Fine-grained RBAC with database and table-level permissions. LDAP/SAML integration available.
Data Management
Automated retention policies, continuous queries for aggregation, and tiered storage for cost optimization.
Get Started
Open Source Installation
docker run -d -p 8000:8000 \
-e STORAGE_BACKEND=local \
-v arc-data:/app/data \
ghcr.io/basekick-labs/arc:latestcurl -X POST http://localhost:8000/api/v1/write/line-protocol \
-H "Authorization: Bearer YOUR_TOKEN" \
-H "Content-Type: text/plain" \
-H "x-arc-database: default" \
--data-binary "cpu,host=server01 value=0.64"curl -X POST http://localhost:8000/api/v1/query \
-H "Authorization: Bearer YOUR_TOKEN" \
-H "Content-Type: application/json" \
-d '{"sql":"SELECT * FROM cpu LIMIT 10"}'Arc Enterprise
For organizations requiring clustering, RBAC, and dedicated support.
Multi-node clustering and high availability
Role-based access control (RBAC) and SSO
Dedicated support and SLAs
Commercial licensing available
enterprise@basekick.net
Ready to handle billion-record workloads?
Deploy Arc in minutes. Own your data in Parquet. Use for analytics, observability, AI, IoT, or data warehousing.
