ELT That Never Loses a Record
Fault-tolerant data pipelines with checkpointing. Resume from failures, never re-sync.
View Core on GitHub$ bizon run config.yml
INFO Pipeline hubspot-to-bigquery started
INFO Resuming from cursor: page 847 of 1,203
INFO Loaded 10,000 contacts → bigquery (1.2s)
INFO Checkpoint saved: page 857
...
✓ Synced 1.2M contacts | 42k rec/min | 0 failures Simple, declarative pipelines
Define your entire pipeline in YAML. No boilerplate, no complexity.
name: hubspot-to-bigquery
source:
name: hubspot
stream: contacts
sync_mode: incremental
authentication:
type: oauth
destination:
name: bigquery
config:
project_id: my-project
dataset_id: crm_data
engine:
backend:
name: postgres # Checkpoint state Why teams choose Bizon
Built for reliability, scale, and visibility—without forcing a specific queue or storage backend.
Checkpoint Recovery
Resume from the last checkpoint when things fail. No full restarts, no re-syncing millions of records.
High Throughput
Designed for billion-record pipelines with lean dependencies and efficient buffering.
Your Infrastructure
Queue agnostic (Kafka, RabbitMQ, Python Queue) with flexible backends (Postgres, BigQuery, SQLite).
Everything you need to orchestrate pipelines
A complete self-hosted platform for managing data pipelines at scale. Open source, no limits.
Web Dashboard
Visual interface to create, monitor, and manage all your pipelines in one place.
REST API
Full programmatic access for automation, CI/CD integration, and custom tooling.
Custom Python Sources
Build your own source connectors with a simple Python interface. No plugins, just code.
Job Scheduling
Cron-based scheduling with concurrent workers and Postgres-backed job queue.
Self-Hosted
Deploy on your infrastructure with Docker or Kubernetes. Full control over your data.
Open Source
GPL-3.0 licensed. No usage limits, no vendor lock-in. Only infrastructure costs (~$5-10/mo).