ELT That Never Loses a Record

Fault-tolerant data pipelines with checkpointing. Resume from failures, never re-sync.

View Core on GitHub
terminal
$ bizon run config.yml

INFO  Pipeline hubspot-to-bigquery started
INFO  Resuming from cursor: page 847 of 1,203
INFO  Loaded 10,000 contacts → bigquery (1.2s)
INFO  Checkpoint saved: page 857
...
 Synced 1.2M contacts | 42k rec/min | 0 failures

Simple, declarative pipelines

Define your entire pipeline in YAML. No boilerplate, no complexity.

config.yml
name: hubspot-to-bigquery

source:
  name: hubspot
  stream: contacts
  sync_mode: incremental
  authentication:
    type: oauth

destination:
  name: bigquery
  config:
    project_id: my-project
    dataset_id: crm_data

engine:
  backend:
    name: postgres  # Checkpoint state

Why teams choose Bizon

Built for reliability, scale, and visibility—without forcing a specific queue or storage backend.

Checkpoint Recovery

Resume from the last checkpoint when things fail. No full restarts, no re-syncing millions of records.

High Throughput

Designed for billion-record pipelines with lean dependencies and efficient buffering.

Your Infrastructure

Queue agnostic (Kafka, RabbitMQ, Python Queue) with flexible backends (Postgres, BigQuery, SQLite).

Bizon Platform

Everything you need to orchestrate pipelines

A complete self-hosted platform for managing data pipelines at scale. Open source, no limits.

Web Dashboard

Visual interface to create, monitor, and manage all your pipelines in one place.

REST API

Full programmatic access for automation, CI/CD integration, and custom tooling.

Custom Python Sources

Build your own source connectors with a simple Python interface. No plugins, just code.

Job Scheduling

Cron-based scheduling with concurrent workers and Postgres-backed job queue.

Self-Hosted

Deploy on your infrastructure with Docker or Kubernetes. Full control over your data.

Open Source

GPL-3.0 licensed. No usage limits, no vendor lock-in. Only infrastructure costs (~$5-10/mo).