Open Source - MIT
Mako
Real-time data pipelines
Declarative framework for orchestrating data pipelines. Configure your sources, transforms and sinks in YAML — Mako handles the rest.
go install github.com/Stefen-Taime/mako@latestFeatures
Declarative
Everything is configured in YAML. No code needed for standard pipelines.
Real-time
Native Kafka, PostgreSQL CDC and HTTP streaming support.
WASM Transforms
Go/Rust/TinyGo plugins compiled to WebAssembly for performant transformations.
Observability
Built-in Prometheus metrics, Grafana dashboards and Slack alerts.
Quick Start
# pipeline.yaml
source:
type: http
url: https://api.example.com/data
format: json
transforms:
- type: field
operations:
- rename: { from: "old_name", to: "new_name" }
- drop: ["unused_field"]
sink:
type: postgres
connection: postgres://user:pass@host/db
table: events
mode: upsertSources
- • HTTP/REST APIs (pagination, OAuth2, rate limiting)
- • JSON, CSV, Parquet files (gzip)
- • Apache Kafka
- • PostgreSQL CDC (Change Data Capture)
- • DuckDB (embedded SQL queries)
Sinks
- • PostgreSQL, Snowflake, BigQuery, ClickHouse
- • S3, Google Cloud Storage
- • DuckDB, Kafka, stdout
Transforms
- • SQL enrichment via DuckDB
- • WASM plugins (Go/Rust/TinyGo)
- • Schema validation (Confluent Schema Registry)
- • Data quality checks
- • PII masking (SHA-256)
- • Field operations (rename, drop, cast, flatten, deduplicate)
Orchestration
DAG-based workflow engine with parallel execution and SQL quality gates.