Ingest data
from anywhere

Ingest data from any source into your data lake or data warehouse, with no code required. Extend if needed with custom code.

Trusted by forward-thinking teams

Internet Society
Formo
Karaca
Papara
Obilet
Workhy
Buluttan
Lessmore
Spektra
Fomo Games
Rotatelab
Talemonster
name: raw.users
type: ingestr
parameters:
  source_connection: postgres
  source_table: 'public.users'
  destination: bigquery

Build data pipelines faster

Built-in connectors, defined with YAML

Bruin is a code-based platform, meaning that everything you do comes from a Git repo, versioned. All of the data ingestions are defined in code, version controlled in your repo.

Multiple platforms
Bruin supports quite a few platforms as built-in connectors. You can ingest data from AWS, Azure, GCP, Snowflake, Notion, and more.
Built on open-source
Bruin's ingestion engine is built on ingestr, an open-source data ingestion tool.
Custom sources & destinations
Bruin supports pure Python executions, enabling you to build your own data ingestion code.
Incremental loading
Bruin supports incremental loading, meaning that you can ingest only the new data, not the entire dataset every time.

Build safer

End-to-end quality in raw data

Bruin's built-in data quality capabilities are designed to ensure that the data you ingest is of the highest quality and always matches with your expectations.

Built-in quality checks
Bruin supports built-in quality checks, such as not_null, accepted_values, and more, all ready to be used in all assets.
Custom quality checks
Bruin allows you to define custom quality checks in SQL, enabling you to define your own quality standards.
Templating in quality checks
Bruin supports templating in quality checks, meaning that you can use variables in your checks, and run checks only for incremental periods.
Automated alerting
Failing quality checks will automatically send alerts to the configured channels, ensuring that you are always aware of the data quality issues.
name: raw.users
type: ingestr

parameters:
  source_connection: postgresql
  source_table: 'public.users'
  destination: bigquery

columns:

  # Define columns along with their quality checks
  - name: status
    checks:
      - name: not_null
      - name: accepted_values
        values:
          - active
          - inactive
          - deleted

# You can also define custom quality checks in SQL        
custom_checks:
  - name: new user count is greater than 1000
    query: |
      SELECT COUNT(*) > 1000 
      FROM raw.users 
      WHERE status = 'active' 
        AND created_at BETWEEN "{{start_date}}" AND "{{end_date}}"

Seamless data ingestion

Copy data between any source and destination

Bruin's ingestion engine is built on ingestr , an open-source data ingestion tool. You can ingest data from any source to any destination with no code required.

Google BigQuery to Snowflake
Google BigQuery to PostgreSQL
Google BigQuery to Amazon Redshift
Google BigQuery to Databricks
Google BigQuery to DuckDB
Google BigQuery to MySQL
Google BigQuery to Microsoft SQL Server
Snowflake to Google BigQuery
Snowflake to PostgreSQL
Snowflake to Amazon Redshift
Snowflake to Databricks
Snowflake to DuckDB
Snowflake to MySQL
Snowflake to Microsoft SQL Server
PostgreSQL to Google BigQuery
PostgreSQL to Snowflake
PostgreSQL to Amazon Redshift
PostgreSQL to Databricks
PostgreSQL to DuckDB
PostgreSQL to MySQL
PostgreSQL to Microsoft SQL Server
Amazon Redshift to Google BigQuery
Amazon Redshift to Snowflake
Amazon Redshift to PostgreSQL
Amazon Redshift to Databricks
Amazon Redshift to DuckDB
Amazon Redshift to MySQL
Amazon Redshift to Microsoft SQL Server
Databricks to Google BigQuery
Databricks to Snowflake
Databricks to PostgreSQL
Databricks to Amazon Redshift
Databricks to DuckDB
Databricks to MySQL
Databricks to Microsoft SQL Server
DuckDB to Google BigQuery
DuckDB to Snowflake
DuckDB to PostgreSQL
DuckDB to Amazon Redshift
DuckDB to Databricks
DuckDB to MySQL
DuckDB to Microsoft SQL Server
MySQL to Google BigQuery
MySQL to Snowflake
MySQL to PostgreSQL
MySQL to Amazon Redshift
MySQL to Databricks
MySQL to DuckDB
MySQL to Microsoft SQL Server
Microsoft SQL Server to Google BigQuery
Microsoft SQL Server to Snowflake
Microsoft SQL Server to PostgreSQL
Microsoft SQL Server to Amazon Redshift
Microsoft SQL Server to Databricks
Microsoft SQL Server to DuckDB
Microsoft SQL Server to MySQL
MongoDB to Google BigQuery
MongoDB to Snowflake
MongoDB to PostgreSQL
MongoDB to Amazon Redshift
MongoDB to Databricks
MongoDB to DuckDB
MongoDB to MySQL
MongoDB to Microsoft SQL Server
Oracle Database to Google BigQuery
Oracle Database to Snowflake
Oracle Database to PostgreSQL
Oracle Database to Amazon Redshift
Oracle Database to Databricks
Oracle Database to DuckDB
Oracle Database to MySQL
Oracle Database to Microsoft SQL Server
SAP HANA to Google BigQuery
SAP HANA to Snowflake
SAP HANA to PostgreSQL
SAP HANA to Amazon Redshift
SAP HANA to Databricks
SAP HANA to DuckDB
SAP HANA to MySQL
SAP HANA to Microsoft SQL Server
Notion to Google BigQuery
Notion to Snowflake
Notion to PostgreSQL
Notion to Amazon Redshift
Notion to Databricks
Notion to DuckDB
Notion to MySQL
Notion to Microsoft SQL Server
Google Sheets to Google BigQuery
Google Sheets to Snowflake
Google Sheets to PostgreSQL
Google Sheets to Amazon Redshift
Google Sheets to Databricks
Google Sheets to DuckDB
Google Sheets to MySQL
Google Sheets to Microsoft SQL Server
Shopify to Google BigQuery
Shopify to Snowflake
Shopify to PostgreSQL
Shopify to Amazon Redshift
Shopify to Databricks
Shopify to DuckDB
Shopify to MySQL
Shopify to Microsoft SQL Server
Gorgias to Google BigQuery
Gorgias to Snowflake
Gorgias to PostgreSQL
Gorgias to Amazon Redshift
Gorgias to Databricks
Gorgias to DuckDB
Gorgias to MySQL
Gorgias to Microsoft SQL Server

Ready to ship reliable data?

Production-ready pipelines without the complexity. Deploy today.