All integrations
Amazon Ads
+
Bruin

Amazon Ads + Bruin

Source

Ingest Amazon Ads data into your warehouse with incremental loading, quality checks, and full lineage. Defined in YAML, version-controlled in Git.

For business teams

What you get

  • Cross-channel ad reporting

    See Amazon Ads spend alongside Google Ads, Facebook, and every other channel — in one place, updated automatically.

  • True ROAS, not estimated

    Join Amazon Ads spend with actual revenue from Stripe or your CRM. Know your real return on ad spend, not what the ad platform tells you.

  • No more manual exports

    Stop downloading CSVs from Amazon Ads. Stakeholders get fresh data every morning without asking anyone.

  • Catch budget anomalies early

    Quality checks flag unexpected spend spikes or zero-impression campaigns before they burn budget.

For data & engineering teams

How it works

  • Incremental sync with lookback

    Bruin handles Amazon Ads's attribution windows automatically. Set lookback days in the connection URI — no custom logic needed.

  • YAML-defined, Git-versioned

    Your Amazon Ads pipeline is a YAML file. Review in PRs, deploy with CI/CD, roll back with git revert.

  • Column-level quality checks

    Validate spend, impressions, and clicks with not_null, unique, and custom SQL checks. Pipeline stops on failure.

  • Multi-destination support

    Land Amazon Ads data in BigQuery, Snowflake, Redshift, or DuckDB. Switch destinations by changing one line.

Before you start

Amazon Ads API access via Amazon Developer account
Active Amazon Advertising account

Step 1

Add your Amazon Ads connection

Connect using Amazon Ads API OAuth2 credentials. Add this to your Bruin environment file — credentials are stored securely and referenced by name in your pipeline YAML.

Parameters

  • client_idAmazon Ads API OAuth2 client ID
  • client_secretAmazon Ads API OAuth2 client secret
  • refresh_tokenOAuth2 refresh token from the authorization flow
  • regionAdvertising API region (NA, EU, FE)
connections:
  amazon_ads:
    type: amazon-ads
    uri: "amazon-ads://?client_id=<your-client-id>&client_secret=<your-client-secret>&refresh_token=<your-refresh-token>&region=<region>"

Step 2

Create your pipeline

Define a YAML asset that tells Bruin what to pull from Amazon Ads and where to land it. This file lives in your Git repo — reviewable, version-controlled, and deployable with CI/CD.

Available tables

campaignsad_groupskeywordsreportsproduct_ads
name: raw.amazon_ads_campaigns
type: ingestr

parameters:
  source_connection: amazon_ads
  source_table: 'campaigns'
  destination: bigquery

Step 3

Add quality checks

Add column-level and custom SQL checks to your Amazon Ads data. If a check fails, the pipeline stops — bad data never reaches downstream models or dashboards.

Catch negative ad spend before it reaches reports
Validate impressions >= clicks on every sync
Flag campaigns with missing IDs or null spend
columns:
  - name: campaign_id
    checks:
      - name: not_null
  - name: spend
    checks:
      - name: not_null
  - name: impressions
    checks:
      - name: not_null

custom_checks:
  - name: no negative ad spend
    query: |
      SELECT COUNT(*) = 0
      FROM raw.amazon_ads_campaigns
      WHERE spend < 0

Step 4

Run it

One command. Bruin connects to Amazon Ads, pulls data incrementally, runs your quality checks, and lands clean data in your warehouse. If a check fails, the pipeline stops — bad data never reaches downstream.

Backfill historical data with --start-date
Schedule with cron or trigger from CI/CD
Full lineage from Amazon Ads to your dashboards
$ bruin run .
Running pipeline...

  amazon_ads_campaigns
    ✓ Fetched 2,847 new records
    ✓ Quality: campaign_id not_null     PASSED
    ✓ Quality: spend not_null           PASSED
    ✓ Quality: no negative ad spend     PASSED
    ✓ Loaded into bigquery

  Completed in 12s

Other Ad Platform integrations

Ready to connect Amazon Ads?

Start for free, or book a demo to see how Bruin handles ingestion, quality, lineage, and scheduling for your entire data stack.