Back to Showcase

Polymarket Pipeline

Jonathan Ramírez Quijada

Daily batch data pipeline that ingests ~15 GB of hourly Parquet files (~750 million rows) from a public archive, lands them in Google Cloud Storage, loads them into BigQuery and enriches each market with category metadata fetched from the Polymarket Gamma API. Bruin was used for the entire data lifecycle: ingestion, transformation, orchestration, data quality and analysis via the AI Data Analyst. The result powers a Looker Studio dashboard tracking platform liquidity, spread quality, and category activity over time.

Share: