HimalayasHimalayas logo
ED
Open to opportunities

Eric Du

@ericdu1

Senior data engineer building reliable streaming and analytics platforms with Python, Spark, Kafka, and dbt.

United States
Message

What I'm looking for

I’m looking for a team where I can build reliable, observable data platforms and streaming pipelines—using strong engineering practices (testing, CI/CD, infrastructure-as-code) and ownership of governance, SLOs, and self-service analytics.

I’m a Senior Data Engineer focused on turning messy, high-volume event data into trustworthy, fast-to-use analytics. At my core, I architect canonical datasets and production pipelines that cut time-to-insight and keep reliability high.

At Robinhood, I unified 50+ disparate event sources to deliver trading, portfolio performance, and user engagement metrics—bringing new metric build times from weeks to under 3 days. I build production-grade pipelines in Python, Spark, and Airflow with unit/integration testing and CI/CD, achieving 99.9% pipeline reliability.

I also design real-time stream processing with Kafka and Flink, ingesting 1M+ events/sec while balancing cost and latency. I’ve driven sub-second enrichment for fraud detection, improved data-platform self-service with dbt + Looker (enabling 100+ analysts), and strengthened reliability with SLOs, anomaly detection, and blameless post-mortems.

Earlier, I engineered HIPAA-compliant ELT pipelines with Snowflake and dbt, building audit-ready governance for EHR data. I scaled platforms dramatically, optimized Airflow and Snowflake performance for sub-second matching, implemented Datadog-based observability and SLAs, and used Terraform and AWS to make deployments repeatable and safe.

Experience

Work history, roles, and key accomplishments

Robinhood logoRO
Current

Senior Data Engineer

Apr 2023 - Present (3 years)

Architected canonical trading/portfolio/user datasets by unifying 50+ event sources, cutting time to build new metrics from weeks to under 3 days and reducing data-to-insight latency. Built Python/Spark/Airflow pipelines with unit/integration testing and Terraform-backed CI/CD for 99.9% reliability, and delivered Kafka+Flink stream processing at 1M+ events/sec with sub-second enrichment that reduc

Spring Health logoSH

Data Engineer

Aug 2016 - Mar 2023 (6 years 7 months)

Built HIPAA-compliant ELT pipelines in Snowflake and dbt for an EHR platform covering 10M+ lives, scaling eligible members from 260K to 10M by optimizing Airflow DAGs and tuning Snowflake performance for sub-second provider matching. Deployed real-time ML-driven precision matching integrations, automated clinical data flows, and implemented Datadog-based observability with SLAs for data freshness

Education

Degrees, certifications, and relevant coursework

University of California - Los Angeles (UCLA) logoUU

University of California - Los Angeles (UCLA)

Bachelor’s Degree of Computer Science, Computer Science

2012 - 2016

Earned a Bachelor's degree in Computer Science at UCLA from 2012 to 2016.

Find your dream job

Sign up now and join over 100,000 remote workers who receive personalized job alerts, curated job matches, and more for free!

Sign up
Himalayas profile for an example user named Frankie Sullivan