We are looking for a Senior Data Engineer to design, develop, and optimize our data infrastructure on Google Cloud Platform (GCP). The role involves architecting scalable pipelines using Databricks, BigQuery, Google Cloud Storage, Apache Airflow, dbt, Dataflow, and Pub/Sub, and enforcing data quality standards using Great Expectations.
Requirements
- 3+ years of professional experience in data engineering, with at least 1 year in mobile data
- Proven track record building and maintaining BigQuery environments and Google Cloud Storage-based data lakes
- Deep knowledge of Apache Airflow for scheduling/orchestration and ETL/ELT design
- Experience implementing dbt for data transformations, RabbitMQ for event-driven workflows, and Pub/Sub + Dataflow for streaming/batch data pipelines
- Familiarity with designing and implementing Data Mart (Data Mach) solutions, as well as using Terraform for IaC
- Strong coding capabilities in Python, Java, or Scala, plus scripting for automation
- Experience with Docker and Kubernetes (K8s) for containerizing data-related services
- Hands-on with CI/CD pipelines and DevOps tools (e.g., Terraform, Ansible, Jenkins, GitLab CI) to manage infrastructure and deployments
- Proficiency in Great Expectations (or similar) to define and enforce data quality standards
- Expertise in designing systems for data lineage, metadata management, and compliance (GDPR, CCPA)
- Strong understanding of OLTP (Online Transaction Processing) and OLAP (Online Analytical Processing) systems
- Excellent communication skills for both technical and non-technical audiences
- High level of organization, self-motivation, and problem-solving aptitude
Benefits
- Growth and career development
- Work-Life balance
- Comprehensive benefits
- Unique Perks
