Design and maintain ETL/ELT pipelines to ingest, transform, and load large datasets into GCP-based data platforms such as BigQuery and Cloud Storage.
Develop and optimise scalable back-end components and modular Python code for data processing, workflow orchestration, API integrations, and automation.
Build and operationalise semantic data layers to support standardised metrics and improved data accessibility.
Utilise GCP services including Cloud Composer, Dataflow, Pub/Sub, and Cloud Functions to create automated and reliable workflows.
Develop transformation workflows using Dataform, including SQLX transformations, automated tests, CI/CD integrations, and documentation.
Collaborate with data scientists, analysts, and business partners to convert data needs into efficient technical solutions.
Implement data quality checks to ensure accuracy, integrity, and performance in all data workflows.
Integrate semantic layers into BI tools and leverage metadata and lineage tools for improved governance.