Description

Role Description & Requirements: 

Location: Australia based (Sydney location preferred, alternatively Brisbane) 

● Lead technical architecture and solution design for large scale data platform solutions 

● Proactively Interact with customer’s data engineering team to provide opinionated guidance, technical product workshops on data and analytics platform architecture, best practices, solution design and development 

● Independently develop solutions, i.e., perform hands on coding for the designed solution, handover developed solution to customer’s technical team 

● Perform pair programming with customer’s data engineers 

● Project manage small to mid sized projects, managing scope, priorities, deliverables, risks/issues, and timelines for successful outcomes 

● Collaborate with cross-disciplinary project team, play the role of workstream or overall project tech lead depending on size and scale of project / program 

 

Minimum Qualifications 

 

● Expertise Required 

○ In-depth knowledge and experience of GCP data and analytics technologies 

○ Designing data platform architecture and design for (a) migrating open source or other public cloud based data platforms to to GCP cloud native technologies (not just lift and shift migration but re-engineering to native services) (b) Designing green field data platforms on cloud 

○ Define solution architecture and detailed design, and perform hands-on implementation for 

■ data pipelines for batch and event driven ingestion and processing of data from a variety of sources such as on-prem files, on-prem databases, APIs, etc. 

■ data pipelines for real time data ingestion and processing 

○ Automatic transpilation of legacy code (HIVE, Teradata, python logic etc.) to BQ SQL 

○ BQ query performance optimisation 

○ CI/CD pipelines for data workloads using Cloud Build, Artifact Registry, Terraform 

○ Data governance solutioning using GCP governance tooling (Dataplex, Data Catalog) 

 

Tools and languages experience Required 

 

Must have 

 

● GCP Dataflow, Dataproc, Pub/Sub, Cloud Composer, Cloud Workflow, BigQuery, Cloud Run, CloudBuild 

● Must have: programming knowledge and willingness to be hands-on - Python, Java 

 

Good to Have 

 

● Experience in computing infrastructure (e.g., servers, databases, firewalls, load balancers, kubernetes) and architecting, developing, or maintaining solutions in virtualized environments.