Description


Seeking a experienced 10+ years GCP Data Engineer with Cost Optimization experience with expertise in Dataproc Serverless, Apache Spark, Google Cloud Composer, and BigQuery to join our innovative data engineering team. As a pivotal member of our team, you will be responsible for architecting, implementing, and managing complex data pipelines and analytics solutions on Google Cloud Platform (GCP).


> Key Responsibilities:
> Data Pipeline Design and Development
> Dataproc Serverless and Spark Expertise
> Cloud Composer Workflow Management
> BigQuery and GCS Integration
> Data Security and Compliance
> Performance and Cost Optimization
>Oozie scheduler and knowledge on Composer also capable to convert Oozie jobs to composer based jobs

Qualifications:
>Bachelor's or Master's degree in Computer Science, Information Technology, or a related field.
> Proven experience as a data engineer with a focus on GCP, Dataproc Serverless, Spark, Cloud Composer, and BigQuery.
> Experience in Data validation pre and post pipeline migration.
> Familiarity with cloud-based data storage solutions, particularly Google BigQuery and Google Cloud Storage.
> Continuously optimize data pipelines, Spark jobs, and BigQuery queries for cost-effectiveness and performance.
> Identify and resolve performance bottlenecks and inefficiencies in data processing
> Strong SQL skills and experience with data modeling.
> Exceptional problem-solving abilities and meticulous attention to detail.

Education

Bachelor's or Master's degree in Computer Science, Information Technology