Description

Job Description – 
Skills: Google BigQuery, DataProc, Data fusion, Apache Airflow, Google Cloud Platform, Python, SQL, CI/CD pipelines
Responsibilities:
Support Data ingestion pipelines, data processing workflows, and data storage solutions using GCP services such as Cloud Storage, Big Query, Dataflow, Composer etc.
Work with other DEs to integrate data solutions into applications.
Manage & configure data integration processes to extract data from various sources, transform and cleanse it, and load it into appropriate data storage systems.
Maintain and support ETL processes to ensure data quality and consistency.
Support efficient data retrieval and analysis.
Monitor data pipelines and processes for data quality issues.
Troubleshoot and resolve data-related problems in a timely manner.
Perform regular data backups and implement disaster recovery procedures.
Collaborate with cross-functional teams to understand data requirements and provide technical guidance and support.
Document data infrastructure, processes, and workflows to ensure knowledge sharing and maintainability.
Optimize data structures and queries for performance and scalability as problem management for long running jobs
Deployments on BigQuery, BigQuery object provision automation, BigQuery capacity management (Proactive/Reactive)
Optimizing and tuning Big Queries, Changes to DDL statements
Admin activities e.g., Create/Drop/Alter table/view, Data loads and fixes.
Provide configuration details to terraform team

Education

Any Graduate