Description

Job description:

- At least 4+ years of relevant experience in DataStage Admin, Development and CDC

- At least 2 years of recent hands-on experience in IBM Data Stage using v11.x ETL tool and rest in other ETL tools.

- Expertise in Data Warehousing ETL Tool DataStage using Components like DataStage Designer, DataStage Director and DataStage Operations Console.

- Efficient in all phases of the development lifecycle, coherent with Data Cleansing, Data Conversion, Performance Tuning and System Testing.

- Expertise in designing DataStage Parallel Jobs, Job Sequencing, Creating Parameter Sets, Environment Variables and Creating Parallel Routines, Data Cleansing and Writing Transformation expressions to derive values and remove unwanted data.

- Efficient in identifying and finalizing the various source file formats and frequencies from different source systems.

- In depth experience in troubleshooting of Datastage jobs and addressing issues like performance tuning.

- Expert level proficiency in UNIX skills

- Experienced in automation of the jobs using Unix scripts.

- Experience in developing of reusable Unix Jobs/components

- Strong understanding of Data warehousing concepts, dimensional Star Schema and Snowflakes Schema methodology.

- Build the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources using Snowflake Cloud Datawarehouse as well as SQL and Azure ‘big data’ technologies

- Work with stakeholders including the Executive, Product, Data and Design teams to assist with data-related technical issues and support their data infrastructure needs.

- 4+years of experience in databases technologies like DB2, Oracle, TERADATA, Hive and HBase.

- Very efficient in developing SQL queries and debugging complex SQL queries.

- Expert level SQL tuning skills in any of the RDBMS.

- Good knowledge on AutoSys scheduler is desired

- Good knowledge on Agile / Scaled Agile principles. Excellent communication skills (both spoken and written)

- Documents all technical and system specifications documents for all ETL processes and perform unit tests on all processes and prepare required programs and scripts.

- Assist in the ongoing development of technical best practices for data movement, data quality, data cleansing and other ETL-related activities

- Design, implement, and continuously expand data pipelines by performing extraction, transformation, and loading activities

- Gather requirements and business process knowledge in order to transform the data in a way that’s geared towards the needs of end users

- Maintain and improve already existing processes

- Ensure that the data architecture is scalable and maintainable

- Work with the business in designing and delivering correct, high quality data

- Investigate data to identify potential issues within ETL pipelines, notify end-users and propose adequate solutions

Education

Any Graduate