At least 4+ years of relevant experience in DataStage Admin, Development and CDC
At least 2 years of recent hands-on experience in IBM Data Stage using v11.x ETL tool and rest in other ETL tools.
Expertise in Data Warehousing ETL Tool DataStage using Components like DataStage Designer, DataStage Director and DataStage Operations Console.
Efficient in all phases of the development lifecycle, coherent with Data Cleansing, Data Conversion, Performance Tuning and System Testing.
Expertise in designing DataStage Parallel Jobs, Job Sequencing, Creating Parameter Sets, Environment Variables and Creating Parallel Routines, Data Cleansing and Writing Transformation expressions to derive values and remove unwanted data.
Efficient in identifying and finalizing the various source file formats and frequencies from different source systems.
In depth experience in troubleshooting of Datastage jobs and addressing issues like performance tuning.
Expert level proficiency in UNIX skills
Experienced in automation of the jobs using Unix scripts.
Experience in developing of reusable Unix Jobs/components
Strong understanding of Data warehousing concepts, dimensional Star Schema and Snowflakes Schema methodology.
Build the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources using Snowflake Cloud Datawarehouse as well as SQL and Azure big data technologies
Work with stakeholders including the Executive, Product, Data and Design teams to assist with data-related technical issues and support their data infrastructure needs.
4+years of experience in databases technologies like DB2, Oracle, TERADATA, Hive and HBase.
Very efficient in developing SQL queries and debugging complex SQL queries.
Expert level SQL tuning skills in any of the RDBMS.
Good knowledge on AutoSys scheduler is desired
Good knowledge on Agile / Scaled Agile principles. Excellent communication skills (both spoken and written)
Documents all technical and system specifications documents for all ETL processes and perform unit tests on all processes and prepare required programs and scripts.
Assist in the ongoing development of technical best practices for data movement, data quality, data cleansing and other ETL-related activities
Design, implement, and continuously expand data pipelines by performing extraction, transformation, and loading activities
Gather requirements and business process knowledge in order to transform the data in a way thats geared towards the needs of end users
Maintain and improve already existing processes
Ensure that the data architecture is scalable and maintainable
Work with the business in designing and delivering correct, high quality data
Investigate data to identify potential issues within ETL pipelines, notify end-users and propose adequate solutions
Any Graduate