Post Job Free

Resume

Sign in

Software Development Data Warehousing

Location:
Frisco, TX
Salary:
130k/anum
Posted:
January 09, 2024

Contact this candidate

Resume:

Sridhar

Email: ad2lxw@r.postjobfree.com

Mobile: 763-***-****

Location: USA, MN

Professional Summary:

** ***** ** ********* ********* in ETL tools like IBM InfosphereDataStage8.x/11.5 and DBT.

Proficient in all aspects of the Software Development Life Cycle (SDLC), encompassing requirements analysis, high and low-level design, development and coding, unit and integration testing, version management, deployment and implementation, end user training, and ongoing maintenance and support.

Strong knowledge in Data Modeling, Dimensional Modeling, and designing Data Warehouse, Data Marts using star and snowflake schemas.

Used DataStage Director and the runtime engine to schedule the jobs, monitor scheduling, and validate its components.

Hands-on expertise with Snowflake, a cloud-based database and Teradata for loading various datasets from relational database management system (RDBMS) sources.

Extracting data from various sources, transforming it, and loading it into the Teradata system.

Used Hashed file to extract and write data and to act as an intermediate file in a job. A hash file is also used as a reference table based on single key field enhancements.

Strong knowledge of best practices and performance tuning of Snowflake and Azure environment.

Extensive hands-on experience with IBM Infosphere Suite versions 11.3, 9.1, 8.7, 8.5, 8.0.1, and 7.5.x, utilizing key components such as DataStage Designer, DataStage Manager, DataStage Director, DataStage Administrator, Information Server, and Parallel Extender.

Proficient in designing scalable, fault-tolerant, and high-performing analytic solutions for Data Lakes, Data Warehouses, and Data Marts, catering to a range of dataset sizes while meeting Data Analytics requirements, access patterns, data management needs, and resource and cost constraints.

Managing metadata, data lineage, and data cataloging.

Experience in Application Design, Data Extraction, Data Acquisition, Data Mining, Development, Implementations and Testing of Data Warehousing and Database Business Systems.

In-Depth understanding of Snowflake Cloud Technology.

Having Good knowledge on AWS, Cloud based services, Python, Spark framework.

Data preprocessing, feature engineering, and model selection using Teradata data.

Experience in Production Support applications involving IBM Infosphere DataStage ETL jobs.

Having good Knowledge in implementation of DataStage Parallel Jobs and sequences to move data from multiple sources like Flat files, Teradata, SQL Server, and Oracle into a common target area such as Data Marts and Data Warehouse.

Unix experience & Gitlab (CI/CD).

Experience in preparing estimates, Resource Planning, task distribution, status tracking, leading the team, taking responsibility and meeting deadlines.

Possesses a deep understanding of ER Modeling, Dimensional Modeling, Star and Snowflake Schemas, and the creation of conceptual, logical, and physical data models using Erwin/Power Designer.

Sound Working Experience in RDBMS like Snowflake, Oracle, Teradata and DB2.

Worked on multiple domains like Banking, Retail, and Insurance.

Ensuring the reliability and correctness of Airflow workflows.

Developing and executing test plans for data pipelines.

Identifying and reporting bugs or issues.

Creating and managing DAGs (Directed Acyclic Graphs) for specific analyses.

scheduled interval and bash commands for the jobs to run.

Work Experience:

Working as a Sr Software engineer with Tech Mahindra from Sep 2011 to till date.

Technical Skills:

ETL Tools : Dbt, IBM Infosphere DataStage 8.x/11.5, Snowflake, Airflow

Data Modeling Tools: Erwin

Data Bases : Snowflake, SQL Server, DB2, Teradata, Oracle.

Languages : SQL, PL/SQL, and Shell Scripting.

Others : Airflow Scheduler, Control - M, Tivoli, Autosys and Run Deck

Educational Qualification & Certification:

Master’s Degree from Acharya Nagarjuna University, India in 2004 (MCA).

Certified as IBM Infosphere DataStage Enterprise Edition 8.5/7.5 Solution Developer.

Bachelor’s Degree from Acharya Nagarjuna University, India in 2001.

Professional Experience:

Client : Scotiabank (Sep.’20- Present)

Project : AML-R/AML-C/EDL/FINRAC

Role : Sr Software Engineer

Responsibilities:

Involved in gathering requirement specifications by having Interactive Sessions with the business.

Defining business requirements and objectives for data workflows.

Extensive experience on migrating the DataStage ETL jobs to dbt models.

Designed and developed different types dbt models like source, stage, INT and load models.

Setting up and maintaining the infrastructure for Airflow (e.g., servers, containers, cloud resources).

Automating deployment processes and managing configurations.

Monitoring system health and performance.

Used DataStage Parallel Extender stages namely Datasets, Sort, Lookup, Change Capture, Funnel, Peek, SCD, and Row Generator.

Proficiency in data analysis and machine learning.

Ability to write Python scripts and utilize relevant libraries.

Involved in Unit testing, System testing to check whether the data is loading into target, which was extracted from different source systems according to the user requirements.

Experience in Building snow pipe for handling Streaming data and Data sharing in Snowflake.

Implementing CI/CD pipelines for Airflow workflows.

Snowflake knowledge on using features such as Zero Copy Clone, Time Travel, User defined functions, etc. Managing security in Snowflake including the creation of custom Roles to control access to Data, Databases, Warehouses etc.

Ensuring the reliability and correctness of Airflow workflows.

Developing and executing test plans for data pipelines.

Environment:

DataStage 11.5, Airflow, Bigdata Hadoop, DB2, SQL Server, UNIX, Snowflake and Control-M.

Client Name : CMIC (Church Mutual Insurance Company)

(Jan 2015 to Aug 2020)

Project : Insurance.

Role : Sr Software Engineer.

Roles and Responsibilities:

Provided solutions and detail designs for new intake requests from business.

Prepared data mapping between sources and target systems.

Involved in all phases of software engineering including requirements analysis, application design, and code development and testing.

Used Stored procedures to process the data to the semantic layer.

Used DataStage jobs to create a common framework to build the source files.

Used Control M to run and monitor the DataStage jobs.

Meeting with the managers from time to time to keep updated.

Working with the business users in requirements gathering and solution options development

Involved in fixing bugs identified during production runs within the existing functional requirements.

Taking Business Sign-off from the Business users for all new requirements and enhancements.

Supporting and implementing data integration processes for multiple sources.

Preparing the Source and Target Mapping documents for IBM InfoSphere DataStage jobs and job sequences.

Developed the Audit control process to track the data and the files.

Reviewing the Code developed for the Quality standards.

Environment:

IBM Infosphere DataStage 11.5, Oracle, Control M, UNIX/Linux, Flat Files, GITHUB, SVN

Client Name : ICBC (Insurance Corporation of British Columbia), Vancouver, CA

(Dec 2013 to Jan 2015)

Project : ICBC ETL and Data Warehousing.

Role : Sr Software Engineer.

Roles and Responsibilities:

Developed several transformations in the process of building a Data warehouse database.

Used parallel jobs for Extraction, Transformation, and Load.

Analyzed and designed source code documentation for investment Data Warehouse.

Used DataStage Designer for developing various jobs to extract, cleanse, transform, integrate, and load data into the Data Warehouse.

Generating unique keys for composite attributes while loading the data into the Data Warehouse.

Used DataStage Manager for managing DataStage repository (view and edit), defining custom routines & and transforms, importing and exporting items between different DataStage systems or exchanging metadata with other data warehousing tools.

Extensively used Parallel Stages like Join, Merge, Lookup, Filter, Remove Duplicates, Funnel, Row Generator, Modify, Change Data Capture, Transformer, Peek, Row Generator etc. for development and debugging purposes.

Extensively worked with Data Stage Job Sequences to Control and Execute Data Stage Jobs and Job Sequences using various Activities and Triggers.

Written SQL Queries used in the balancing for getting count through links in the jobs.

Extracting the data from the tables and analyzing the missing customers, Accounts and the relation.

Extracting the data from the database for reporting by writing a large SQL Query.

Environment:

IBM WebSphere DataStage 8.5.x, Oracle 9i, SQL Server, DB2 UDB, Teradata, Toad, Control M .

Client Name : Bell Canada (Jan 2011 to Nov 2013)

PROJECT : TED Data Mart Creation

Role : ETL Developer

Roles and Responsibilities:

Performed Rules analysis using an information analyzer to check the data quality.

Analyzing data stored in Teradata databases to extract insights and support decision-making.

Conducting ad-hoc data analysis and generating meaningful reports for stakeholders.

Data mining and statistical analysis to identify trends and patterns.

Performed advanced analysis to profile and assess source system data.

Performed admin activities like user management and migration management.

Developing high-level design documents and templates to follow a specific design approach.

Involved in development activities to design and develop DataStage jobs and sequencers.

Used DataStage Designer for developing various jobs to extract, cleanse, transform, integrate, and load data into the Data Warehouse.

Extensively used Parallel Stages like Join, Merge, Lookup, Filter, Remove Duplicates, Funnel, Row Generator, Modify, Change Data Capture, Transformer, Peek, Row Generator etc. for development and debugging purposes.

Loading large volumes of data files to the integrated data warehouse using DataStage jobs.

Analyzing the failure of jobs and finding the root cause for the failure & and long execution of jobs.

Performed code unit testing by stimulating various business cases.

Parameterized DataStage jobs to allow portability and flexibility during runtime.

Used DataStage Director to run and monitor the jobs performed.

Created UNIX scripts to run the DataStage jobs and ftp the files.

Designing and developing Teradata databases

Configuring and testing Teradata systems

Troubleshooting Teradata systems

Liaising with Teradata support staff and other technical teams

Working with the business users in requirements gathering and solution options development

Involved in fixing bugs identified during production runs within the existing functional requirements.

Environment:

IBM Infosphere DataStage and Quality Stage (version 8.7), IBM Information Analyzer (v8.7), Teradata 7.0, Unix.



Contact this candidate