Post Job Free
Sign in

Data Engineer Entry

Location:
Austin, TX
Posted:
May 23, 2025

Contact this candidate

Resume:

KARTHIK GEEDIPALLI @: **************@*****.*** Phone: 908-***-**** LinkedIn: https://www.linkedin.com/in/kgeedipalli/

Professional Summary

Around 14 years of IT experience in the Analysis, Design, Development, Testing of ETL methodologies and Implementation of business application systems for Insurance, Retails, Health care, Financial, and Telecom Sectors.

Experience in implementation of Business Intelligence solutions using Data Warehouse/Data Mart Design, OLAP, BI, Client/Server applications.

Experience in IBM Info Sphere Information Server 11.7/11.5/11.3/8.7 components such as DataStage Designer, DataStage Manager, DataStage Director and DataStage Administrator.

Experience of using Informatica 10.2/9.6/9.1 PowerCenter Client tools - Mapping Designer, Repository manager, Workflow Manager/Monitor and Server tools like Informatica Server, Repository Server manager.

Experience in working with BI tools like ThoughtSpot and SSRS.

Experience in using Automation Scheduling tools like Control-M, Zena, Opswise & Autosys.

Worked on Data migration projects, ensuring data integrity and optimal performance.

Worked on Data Stage migration project from version 9.1 to 11.3

Used UCD & Gitlab as Version Control and for Code migration to different environments.

Good experience in using agile project planning tools like JIRA & RTC.

Experience in writing UNIX shell scripts and automation of the ETL processes using UNIX shell scripting.

Developed both Server & Parallel jobs using different processing stages like Transformer, Aggregator, Lookup, CDC, Change Apply and Filter.

Worked with different Informatica transformations like Aggregator, Lookup, Joiner, Filter, Router, Update strategy, Transaction Control, Union, Normalizer, SQL in ETL development.

Experience in Dimensional Modeling using Star and Snowflake Schema, Identifying Facts and Dimensions, Physical and logical data modeling.

Experience in coding using SQL, PL/SQL Procedures/Functions, Triggers and Packages

Responsible for Preparation of Design Documents / Test Case Specifications / Performance Review and Coding and getting it signed off from client.

Interacted well with application Developers, Users, Project Managers, and Business Analysts to review the requirement specifications.

Experience in all phases of Data warehouse development from requirements gathering for the data warehouse to develop the code, Unit Testing and Documenting.

Technical Skills

ETL Tools: IBM Info sphere DataStage 11.7/11.5/11.3/8.7/8.5(Designer, Director and Administrator), Informatica Power Centre 10.2/9.6/9.1 (Designer, Repository manager, Workflow Manager/Monitor).

Scheduling tools: Control-M, DS Director, Zena, Autosys, and Opswise.

Database: Oracle, IBM DB2, Netezza, SQL Server, Snowflake Cloud DB, Salesforce.

Version Control: UCD, Gitlab.

BI Tools – ThoughtSpot, SSRS.

Data Warehousing: Snow-Flake Modeling, Fact and Dimensions, Physical and Logical Data Modeling and Erwin.

Project Planning Tools: JIRA and RTC (Rational Team Concert)

Education & Certifications

Bachelor of Technology in Computer Science and Engineering. 2003-07. India.

Masters in Computer Science. 2008-09. State University of New York. USA.

Snowflake- Snow Pro Core Certification. March 2023.

Snowflake - Hands-On Essentials: Data Warehousing Workshop. July 2024

Snowflake - Hands-On Essentials: Collaboration, Marketplace & Cost Estimation Workshop. Sept 2024

Snowflake - Hands-On Essentials: Data Application Builders Workshop. Sept 2024

Professional Experience

USAA Sep ’18 – Present.

Phoenix. AZ

Sr. ETL Consultant

USAA (United Services Automobile Association) offers a range of personal property and casualty (P&C) insurance. USAA proactively monitor, identify, and act on MSRs (Member Service representatives) and their leaders’ behaviors that may be contrary to the USAA Mission. With the implementation of this new project SSGP, this will help USAA to anticipate, identify, and govern sales and service practices. This helps in identifying Fraud and Ghost accounts and helps in providing better customer service experience.

Roles and Responsibilities:

Used DataStage 11.7/11.5 to transform data from different product platforms into standardized data.

Worked with technical Team in evaluating applications systems or evaluating requests for proposals and applications systems to perform the ETL process.

Used Informatica 9.6 to extract P&C Data from NZ Data sources and scheduled these jobs using Control-M. Upgraded to Informatica 10.2 to connect to Snowflake Cloud Database and load Data to Cloud.

Create, design, and optimize ETL workflows using IICS to integrate data from various sources.

Integrated data quality checks within the ETL workflows using IICS Data Quality features to cleanse, standardize, and validate data.

Optimize real-time and batch data processing workflows using IICS Task flows, Mapping Configurations, and Parameterized Mappings.

Created Datamart and created different tables and structures to store data we fetch from different sources.

Used Control-M for job scheduling and used both Unix and Python scripts to create tables and export them to different tiers of environments.

Used Gitlab scripts to promote code and make changes to our Datamart tables and structures.

Created a Target Datamart in DB2 using PL/SQL objects to improve performance.

Worked with user and process requirements and ensures those requirements can be achieved through high quality deliverables.

Used UCD and Gitlab for code deployment and as Version control tools.

Performed UAT Testing for code to check end to end performance of our ETL process.

Performed data validation, cleansing, and standardization using Informatica IDMC Data Quality.

Worked as a resource to the business and as a technical resource to cross functional third party and internal team members on highly complex design/code reviews.

Analyzed and troubleshooted complex problems and recommended solutions or practices relative to root cause analyses and identification of solutions for improving system performance and availability.

Monitored ETL job performance using IDMC Monitor, IICS Task Flows, and error handling mechanisms.

Used standard data modeling, data flow, and created ETL solutions using data integration tools for interfacing between source application and the Enterprise Data Warehouse.

Created Informatica mappings using transformations like lookup, joiner, Rank, Aggregator, Lookup, Source Qualifier and Router transformations to extract, transform and load data mart area.

Provided ETL technical specifications, designs, Data Movement process, implements and provide optimal solutions and integrates existing jobs and utilizes them to maximize object reuse.

Performed problem assessment, resolution and documentation in existing ETL mapping and workflows.

Environment: DataStage 11.7/11.5, Informatica 9.6/10.2, Oracle, DB2, Netezza, Snowflake Cloud DB, PL/SQL, Salesforce, Python, JIRA, RTC, Control-M,, UCD and Gitlab.

Lands’ End Jan’18 – Aug’18.

Dodgeville. WI

Sr. ETL Consultant

This project involves Conversion of forecasting and merchandise data from Legacy System to ERP platform and using this Data to build a new EDW warehouse that will be used to support business and sales requirements. We used both Oracle and DB2 databases to fetch Data from different modules and load the Data into Lands’ End BI Enterprise Data Warehouse. From here Business Users and Retail stores can fetch reports related to sales, forecast and Inventory.

Roles and Responsibilities:

Used Informatica 9.6/10.2 to transform data from different product platforms into standardized data.

Using Informatica PowerCenter created mappings and mapplets to transform the data according to the business rules.

Participated in requirements gathering and created Source to Target mappings for development.

Prepared Functional Design, technical design and mapping documents depicting the ETL process involved in the project.

Co-coordinating with client managers, business architects and data architects for various sign offs on data models, ETL design docs, testing docs, migrations and end user review specs.

Used principles of Normalization to improve performance and used PL/SQL in ETL Code.

Worked on preparing integration test case plans and test scenarios along with testing.

Worked with business team for analyzing the Source Data & building the design documents.

Used ASG-Zena to schedule both UNIX and Informatica jobs.

Worked closely with Mainframe Legacy Team in optimizing Mainframe jobs to Informatica jobs.

Worked with DBAs in creation of new table structure and Indexes for better access of Data.

Experience in generating and interpreting mapping documentation and translating into detailed design specifications using ETL code.

Involved in creating database connectivity and executing queries in parallel job execution. Used Zena to schedule and running the jobs, testing, and debugging its components & monitoring performance.

Worked with job scheduling team and Prepared ETL job run dependency list by considering the load and availability of various systems.

Environment: Informatica 9.6/10.2, Oracle 11g/12c, PL/SQL, DB2, Zena, JIRA, TOAD, Mainframe, XML files.

TRACFONE Jan’17 – Dec’17.

Miami. FL

ETL Consultant

This project involves Conversion of DataStage Version from 8.7 to 11.5. We have a Tracfone Revenue Project in which we use BRM/Oracle to fetch all Revenue and Promotion details of Tracfone Billing. We used both Oracle and DB2 Databases to fetch Data from different modules to calculate revenue based on different brands of services provided by Tracfone. Also worked on third party reporting Tool – Thought Spot, which is helpful in viewing all reports at any level.

Roles and Responsibilities:

Used DataStage 11.5 and 8.7 to transform a variety of financial transaction files from different product platforms into standardized data.

Designing ETL jobs incorporating complex transform methodologies using DataStage tool resulting in development of efficient interfaces between source and target systems.

Worked on BI tool – ThoughtSpot, a new Data analytical platform.

Used DataStage Netezza Enterprise stage to load data, utilizing the available processors to achieve job performance.

Used SQL, PL/SQL programs to retrieve data using cursors and exception handling.

Participated in requirements gathering and created Source to Target mappings for development.

Used DataStage Director to validate, run and monitor the DataStage jobs.

Involved in creating database connectivity and executing queries in parallel job execution. Used Director to schedule and running the jobs, testing, and debugging its components & monitoring performance.

Extensively involved with business team for analyzing the Source Data & building the design documents

Prepared ETL job run dependency list by discussing with scheduling team and java extracts team and by considering the load and availability of various systems.

Prepared mapping documents, technical design document and process flow documents using Visio.

Prepared integration test case plans and test scenarios along with testing.

Environment: DataStage 11.5, 11.3 and 8.7, Oracle 11g, PL/SQL, Netezza, Flat files, Autosys, UNIX, ThoughtSpot

GEICO Mar’15 – Dec’16.

Macon. GA.

ETL Consultant

This project involves the Conversion of GEICO Billing System from Mainframe platform to .NET Platform. So, we used to Extract Auto, P&C billing Data from Mainframe DB2 Database and compare the Data with Reports from .NET based SQL Database. We used both SSRS reporting and DataStage for all Extractions and Comparison of Data from both platforms and Compare Data in Prod and Non-Prod environments.

Roles and Responsibilities:

•Used DataStage as an ETL tool to extract data from sources systems, loaded the data into the ORACLE database.

•Designed and Developed DataStage Jobs to Extract Auto, P&C billing data from heterogeneous sources, applied transform logics to extracted data and Loaded into Data Warehouse Databases.

•Used Director and its run-time engine to schedule running the solution, testing, and debugging its components, and monitoring the resulting executable versions on ad hoc or scheduled basis.

•Developed complex store procedures using input/output parameters, cursors, views, triggers, and complex queries using temp tables and joins.

•Converted complex job designs to different job segments and executed through job sequencer for better performance and easy maintenance.

•Maintained Data Warehouse by loading dimensions and facts as part of project. Also worked for different enhancements in FACT tables.

•Created shell script to run DataStage jobs from UNIX and used control-M as scheduling tool.

•Coordinate with team members and administer all onsite and offshore work packages.

•Performed performance tuning of the jobs by interpreting performance statistics of the jobs developed.

•Documented ETL test plans, test cases, test scripts, and validations based on design specifications for unit testing, system testing, functional testing, prepared test data for testing, error handling and analysis.

•Developed Test Plan that included the scope of the release, entrance and exit criteria and overall test strategy. Created detailed Test Cases and Test sets and executed them manually.

Environment: IBM Info sphere DataStage 8.5/9.1, Oracle 11g, PL/SQL, SQL Server, Flat files, UNIX, TOAD, Control-M.

LABCORP Jun’13 – Feb’15.

Burlington. NC.

ETL Developer

This project involves Working on Database of a Patient Health portal of LabCorp. This Application – ‘Patient portal’ is used by patients to register, see their reports and pay bills online. On the other end, LabCorp internal three tier employees use this to update the Test Reports of Labs into this portal. This Application developed on Java and used Oracle and Teradata DB. We use to extract reports of all Patient details, their Billing and Health reports.

Roles and Responsibilities:

Requirement analysis and gathering to provide technical and architectural support to the team.

Documented the Purpose of Mapping as to facilitate the personnel to understand the process and incorporate the changes as and when necessary.

Used DataStage Designer to develop processes for Extracting, Cleansing, Transforming, Integrating, and Loading data into Data warehouse.

Worked on logical and physical design of Data warehouse, identified sources/targets, and analyzed source data for dimensional modeling.

Worked and followed all HIPPA rules while creating Data warehouse and maintaining it.

Developed the data warehouse repository using DataStage Manager by importing the source and target database schemas.

Integrate structured and semi-structured data from QNXT Claims, Enrollment, and Provider modules into enterprise data warehouses.

Extracted and loaded QNXT membership, claims, provider, and billing data into data warehouses and downstream applications.

Extensively designed, developed, and implemented Parallel Extender jobs using Parallel Processing (Pipeline and partition) techniques to improve job performance while working with bulk data sources.

Extensively worked with architects and proposed solutions in building common design approach for building the Job control and error recording tables

Created Naming Standards Document and Low-Level Design Documents and used them across the Projects.

Defect Tracking, unit testing, defect reporting, analyzing results and documentation.

Environment: IBM DataStage 8.5/9.1, Microsoft SQL 2005/2008, IBM DB2, PUTTY, WINSCP and HP Quality Center

ASURION Jul’11 – May’13.

Nashville. TN.

ETL Developer

This project involves the support and maintenance of a Claim Management System. Customers and internal customer care executives, respectively use the Applications – ICARE and IWEB. The Source and Target included Oracle, DB2 and SQL Server. We use to Extract Data of all Claims and Customer details and created jobs to fetch hourly reports.

Roles and Responsibilities:

Experience in testing mobile features in different environments, which includes Process of new claims, and payment related features.

Conducted source information profiling using Quality Stage Investigation stage and Information Analyzer

Developed parallel jobs to load data into XML files using XML input and XML output stages.

Extensively used Quality Stage patterns like CASS, Standardize, Match Frequency, etc. to standardize name, address, and consumer information.

Ensured Data Quality and similarity matches using Quality Stage Patterns

Implemented Surrogate Key Generator Stage, Slowly Changing Dimension Stage etc.

Worked on the DataStage server in both AIX and UNIX simultaneously using virtual machine.

Created the Enterprise Data Warehouse Data Migration ETL design document utilizing Info Sphere DataStage and Quality Stage

Created the Quality Stage match specifications while training the employees in Quality Stage and DataStage concepts and programming.

Created project plan, monitored the progress on all tasks while working closely with the data modelers and data base administrators to create the conceptual, logical, and physical design of the Enterprise Data Warehouse.

Implemented Slowly Changing Dimension (SCD) as per business requirement for saving history using WebSphere DataStage 8.1 for dimension tables in Star Schema

Used DataStage Director and the runtime engine to schedule running the server jobs, monitoring scheduling, and validating its components.

Extensively worked on Quality Stage in Information Server 8.5 for Data Cleansing and Standardizing

Environment: IBM DataStage 8.0.1, IBM AIX 5.2, Oracle 10g, XML files, sequential flat files, TOAD.

AT&T May’10 – Jun’11.

Earth City, MO.

ETL Developer

AT&T is a one of the leading Tele-communication companies providing largest wireless voice and data services. DAWN (Detailed Agent Workflow Navigation) is a custom-Built Web Application, developed to give access to Agent to login, verify the account information, to update personal details, view payment history and bill details. This system is developed on .NET platform and SQL DB’s. We extracted all Daily and Weekly reports using DataStage and share all details on Internal SharePoint.

Roles and Responsibilities:

Worked on End-to-end implementation, maintenance, optimizations and enhancement of the application.

Worked on various phases in the project like project estimation, Requirement Gathering and Analysis, Designing, Coding, User Acceptance Testing to deployment in production, including post implementation support.

Assisting the Management and Quality team (QMS) by providing various Metrics and data for the project.

This is important as View360 projects are very process oriented and emphasize on Quality standards.

Handling Change requests - Understanding the application workflow in the existing jobs on DataStage and applying new changes to them, testing & Review activities.

Adhering to the Process and creating and posting all the required documents and deliverables like the UTC document, Review Checklist, etc.

Created a data dictionary to map the business requirements to attributes to design logical data model.

Developed Data validation and error handling jobs.

Assisting the Project Management team in providing the estimation of projects, so we could plan estimate the effort and time by when we could deliver the best solution.

Extensively worked with DataStage Shared Containers for Re-using the Business functionality

Worked on performance tuning to address very critical and challenging issues.

Co-ordinate with SIT and UAT team to fix the Test Problem Reports

Environment: DataStage 7.5, Oracle 9i/10g, DB2, flat files, PL/ SQL, Windows 2003 Server, SQL Server.



Contact this candidate