RAJESH BULLEDDULA [pic]
abnfn9@r.postjobfree.com Mobile:
PROFESSIONAL SUMMARY
. Over 6 years of IT experience in design, development and
implementation of data integration, client/server and Oracle
development applications
. 5 years of Extraction, Transformation and Loading (ETL) experience
using IBM Websphere DataStage 8.0.1/8.0, Ascential DataStage
7.5/7.0/6.x/5.x (Administrator, Director, Manager, Designer), Parallel
Extender, Quality Stage (Integrity) and development of Data
Warehouse/Data mart Applications
. Expertise in Dimensional data modeling, Star schema modeling, Snow-
Flake modeling, identification of fact and dimension tables,
Normalization, Physical and Logical data Modeling using Erwin and
Oracle Warehouse Builder to implement Business Intelligent systems
. Experience in Full Software Development Life Cycle (SDLC) in
Collecting Requirements, Design, Coding, Unit Test Plan Preparations,
System Testing and Project documentation
. Experience in both Parallel extender jobs and Server Jobs in
DataStage.
. Designed and developed jobs using Parallel Extender for splitting bulk
data in to subsets and to dynamically distribute to all available
nodes to achieve best job performance
. Strong knowledge in scheduling DataStage jobs using Crontab as well as
preparing Test Plans and Testing of Batch programs - changes to
scheduling (Autosys and Jobtrac)
. Excellent experience in troubleshooting of DataStage jobs and
addressing production issues like performance tuning and enhancements.
. Excellent knowledge of studying the data dependencies using Metadata
stored in the DataStage Repository
. Experience in Integration of various data sources like UDB, Sybase,
Oracle, TeraData, DB2UDB, SQL Server, Sybase, and Ms-Access
. Experience in creating reports using Mainframe SAS and Unix SAS
. Hands on experience in SQL Loader for data migration from Legacy
system
. Experienced in UNIX Shell scripting (Korn, BASH, and CSH) as part of
file manipulation
. Good understanding of Cognos, Microstrategy and Business Objects OLAP
tools
. Extensive experience in client-server and internet application
development using Oracle, MS SQL server, PL/SQL, stored procedures,
Functions, Triggers, ODBC and Visual Basic
. Experience in UNIX, Linux, Z/OS operating Systems
. Profound knowledge in ASP, C#, VB6, XML, Java, C++ and .NET
. Experienced in translating user requirements to technical
specifications and writing system specifications
. Strong understanding of business processes & its interface with IT.
Highly organized and detail-oriented professional with strong
technical skills.
. Excellent communication skills and interpersonal abilities with
maximum contribution to attain the team goal
TECHNICAL SKILLS
ETL Tools : IBM Websphere DataStage 8.0.1/8.0, Ascential
DataStage 7.5/7.1/6.0/5.2/5.1
(Parallel Extender, Quality Stage)
OLAP Tools : Business Objects 6.1/5.1/5.0/4.1 (Supervisor,
Designer, BO Reporter, WEBI,
BCA, Info View), Cognos, Microstrategy 7.2/7.1
Databases : Oracle 11g/10g/9i/8i/8.x, MS SQL Server
7.0/6.5, DB2 UDB, MS Access 7.0,
TeraData, ODBC, PeopleSoft CRM 8.4/8.9, Sybase
Database Tools : SQL* Plus 9.2, SQL Loader 9.2, SQL Navigator
4.1, TOAD 7.5, DB2 Import, DB2
Export
Data Modeling Tools : Erwin 4.1/3.5, ER/Studio 3.5, Microsoft
Visio 2000
Other Tools : Autosys 4.5, Rational ClearCase and
ClearQuest, MS Visual Studio
Operating Systems : Windows 7/Vista/XP/2000/NT/98/95, Unix,
Sun Solaris 5.8, IBM AIX 5.x, OS390,
z/OS
Prog. Languages : COBOL, SAS, SQL, Transact SQL, PL/SQL, Unix
Shell scripting, C, C++, Java,
VB 5.0
Web Development : J2EE, Applets, Servlets, EJB, XML, HTML 4.0,
ASP
Educational Qualification
. Bachelors in Electronics and Communication Engineering (JIIT
University, INDIA) GPA: 3.7
CERTIFICATIONS
. Oracle PL/SQL Developer Certified Associate (OCA)
PROFESSIONAL EXPERIENCE
Freddie Mac, McLean, VA
May 2009 - Present
ETL Developer
Freddie Mac is a shareholder owned company whose people are dedicated to
lowering the costs and increasing access to quality housing for more of
America's families. The corporate Data Warehouse (CDW) is a centralized
data repository that makes accessible historical and current business
relevant data as of the prior business day from disparate data sources
residing within the company transaction processing and end-user systems
into one relational database for decision support, management reporting, ad-
hoc and trend analysis, modeling and data mining purposes. The data
warehouse can be made up of both detailed, summary and aggregate point-in-
time data and is founded upon the historical archive of detailed data that
is needed to support today's business requirements, as well as the
unanticipated business questions of the future. The objective of this
project is to provide an integrated Mainframe IMS data from various sources
like PE (Project Enterprise), Midas (Mortgage Information Direct Access
System) to Corporate Data warehouse.
Responsibilities:
. Designed and Created Parallel Extender and sequencer jobs which
distribute the incoming data concurrently across all the nodes, to
achieve the best performance
. Developed COBOL/SAS/JCL programs to retrieve IMS Data and load the
Data to CDW Tables
. Used Mainframe and Unix SAS to generate various DQ reports
. Used SAS to compare Source Data and Target Database
. Used the DataStage Designer to develop various jobs for extracting,
cleansing, transforming, integrating, debugging, and loading data into
data warehouse
. Used the DataStage Director and Manager for monitoring, validating and
debugging DataStage components
. Integrated data from several different legacy sources and loaded it to
the base tables
. Created Autosys JIL (BOX, CMD, FWT) Scripts to schedule Unix Shell
scripts
. Performed Unit Testing, Integration Testing and User Acceptance
Testing (UAT)
. Created process flows diagrams using Microsoft VISIO
. Performed Tuning of the ETL programs to increase the processing and
loading efficiency
. Extracted data from Mainframe Flat files, Unix Flat Files, Sybase, DB2
legacy system and loaded into the Corporate Data Warehouse
. Used Mainframe Change Management (MCM) to manage projects, production
hotlines on Mainframe
. Designed and developed Code table extract process
. Used Clear Case/Clear Quest for the versioning and migration of code
to production
. Monitor, debug, resolve and support troubleshooting of production
issues
Environment:
Ascential DataStage 7.5.1, DB2UDB 8.0, SAS, Sybase, Autosys, Cobol, JCL,
MVS, SQL, PL/SQL, Shell Scripts, UNIX, Windows XP, OS/390, AIX 5.2,
Jobtrac, Clearcase and Clear quest, Rapid SQL.
Blue Cross Blue Shield, Berlin, VT
Jan 2009 - Apr 2009
ETL Developer
The Blue Cross and Blue Shield Association (BCBSA) is a national federation
of 39 independent, community-based and locally operated Blue Cross and Blue
Shield companies. As the nation's oldest and largest family of health
benefits companies, the Blue Cross and Blue Shield Association prides
itself on being the most recognized brand in the health insurance industry
along with many other celebrated milestones. As a developer, was involved
in loading the data marts for the provider, members & claims. Involved in
extracting data from their legacy system sources and loading into a Data
Mart. The major job involved in cleansing the data and transforming the
data to the staging area then loading the data in the Data Mart for
analyzing business performance at various stages.
Responsibilities:
. Involved in gathering Business Requirements for reports and came up
with standard Requirement gathering documents
. Designed the mappings between sources (external files and databases)
to staging and target warehouse database
. Used the DataStage client tools to develop processes for extracting,
cleansing, transformation, integrating and loading data into data
warehouse database (DB2 UDB)
. Implemented the Slowly Changing Dimension, SCD Type-II strategy
in DataStage, for the OLAP Dimension tables storing history data
. Designed and developed several extract programs to send files to third
party vendors
. Designed and developed Code table extract process
. Created DataStage jobs using DataStage Designer and extracted data
from various sources, transformed data according to the requirement
and loaded into data warehouse schema
. Involved in tuning DataStage jobs for better performance
. Used the DataStage Director and the runtime engine to schedule running
the solution, testing and debugging its components and monitoring the
resulting executable versions
. Create master controlling sequencer jobs using the DataStage Job
Sequence
. Used Clear Case/Clear Quest for the versioning and migration of code
to production
Environment:
IBM Information Server 8.0, SAS, DB2 UDB 9.0, Linux, DB2 Control Center,
SQL, PL/SQL, Shell Scripts, UNIX, Windows XP, OS/390, Jobtrac, Clearcase
and Clear quest, Rapid SQL.
Freddie Mac, McLean, VA
Feb 2008 - Dec 2008
ETL Developer
Freddie Mac is a shareholder owned company whose people are dedicated to
lowering the costs and increasing access to quality housing for more of
America's families. The objective of this project is to provide an
integrated data from various DB2 database sources in LP (Loan prospector),
SAP (Security Accounting Project) to FAS 140 in Corporate Data warehouse.
Responsibilities:
. Designed and Created Parallel Extender and sequencer jobs which
distribute the incoming data concurrently across all the nodes, to
achieve the best performance
. Used Mainframe SAS to extract data from DB2 and Sybase Tables to
create Load Ready Files
. Used SAS to compare Source Data and Target Database
. Used the DataStage Designer to develop various jobs for extracting,
cleansing, transforming, integrating, debugging, and loading data into
data warehouse
. Used the DataStage Director and Manager for monitoring, validating and
debugging DataStage components
. Integrated data from several different legacy sources and loaded it to
the base tables
. Created Autosys JIL (BOX, CMD, FWT) Scripts to schedule Unix Shell
scripts
. Performed Unit Testing, Integration Testing and User Acceptance
Testing (UAT)
. Performed Tuning of the ETL programs to increase the processing and
loading efficiency
. Extracted data from Unix Flat Files, Sybase, DB2 legacy system and
loaded into the Corporate Data Warehouse
. Imported/Exported data between Development and Production using DB2
export/import utility
. Designed and developed Code table extract process
. Used Clear Case/Clear Quest for the versioning and migration of code
to production
. Monitor, debug, resolve and support troubleshooting of production
issues
Environment:
Ascential DataStage 7.5.1, DB2, DB2UDB 8.0, SAS, Sybase, Autosys, SQL,
PL/SQL, Shell Scripts, UNIX, Windows XP, OS/390, AIX 5.2, Jobtrac,
Clearcase and Clear quest, Rapid SQL.
Verizon Business, Silver Spring, MD
Jun 2007 - Jan 2008
ETL Developer
The Verizon Business Data warehouse (VzBDW) is to provide billing
information in a standard format for the various Verizon billing systems
across all Verizon jurisdictions. The main purpose of this project is to
integrate the Data from MCI systems to the existing Verizon system.
Responsibilities:
. Gathered Business Requirements by working closely with Business users
. Prepared design documents and mapping documents source system
(Mainframe) and the warehouse tables (TeraData) and Designed the jobs
based on the understanding the data model table relationship
. Extensively worked on Data Extraction, Transforms, Loading and
Analysis
. Developed Job sequencers for executing DataStage jobs
. Developed jobs in Parallel Extender using different stages like Join,
Transformer, External filter, Row generator, Column generator, Merge,
Lookup, Remove Duplicates, Copy, Filter, Funnel, Dataset, Lookup File
Set, Change Data Capture, Modify, and Aggregator
. Used Parallel Extender for distributing load among different nodes by
implementing pipeline and partitioning of data in parallel extender
. Used different Partition methods and collecting methods for
implementing parallel Processing
. Created Parameters and Environment variables to run the jobs
. Importing/Exporting the DataStage projects and taking the backup
. Integrated data from various sources into the staging area in data
warehouse for integrating and cleansing data
. Extensively used all components of DataStage (Manager, Designer and
Director) for various development and support activities
. Used Clear Case/Clear Quest for the versioning and migration of code
to production
. Loading and Validating Data from mainframe flat files to TeraData
Tables
. Involved in coding of scheduling module using java, servlets and jsp
Environment:
Ascential DataStage 7.5, Shell Scripts, TeraData, UNIX, Windows XP,
TeraData SQL Assistant, Java and Sun Solaris 5.1, SQL, PL/SQL, OS/390, AIX
5.2, Jobtrac, Clearcase and Clear quest, Rapid SQL.
Freddie Mac, McLean, VA
Feb 2006 - May 2007
ETL Developer
Freddie Mac is a shareholder owned company whose people are dedicated to
lowering the costs and increasing access to quality housing for more of
America's families. The objective of this project is to provide an
integrated Mainframe IMS data from various sources like PE (Project
Enterprise), Midas (Mortgage Information Direct Access System) to Corporate
Data warehouse.
Responsibilities:
. Designed and Created Parallel Extender and sequencer jobs which
distribute the incoming data concurrently across all the nodes, to
achieve the best performance
. Developed COBOL/SAS/JCL programs to retrieve IMS Data and load the
Data to CDW Tables
. Used Mainframe and Unix SAS to generate various DQ reports
. Used SAS to compare Source Data and Target Database
. Used the DataStage Designer to develop various jobs for extracting,
cleansing, transforming, integrating, debugging, and loading data into
data warehouse
. Used the DataStage Director and Manager for monitoring, validating and
debugging DataStage components
. Integrated data from several different legacy sources and loaded it to
the base tables
. Created Autosys JIL (BOX, CMD, FWT) Scripts to schedule Unix Shell
scripts
. Performed Unit Testing, Integration Testing and User Acceptance
Testing (UAT)
. Performed Tuning of the ETL programs to increase the processing and
loading efficiency
. Extracted data from Mainframe Flat files, Unix Flat Files, Sybase, DB2
legacy system and loaded into the Corporate Data Warehouse
. Used Mainframe Change Management (MCM) to manage projects, production
hotlines on Mainframe
. Designed and developed Code table extract process
. Used Clear Case/Clear Quest for the versioning and migration of code
to production
. Monitor, debug, resolve and support troubleshooting of production
issues
Environment:
Ascential DataStage 7.5.1, DB2UDB 8.0, SAS, Sybase, Autosys, Cobol, JCL,
MVS, SQL, PL/SQL, Shell Scripts, UNIX, Windows XP, OS/390, AIX 5.2,
Jobtrac, Clearcase and Clear quest, Rapid SQL.
American International Group Inc, (AIG), NY
Sep 2005 - Jan 2006
ETL Developer
AIG is a leading insurance corporation with Various Investment programs.
The project was to maintain the Customer information Database for
supporting the centralized SAP system. The main aim of the project is to
develop web services in DataStage to maintain Customer information through
different systems.
Responsibilities:
. Re-architecture the Entire design due to performance issues
. Developed DataStage code with an approach to make them as web services
. Developed special strategies of DataStage development in order to
integrate it as a batch process and also as a RTI service
. Extensively worked on the Data Model to decrease the complexity of
queries
. Created a shared container approach in order to make the code more
visible
. Involved in developing a system plan to atomize the batch runs
depending on profit center requirements
. Extensively used Quality stage for data Cleansing and to standardize
address information
. Designed documents for maintaining best practices of DataStage
. Responsible for developing test data and stress Test analysis
. Designed complex logics in DataStage with out hash files or Flat files
in order to improve the performance
. Wrote Unix Shell Scripts to automate the process
Environment:
Ascential DataStage 7.5 (Administrator, Manager, Designer, Director,),
COBOL, JCL, SAS, Windows NT, Oracle 10g, SQL, PL/SQL, SQL Server 7.0, DB2,
Quality Stage, Web services Pack, Real Time Integration services
Daimler Chrysler, Detroit, MI
May 2004 - Aug 2005
ETL Developer
Daimler Chrysler Inc., with headquarters in Detroit, Michigan, USA is among
the three automobile giants of USA. This project aims at providing
maintenance, production support and enhancement to Chrysler Corporation. My
applications are related with MOTOR PARTS (MOPAR) of Chrysler Corporation.
Responsibilities:
. Prepared mapping documentation based on requirement specification.
. Redesigned the existing jobs with a different logical approach to
improve the performance.
. Involved in the creation of jobs using DataStage Designer to validate,
schedule, run and monitor the DataStage jobs.
. Involved in designing the process for loading the data from all the
source systems to Operational Data Store.
. Designed Jobs, which extracts data from multiple source systems and
Flat Files and transform the data and create Dat files.
. Create Sequential Files for the Lookup Code Tables and Loaded Directly
to the DB2 Data Base.
. Involved in developing Shell scripts for loading Dat files into DB2.
. Designed jobs with Stages like LinkCollecter and LinkPartitioner for
Initial data movement run to handle the very huge volumes of Data
. Used the DataStage Designer to develop processes for extracting,
cleansing, transforming, integrating, and loading data into data
warehouse database.
. Wrote Unix Shell Scripts to automate the process.
. Prepared documentation for addressing the referential integrity
relations in between the tables at ETL level and monitoring of jobs
Environment:
Ascential DataStage 6.0 (Parallel Extender), SQL Loader 9.2, TOAD-7.5, SQL*
Plus 9.2, Windows NT, Sun Solaris 5.8, Erwin 3.5, DB2 and QMF
HCL Technologies, INDIA
Aug 2003 - Apr 2004
System Analyst
Domain Party Associates - Development of Reusability Components
This project involves in Developing of new components, which can be used as
reusable components by other domain areas. This will allow the promotion
of these components to a broader scale and a reduction of redundant
functionality. This, in turn, will reduce the amount of code for service
and maintenance, lowering the expenses and improving efficiency.
Responsibilities:
. Designing of new components (3 Modules) which includes design of
Technical Spec document, design of Physical and Logical Interface
document
. Coding, Preparing unit Test Case Design Document and Unit Testing
. Issuing the Components to Production
. Supporting during System and Integration Testing
. Production support
. Worked on programs for scheduling data loading and transformations
using DataStage from legacy systems to Oracle 8I using SQL*Loader and
PL/SQL
. Involved in creation of jobs using DataStage Designer to validate,
schedule, run and monitor the DataStage jobs
. Involved in designing the procedures for getting the data from all
systems to Data Warehousing system
. Implementation of Surrogate key by using Key Management functionality
for newly inserted rows in Data warehouse
. Exporting the universe to the Repository to make resources available
to the users
Environment:
Ascential DataStage 6.0/6.1, SQL, Business objects 5.1.3, Microsoft Visio,
Shell Scripts, UNIX, Windows 2000, IBM 3090, COBOL, PL/1, MVS JCL, DB2 on
MVS/ESA, TSO/ISPF, TSO DBX, SAS, TSO DB2MENU and Xpeditor