Summary
Around * years of experience in the field of Software Quality Assurance and
Testing with a strong exposure to all phases of Software Development Life Cycle
(Initiation, Design, Development, Testing and Implementation) and Software
Testing Life Cycle (Test Planning, Test Execution and Test Analysis)
3 years of experience as QA Lead/Lead Tester in handling multi-million dollar
project.
3 years of extensive experience in performing data warehousing testing using
Informatica.
Gathered, reviewed, understood and performed gap analysis on Business and
Functional Requirements and Use case requirements
Understood the project requirements, created test strategies, test plans, test
scenarios and test conditions/cases
Ability to organize test strategy and test plan review session with the stake
holders and end users
Established communication bridges with development, implementation and end user
teams
Created QA project plan based on master project plan created by the project
manager
Organized knowledge transfer sessions from the business team to better understand
the applications that are involved in the project
Performed Test planning, Test execution and Test analysis using Quality Center
and have an ability to import requirements and test cases from excel or word
Managed a team of 15 QAs on both onsite and offshore capacity
Coordinated and managed all the testing tasks from both onsite and offshore teams
Organized testing status meetings and provided inputs to the project manager and
QA manager
Involved in the projects risk analysis meeting and provided testing related risk
to the project manager
Ensured that all the testing related risks are met by coordinating with
development, DBA, infrastructure and deployment teams
Created WBS - work breakdown structure and RACI - responsible accountable
consulted and informed documents by outlining all the tasks to be completed for
this project and reviewed with the team
Created daily testing checklist to make sure that testing is on track as per the
time lines and deadlines
Coordinated and performed testing on applications that are in scope of the
project using both automated and manual testing on Web based, Client/Server and
Mainframes applications in Windows, UNIX, Mainframes and Web environments
Executed testing as per the plan using different testing techniques like Sanity,
Functional, Black Box, System, Penetration, Interoperability, Integration, End to
End, Adhoc, Smoke, Cross Browser, Cross Platform, Regression, Batch, Database
(Data model, Data Mapping, Data Validation, Data Integrity), User Acceptance,
Load, Stress and Performance Testing on client server, online and mobile
applications
Created automation test strategy and used HP Mercury Quick Test Professional for
creating automation regression suites using VB Scripting
Maintained automation regression testing suites for multiple projects
Experienced in working with Multiple browsers and Cross Platform environments
Extensive Experience in UNIX (AIX/Solaris/ HP-UX 11.x), and Windows Operating
system
Extensively used SQL and PL/SQL to write Stored Procedures, Functions, Packages
Goal oriented, self-motivated with excellent communication, analytical and
interpersonal skills
Ability to work in a team oriented, collaborative environment, customer oriented,
highly versatile, innovative, with excellent interpersonal, problem solving,
trouble shooting and communication skills
Quick learner and adaptive to new and challenging technological environments
Education
Bachelors in Computer Science & Engineering, JNTU University, India 2005
Experience
Alberta Blue Cross, Edmonton, AB
Jun 2012- Till date
QA Lead/Lead Tester,
Administration Redevelopment System (ASR-Sales) is an enhancement application
designed and developed to replace the existing Application. The application is
designed with robust features to satisfy the organization's goal to improve the
customer service experience and the way the handling of the current loads in
processing claims and administration.
Worked closely with the team management in creating the Master Test Plan, Test
Strategy and Resource Planning, and Risk mitigation.
Identified the complexities involved in the spaghetti architecture and created
the test scenarios
Involved in writing the test cases in the Test Plan of the Application Lifecycle
Management (ALM 11 tool)
Created Test conditions and mapped the requirements traceability from the
Requirements folder to the Test cases in the Test Plan to ensure the coverage is
completed.
Worked across the teams in setting up the test data to facilitate the test
execution
Assigned tasks and worked along with the team members to perform the Sanity
testing, system testing, system integration testing etc.
Created automation test strategy and used Quick Test Professional (QTP) for
creating automation regression suites using VB Scripting
Involved in the test case execution in the Test Lab and logged the defects
accordingly
Worked with development team to ensure the defect is fixed and a follow-up with
the team members to retest and proceed with the action on the defect based on its
nature of pass or failure.
Worked on the XML file cloning to create the sample data to test the application.
Conducted weekly team meetings to identify the road blocks while writing the test
cases, test case executions, defect fixation and reported to the team management
immediately.
Tasks were assigned such that during the progression of the test execution if the
application is down, a parallel task is assigned to ensure the productivity and
time of the resource is utilized to the maximum based on the priorities.
Involved in the defect triage meetings and laid plans in finding the solutions
Since the application comprises of multiple projects, proper handling of the
Service Requests (SRs) is a major task and should work closely with the inter
project team managements.
Coordinated the BPV (Business Process Validation) with business sponsors of other
project teams involved as part of the application.
Performed system integration testing with other projects and smoke testing to
ensure the accuracy of the application.
Worked closely with the UAT Team of business users using the production data and
published the reports on a daily basis.
Environment: HP Application Lifecycle Management 11.0, HP Quick Test
Professional, UNIX (HP-UX), Oracle 11g, Windows 7, Java, Putty, SQL Developer
(Query Tool), MS Office Suite, MS Visio, Apache, Weblogic, Snag It.
Ministry of Education, Calgary, AB
Jun 2010 - Mar 2012
QA Lead,
Approval Framework data warehousing process is a combination of legacy and
current data sources. The approval framework (DF) is a decision-making process
designed to assist school boards in making informed decisions, based on what
their analytical data, about areas that require special interest and
concentration to provide better results and performance from the students.
Performed Business Analysis to plan and create the Master Test Plan.
As the QA Stake holder, conducted meetings with business users to create the UAT
and Use cases.
Worked as a QA in testing the Data Migration project
Understood the complex architecture of the project
Identified different data sources that are in scope the project and understood
its process
Organized and understood the complete end to end business process
Involved in the project as QA Lead in managing and performed testing on the
complete data warehousing process
Involved in the documentation of master test strategy with the inputs from the QA
manager
Published the draft version of the master test strategy to the project members
for review and sign-off
Provided test estimates to the project manager based on testing scope and tasks
Identified with work effort and created Work Breakdown Structure and published to
the team members
Performed test management and Defect management using HP Mercury Quality Center
Created detailed Manual Test Scenarios and Test cases according to the
Business/Functional requirements
Tested various Informatica mappings with Tuning Techniques to improve the
mapping/session performance
Performed Data testing, Transformation testing, Business rules and Data quality
Wrote complex SQL Queries to incorporate ETL logic and to validate Test Cases
Performed functionality testing on both online and client server application
using converted data
Checked the results against to the expected results and checked and validate the
Data validity
Involved in data warehouse testing by checking ETL procedures/mappings
Worked on understanding of ETL plan and ETL mappings for the data flow
information
Also involved in working of Transformation rules in the ETL tool
Worked with Informatica mapping ETL plan for analyzing the business rules are
mapped and designed properly
Involved in testing of output of transformations data in the destination tables
Worked extensively on Source Qualifier, Sorter, Aggregator, Normalizer,
Expression, Filter, Router, Update strategy, Lookup, XML, Stored Procedure,
Sequence Generator, and Union Transformations.
Tested the complex ETL mappings and configured the sessions and workflows to
populate the Decision Frame worktables in the Data Warehouse.
Worked with Star Schema model to design dimensional and fact tables.
Performed Manual testing for Functional, Integration, End to End and UAT testing
for Data Migration project
Reported defects with proper priorities and severities using HP Mercury Quality
Centre
Generated and published customized defect reports to the project team and
organized defect triage meetings
Environment: Informatica Power Center 9.0, Business Objects Crystal Reports, HP
Mercury Quality Center 10.0, HP Mercury Quick Test Professional 10.0, Mercury
LoadRunner 9.0, UNIX (HP-UX), Oracle 11g, Windows XP Professional, Flat files,
SQL Server 2008, SQL Developer (Query Tool), Unix
AMA - Alberta Motor Association Insurance, Edmonton, AB
Jan 2009 - May 2010
Senior QA Analyst, Marketing Analytics Portal System
AMA Marketing Analytics System will deliver a business intelligence tool, which
enables sales and marketing department to capture and leverage information to
support and provide insights into the business. The solution involves the
extraction of customer level data; merging with outside lists and demographics;
and includes data from AXA transactional systems; results of sales efforts from
call center and sales systems, as well as other vendor services.
Involved in the project as ETL Informatica QA Analyst in testing the Marketing
Analytics System
Requested overview of the complete business process from the business users of
AXA insurance
Identified all the project documented and gathered MAS project strategy and
understood the conversion process and conversion time lines
Worked closely with project manager and QA manager to understand the deliverables
Understood the functional and business requirements of the project by organizing
review sessions with the Business analyst
Involved in the creation of master test plan by outlining test assumptions,
risks, timelines, roles and responsibilities, build management, entry/exit
criteria and deliverables
Worked closely with the ETL and Development team to understand the business
process
Created detailed Test Scenarios and Test cases according to the
Business/Functional requirements for the MAS application
The solution provides the ability to integrate external list purchases
maintaining the required data quality against AXA's information for use in the
marketing process
Tested the Sales & Marketing Data mart, which includes the data model, business
rules, Extract/Transformations/Load procedures and schedules
Tested the common definitions using the business rules for all business regions
reporting on member, lead or prospect data
Validated the forecast pipeline activity, which defines when a Lead becomes the
member by generating reports
Performed System integration testing with the Lead Management System (LMS)
historical data that is critical for campaign management (primarily campaign
related measures based on the campaign history needs) will be migrate to the MAP
environment
Verified and tested the Informatica transformations like Expression, connected
and unconnected Lookups, Source Qualifier, Filter, Aggregator, sorter, Sequence
Generator, Normalizer, Router, Joiner, Stored Procedure etc
As part of performance testing validated and verified the Performance tuning, ETL
Procedures and processes
Tested ETL mappings using Informatica Power Center to move information from
multiple sources into staging tables and then to common consolidated data area of
Data Warehouse and then to Flat files as requested by the vendor
Executed PL/SQL Procedures and Functions for Procedure transformations also
created and used different tasks like Decision and email tasks
Performed Analytics reports testing based on the requirements
Identified and reported data mapping, data integrity, data corruption issues to
the ETL and Development teams
Used File Aid, TSO to perform data validation checks
Provided daily and weekly status to Test Manager
Environment: Informatica Power Center 8.6, HP Mercury Quality Center 9.0, HP
Mercury Quick Test Professional 9.0, Mercury LoadRunner 9.0, Cognos, Windows XP,
UNIX, Oracle 11g, SQL Server 2008, MS Excel (Source), Flat files (End Targets),
TOAD (Query Tool), PL/SQL.
GSK Pharmaceutical, Montreal, QC Jul 2007 - Dec 2008
Senior QA Analyst, Online Oracle Siebel Sales Application
Siebel Sales application is used by all the GSK organization throughout the world
to project and forecast sales. Siebel Sales application collects data from
different markets (Asia, Africa, Europe) data sources and loaded into historical
database. Using the historical data analytics processing is performed to generate
multiple reports that will be used by the marketing and sales department to
analyze the data.
Identified all the project documents and gathered them from Microsoft SharePoint
Requested walkthrough sessions from the development team to understand the
overall process of the Oracle Siebel Sales process
Involved in requirement discussion to understand the priorities of the
requirements from the business side
Involved in creating test strategy and test plan based on the requirements and
published to the project team
Drafted detailed testing scenarios to cover all the requirements and published
these scenarios to the DBA for the creation of test data and test beds
Understood and analyzed the market business data from different countries for
creation of test data
Involved in the test data creation for specific type of testing
Created detailed test cases in Excel and imported into Quality Center using
export process
Imported requirements from word documents into Quality Center
Mapped test cases to specific requirements to ensure requirement test coverage
Included test data parameters during test case creation in Quality Center
Executed the test cases for sanity, functionality, integration, security,
backend, shakedown, end to end, regression and performance testing
Published test planning and test execution reports generated from Quality Center
to the management
Identified defects and reported to the development team with defect details and
coordinated with the development team on regular basis to understand the status
of the defects
Performed UAT testing as requested by the business support team
Performed Data conversion, Quality and Verification testing on the data from
different data sources.
Developed UNIX (Korn shell) scripts for ftp'ing files, to execute the SQL
scripts, Stored Procedures etc.
Used SQL Loader to load data from flat files to the Database tables for testing
purpose
Scheduled the UNIX jobs/batches comprising of Informatica workflows, Oracle SQL
scripts, and ftp process using the control-m job scheduler.
Updated all the test documentation with the actual results and stored them in
release binder for auditing purpose
Environment: Informatica Power Center 8.5/8.1, HP Mercury Quality Center 9.0, HP
Mercury Quick Test Professional 9.0, Mercury LoadRunner 9.0, Informatica Power
Exchange 8.x, Crystal Reports, Mainframes, UNIX (SUN-Solaris), Oracle Siebel 6.0,
Oracle 10g, Windows XP Professional, Flat files, SQL Server 2005, SQL Loader,
Serena Version Manager, Toad (Query Tool), Control-m scheduler, Korn shell
Scripting.
Canadian Western Bank, AB
Mar 2006 - Jun 2007
QA Analyst, Banking Portal Application
Banking Portal Application is used by the internal and external employees of the
bank to access the banking products, services, rates and apply for multiple
credit products offered by the bank. The banking portal application is integrated
with Vignette content management, which is used for centralizing the entire
content that gets published on to the application.
Reviewed and understood the Banking portal application requirements
Requested the BA and PM to organize a walkthrough and an overview from the
business to better understand the application and project
Involved in the creation of test plan with the inputs from the stake holders and
manager
Used Quality Center to manage the complete end to end process of testing
Administered the projects from Quality Center from testing perspective
Imported and created requirements from the word documents into the Requirements
module
Documented detailed test cases in the test plan tab and mapped the test cases to
the requirements to ensure the coverage of testing with the requirements and to
ensure the traceability
Created different test sets based on priority and business test and mapped the
specific test cases into the test sets
Executed the test cases from different test sets and reported defects in the
defects tab with proper priorities and severities
Produced daily test execution and defect status reports to the project team
Executed test cases for functionality, system integration and end to end testing
Involved in the creation of UAT test plan, UAT test scenarios and UAT test cases
Performed backend database testing on oracle database using SQL
Executed test cases to check the content and integration with the Vignette
content management server
Participated in the daily defect status and test execution status meetings
Environment: ASP, ASP.NET, XML, Weblogic, HP Mercury Quality Center 8.0, HP
Mercury Quick Test Professional 8.0, Mercury LoadRunner 8.0, NUnit, Vignette,
Oracle 10g, SQL Plus, Toad, XML, UNIX Shell Scripting.
Hartford Life Insurance, CT
Oct 2005 - Feb 2006
QA Analyst, Claims Management System
Claims management system is a web based application used by the customer and
underwriters to report, view and process the claims that are reported by the life
insurance policy customers. Online Claims Management system is integrated with
printfree application that will be used by the underwriter to print and send the
office documents to the customer
Created test cases in excel and mapped the test cases to the requirements using
RTM - requirement traceability matrix
Provided the test cases to the BA for review and sign off
Stored test case in specific folders based on specific types of testing to test
execution purpose
Performed System integration testing between Claims Management System and FASAT.
Performed System and Functionality Testing on FASAT application
Produced daily test execution and defect status reports to the project team
Performed sanity, functionality, system integration, end to end, user acceptance,
security, navigation, security, content, regression, load performance and stress
testing
Participated in the daily defect status and test execution status meetings and
published the reports
Environment: Java, Java, Java Script, J2ee, EJB, HP Mercury Quality Center 8.0,
HP Mercury Quick Test Professional 8.0, Mercury LoadRunner 8.0, WebSphere, FASAT,
MQ Series, Oracle 9i, Oracle Reports, XML, PL/SQL, SQL, Windows NT, UNIX (HP-UX
11.x), Unix Shell Scripting.
Skills
Testing Tools: HP Mercury Quick Test Professional 10.0, HP Mercury Quality Center
10.0, HP Mercury LoadRunner 10.0, FASAT
Data warehousing: Informatica Power Center 6.x/7.x/8.X/9.0, Power Exchange,
Repository Server Administrator Console, Star Schema, Snowflakes, FACT and
Dimensions tables
Databases: Oracle 11g/10g/9i/8i, T-SQL, MS-SQL SERVER 7.0/2000/2005, Teradata,
TOAD 8.0/7.1, Winsql, SQL*Plus, SQL Developer, Squirrel SQL
Languages: C, C++, Java, Java Script, J2ee, EJB, JUnit, NUnit, Weblogic,
WebSphere, MQ Series, ASP, ASP.NET, XML and CSS.
Reporting Tools: Cognos, Crystal, MS Access Reports, Oracle reports 6i
Environment: UNIX (Solaris, AIX, HP-UX), LINUX, Windows 9x/2000/NT/XP/VISTA
Other Tools: MS Visio, Legacy (COBOL 3.x/2.x), Control-m, AutoSys, Redwood