ANNA SUNDARAM
abnlp4@r.postjobfree.com
Location: Omaha, NE
Masters in Computer Engineering, with over eight years of experience involving quality
assurance of Client/Server and web based applications. Possess exemplary professional
skills in testing methodologies and software development and was involved in all
phases of the Software Development Life Cycle. Performed different dimensions of
testing, Functionality, Usability, Reliability, Performance, and Regression testing. I am
experienced in both Manual and Automated testing techniques. Proficient in the use of
Mercury testing tools such as Quality Center, QTP, and Loadrunner. Extensively used
load runner to create Vugen scripts to test web, database and client server applications.
Created Vugen scripts and updated the scripts using parameterization and correlation.
Executed test scenarios and performed application tuning to get the desired results. I
am a dynamic and assertive team player with strong commitment.
EDUCATION:
M.S (Computer Engineering), Wright State University, Dayton, Ohio.
Certifications:
HP Certified - HP Performance center v9 - Accredited Integration Specialist (AIS)
HP Virtual User Generator Software
•
HP LoadRunner Software
•
HP Certified - HP Quality Center 9.2 Software
Brainbench Certified
1. Software Quality Assurance Certificate
2. Software Testing Certificate
3. Business Communication Certificate
4. Time Management (U.S.) Certificate
SKILLSET:
GUI /Tools Visual Basic 6.0
Technologies ODBC, SQL Server, Weblogic Server, WebSphere, Apache, J2EE
Automation WinRunner, Loadrunner, Performance Center, QTP, Silk
Tools Performer, Sitescope, HP Diagnostics
Defect Tracking Rational Clear Quest, Test Director, Quality Center
Tools
Protocols HTTP, Web Services, Siebel-Web, Citrix Multi Protocol, COM
DCOM, JAVA, Winsock
Packages Microsoft Office 2000, MS Visio, MS Project.
TD Ameritrade, NE July 2007 – Present
Performance Engineer Lead
Project involved in testing a web based application, which enables customers to access
accounts and perform online trading.
Application accessed by customers to buy/sell stocks based on the market/limit. The
online application gives the users the ability to perform transactions, save orders, place
order depending on the user need.
Responsibilities:
• Participated in all the phases of SDLC, starting from requirement, design,
development, Testing and implementation phase.
• Analyzed and understood Business requirements and procedures and
formed detail test plan and test cases. Conducted Walkthroughs and
review meetings with Business Analysts & Development team. Participated
in the design, code and test plan walk through with the team members.
• Responsible for creating the performance test plan and reviewing it with
the Business analyst and the development team and getting their approval
on the same.
• Provide management with metrics, reports, and schedules as necessary and
participated in the design walkthroughs and meeting.
• Deploying and managing the appropriate testing framework to meet the
testing mandate
• Spoke with the Application development and the business to gather
information on the current and the expected volume of users, orders.
• Executed performance test on the existing hardware to confirm the
scalability of the application.
• Planning, deploying and managing the testing effort for the given
engagement/release.
• Effectively leading the testing team, Work allocation for onsite/offshore
testing team as per the new projects
• In-depth knowledge in the analysis of application's higher-risk aspects, and
determining scope and limitations of tests
• Planned the load by analyzing Task distribution diagram, Transaction
Profile and User profile and executed Performance Testing using
Load runner
• Involved in studying the server statistics and determine the capacity on the
system
• Using Load runner, created scenarios for Load and Performance testing
with different host systems and also configured the test environment.
Generated graphs and studied them to monitor the software performance.
• Verifying the Vugen Scripts and Scenarios created by the team members
before test execution
• Used secure CRT to setup performance monitors and reported the CPU and
memory peaks to the AD team.
• Created load runner Vugen Scripts and used Correlation to parameterize
dynamically changing parameters like Session IDs. Used Rendezvous
points, Load balancing and IP spoofing to load test specific transactions.
• Responsible for setting up sitescope monitors to drill down different layers
of the application to identify the performance bottleneck
• Conducted discussion with Development team to decide on the hardware
required, based on the performance test, to support the future user growth.
• Analyzed various graphs generated by Loadrunner Analyzer including
Database Monitors, Network Monitor graphs, User Graphs, Error Graphs,
Transaction graphs and Web Server Resource Graphs.
• Used Sitescope to gather server information Memory and CPU usage.
• Wrote and tracked the defects using Quality Center and communicated
with the developers.
• Executing the scenarios, analyzing the Graphs and coordinating with the
DBAs and Network Admin to ensure optimum performance.
• Presenting test results and analysis to the application and development
team.
• Provide weekly updates to the client and application team based on the test
results and analysis.
• Involved in Database tuning to enhance the application performance.
Environment: Loadrunner 9.1, 9.5, Perfomance center 9.5, Sitescope, Quality Center,
Unix, Windows, JAVA, Jboss, Weblogic, Oracle10g, XML, SQL Server, MS Access,
MS Vision, MS Project, VB, J2ee analysis, HTML.
JP Morgan Chase, IL August 2007 – July 2009
QA Performance Test Lead
Project was testing the Strategic Know Your Customer system and Pega Rules process
commander. The main feature of the application is that the generate package with
required templates and products. The application was also used to create KYCs. The
KYCs were created based on the customer entitlements and requirements.
The Process commander system was mainly used to create cases depending upon the
customer needs and accounts. The system creates messages depending upon the cases
that are being created. Customer can create correspondence and adjustments as required
and can also set SLAs for the cases. External customer can transfer cases to different
groups and can also amend, cancel and recall their cases.
Responsibilities:
• Responsible for defining the scope of the project gathering business
requirements and documents them textually or within models.
• Interviewing business area experts, asking detailed questions and carefully
recording the requirements in a format that can be reviewed and understood by
both business people and technical people.
• Co-ordinate testing activities with different team and ensure that the project
schedule is met without any issues.
• Defined input requirements to document the Business Requirements
Specifications and developed the Requirements Traceability Matrix.
• Worked on the improvement of QA Process by reviewing and evaluating
existing practices with standard testing guidelines.
• Responsible for co-coordinating with the offshore team to complete the project
within the deadlines.
• Provide management with metrics, reports, and schedules as necessary and
participated in the design walkthroughs and meeting.
• Involved in managing an offshore and onsite team, responsible for
performance and manual testing.
• Planned the load by analyzing Task distribution diagram, Transaction
Profile and User profile and executed Performance Testing using
Load runner
• Involved in studying the server statistics and determine the capacity on the
system
• Using Load runner, created scenarios for Load and Performance testing
with different host systems and also configured the test environment.
Generated graphs and studied them to monitor the software performance.
• Verifying the Vugen Scripts and Scenarios created by the team members
before test execution
• Have created load runner test scripts in C language and have updated the
test scripts by using if loops and randomization functions.
• Worked on creating scripts for different protocols like web, web services
(XML, SOAP), JAVA, ODBC, Citrix and Siebel.
Worked with the development team on MQ series projects to release
•
messages during performance test execution.
• Worked with the development team to sort the issues due to proxy and
firewall settings to achieve the exact performance of the application.
• Responsible for setting up monitors in controller, sitescope and interscope
to monitor the CPU and memory utilization.
• Using HP Diagnostics to identify the root-cause of the transaction that
caused the slowness in the system.
• Conducted Regression and functional testing for different applications.
• Was involved in Capacity Planning for Enterprise release applications, to
calculate the future growth in the transaction volume and concurrent users
on the system
• Conducted discussion with Development team to decide on the hardware
required, based on the performance test, to support the future user growth.
• Written commands in SQL to verify backend data and have also monitored
CPU and memory usage on SQL and Oracle servers.
• Analyzed various graphs generated by Loadrunner Analyzer including
Database Monitors, Network Monitor graphs, User Graphs, Error Graphs,
Transaction graphs and Web Server Resource Graphs.
• Used Sitescope to gather server information Memory and CPU usage.
• Wrote and tracked the defects using Quality Center and communicated
with the developers.
• Tested the pre-prod servers to make sure, production has the similar performance
compared to the performance test environment.
• Conducted weekly meetings with Project Head, Business and development
teams.
• Executing the scenarios, analyzing the Graphs and coordinating with the DBA’s
and Network Admin to ensure optimum performance.
• Provide weekly updates to the client and application team based on the test
results and analysis.
Environment: Windows NT, IIS, Citrix, ODBC, JAVA, C, C++, J2EE, Weblogic,
Apache, Oracle10g, 11i,.NET, XML, Quality Center, Load runner 8.1, 9.4, HP
Diagnostics, Sitescope, QTP, UNIX, Weblogic, Websphere, Apache, SQL Server, MS
Office, MS Access, MS Vision, MS Project, VB, J2ee analysis, HTML.
Washington Mutual, IL January 2006 – September 2007
QA Performance Test Lead/Tester
The project involved testing client/server and web-based application developed for the
Home Loans Department.
Project was a Web-based application, which can be accessed by the brokers and loan
officers and users from the different locations. Users can request the loan online and it
consisted of four modules funding, pay down, transfer and shipping. Loans can be
transferred and paid down online using this application. Worked on risk and collateral
related applications.
Responsibilities:
Studied the URS document and created the Functional Requirement Specification
•
document.
• Worked according to the activities laid down in each phase of Software
development life cycle and Coordinated the environment setup for Testing.
Meet with client groups to determine performance requirements and goals and
•
determine test strategies based on requirements and architecture.
• Created the strategy document that defines the test environment, phases of
testing, time lines of different phases of testing, entrance and exit criteria into
different phases of testing and resources required to conduct the effort.
• Using LoadRunner, created vugen scripts and scenarios for Load and
Performance testing with different host systems and also configured the test
environment. Generated graphs and studied them to monitor the software
performance.
• Created Vugen Scripts in C and have used certain C++ functions to update the
baseline script.
• Have written TSL scripts for test automation for Ajax web based applications.
• Used Correlation to parameterize dynamically changing parameters like Session
ID’s.
• Monitored resources to identify performance bottlenecks, analyze test results
and report the findings to the clients, and provide recommendation for
performance improvements as needed.
• Identified functionality and performance issues, including: deadlock conditions,
database connectivity problems, and system crashes under load.
• Created SQL queries to verify the Database response and to verify if the
response time was within the SLA.
• Used TOAD to create DB commands and validate SQL data base.
• Provided management and vendor with analysis reports and recommendations,
which resulted tuning of the application. Vertical scaling and garbage collection
were performed. Communicated with the vendor to resolve issues.
• Confirmed the scalability of the new servers and application under test after the
architecture redesign.
• Tested the pre-prod servers to make sure, production has the similar performance
compared to the performance test environment.
• Conducted weekly meetings with Project Head, Business and development
teams.
• Executing the scenarios, analyzing the Graphs and coordinating with the DBA’s
and Network Admin to ensure optimum performance.
Environment: Windows NT, Citrix, Oracle 2-tier, Loadrunner 8.0, 8.1, Quality Center,
QTP, XML, SOAP, C, C++, COM/DCOM, SQL, ODBC, Mainframe, Weblogic, UNIX,
Websphere, JBOSS, Apache, SQL Server, Ajax, MS Office, MS Access, MS Vision,
MS Project, VB, J2ee analysis, HTML.
MetLife, NJ January 2004 – December 2005
QA Performance Test Lead/Tester
The project involved testing the client/server and web-based application developed for
Claim Management System, online mortgage and Insurance application.
Project was a Web-based application, which can be accessed by the brokers and loan
officers and users from the different locations. This online application gives the users
the ability to use various tools like mortgage qualification, mortgage payment, refinance,
debt consolidation and monthly payment calculators. It also allows the customer to
View the Account status, Asset allocation chart, current market indices, trade equities
online.
Client-Server application, Citrix, which was a Insurance Claim Management System,
Where users can log in and check the status of the logged claims, add in new claims and
track the claim information based on the Claim ID and manages the claims.
Responsibilities:
• Studied the URS document and created the Functional Requirement Specification
document.
• Worked according to the activities laid down in each phase of Software
development life cycle and Coordinated the environment setup for Testing.
• Meet with client groups to determine performance requirements and goals and
determine test strategies based on requirements and architecture.
• Identified and classified Manual and Automated test cases by isolating the
repetitive actions.
• Developed detailed Manual Test cases and Scenarios. Studied Requirements
and designed manual test cases accordingly.
• Identifying the functional test cases for Regression Testing and automated these
Test Scripts using QTP.
• Installed the Citrix client to talk with the Citrix server and record the traffic
going back and forth.
• Created Database Vuser scripts to simulate client activities and performed Load,
Stress and Performance test using LoadRunner/Performance Center
• Generated and Created VuGen/Vuser scripts using Vuser Generator and
Created Scenarios in LoadRunner Controller
• Analyzed LoadRunner/Performance Center test result Involved in Preparing
Test Plan and Test Cases based on the analysis of the business requirements.
• Used QTP to develop scripts to perform Functionality and GUI testing.
• Tested Peoplesoft payables for response time requirements, which streamlines
accounts payable operations and enhances supplier relationships.
• Inserted rendezvous points in order to simulate heavy loads for conducting
Load Testing.
• Used ramp-up and ramp-down to simulate real-time scenarios.
• Analyzed the results using LoadRunner’s Online Monitors and Graphs to
identify bottlenecks in the server resources.
• Identified functionality and performance issues, including: deadlock conditions,
database connectivity problems, and system crashes under load.
• Provided management and vendor with analysis reports and recommendations,
which resulted tuning of the application. Vertical scaling and garbage collection
were performed. Communicated with the vendor to resolve issues.
• Confirmed the scalability of the new servers and application under test after the
architecture redesign.
• Conducted weekly meetings with Project Head, Business and development
teams.
• Executing the scenarios, analyzing the Graphs and coordinating with the DBA’s
and Network Admin to ensure optimum performance.
Environment: Windows NT, Citrix, QA Load, WinRunner, LoadRunner 8.1, Quality
Center, Performance Center, Peoplesoft Payables, J2ee analysis,Oracle DB, QTP, Site
Scope, MS Office, MS Access, MS Vision, MS Project
Charter One, OH June 2001 – December 2003
Sr. Quality Assurance Analyst
Project was testing the online insurance handling, finance and Banking module. It helps
people find the right insurance for their individual needs. Online Insurance requires the
user to login with a valid user id and password to authenticate the person with a secure web
log-in. After authentication, they can learn about insurance, get a quote, retrieve a quote or
buy an online insurance policy. Customers can access their account information online, get
details of previous transactions, view statements for the last 2 years and they can also file a
claim or pay their monthly payments online.
Responsibilities:
• Participated in all the phases of SDLC, starting from requirement, design,
development, Testing and implementation phase.
• Supervised the testing process from post development through build, system
and integration testing.
Created Test Plans and test cases using Test Director. Organized test cases into
•
test sets for the purpose of execution.
• Worked according to the activities laid down in each phase of Software
development life cycle and Coordinated the environment setup for Testing.
• Gathered and created test data like pre-test, input, regression, actual and
expected data.
• Written Test Cases for Functional, Integration, Regression testing for different
positive, negative and boundary scenarios.
• Extensively created Data Driven Test scripts to read data from text files for
testing. Checked for data boundaries/limits, incorrect input of data.
• Performance testing was done using LoadRunner by planning the load by
analyzing Task distribution diagram, Transaction Profile and User profile,
creating virtual users and analyzing different reports.
• Involved in creating Load test scenarios using LoadRunner to bombard the
server with virtual user requests and tests the performance under stress
conditions.
• Extensive Parameterization and correlation of the VuGen scripts to ensure the
real time load conditions. Used Rendezvous points to load test specific
transactions.
• Used LoadRunner sophisticated analysis module to drill down to determine the
specific source of bottlenecks within the application architecture and generate
actionable reports.
• Monitored Performance Measurements such as end-to-end response time,
network and server response time, server response time, middleware-to-server
response time.
• Executing the scenarios, analyzing the Graphs and coordinating with the DBA’s
and Network Admin to ensure optimum performance.
• Followed up on the bugs logged with developer and retest the test cases.
Environment: Windows NT, XML, SOAP, Test Director, WinRunner, QTP,
LoadRunner, FlexPLM, MS Office, MS Vision, MS Project..
DP Solutions India Pvt. Ltd., Chennai March 1999 - January 2000
System Analyst
Project included various modules related to customer’s history maintenance, plan
configurations, automatic and manual invoices generation and customer service.
Responsibilities:
Responsible for defining the scope of the project gathering business
•
requirements and documents them textually or within models.
Interviewing business area experts, asking detailed questions and carefully
•
recording the requirements in a format that can be reviewed and understood by
both business people and technical people.
Defined input requirements to document the Business Requirements
•
Specifications and developed the Requirements Traceability Matrix.
Conducted walkthroughs with stakeholders to design complete and detailed test
•
plans and test cases.
Analyzed the system requirements and developed detailed test plan, test cases
•
and test scripts for functionality and GUI testing
Implemented and monitored Individual Development Plans focusing on total
•
performance, including both quality and productivity.
Interfaced with the developers to provide all kind of support and resources
•
needed to resolve technical issues.
Environment: Windows NT 4.0, Test Director, WinRunner, MS Excel, Visual Basic,
MS Access, Oracle.
EDUCATION:
M.S (Computer Engineering), Wright State University, Dayton, Ohio.