Post Job Free

Resume

Sign in

Quality Assurance Data

Location:
Sunnyvale, CA
Posted:
October 14, 2014

Contact this candidate

Resume:

Shilpi Gupta

Summary:

. About * years of experience in IT industry that includes about 2+ years

of as Hadoop Develper (with knowledge of Sqoop, Pig, Hive, HBase, Flume)

. 7 years of diversified experience in Information Technology including

PL/SQL, System Analysis, Development. Strong Database and Testing

concepts.

. Experience in developing, designing, testing and maintaining object

oriented application.

. Expertise in Hadoop echo systems HDFS, Map Reduce, Pig, Sqoop and Hive

for scalability, distributed computing and high performance computing.

. Strong knowledge on creating and monitoring Hadoop clusters on EC2, VM,

CDH3, CDH4 Cloudera Manager on Linux, Ubuntu OS etc.

. Experience in developing PIG Latin Scripts and using Hive Query Language

for data Analytics.

. Strong knowledge in NOSQL column oriented databases like HBase and its

integration with Hadoop cluster

. Good working experience using Sqoop to import data into HDFS from RDBMS

and vice-versa.

. Good knowledge in using job scheduling and monitoring tools like Oozie

and ZooKeeper

. Well versed in Core Java.

. Extensive experience in Software development life Cycle (SDLC)

. Having Good knowledge on Single node and Multinode Cluster

Configurations.

. Knowledge of MS SQL and MS access.

. Sound network security and management concepts.

. A very good team player with the ability to work independently

. A competent, co-operative and result driven IT consultant possessing

excellent interpersonal, written and oral communication skills

. Good analytical and debugging skills; systematic approach to programming

tasks; good interpersonal skills

Skill Set:

Specialties Hadoop, Big Data, PL / SQL Database, Oracle E Business

suite (Oracle iRecruitment) Data Integration, Data

Quality, Databases, Quality Control and Quality Assurance

Databases Oracle, MySql, HBase

Languages Java, SQL, Sqoop, Hadoop, Pig, Eclipse, Hive, YARN,

Zookeeper, oozie, HTML, XML, Javascript,

Operating System Ms-Windows & UNIX

Languages SQL, PL/SQL, Shell Scripting

FTP Tools FileZilla, Putty

Other tools SQL Developer, TOAD 7.6/8.5, MS Office (Access, Excel,

Power Point, Word)

Certifications

Working towards Cloudera certification (CCD-410)

Professional Experience:

Macy's, SFO, CA

Feb 14 - Present

Role: Hadoop Developer

Maintaining the customer member details and rewards points transaction are

very difficult in terms of storage and processing. Member loyalty

management system is replacing the existing reward management system which

is developed as a web service provider with the help of database sharing.

Aim of this system is to reduce the response time of web service. This

system is designed with HBase storage handler and later planning to remove

some BI reports generation using Hive. The solution is based on the open

source Big Data s/w Hadoop. The data will be stored in Hadoop file system

and processed using Map/Reduce jobs.

Responsibilities:

. Application installation of Hadoop, Hive, MapReduce & Sqoop

. Involved in moving all log files generated from various sources to HDFS

for further processing through Flume.

. HDFS support and maintenance and Adding/Removing a Node, Data

Rebalancing.

. Developed MapReduce application using Hadoop, MapReduce programming and

Hbase.

. Created Hive tables to store the processed results in a tabular format.

. Involved in developing the Pig scripts

. Involved in developing the Hive Reports.

. Involved in writing Mapreduce programs and tested using MRUnit.

. Analyzed large amounts of data sets to determine optimal way to aggregate

and report on it.

. Participate in requirement gathering and analysis phase of the project in

documenting the business requirements by conducting workshops/meetings

with various business users.

Platform: Hadoop, Map Reducer, Hive, Pig

Sears, Hoffman Estates, IL

Mar 13 - Dec 13

Role: Hadoop Developer / Support

The purpose of the project is to perform the analysis on the Effectiveness

and validity of controls and to store terabytes of log information

generated by the source providers as part of the analysis and extract

meaningful information out of it. The solution is based on the open source

Big Data software Hadoop. The data will be stored in Hadoop file system and

processed using Map Reduce jobs, which intern includes getting the raw

data, process the data to obtain controls and redesign/change history

information, extract various reports out of the controls history and Export

the information for further processing.

Responsibilities:

. Involved in Design and Development of technical specifications using

Hadoop technology.

. Involved in moving all log files generated from various sources to HDFS

for further processing.

. Written the Apache PIG scripts to process the HDFS data.

. Created Hive tables to store the processed results in a tabular format.

. Monitoring Hadoop scripts which take the input from HDFS and load the

data into Hive.

. Created external tables in Hive.

Sprint, Bolingbrook, IL

Role: Java/Hadoop Developer DEC 2011 to Feb 2013

Sprint, is a United States telecommunication holding company that provides

wireless services and is also a major global Internet carrier.

Responsibilities:

. Analyzed Hadoop clusters, other analytical tools used in big data like

Hive, Pig and databases like HBase.

. Used Hadoop to build scalable distributed data solution

. Extracted feeds form social media sites such as Facebook, Twitter using

Flume

. Used Sqoop extensively to ingest data from various source systems into

HDFS.

. Extracted feeds form social media sites such as Facebook, Twitter using

Flume

. Used Sqoop extensively to ingest data from various source systems into

HDFS.

. Written Hive queries for data analysis to meet the business requirements

. Created Hive tables and worked on them using Hive QL.

. Installed cluster, worked on commissioning & decommissioning of Datanode,

Namenode recovery, capacity planning, and slots configuration

. Assisted in managing and reviewing Hadoop log files

. Assisted in loading large sets of data (Structure, Semi Structured,

Unstructured)

. Implemented Hadoop cluster on Ubuntu Linux

. Installed and configured Flume, Sqoop, Pig, Hive, HBase on Hadoop

clusters

. Managed Hadoop clusters include adding and removing cluster nodes for

maintenance and capacity needs.

. Wrote test cases in JUnit for unit testing of classes

. Involved in templates and screens in HTML and JavaScript.

Dell, Plano, TX Jan 2010 to Nov2011

Project Name : - Darwin - Item Authority Audit (IAA) 1.2

Department : - Catalog Quality

Role : - QA- Product Ads Auditor

Project Description:

Item Authority Audit Tool delivers an Item Authority Quality Audit that

measures the quality of the customer experience in terms of product

uniqueness. This audit will focus on the Search Result/Browse experience

where defects in Item Authority have the most customer impact. Using this,

a random sampling of terms will be taken and each term's search results

will be manually audited for item authority defects. Thus you will enable

both an absolute quality metric and customer impact metrics in the form of

"spurious traffic," search impressions and detail page hits to non-

authoritative ASINs. The goal of this tool is efficiently provide an

accurate data page quality.

Roles and Responsibilities:

. Understanding the Functional Requirements and Specifications.

. Creation of Test Cases and execution of Test cases.

. Experience in reviewing the Test Case documents.

. Performed Functional, Integration, System Testing.

. Identifying the test cases for automation.

. Responsible for sending daily and weekly status report.

Freecharge. Bangalore

Oct 07 - Dec 09

Role: Quality Assurance (QA)

Freecharge.in is a web and mobile application which provides online

recharge service for Mobile, DTH & Data card to the customers. For making

online recharge, users need to register on freecharge site and give number

details and bank card details. In freecharge site, users can also select

different varieties of coupons and also some cross deals. Freecharge also

has android App so customers can install and use it for recharge purpose.

Responsibilities:

. Understanding the Functional Requirements and Specifications.

. Creation of Test Cases and execution of Test cases.

. Experience in reviewing the Test Case documents.

. Performed Functional, Integration, System Testing.

. Identifying the test cases for automation from manual test case list.

. Writing and executing the automation test scripts.

. Performed Defect logging and Tracking.

. Used SQL queries for database check.

. Responsible for sending daily status report.

US Airways, Tempe, AZ Mar 06 to

July 07

Role: PL SQL Programmer

Oracle E-Business Suite

Responsibilities:

. Design and Developing of workflow procedures.

. Has done many migrations of project's application interfaces.

. Was responsible for handling the Billing application and related Change

requests.

. Implemented PL and SQL queries, triggers and Stored Procedures as per the

design and development related requirements of the project.

. Has done end to end testing of the applications and Error casual

analysis.

. Has done proper documentation of the application I've developed, for

better future understanding.

. Prepare project-phase plans and implementation schedule, plan and conduct

meeting and workshops relative to the project plans, and communicate day

to day project status to all levels of management.

[pic]



Contact this candidate