Shilpi Gupta
Summary:
. About * years of experience in IT industry that includes about 2+ years
of as Hadoop Develper (with knowledge of Sqoop, Pig, Hive, HBase, Flume)
. 7 years of diversified experience in Information Technology including
PL/SQL, System Analysis, Development. Strong Database and Testing
concepts.
. Experience in developing, designing, testing and maintaining object
oriented application.
. Expertise in Hadoop echo systems HDFS, Map Reduce, Pig, Sqoop and Hive
for scalability, distributed computing and high performance computing.
. Strong knowledge on creating and monitoring Hadoop clusters on EC2, VM,
CDH3, CDH4 Cloudera Manager on Linux, Ubuntu OS etc.
. Experience in developing PIG Latin Scripts and using Hive Query Language
for data Analytics.
. Strong knowledge in NOSQL column oriented databases like HBase and its
integration with Hadoop cluster
. Good working experience using Sqoop to import data into HDFS from RDBMS
and vice-versa.
. Good knowledge in using job scheduling and monitoring tools like Oozie
and ZooKeeper
. Well versed in Core Java.
. Extensive experience in Software development life Cycle (SDLC)
. Having Good knowledge on Single node and Multinode Cluster
Configurations.
. Knowledge of MS SQL and MS access.
. Sound network security and management concepts.
. A very good team player with the ability to work independently
. A competent, co-operative and result driven IT consultant possessing
excellent interpersonal, written and oral communication skills
. Good analytical and debugging skills; systematic approach to programming
tasks; good interpersonal skills
Skill Set:
Specialties Hadoop, Big Data, PL / SQL Database, Oracle E Business
suite (Oracle iRecruitment) Data Integration, Data
Quality, Databases, Quality Control and Quality Assurance
Databases Oracle, MySql, HBase
Languages Java, SQL, Sqoop, Hadoop, Pig, Eclipse, Hive, YARN,
Zookeeper, oozie, HTML, XML, Javascript,
Operating System Ms-Windows & UNIX
Languages SQL, PL/SQL, Shell Scripting
FTP Tools FileZilla, Putty
Other tools SQL Developer, TOAD 7.6/8.5, MS Office (Access, Excel,
Power Point, Word)
Certifications
Working towards Cloudera certification (CCD-410)
Professional Experience:
Macy's, SFO, CA
Feb 14 - Present
Role: Hadoop Developer
Maintaining the customer member details and rewards points transaction are
very difficult in terms of storage and processing. Member loyalty
management system is replacing the existing reward management system which
is developed as a web service provider with the help of database sharing.
Aim of this system is to reduce the response time of web service. This
system is designed with HBase storage handler and later planning to remove
some BI reports generation using Hive. The solution is based on the open
source Big Data s/w Hadoop. The data will be stored in Hadoop file system
and processed using Map/Reduce jobs.
Responsibilities:
. Application installation of Hadoop, Hive, MapReduce & Sqoop
. Involved in moving all log files generated from various sources to HDFS
for further processing through Flume.
. HDFS support and maintenance and Adding/Removing a Node, Data
Rebalancing.
. Developed MapReduce application using Hadoop, MapReduce programming and
Hbase.
. Created Hive tables to store the processed results in a tabular format.
. Involved in developing the Pig scripts
. Involved in developing the Hive Reports.
. Involved in writing Mapreduce programs and tested using MRUnit.
. Analyzed large amounts of data sets to determine optimal way to aggregate
and report on it.
. Participate in requirement gathering and analysis phase of the project in
documenting the business requirements by conducting workshops/meetings
with various business users.
Platform: Hadoop, Map Reducer, Hive, Pig
Sears, Hoffman Estates, IL
Mar 13 - Dec 13
Role: Hadoop Developer / Support
The purpose of the project is to perform the analysis on the Effectiveness
and validity of controls and to store terabytes of log information
generated by the source providers as part of the analysis and extract
meaningful information out of it. The solution is based on the open source
Big Data software Hadoop. The data will be stored in Hadoop file system and
processed using Map Reduce jobs, which intern includes getting the raw
data, process the data to obtain controls and redesign/change history
information, extract various reports out of the controls history and Export
the information for further processing.
Responsibilities:
. Involved in Design and Development of technical specifications using
Hadoop technology.
. Involved in moving all log files generated from various sources to HDFS
for further processing.
. Written the Apache PIG scripts to process the HDFS data.
. Created Hive tables to store the processed results in a tabular format.
. Monitoring Hadoop scripts which take the input from HDFS and load the
data into Hive.
. Created external tables in Hive.
Sprint, Bolingbrook, IL
Role: Java/Hadoop Developer DEC 2011 to Feb 2013
Sprint, is a United States telecommunication holding company that provides
wireless services and is also a major global Internet carrier.
Responsibilities:
. Analyzed Hadoop clusters, other analytical tools used in big data like
Hive, Pig and databases like HBase.
. Used Hadoop to build scalable distributed data solution
. Extracted feeds form social media sites such as Facebook, Twitter using
Flume
. Used Sqoop extensively to ingest data from various source systems into
HDFS.
. Extracted feeds form social media sites such as Facebook, Twitter using
Flume
. Used Sqoop extensively to ingest data from various source systems into
HDFS.
. Written Hive queries for data analysis to meet the business requirements
. Created Hive tables and worked on them using Hive QL.
. Installed cluster, worked on commissioning & decommissioning of Datanode,
Namenode recovery, capacity planning, and slots configuration
. Assisted in managing and reviewing Hadoop log files
. Assisted in loading large sets of data (Structure, Semi Structured,
Unstructured)
. Implemented Hadoop cluster on Ubuntu Linux
. Installed and configured Flume, Sqoop, Pig, Hive, HBase on Hadoop
clusters
. Managed Hadoop clusters include adding and removing cluster nodes for
maintenance and capacity needs.
. Wrote test cases in JUnit for unit testing of classes
. Involved in templates and screens in HTML and JavaScript.
Dell, Plano, TX Jan 2010 to Nov2011
Project Name : - Darwin - Item Authority Audit (IAA) 1.2
Department : - Catalog Quality
Role : - QA- Product Ads Auditor
Project Description:
Item Authority Audit Tool delivers an Item Authority Quality Audit that
measures the quality of the customer experience in terms of product
uniqueness. This audit will focus on the Search Result/Browse experience
where defects in Item Authority have the most customer impact. Using this,
a random sampling of terms will be taken and each term's search results
will be manually audited for item authority defects. Thus you will enable
both an absolute quality metric and customer impact metrics in the form of
"spurious traffic," search impressions and detail page hits to non-
authoritative ASINs. The goal of this tool is efficiently provide an
accurate data page quality.
Roles and Responsibilities:
. Understanding the Functional Requirements and Specifications.
. Creation of Test Cases and execution of Test cases.
. Experience in reviewing the Test Case documents.
. Performed Functional, Integration, System Testing.
. Identifying the test cases for automation.
. Responsible for sending daily and weekly status report.
Freecharge. Bangalore
Oct 07 - Dec 09
Role: Quality Assurance (QA)
Freecharge.in is a web and mobile application which provides online
recharge service for Mobile, DTH & Data card to the customers. For making
online recharge, users need to register on freecharge site and give number
details and bank card details. In freecharge site, users can also select
different varieties of coupons and also some cross deals. Freecharge also
has android App so customers can install and use it for recharge purpose.
Responsibilities:
. Understanding the Functional Requirements and Specifications.
. Creation of Test Cases and execution of Test cases.
. Experience in reviewing the Test Case documents.
. Performed Functional, Integration, System Testing.
. Identifying the test cases for automation from manual test case list.
. Writing and executing the automation test scripts.
. Performed Defect logging and Tracking.
. Used SQL queries for database check.
. Responsible for sending daily status report.
US Airways, Tempe, AZ Mar 06 to
July 07
Role: PL SQL Programmer
Oracle E-Business Suite
Responsibilities:
. Design and Developing of workflow procedures.
. Has done many migrations of project's application interfaces.
. Was responsible for handling the Billing application and related Change
requests.
. Implemented PL and SQL queries, triggers and Stored Procedures as per the
design and development related requirements of the project.
. Has done end to end testing of the applications and Error casual
analysis.
. Has done proper documentation of the application I've developed, for
better future understanding.
. Prepare project-phase plans and implementation schedule, plan and conduct
meeting and workshops relative to the project plans, and communicate day
to day project status to all levels of management.
[pic]