Post Job Free

Resume

Sign in

Hadoop Developer

Location:
Hyderabad, Telangana, India
Posted:
March 08, 2016

Contact this candidate

Resume:

SHANKAR RAVULA

+91-810******* E-Mail: actt92@r.postjobfree.com

Professional Summary:

2+ years of overall IT experience in Application Development in Sql and Big Data Hadoop.

1.2 years of exclusive experience in Hadoop and its components like HDFS, Map Reduce, Apache Pig, Hive, Sqoop, HBase.

Extensive Experience in Setting Hadoop Cluster

Good working knowledge with Hive and Apache Pig

Involved in writing the Pig scripts to reduce the job execution time

Very well experienced in designing and developing both server side and client side applications.

Excellent communication, interpersonal, analytical skills, and strong ability to perform as part of team.

Exceptional ability to learn new concepts.

Hard working and enthusiastic.

Knowledge on MongoDB.

Professional Experience:

Currently Working as a Software Engineer in Infosys Ltd, Hyderabad, India since November’ 2013.

Qualifications:

B.Tech

In Jawaharlal Nehru Technological University Hyderabad

Technical Skills:

Languages

MapReduce, Pig, Sqoop, Pig, Hive, Hbase.

Framework

Hadoop.

Version Control / Tracking Tool

SVN.

Database

SQL (DDL, DML, DCL).

Operating Systems

Windows7, Windows XP, Windows8.

Project Details:

PROJECT #1:

Project Name : Target – Web Intelligence

Client : Target Minneapolis, Minnesota, USA.

Environment : Hadoop, Apache Pig, Hive, SQOOP, MySQL.

Duration : Nov 2014 to till Date

Role : Hadoop Developer

Description:

This Project is all about the rehousting of their (Target) current existing project into Hadoop platform. Previously Target was using mysql DB for storing their competitor’s retailer’s information.[The Crawled web data]. Early Target use to have only 4 competitor retailers namely Amazon.com, walmart.com etc….

But as and when the competitor retailers are increasing the data generated out of their web crawling is also increased massively and which cannot be accomodable in a mysql kind of data box with the same reason Target wants to move it Hadoop, where exactly we can handle massive amount of data by means of its cluster nodes and also to satisfy the scaling needs of the Target business operation.

Roles and Responsibilities:

Moved all crawl data flat files generated from various retailers to HDFS for further processing.

Written the Apache PIG scripts to process the HDFS data.

Created Hive tables to store the processed results in a tabular format.

Developed the sqoop scripts in order to make the interaction between Pig and MySQL Database.

Involved in gathering the requirements, designing, development and testing

Writing the script files for processing data and loading to HDFS

Writing CLI commands using HDFS.

Developed the UNIX shell scripts for creating the reports from Hive data.

Completely involved in the requirement analysis phase.

Analyzing the requirement to setup a cluster

Created two different users (hduser for performing hdfs operations and map red user for performing map reduce operations only)

Ensured NFS is configured for Name Node

Setting Password less hadoop

Setting up cron job to delete hadoop logs/local old job files/cluster temp files

Setup Hive with MySQL as a Remote Metastore

Moved all log/text files generated by various products into HDFS location

Written Map Reduce code that will take input as log files and parse the logs and structure them in tabular format to facilitate effective querying on the log data

Created External Hive Table on top of parsed data.

PROJECT #2:

Title : Project Management Information System (PMIS)

Client : DMDE, Secunderabad

Duration : Dec 2013 to Nov 2014

Team size : 3

Environment : Oracle 9i, Windows, Reports.

Role : SQL Developer

Project Description:

Project Management Information system is a Database System which is used for maintaining the entire Database regarding Naval Projects developed by DMDE. This Project Processing Lifecycle of Products from gathering of Technical Requirements to Final Production which includes Master data, financial data, Purchase Orders, Quality Assurance report etc. The centralized data is maintained in the oracle database. The end users will be able to take the Business Decisions based on our different Forecast reports. Finally the data is maintained in different staging servers. The data warehouse systems will be maintained internally for Business analysis and ongoing customization.

Responsibilities:

Requirement gathering and worked according to the CR.

Worked under the senior level guidance.

Data validation/Reconciliation report generation.

Code Development as per the client requirements.

Involved in the development backend code, altered tables to add new columns, Constraints, Sequences and Indexes as per business requirements.

Perform DML, DDL Operations as per the Business requirement.

Creating views and prepares the Business Reports.

Resolved production issues by modifying backend code as and when required.

Used different joins, sub queries and nested query in SQL query.

Involved in creation of sequences for automatic generation of Product ID.

Created Database Objects like tables, Views, sequences, Synonyms, Stored Procedures, functions, Packages, Cursors, Ref Cursor and Triggers.

Testing of code functionality using testing environment.



Contact this candidate