PHANI
Hadoop Developer
Email ID: *********@******.***
Ph: 732-***-****
PROFESSIONAL SUMMARY
• Over 7 years of experience with emphasis on Big Data Technologies, Development and Design of Java based
enterprise applications.
• Extensive experience in development of Big Data projects using Hadoop, Mapreduce, Pig,Hive and other open
source tools/technologies.
• Experience in installation, configuration, supporting and managing Hadoop clusters.
• Implemented standards and processes for Hadoop based application design and implementation.
• Responsible for writing MapReduce programs using Java.
• Logical Implementation and interaction with HBase.
• Developed MapReduce jobs to automate transfer of data from HBase.
• Performed data analysis using Hive and Pig.
• Loadedstreaming log data from various webservers into HDFS using Flume.
• Successfully loaded files to Hive and HDFS from Oracle and SQL Server using SQOOP.
• Assist with the addition of Hadoop processing to the IT infrastructure.
• Worked in Multiple Environment in installation and configuration.
• Document and explain implemented processes and configurations in upgrades.
• Support development, testing, and operations teams during new system deployments.
• Evaluate and propose new tools and technologies to meet the needs of the organization.
• Experience in using Sqoop, Oozie and Cloudera Manager.
• Good Knowledge on Hadoop Cluster architecture and monitoring the cluster.
• Implemented stand-alone installation, file system management, backups, process control, user administration and
device management in a networked environment.
• An excellent team player and self-starter with good communication skills and proven abilities to finish tasks before
target deadlines.
TECHNICAL SKILLS
Programming Languages : Java 1.4, C++, C, SQL, PIG, PL/SQL.
Java Technologies : JDBC.
Frame Works : Jakarta Struts 1.1, JUnit and JTest, LDAP.
Databases : Oracle8i/9i, NO SQL (HBase),MY SQL,MS SQL server.
IDE’s & Utilities : Eclipse and JCreator, NetBeans.
Web Dev. Technologies : HTML, XML.
Protocols : TCP/IP, HTTP and HTTPS.
Operating Systems : Linux, MacOS, WINDOWS 98/00/NT/XP.
Hadoop ecosystem : Hadoop and MapReduce, Sqoop, Hive, PIG, HBASE, HDFS, Zookeeper, Lucene,
Sun Grid Engine Administration
EDUCATION
• Bachelors in Computer Science.
PROFESSIONAL EXPERIENCE
Wells Fargo Home Mortgage, Minneapolis, MN Mar 2013 – Till
Date
Hadoop Developer
Wells Fargo & Company is an American multinational diversified financial services company. The CORE project deals
with improving end-to-end approach to real estate-secured lending, the overall customer experience and achieving the
vision of satisfying all the customers’ financial needs.
The purpose of the project is to build an enterprise big data platform that would be used to load, manage and process
terabytes of transactional data, machine log data, performance metrics, and other adhoc data setsand extract meaningful
information out of it. The solution is based on the ClouderaHadoop.
Responsibilities
• Worked on analyzing, writing Hadoop Mapreduce jobs using Java API, Pig and Hive.
• Responsible for building scalable distributed data solutions using Hadoop.
• Involved in loading data from edge node to HDFS using shell scripting.
• Worked on installing cluster, commissioning & decommissioning of datanode, namenode high availability, capacity
planning, and slots configuration.
• Created HBase tables to store variable data formats of PII data coming from different portfolios.
• Implemented a script to transmit sysprin information from Oracle to Hbase using Sqoop.
• Implemented best income logic using Pig scripts and UDFs.
• Implemented test scripts to support test driven development and continuous integration.
• Worked on tuning the performance Pig queries.
• Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required.
• Responsible to manage data coming from different sources.
• Load and transform large sets of structured, semi structured and unstructured data
• Experience in managing and reviewing Hadoop log files.
• Job management using Fair scheduler.
• Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI
team.
• Responsible for cluster maintenance, adding and removing cluster nodes, cluster monitoring and troubleshooting,
manage and review data backups, manage and review Hadoop log files.
• Installed Oozie workflow engine to run multiple Hive and pig jobs.
• Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.
• Supported in setting up QA environment and updating configurations for implementing scripts with Pig and Sqoop.
Environment
Hadoop, HDFS, Pig, Sqoop, HBase, Shell Scripting, Ubuntu, Linux Red Hat.
Dow Chemical Company, Midland, MI Nov 2011 – Feb
2013
Hadoop Developer
Philadelphia is the home for Dow’s Advanced Materials Division, which is the umbrella for some of the company’s most
important high-growth specialty businesses. Five businesses make up the division – Coatings, Building and Construction,
Paper and Textiles, Specialty Packaging and Films and Separations Technologies, and Electronic Materials – that serves
the fastest-growing end markets such as water, electronics, food, pharmaceuticals, health care, paints and more. .
Responsibilities
• Involved in review of functional and non-functional requirements.
• Facilitated knowledge transfer sessions.
• Installed and configured Hadoop Mapreduce, HDFS, Developed multiple MapReduce jobs in java for data cleaning
and preprocessing.
• Importing and exporting data into HDFS and Hive using Sqoop.
• Experienced in defining job flows.
• Experienced in managing and reviewing Hadoop log files.
• Experienced in running Hadoop streaming jobs to process terabytes of xml format data.
• Load and transform large sets of structured, semi structured and unstructured data.
• Responsible to manage data coming from different sources.
• Got good experience with NOSQL database.
• Supported Map Reduce Programs those are running on the cluster.
• Involved in loading data from UNIX file system to HDFS.
• Installed and configured Hive and also written Hive UDFs.
• Involved in creating Hive tables, loading with data and writing hive queries which will run internally in map reduce
way.
• Gained very good business knowledge on health insurance, claim processing, fraud suspect identification, appeals
process etc.
• Developed a custom File System plug in for Hadoop so it can access files on Data Platform.
• This plugin allows Hadoop MapReduce programs, HBase, Pig and Hive to work unmodified and access files
directly.
• Designed and implemented Mapreduce-based large-scale parallel relation-learning system
• Extracted feeds form social media sites such as Facebook, Twitter using Python scripts.
• Setup and benchmarked Hadoop/HBase clusters for internal use
• Setup Hadoop cluster on Amazon EC2 using whirr for POC.
• Wrote recommendation engine using mahout.
Environment
Java 6, Eclipse, Oracle 10g, Sub Version, Hadoop, Hive, HBase, Linux,, MapReduce, HDFS, Hive, Java (JDK
1.6), Hadoop Distribution of HortonWorks, Cloudera, MapReduce, DataStax, IBM DataStage 8.1, Oracle 11g / 10g,
PL/SQL, SQL*PLUS, Toad 9.6, Windows NT, UNIX Shell Scripting.
Kraft Foods, Chicago, IL Aug 2010 – Oct
2011
Hadoop Developer
Kraft Foods, Inc. has approximately 140,000 diverse employees around the world and is mainly involved in creating food
products all around United States.
Responsibilities
• Responsible for architecting Hadoop clusters.
• Assist with the addition of Hadoop processing to the IT infrastructure.
• Perform data analysis using Hive and Pig.
• Load log data into HDFS using Flume, Kafka.
• Monitoring Hadoop cluster using tools like Nagios,Ganglia and Cloudera Maneger.
• Automation script to monitor HDFS and HBase through cronjobs.
• Plan, design, and implement processing massive amounts of marketing information, complete with information
enrichment, text analytics, and natural language processing.
• Prepare multi-cluster test harness to exercise the system for performance and failover.
• Develop high-performance cache, making the site stable and improving its performance.
• Create a complete processing engine, based on Cloudera's distribution, enhanced to performance.
• Administrative support for parallel computation research on a 24-node Fedora/ Linux cluster.
• Build and support standard-based infrastructure capable of supporting tens of thousands of computers in multiple
locations.
• Negotiated and managed projects related to designing and deploying tis architecture.
Environment
Hive, Pig, HBase, Zookeeper, Sqoop, Java, JDBC, JNDI, Struts, Maven, Trac, Subversion, JUnit, SQL language, spring,
Hibernate, Junit, Oracle, XML, AltovaXmlSpy, Putty and Eclipse.
John Deere, Moline, IL Oct 2009 – Jul
2010
J2EE Developer
Worked on a web-based application which is basically designed to provide ease and comfort to the customers so that they
can view current status of their various orders, its shipment details, invoice details and current outstanding status online. If
customer is a corporate customer, they can also view order status and other details for their various subordinate customers.
This application also provides online information of various products.
Responsibilities
• Involved in Presentation Tier Development using JSF Framework and ICE Faces tag Libraries.
• Involved in business requirement gathering and technical specifications.
• Implemented J2EE standards, MVC2 architecture using JSF Framework.
• Implementing Servlets, JSP and Ajax to design the user interface.
• Extensive experience in building GUI (Graphical User Interface) using JSF and ICE Faces.
• Developed Rich Enterprise Applications using ICE Faces and Portlets technologies.
• Experience using ICE Faces Tag Libraries to develop user interface components.
• Used JSF, JSP, Java Script, HTML, and CSS for manipulating, validating, customizing, error messages to the User
Interface.
• Used EJBs(Session beans) to implement the business logic, JMS for communication for sending updates to various
other applications and MDB for routing priority requests.
• All the Business logic in all the modules is written in core Java.
• Wrote WebServices using SOAP for sending and getting data from the external interface.
• Developed a web-based reporting for monitoring system with HTML and Tiles using Struts framework.
• Middleware Services layer is implemented using EJB(Enterprise Java Bean - stateless) in WebSphere environment.
• Used Design patterns such as Business delegate, Service locator, Model View Controller, Session façade, DAO .
• Funds Transfers are sent to another application using JMS technology asynchronously.
• Involved in implementing the JMS (Java messaging service) for asynchronous communication.
• Involved in writing JMS Publishers to post messages.
• Involved in writing MDB(Message Driven Beans) as subscribers.
• Created Stored procedures using PL-SQL for data modification (Using DML insert, update, delete) in Oracle
• Interaction with Oracle database is implemented using Hibernate.
Environment
J2EE, EJB, JSF, ICE Faces, EJB, WebServices, XML, XSD, Agile, Microsoft Visio, Clear Case, Oracle 9.i/10.g,
Weblogic8.1/10.3,RAD, LOG4j,Servlets, JSP, Unix.
ValTech Solutions, Addison, TX Oct 2007 – Sep
2009
J2EE Developer
Project: OTS
OTS is a Transaction System where the drug tests are automated online, this allows big employers like Wal-Mart, Walt
Disney to save their time and speed up the employee hiring process, the basic operations are CRUD (create, read, update,
delete) on a COC form (Chain of Custody form) and Advanced search capabilities.
Responsibilities
• Involved in designing the application and prepared Use case diagrams, class diagrams, sequence diagrams.
• Developed Servlets and JSP based on MVC pattern using Struts Action framework.
• Used Tiles for setting the header, footer and navigation and Apache Validator Framework for Form validation.
• Using Resource and Properties files for i18n support.
• Involved in writing Hibernate queries and Hibernate specific configuration and mapping files.
• Used Log4J logging framework to write Log messages with various levels.
• Involved in fixing bugs and minor enhancements for the front-end modules.
• Used JUnit framework for writing Test Classes.
• Used Ant for starting up the application server in various modes.
• Used Clear Case for version control.
• Used SDLC Life Cycle
Environment
Java JDK1.4, EJB2.x, Hibernate 2.x, Jakarta Struts 1.2, JSP, Servlet, JavaScript, MS SQL Server 7.0, Eclipse3.x,
Websphere 6, Ant, Windows XP, Unix, Excel Macro Development.
Multipoint Technologies Limited, Hyderabad, India Feb 2006 – Sep 2007
J2EE Developer
Project: EMRS
Electronic Medical Record System (EMRS) is complete solution designed for an enterprise wide-area networked
environment to implement a totally paperless office by allowing patient records to be accessed at the actual point-of-care
via a wireless pen pad or stationary PC and by managing all clinical, administrative and follow-up processes
electronically.
Responsibilities
• Involved in Requirement Analysis, Development and Documentation.
• Used MVC architecture (Jakarta Struts framework) for Web tier.
• Participation in developing form-beans and action mappings required for struts implementation and validation
framework using struts.
• Development of front-end screens with JSP Using Eclipse.
• Involved in Development of Medical Records module. Responsible for development of the functionality using Struts
and EJB components.
• Coding for DAO Objects using JDBC (using DAO pattern).
• XML and XSDs are used to define data formats.
• Implemented J2EE design patterns value object singleton, DAO for the presentation tier, business tier and Integration
Tier layers of the project.
• Involved in Bug fixing and functionality enhancements.
• Designed and developed excellent Logging Mechanism for each order process using Log4J.
• Involved in writing Oracle SQL Queries.
• Involved in Check-in and Checkout process using CVS.
• Developed additional functionality in the software as per business requirements.
• Involved in requirement analysis and complete development of client side code.
• Followed Sun standard coding and documentation standards.
• Participation in project planning with business analysts and team members to analyze the Business requirements and
translated business requirements into working software .
• Developed software application modules using disciplined software development process.
Environment
Java, J2EE, JSP, EJB, ANT, STRUTS1.2, Log4J, Weblogic 7.0, JDBC, MyEclipse, Windows XP, CVS, Oracle.