Post Job Free

Resume

Sign in

Data Developer

Location:
Austin, TX
Posted:
June 05, 2014

Contact this candidate

Resume:

Raghavender

(Hadoop Developer)

Ph: 732-***-****

Professional Summary:

• Seven years of experience with emphasis on Big Data Technologies, Development and

Design of Java based enterprise applications.

• Extensive experience in development of Big Data projects using Hadoop, Hive and other open

source tools/technologies

• Experience in installation, configuration, supporting and managing Hadoop clusters.

• Implemented in setting up standards and processes for Hadoop based application design and

implementation.

• Responsible for writing MapReduce programs.

• Logical Implementation and interaction with HBase.

• Developed MapReduce jobs to automate transfer of data from HBase.

• Perform data analysis using Hive and Pig.

• Load log data into HDFS using Flume.

• Gained good knowledge on creating strategies on risky transactions.

• Successfully loaded files to Hive and HDFS from MongoDB.

• Assist with the addition of Hadoop processing to the IT infrastructure.

• Worked in Multiple Environment in installation and configuration.

• Document and explain implemented processes and configurations in upgrades.

• Support development, testing, and operations teams during new system deployments.

• Evaluate and propose new tools and technologies to meet the needs of the organization.

• Experience in using Sqoop, ZooKeeper and Cloudera Manager.

• Good Knowledge on Hadoop Cluster architecture and monitoring the cluster.

• 24/7 operational support to production servers and related infrastructure clusters.

• Experience in Administering, Installation, configuration, troubleshooting, Security, Backup,

Performance Monitoring and Fine-tuning of Linux Redhat.

• Implemented stand-alone installation, file system management, backups, process control, user

administration and device management in a networked environment.

• Worked on debugging tools such as Dtrace, Struss and Top. Expert in setting up SSH, SCP,

SFTP connectivity between UNIX hosts.

• An excellent team player and self-starter with good communication skills and proven abilities to

finish tasks before target deadlines.

Technical Skills:

• Programming Languages : Java 1.4, C++, C, SQL, PIG, PL/SQL.

• Java Technologies : JDBC.

• Frame Works : Jakarta Struts 1.1, JUnit and JTest, LDAP.

• Databases : Oracle8i/9i, NO SQL (HBase),MY SQL,MS SQL server.

• IDE’s & Utilities : Eclipse and JCreator, NetBeans.

• Web Dev. Technologies : HTML, XML.

• Protocols : TCP/IP, HTTP and HTTPS.

• Operating Systems : Linux, MacOS, WINDOWS 98/00/NT/XP.

• Hadoop ecosystem :Hadoop and MapReduce, Sqoop, Hive, PIG, HBASE, HDFS,

Zookeeper, Lucene, Sun Grid Engine Administration

Education:

Bachelors in Computer Science and Information technology, JNTU,Hyderabad, India

Professional Experience:

DELL, Austin, TX Jan 2013 –Present Hadoop Developer

The purpose of the project is to store terabytes of log information generated by the company websites

and extract meaningful information out of it. The solution is based on the open source BigData s/w

Hadoop. The data will be stored in Hadoop file system and processed using Map/Reduce jobs which

includes getting the raw html data from the websites, process the html to obtain product and pricing

information, extract various reports out of the product pricing information and export the information for

further processing.

Responsibilities:

• Worked on analyzing Hadoop cluster and different big data analytic tools including Pig,

Hbasedatabase and Sqoop.

• Responsible for building scalable distributed data solutions using Hadoop.

• Implemented nine nodes CDH3 Hadoop cluster on Red hat LINUX.

• Involved in loading data from LINUX file system to HDFS.

• Worked on installing cluster, commissioning & decommissioning of datanode, namenode

recovery, capacity planning, and slots configuration.

• Created HBase tables to store variable data formats of PII data coming from different portfolios.

• Implemented a script to transmit sysprin information from Oracle to Hbase using Sqoop.

• Implemented best income logic using Pig scripts and UDFs.

• Implemented test scripts to support test driven development and continuous integration.

• Worked on tuning the performance Pig queries.

• Worked with application teams to install operating system, Hadoop updates, patches, version

upgrades as required.

• Responsible to manage data coming from different sources.

• Involved in loading data from UNIX file system to HDFS.

• Load and transform large sets of structured, semi structured and unstructured data

• Cluster coordination services through Zookeeper.

• Experience in managing and reviewing Hadoop log files.

• Job management using Fair scheduler.

• Exported the analyzed data to the relational databases using Sqoop for visualization and to

generate reports for the BI team.

• Responsible for cluster maintenance, adding and removing cluster nodes, cluster monitoring and

troubleshooting, manage and review data backups, manage and review Hadoop log files.

• Installed Oozie workflow engine to run multiple Hive and pig jobs.

• Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.

• Supported in setting up QA environment and updating configurations for implementing scripts

with Pig and Sqoop.

Environment: Hadoop, HDFS, Pig, Sqoop, HBase, Shell Scripting, Ubuntu, Linux Red Hat.

Comcast, Philadelphia, PA Sep 2011 – Dec 2012

Hadoop Developer

Comcast is the telecommunications company in the United States and is recognized as a leader in the

network services market by technology industry analyst firms. In ad targeting, we collected the data

from all the customers and our team performed the analysis by using various traits such as

demographics, purchase history etc, and did the advertising based on the results from the analysis.

Responsibilities:

• Involved in review of functional and non-functional requirements.

• Facilitated knowledge transfer sessions.

• Installed and configured Hadoop Mapreduce, HDFS, Developed multiple MapReduce jobs in

java for data cleaning and preprocessing.

• Importing and exporting data into HDFS and Hive using Sqoop.

• Experienced in defining job flows.

• Experienced in managing and reviewing Hadoop log files.

• Extracted files from CouchDB through Sqoop and placed in HDFS and processed.

• Experienced in running Hadoop streaming jobs to process terabytes of xml format data.

• Load and transform large sets of structured, semi structured and unstructured data.

• Responsible to manage data coming from different sources.

• Got good experience with NOSQL database.

• Supported Map Reduce Programs those are running on the cluster.

• Involved in loading data from UNIX file system to HDFS.

• Installed and configured Hive and also written Hive UDFs.

• Involved in creating Hive tables, loading with data and writing hive queries which will run internally

in map reduce way.

• Gained very good business knowledge on health insurance, claim processing, fraud suspect

identification, appeals process etc.

• Developed a custom File System plug in for Hadoop so it can access files on Data Platform.

• This plugin allows Hadoop MapReduce programs, HBase, Pig and Hive to work unmodified

and access files directly.

• Designed and implemented Mapreduce-based large-scale parallel relation-learning system

• Extracted feeds form social media sites such as Facebook, Twitter using Python scripts.

• Setup and benchmarked Hadoop/HBase clusters for internal use

• Setup Hadoop cluster on Amazon EC2 using whirr for POC.

• Wrote recommendation engine using mahout.

Environment:Java 6, Eclipse, Oracle 10g, Sub Version, Hadoop, Hive, HBase, Linux,, MapReduce,

HDFS, Hive, Java (JDK 1.6), Hadoop Distribution of HortonWorks, Cloudera, MapReduce, DataStax,

IBM DataStage 8.1, Oracle 11g / 10g, PL/SQL, SQL*PLUS, Toad 9.6, Windows NT, UNIX Shell

Scripting.

Prudential Financial, Newark, NJ Jun 2010 – Aug2011

Hadoop Developer

Prototyped Track and Trace for Insurances, using Cloudera Hadoop, REST, XML fine-grain access

control with certificates, with capacity of 1,000+ transactions per second, with background processes to

verify chain-of-custody and fraud prevention.

Responsibilities:

• Responsible for architecting Hadoop clusters.

• Assist with the addition of Hadoop processing to the IT infrastructure.

• Perform data analysis using Hive and Pig.

• Load log data into HDFS using Flume, Kafka.

• Monitoring Hadoop cluster using tools like Nagios,Ganglia and Cloudera Maneger.

• Automation script to monitor HDFS and HBase through cronjobs.

• Plan, design, and implement processing massive amounts of marketing information, complete

with information enrichment, text analytics, and natural language processing.

• Prepare multi-cluster test harness to exercise the system for performance and failover.

• Develop high-performance cache, making the site stable and improving its performance.

• Create a complete processing engine, based on Cloudera's distribution, enhanced to

performance.

• Administrative support for parallel computation research on a 24-node Fedora/ Linux cluster.

• Build and support standard-based infrastructure capable of supporting tens of thousands of

computers in multiple locations.

• Negotiated and managed projects related to designing and deploying tis architecture.

Environment:Hive, Pig, HBase, Zookeeper, Sqoop, Java, JDBC, JNDI, Struts, Maven, Trac,

Subversion, JUnit, SQL language, spring, Hibernate, Junit, Oracle, XML, AltovaXmlSpy, Putty and

Eclipse.

CSX Transportation, Albany, NY Jul 2009 – May 2010

J2EE Developer

Yard Enterprise System: CSX Transportation is a Class I railroad in the United States and owns 21,000

route miles. The project is designed to implement two levels of data validation. The project dealt with

designing and implementing the frontend in order to maintain the route information and keep track of

the cargo being delivered by trains to various destinations. The main aim of the project was to increase

the efficiency of cargo delivery by reducing the number of trips and the number of trains to be used with

the help of a common interface.

Responsibilities:

• Involved in Presentation Tier Development using JSF Framework and ICE Faces tag Libraries.

• Involved in business requirement gathering and technical specifications.

• Implemented J2EE standards, MVC2 architecture using JSF Framework.

• Implementing Servlets, JSP and Ajax to design the user interface.

• Extensive experience in building GUI (Graphical User Interface) using JSF and ICE Faces.

• Developed Rich Enterprise Applications using ICE Faces and Portlets technologies.

• Experience using ICE Faces Tag Libraries to develop user interface components.

• Used JSF, JSP, Java Script, HTML, and CSS for manipulating, validating, customizing, error

messages to the User Interface.

• Used EJBs(Session beans) to implement the business logic, JMS for communication for

sending updates to various other applications and MDB for routing priority requests.

• All the Business logic in all the modules is written in core Java.

• Wrote WebServices using SOAP for sending and getting data from the external interface.

• Developed a web-based reporting for monitoring system with HTML and Tiles using Struts

framework.

• Middleware Services layer is implemented using EJB(Enterprise Java Bean - stateless) in

WebSphere environment.

• Used Design patterns such as Business delegate, Service locator, Model View Controller,

Session façade, DAO.

• Funds Transfers are sent to another application using JMS technology asynchronously.

• Involved in implementing the JMS (Java messaging service) for asynchronous communication.

• Involved in writing JMS Publishers to post messages.

• Involved in writing MDB(Message Driven Beans) as subscribers.

• Created Stored procedures using PL-SQL for data modification (Using DML insert, update,

delete) in Oracle

• Interaction with Oracle database is implemented using Hibernate.

Environment : J2EE, EJB, JSF, ICE Faces, EJB, WebServices, XML, XSD, Agile, Microsoft

Visio, Clear Case, Oracle 9.i/10.g, Weblogic8.1/10.3,RAD, LOG4j,Servlets, JSP, Unix.

Infotech/Vonage (Voice cards), India Jun 2007 – Jul 2009

J2EE Developer

The Telephone card portal will enable users to purchase voice Cards and Pre-paid Cards to call

different countries. The portal will have the ability to compare the various voice Cards available in the

Market and allow users to make their own choices. The main aim of designing this portal is to enable

our users to compare the prices and tariff structures of the available cards in the market and then

choose and buy online the right card, which suits their purpose. The portal is designed in such a fashion

wherein the Administrator of this site can add/modify/delete database information, contents and pictures

on some areas of the site.

Responsibilities:

• Involved in designing the application and prepared Use case diagrams, class diagrams,

sequence diagrams.

• Developed Servlets and JSP based on MVC pattern using Struts Action framework.

• Used Tiles for setting the header, footer and navigation and Apache Validator Framework for

Form validation.

• Using Resource and Properties files for i18n support.

• Involved in writing Hibernate queries and Hibernate specific configuration and mapping files.

• Used Log4J logging framework to write Log messages with various levels.

• Involved in fixing bugs and minor enhancements for the front-end modules.

• Used JUnit framework for writing Test Classes.

• Used Ant for starting up the application server in various modes.

• Used Clear Case for version control.

• Used SDLC Life Cycle

Environment: Java JDK1.4, EJB2.x, Hibernate 2.x, Jakarta Struts 1.2, JSP, Servlet, JavaScript, MS

SQL Server 7.0, Eclipse3.x, Websphere 6, Ant, Windows XP, Unix, Excel Macro Development.

Magna Infotech Pvt Ltd., Bangalore, India Sep 2006 – Jun 2007

J2EE Developer

Universal Banking Solution is enhanced account information for the corporate banking customers

across the globe. This web portal provides customer consolidated account information at any point of

time. Customer can view his account balance, consolidated balance, historical balance, historical

consolidated balance, transaction summary, forward balance and unclear values.

Responsibilities:

• Involved in Requirement Analysis, Development and Documentation.

• Used MVC architecture (Jakarta Struts framework) for Web tier.

• Participation in developing form-beans and action mappings required for struts implementation

and validation framework using struts.

• Development of front-end screens with JSP Using Eclipse.

• Involved in Development of Medical Records module. Responsible for development of the

functionality using Struts and EJB components.

• Coding for DAO Objects using JDBC (using DAO pattern).

• XML and XSDs are used to define data formats.

• Implemented J2EE design patterns value object singleton, DAO for the presentation tier,

business tier and Integration Tier layers of the project.

• Involved in Bug fixing and functionality enhancements .

• Designed and developed excellent Logging Mechanism for each order process using Log4J.

• Involved in writing Oracle SQL Queries.

• Involved in Check-in and Checkout process using CVS.

• Developed additional functionality in the software as per business requirements.

• Involved in requirement analysis and complete development of client side code.

• Followed Sun standard coding and documentation standards.

• Participation in project planning with business analysts and team members to analyze the

Business requirements and translated business requirements into working software .

• Developed software application modules using disciplined software development process.

Environment: Java, J2EE, JSP, EJB, ANT, STRUTS1.2, Log4J, Weblogic 7.0, JDBC, MyEclipse,

Windows XP, CVS, Oracle.



Contact this candidate