Post Job Free
Sign in

Data Sales

Location:
Hyderabad, Telangana, India
Posted:
June 14, 2016

Contact this candidate

Resume:

Vineela Pachava

Mobile: 248-***-****

************@*****.***

Professional Summary:

Around Seven years of professional IT experience with Four years of Big dataHadoop ecosystems experience in ingestion, storage, querying, processing and analysis of Big data.

Excellent understanding / knowledge of Hadoop architecture and various components such as HDFS, JobTracker, TaskTracker, NameNode, DataNode and MapReduce programming paradigm.

Working experience in MapReduce programming model and Hadoop Distributed File Systems (HDFS).

Hands on experience in installing, configuring, and using Hadoop ecosystem components like Hadoop MapReduce, HDFS, Oozie, Hive, Sqoop, Pig, Spark and Flume.

Developed scripts, numerous batch jobs to schedule various Hadoop programs.

Experience in analyzing data using HiveQL, PIG Latin, and custom MapReduce programs in Java.

Hands on experience with Apache Spark and exposure to stream data processing Storm.

Worked on MapReduce programs using Python libraries.

Worked on importing and exporting data from different databases like Oracle, Teradata into HDFS and Hive using Sqoop.

Strong experience in collecting and storing stream data like log data in HDFS using Apache Flume.

Well experienced with implementing Join operations using PIG Latin.

Involved in writing data transformations, data cleansing using PIG operations.

Experience in data management and implementation of Big Data applications using Hadoop frameworks.

Experience in understanding the security requirements for Hadoop and integrate with Kerberos authentication and authorization infrastructure.

Experience in designing, developing and implementing connectivity products that allow efficient exchange of data between our core database engine and the Hadoop ecosystem.

Strong working experience in application development using JDBC, Servlets, JSP, Spring Framework, Spring MVC, Hibernate 3.0 ORM.

Expertise in J2EE Design Patterns like MVC, AO, DTO, Front controller, Fascade and Singleton.

Proficient in writing ANT, Maven builds script to automate the application build and deployment.

Expertise in writing Junit and unittest test cases for different kind of applications and also development experience in mock object to simulate real objects.

Experienced in configuring and deploying over Oracle WebLogic, WebSphere and Tomcat server.

Worked in TCP/IP and UDP based network programming (sockets, datagrams and multicasting) in Java as well as in Python.

Experience in using Python libraries NumPy, Scipy, Pandas.

Techno-functional responsibilities include interfacing with users, identifying functional and technical gaps, estimates, designing custom solutions, development, leading developers, producing documentation, and production support.

Excellent interpersonal and communication skills, creative, research-minded, technically competent and result-oriented with problem solving and leadership skills.

Active participate in team to accomplish targets.

Education:

Bachelors of Engineering, JNT University, Anantapur, India.

Technical Skills:

Hadoop/Big Data

HDFS, MapReduce, Hive, Flume, Pig, Zookeeper, Impala HBase, Sqoop, Oozie, Spark

Programming Languages

C, C++, Java, Python, Pig Latin, Hive QL, Shell Script, SQL, PL/SQL

Methodologies

Agile, Design Patterns (Core Java and J2EE)

Testing API

Junit, unittest

Frameworks

MVC, Hibernate, Spring, Hadoop

Databases

Oracle 7.x, 8.1.6, SQL Server7.0, MySQL, HBase

Web Related

Servlets, JDBC, JSP, HTML, XML, JavaScript, AJAX, Web Services

Network Protocols

TCP/IP, UDP, HTTP, DNS, ICMP, DHCP

Employment Experience:

CREDIT ACCEPTANCE, Southfield, Michigan, USA, June’2014 – PRESENT

Project:Credit Acceptance is an indirect auto finance company, working with car dealers. We as a team take advantage of available data, implemented Hadoop on Data warehouse environment to build the risk analysis and predictive modeling strategies using map reduce programand cross verified with SAS system for large number of data sets and build analytical reports based on historical data for different pricing strategies.

Hadoop Developer

Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required.

Experience with Amazon web services: EC2, S3, and EMR (Elastic Map Reduce) equivalent cloud computing approaches.

Installed and configured Flume, Hive, Pig, Sqoop and Oozie on the Hadoop cluster.

UsedHiveQL to store data into tables and to perform queries.

Used Hive to transform large datasets into required format.

Handling structured and unstructured data and applying ETL processes.

Used Pig for data deduplication.

Used Distributed Cache to store reference data and Improve performance of MR.

Extensively used Map-Reduce to process the available terabytes of data.

Worked extensively in creating map reduce jobs to the financial services such as passing the raw data, populating staging tables and storing the refined data in partitioned tables.

Worked on merging and updating the data using MapReduce programs in both Java and Python.

Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.

Used Spark API over Cloudera Hadoop YARN to perform analytics on data in Hive.

Involved in converting Hive/SQL queries into Spark transformations using Spark RDDs, Python.

Exposure to stream data processing Storm.

Integrated Oozie with the rest of the Hadoop stack supporting several types of Hadoop jobs out of the box (like Java MapReduce, Pig, Hive, Sqoop) as well as system specific jobs (such as Java programs and shell scripts).

Involved in installing and configuring Kerberos for the authentication of users and Hadoop daemons.

Developed projects using Maven to build and deploy.

Environment: CDH5, Hive, Sqoop, Pig, Oozie, Cloudera Manager, Java, Linux, Eclipse, Maven

IMSHEALTH, Collegeville, Pennsylvania, USA, May’ 2012 – May’2014

Project:The Project explains how Customer Loyalty scoreboard (CLS), Patient Benchmark Integration Portal (PBI) aligns to the Supplier services Process.

Java/Hadoop Developer

The Premier Provider Application is a Supplier Services application.

Supplier Services application frameworks used Spring Framework, Spring MVC, Hibernate Object/Relational Mapper.

Implemented authentication and authorization for the web layer using the Spring Security.

Used JSP, Java Script, HTML5, and CSS for manipulating, validating, customizing, error messages to the User Interface.

The application uses the Hibernate as the JavaPersistanceAPI to perform all the database transactions.

Used the Log4j as the logging tool and unittest framework in python.

Wrote python scripts to parse XML documents and load the data in database.

Worked on importing the data and generating pivot tables using Pandas and weekly reports for subscribers.

Productively analyze and report data using Python (Pandas, NumPy, matplotlib).

Implemented SOAP Web services and Restful.

Experience in using collections in Python for manipulating and looping through different user defined objects.

Developed Merge jobs in Python to extract and load data into database.

Worked this medical group streamline business processes by developing, installing and configuring Hadoop ecosystem components that moved data from individual servers to HDFS.

Installed and configured MapReduce, HIVE and the HDFS; implemented CDH3 Hadoop cluster. Assisted with performance tuning and monitoring.

Created reports for the BI team using Sqoop to export data into HDFS and Hive.

Designed multiple MapReduce jobs in Java for data cleaning and preprocessing.

Assisted with data capacity planning and node forecasting.

Involved in creating partitioned Hive tables, loading with data and writing hive queries that will run internally in map reduce way.

Attended workshop on Spark which is on initial release with the IT team for new innovations.

Environment:RAD 7, Java 6.0, Python, Pandas, Spring MVC Framework, Hibernate, Oracle 9i, TOAD, UNIX scripting, log4j, unittest, Oracle web logic, Webservices, HDFS, Hadoop, Sqoop, Pig, Hive.

MERCK, INDIA,Feb’ 2011 - April’ 2012

Project:This is the heart of the Supply Chain Management process of the WALGREENS. It mainly deals as the central repository for various sources of data such as Integrated Forecasting, Inventory metrics, BusinessPlan, DataMart, Data Warehouse and various DC Boxes.

Java/J2EE Developer

The Application is developed based on the Agile Development process, so responsible for gathering requirements from the clients regularly.

Developed the Application as a Reusable and highly customizable set of agile, stand alone components.

The Web Application is developed using Spring MVC and used Spring Dependency Injection and AOP.

The GUI is developed using the Servlets, JSP, JavaScript, HTML, CSS and AJAX.

Used the Built in Tag Libraries in the Framework, which ease the code development.

Performed dependency injection and configured spring beans for modules Spring IOC.

Used Maven to build and package the application.

Enhance and modify the presentation layer and GUI framework that are written using JSP and client-side validations done using JavaScript & design enhanced wireframe screens.

Excessive work in writing SQL Queries, Stored procedures using TOAD.

Deployed the Application on Tomcat server.

Wrote unit and system test cases for modified processes and Continuous Integration with the help of QC team and Configuration team on timely manner.

Using the log4j logging tool for logging all our jobs.

Using the Web Services for User Authentication for enabling the Authentication for the report to be accessible as External.

SOAP based XML messages are used for communicating with the Web Services.

Have been using the CVS as the version Control Tool.

Environment:Java 5.0, Spring MVC Framework, JSP, Java Script, HTML, CSS, AJAX, Tag Libraries, Oracle 9i, SQL Server, log4j, Tomcat, Web Services, SOAP, XML, CVS, Selenium.

ASHOK LEYLAND, INDIA, July’2009 – Jan’2011

Project:Ashok Leyland ERP System for Automobile Industry: Domestic Vehicle SalesIt is a comprehensive ERP system for Automobile Industry. This support project consists of 5 major modules: ALPAY (HR & Payroll), ALCORP (Corporate), ALMAP (Manufacturing), ALFIN (Finance), AL NPS (Non Production), ALMARK (Sales & Marketing). The ALMARK Sales & marketing module consists of vehicle sales, engine sales, spare parts sales and service & warranty. The Vehicle Sales module covers following activities:

RO/AO will raise Vehicle Release Authorization (VRA) i.e. providing the authorization to the RSO Champion to bill the vehicle.

Temporary invoice generation, approved by RO.

Java Developer

Work with developers, Business and Release Managers to discussed ways to fix the defects.

Included Log4j for debugging and testing the applications.

Wrote Specification for the development.

Participated in Server side and Client side programming.

Wrote SQL stored procedures, used JDBC to connect to database.

Created a user Interface through HTML and JSP.

Involved in developing various Servlets.

Tested and validated the application on different testing environments.

Environment:Windows NT, Java (JDK 1.2), HTML, JSP, Servlets, Eclipse 3.0, Oracle 8.0, JDBC, and Apache Tomcat.



Contact this candidate