Srujan Bala ************@*****.*** contact: 847-***-****
SUMMARY:
Around 6years of extensive hands-on experience in IT industry. Currently working on Big Data/Hadoop ecosystem tools.
Experience in new Hadoop 2.0 architecture YARN and developing YARN Applications.
Good experience in processing unstructured, Semi-structured and Structured data.
Thorough understanding of the HDFS, Map Reduce framework and experience in developing Map Reduce Jobs.
Experienced in building highly scalable Big-data solutions using Hadoop and multiple distributions like Cloudera, Horton works and NoSQL platforms.
Good Exposure on Map Reduce programming using Java, PIG Latin Scripting and Distributed Application and HDFS.
Hands-on experience on major components in Hadoop Ecosystem including Hive, HBase, HBase-Hive Integration, PIG, Sqoop, Flume& knowledge of Mapper/Reduce/HDFS Framework.
Experience in Loading Tuple shaped data into Pig and Generate Normal Data into Tuples. Ability to build User-Defined Functionalities (UDFs) both PIG and HIVE.
Ability to move the data in and out of Hadoop RDBMS, No-SQL, UNIX and Mainframe from various systems using SQOOP and other traditional data movement technologies.
Worked on HBase Shell, CQL and HBase API as part of the proof of concept.
Having Knowledge in Apache Spark.
Having good knowledge on TDD and JENKINS.
Good knowledge of Collections package, Exception Handling and OOP’s Concepts.
Having knowledge on Low Level Design and Use Case Diagrams.
Having good work experience in developing web applications using Java/J2EE.
Good database knowledge and working experience in Oracle.
Experience in Hadoop Distributions like Cloudera, Horton Works.
Having hands on experience in Data Warehousing to design and loading tables with large data, and can develop the enterprise levels of data.
Maintained, audited and built new clusters for testing purposes using the cloudera manager.
Good understanding of agile methodologies, Test Driven Development and continuous integration.
Analyzing the data sets and reporting them.
Implemented Cluster for NoSQL tools Cassandra,MongoDB as a part of POC to address Hbase limitations.
Technical Skills
Big Data Ecosystems : Hadoop, MapReduce, HDFS, HBase, Hive, Pig, Sqoop,Kafka, Oozie, MongoDB, Cassandra.
Languages : C, Core Java, Unix, SQL, Python, R.
J2EE Technologies : Servlets, JSP, JDBC, Java Beans.
Methodologies : Agile, UML, Design Patterns (Core Java and J2EE).
Monitoring and Reporting : Custom Shell scripts.
NoSQL Technologies : Hbase, Cassandra, MongoDB.
Frameworks : MVC, Struts, Hibernate, And Spring.
Operating Systems : Windows XP/Vista/7, UNIX, Linux.
Web Servers : WebLogic, WebSphere, Apache Tomcat.
WORKEXPERIENCE:
Vibrant Media, NYC, NY Jun 2015 – till date
Hadoop Developer
Description: It is an advertising organization. It analyzes content like social media, click streams, page view, video etc. in real time to ensure maximum ad relevance. The new data is then joined with existing structured data sets for deeper analysis and targeted promotions. With the analytic approach, they understand where and when to deliver rich content experiences that capture and engage consumers.
Responsibilities:
Importing and exporting data into HDFS and Hive using Sqoop and Kafka.
Develop different components of system like Hadoop process that involves Mapreduce & Hive.
Developed interface for validating incoming data into HDFS before kicking off Hadoop process.
Written hive queries using optimized ways like window functions, customizing Hadoop shuffle & sort parameters, ORC file format.
Worked on tuning Hive and Pig to improve performance and solve performance related issues in Hive and Pig scripts with good understanding of Joins, Group and aggregation and how it translates to Map Reduce jobs.
Developing map reduce programs using Combiners, Sequence Files, Compression techniques, Chained Jobs, multiple input and output API.
Experience working on node tools which offer several commands to return Cassandra metrics pertaining disk usage.
Experience in creating tables, dropping and altered at run time without blocking updates and queries using Cassandra.
Worked on pre-processing the logs and semi structured content stored on HDFS using PIG and the processed data is imported into Hive warehouse which enables business analysts to write Hive queries.
Experience in managing and reviewing Hadoop log files.
Worked on Unix shell scripts for business process and loading data from different interfaces to HDFS.
Developed suit of Unit Test Cases for Mapper, Reducer and Driver classes using MR Testing library.
Involved in creating Hive tables, loading with data and writing hive queries.
Hands on experience in eclipse, VPN, Putty, winSCP, VNCviewer, etc.
Environment: Red Hat Enterprise Linux 5, Hadoop 1.0.4, Map Reduce, Hive 0.10, PIG, Shell Script, SQOOP 1.4.3, Eclipse, Java SDK 1.6.
Ascena retail group, NJ June 2014 - May2015
Hadoop Developer.
Description: Data Migration project using Hadoop ecosystems components from traditional data warehousing and BI system for retail domain. Involved in extracting data from different servers and dump the data into Hadoop cluster to generate reports for analysis. The initial areas of Data Migration were MDM, Planning, EDW and Analytics.
Responsibilities:
Developed Mapreduce program to convert mainframe fixed length data to delimited data.
Used Pig Latin to apply transaction on systems of record.
Experience on Hadoop Cluster monitoring tools like Nagios, Ganglia, and Cloudera Manager.
Extensively worked with Cloudera Distribution Hadoop, CDH 5.x, CDH4.x .
Developed Pig scripts and UDFs extensively for Value Added Processing (VAPs).
Design and developed custom Avro storage to use in Pig Latin to load and store data.
Worked on Cassandra which has Hadoop integration with Map reduce support.
Experience in managing and reviewing Hadoop log files.
Actively involved in design analysis, coding, and strategy development.
Developed SQOOP commands to pull data from Teradata and push to HDFS.
Developed Hive scripts for implementing dynamic partitions and buckets for retail history data.
Streamlined Hadoop jobs and workflow operations using Oozie workflow and scheduled through AutoSys on a monthly basis.
Developed MapReduce to generate sequence id in Hadoop.
Developed Pig scripts to convert the data from Avro to text file format.
Developed Pig scripts and UDF's as per the business rules.
Designed and developed read lock capability in HDFS.
Developed Hive scripts for implementing control tables logic in HDFS.
End-to-end implementation with Avro and Snappy.
Provided production support in my initial stages for the product which is already developed.
Created POC for Flume implementation.
Helping other teams to get started with the Hadoop ecosystem.
Health Services Portal, NC Jan 2013-May 2014
Java Developer
Description: The Health Services Portal (Belgacom Group) is Belgium's reference provider of integrated telecommunication services. It offers a complete quadruple-play solution that integrates fixed as mobile telephony, Internet. Health Service Portal is a web application which acts as a portal between service providers and service consumers. It has three Modules Consumers (Users), Providers, and Admin. Admin can approve and reject the Consumers and Providers Requests at the time of Signup. English is the default Language. It provides local language support for consumers including Hindi, Chinese, Japanese, Korean, Thai, Arabic and German.
Responsibilities:
Understanding the business requirements and developed code for module of the application.
Developing the application based on MVC Architecture, and implemented Action classes.
Implemented Model Classes and Struts2 tags as views.
Implemented mapping files for corresponding tables using Hibernate 3.0 in developing the Project.
Involved in Adding Server side Validations.
Creating unit test case documents.
Developed business components to process requests from the user and used the Hibernate to retrieve and update patient information.
Worked with database Objects using TOAD and SQL Navigator for development and administration of various relational databases.
Wrote and used Java Bean classes, JSP, Stored Procedures and JSP custom tags in the web tier to dynamically generate web pages.
Environment: Java5, Struts2.x, Hibernate3.x, Oracle, JSP, JBoss, SVN, Eclipse Html.
Apollo, Hyderabad, India Feb 2011 – Dec 2012
Role: Java/J2EE Developer
Description: The Pharmacy solution is specifically engineered to meet Pharmacy's needs for high-performance products and services. These products and services comprise a single, integrated solution not available from a group of technology providers. The strategy integrates task-oriented workflow, medication therapy management (MTM) programs, centralized files, retail-based fulfillment and central fill, an electronic pharmacy record, and e-prescription technology.
Responsibilities:
Utilized Agile Methodologies to manage full life-cycle development of the project.
Implemented MVC design pattern using Struts Framework.
Form classes of Struts Framework to write the routing logic and to call different services.
Created tile definitions, Struts-config files, validation files and resource bundles for all modules using Struts framework.
Developed web application using JSP custom tag libraries, Struts Action classes and Action. Designed Java Servlets and Objects using J2EE standards.
Used JSP for presentation layer, developed high performance object/relational persistence and query service for entire application utilizing Hibernate.
Developed the XML Schema and Web services for the data maintenance and structures.
Developed the application using Java Beans, Servlets and EJB’s.
Created Stateless Session EJB’s for retrieving data and Entity Beans for maintaining User Profile.
Used WebSphere Application Server and RAD to develop and deploy the application.
Worked with various Style Sheets like Cascading Style Sheets (CSS).
Designed database and created tables, written the complex SQL Queries and stored procedures as per the requirements.
Involved in coding for JUnit Test cases, ANT for building the application.
Environment:Java/J2EE, Oracle 10g, SQL, PL/SQL, JSP, EJB, Struts, Hibernate, WebLogic 8.0, HTML, AJAX, Java Script, JDBC, XML, JMS, XSLT, UML, JUnit, log4j, My Eclipse 6.0
GE Energy, Hyderabad, India March 2010-Jan2011
Junior Java Developer
Responsibilities:
Primary responsibilities included the development of the code using core Java and web Development skills.
Use Struts and JavaScript for web page development and front-end validations.
Fetch and process customer related data using Mercator (IBM WTX) as interface between Apollo workstation with Mainframes.
Created Servlets, JSPs and used JUnit framework for unit testing.
Developed EJBs, DAOs, stored Procedures and SQL queries to support system functionality.
Application design and documentation UML system use cases, class, sequence diagrams developed using MS Visio.
Use Ant scripts to automate application building and deployment processes.
Support Production/Stage application defects, track and document using Quality Center.
Implemented various Unix Shell Scripts as per the internal standards.
Environment: Java 1.4.2, Struts 1.2, Java script, JDBC, CVS, Eclipse, Web logic Server 9.1, Oracle 9i, Toad, Linux.