Post Job Free

Resume

Sign in

Manager Data

Location:
Alpharetta, GA
Posted:
April 17, 2020

Contact this candidate

Resume:

Lalitha Ponnapati adcuno@r.postjobfree.com

Mobile : 470-***-****

Professional Summary

Having 9 + years of IT Experience in Bigdata and Dataware housing domain.

Having 3 + years of working experience on Hadoop Admin.

Strong technical expertise in Hadoop Administration (BIG DATA Administration) with Hadoop 2.0 Multi-Node cluster Set up, Name node (HDFS) High Availability, HDFS Federation, MRv2 and YARN Framework, Apache Oozie workflow scheduler, Hadoop cluster Implementation, Cloudera Manager, Hortonworks Ambari, Rstudio, jupyter Notebook, Zeppelin, Python, autosys, Shell Script, SQL Scripts

Expertise in leading Mainframe applications using COBOL-JCL-DB2-CICS-VSAM

Developed and Maintained COBOL-DB2, CICS-DB2 and CICS web services applications

Experienced in DB2 performance tuning using DB2 explain followed by Query optimization

Experienced in testing Web Services using SOAP

Developed and Maintained applications written in Ezytrieve, SAS, Telon

Proficient in using Mainframe tools like Xpediter, FileAid, FileMaster, Endevor, ESP scheduler, NDM, DFSORT, SYNCSORT Telon

Experienced in providing production support of Mainframe based applications

Excellent in system analysis and understanding client requirements

Practiced standard software development processes like - Estimation (Complexity

point, Function point estimation technique),Preparing High Level Design, Detailed Design, Coding, Testing(Unit Testing, Regression Testing, Integration Testing, User Acceptance Testing) and Implementation.

Prepared quality artifacts like Requirement Traceability Matrix, Detail Design Document, Test Case, Test Logs, and Defect Causal Analysis etc.

Well aware of SDLC, Change Management, Version Control, Lean Techniques, Agile Methodology

Provided technical training to junior resources. Supervised team members' work and ensured defect free, timely delivery to clients.

Proposed and developed Mainframe based tools to automate daily processes and tune existing system performance.

Technical Skills:

Hadoop / Big Data

HDP 3.1.0, HWX Ambari, HDFS, MapReduce, Yarn, Hive, Pig, HBase, Sqoop, Flume, Oozie, Kerberos, Spark, Scala, Zookeeper, Splunk,YUM, Solar, Kafka, Apache Ranger, Bigtop, Big Insights, Cloudera Manager, Nagios, Ganglia, Graphite, Strom, Unix, Shell Scripting, Mongo DB, Rstudio, jupyter Notebook, Zeppelin, Python, autosys

Programming languages

R Studio, Python, Spark, Scala, Pl/1, Cobol, Jcl, Sql, Rexx Asm, Pls, Clist

ETL

Informatica, Datastage, Cognos.

Database

Postgresql, Oracle, HBase, DB2, IMS/DB

Operating Systems

Windows, Ubuntu, Centos, Red Hat Linux, UNIX, MVS

Work Experience:

Client: MAERSK Oct ’17 to Till date Role : Hadoop Administrator

Scope : Maersk Line Apmoller : Maersk Line is one of the leading liner shipping companies in the world, serving customers all over the globe. Maersk has activities in a variety of business sectors, primarily within the transportation and energy sectors. It is the largest container ship operator and supply vessel operator in the world. Maersk is based in Copenhagen, Denmark, with subsidiaries and offices in more than 135 countries worldwide.

Design Hadoop Cluster environment from the scratch to support Multi-tenant architecture process.

Build Hortonworks distribution Hadoop Cluster environment based on architecture design (number of master nodes, data nodes, client nodes and network setting)

Configure Active Directory/LDAP, MIT Kerberos and SSL configuration on Cluster to provision authorization and for multiple users.

Enable High availability server architecture process for Name Services: Hadoop hdfs metadata, Hive Services and resource managers to run business as usual without cluster down.

Monitor Cluster alerts, Node and data disk failures and perform decommission and commission servers

Worked on Ambari upgrades and resolve the upgrade issues.

Analyzing Hadoop cluster and different Big Data analytic tools including Hive, HDFS, HBase, Spark and Scala.

Importing and exporting data into HDFS and Hive using SQoop from HDFS to Relational Database Systems (RDBMS) and vice-versa.

Worked on HDP 3.1 upgrade and resolved most of the upgrade issues and closely work with HWX/CDH teams to do bug fixes.

Execute Agile/Scrum methodology to develop the application by performing everyday standups to discuss status update of the stories and roadblocks. Use Version One as agile project management software for managing the backlogs, stories, goals and incidents.

Environment: Redhat Linux, Hortonworks distribution (HDFS, YARN, Ambari, Hive/HS2, Sqoop, Spark, Hbase, Ranger etc) PostgreSQL, Kerbros, Ranger, Rstudio, jupyter Notebook, Zeppelin and Python.

Client: Lenovo, Morrisville, NC. Oct ’16 to Sep' 17

Role: Hadoop Administrator

Scope: Lenovo supplies replacement parts to Customer Engineers and Customers all over the world. Customer Engineers are the group of people who fix computers under warranty period.

Responsibilities:

Setup a Multi Node Cluster. Plan and Deploy a Hadoop Cluster using Cloudera Manager.

Deployed multi-node development, testing and production Hadoop clusters with different Hadoop components (HIVE, PIG, SQOOP, OOZIE, FLUME, HCATALOG, ZOOKEEPER) using Cloudera Manager.

Configured Capacity Scheduler on the Resource Manager to provide a way to share large cluster resources.

Deployed Name Node high availability for major production cluster and designed automatic failover control using zookeeper and quorum journal nodes.

Configured Oozie for workflow automation and coordination.

Good experience in troubleshoot production level issues in the cluster and its functionality.

Backed up data on regular basis to a remote cluster using distcp.

Regular Ad-Hoc execution of Hive and Pig queries depending upon the use cases.

Regular Commissioning and Decommissioning of nodes depending upon the amount of data.

Experience in Disaster Recovery and High Availability of Hadoop clusters/components.

Diligently teaming with the infrastructure, network, database, application and business intelligence teams to guarantee high data quality and availability.

Diagnose and resolve performance issues and scheduling of jobs using

Experience designing data queries against data in the HDFS environment using tools such as Apache Hive.

Imported data from MySQL server to HDFS using Sqoop.

Manage the day-to-day operations of the cluster for backup and support.

Used the RegEx, JSON and Avro SerDe’s for serialization and de-serialization packaged with Hive to parse the contents of streamed log data.

Writing shell scripts to automate the administrative tasks.

Implemented Hive custom UDF’s to integrate the Weather and geographical data with business data to achieve comprehensive data analysis.

Used Oozie workflow engine to manage interdependent Hadoop jobs and to automate several types of Hadoop jobs such as Java map-reduce, Hive and Sqoop as well as system specific jobs.

Worked along with the Hadoop Operations team in Hadoop cluster planning, installation, maintenance, monitoring and upgrades.

Environment: Cloudera Manager, HDFS, Hbase, Hive, Pig, Kerberos, Sqoop, Knox, Unix, Shell Script, Linux, Ranger,Yarn, Apache Oozie workflow, Flume, Zookeeper.

Client: UPS, Atlanta, GA Apr ’10 to Aug’ 16

Role: Mainframe Developer

Scope : eTT Tracking is a global web-based internal UPS system that provides Customer Service Representatives (CSRs) with search capabilities to Package Data, Delivery Data, Claims, COD Payout, and Damages/Tracers. eTT Tracking also provides access to various external UPS applications such as Will Call Locator, Time-in-Transit, Void Shipment, and Guaranteed Service Refund

Analysis of the programs for requested Crs.

Managing all faces of project life cycle- design, development, testing and deployment.

Developed RAD (Receive and Deliver) application to receive the trade settlement status from DTC.

Developed a dash board to resolve the open items.

Moving the code to UAT for client testing and resolving the issues on priority basis.

Developed a pro-play tool to make sure no functional issues with latest code.

Communicate with clients and department heads to gather requirements.

Ensure the technical specifications and set deadlines.

Work directly with IT teams throughout development, deployment and testing.

Prepare reports and ensure they are error less.

Develop the post edge application to receive the trade confirmation messages from Impact system

Having good knowledge on clearance and settlement activities

Data analysis to identify the data inconsistencies and fixing the same.



Contact this candidate