Hari Babu Daddala
In pursuit of challenging assignments that would facilitate the maximum utilization and application of my broad skills and expertise in making a positive difference to the organization. Summary
Result oriented professional with 4+ years of experience on Business Intelligence and Data Warehousing using ETL tool Informatica, SSIS, SAP DS (Data Services) and Hadoop Administration on Hortonworks Data Platform (HDP).
Having 3 years of experience on Hadoop administration and its eco system components YARN, MapReduce, Sqoop, Pig, Hive, HBase, Oozie, Zookeeper.
Responsible for syncing data to multiple reporting mini clusters (ponds) like sales, Inventory, SCP and Partner information for analytics.
Configured various property files like core-site.xml, hdfs-site.xml, mapred- site.xml and hadoop-env.xml based upon the job requirement.
Responsible for subscribing data to multiple reporting platforms like Teradata, Oracle, SQL Server and HANA DB.
Loading summarized data back to RDBMS based on the data sets and granularity of data for real-time reports.
Monitoring the Oozie jobs to make sure the jobs are getting completed within SLA response and resolution times and ensure cases are regularly updated.
Managing the MIM calls when the major issues are happened and sharing the MOM notes to leadership teams.
Diligently teaming with the infrastructure, network, database, application and business intelligence teams to guarantee high data quality and availability.
Involving with various reporting teams for migrating data from the RDBMS/respective reporting DB’s data source to Hadoop distributed File System Environment and vice versa for analytics and reporting purpose based on retention policy.
Experience in commissioning and decommissioning of nodes, keeping track of jobs, monitoring critical alerts, configuring high availability, configuring data backups and data purge.
Handling Ganglia & Nagios alerts to handle CPU, Memory, Storage and Network.
Developing Automation scripts for process improvements as a part of automation process.
Creating the snapshot and increasing the Quota setting.
Data copy from one cluster to another or cluster using distcp utility.
Maintaining the root cause analysis & issues tracker sheet.
Worked in 24x7 environment for production support in an on-call rotation.
Excellent problem-solving skills with strong technical background and good interpersonal skills. Quick learner and excellent team player, ability to meet deadlines and work under pressure.
July 2017 - Present
Dell International Services Pvt Ltd, Bangalore
Nov 2014 – July 2017
NTT Data, Bangalore
Completed B.Tech from Jawaharlal Nehru Technological University, Anantapur.
Apache Hadoop, YARN,Hive, Sqoop, Pig, HBase, Horton Works
Informatica Power Center 8.X/9.X,SSIS
MySQL, Oracle, Teradata, HANA, SQL Server, Greenplum
Tableau, HANA BOBJ, Power BI, TOLAP Cube, SSRS
Ambari, Nagios, Informatica Monitor
Toad, Web I, Service Now, Intelligence Center
Oozie and Control M
Key Carrier Achievements
On the Spot Award in FY19Q1
Team award in FY19 Q3
Silver award in FY20 Q1
Project # 1:
Project : Business Management System (BMS)
Client : Dell Technologies Inc.
Environment : HDFS, Map Reduce HIVE, PIG, Sqoop,SAP DS, Informatica,TOLAP Cube,BOBJ, Oozie and Control M
Database : My Sql, HANA, Teradata, Oracle, SQL Server,
Duration : July-2017 to till date
BMS Reporting solution which enables a more flexible, faster global standard reporting. Simplifying navigation, analysis and drill down from GM to Order to Rep Level. Focuses on Reporting, Analytics, Data Stewardship, Business Analytics, and SFDC optimization to enable faster, more proactive and predictable decision making for CSO leadership and sales organizations.
Roles and Responsibility:
Hands on experience in installing, configuring Hadoop ecosystem components like Hadoop Map Reduce, HDFS, HBase, Oozie, Hive, Sqoop.
Expert in dealing with structured, semi-structured and unstructured data in Hadoop eco system.
Installation and configuration, Hadoop Cluster and Maintenance, Cluster Monitoring and Troubleshooting and Transform data from RDBMS to HDFS
Work along with the Service Providers to resolve the tickets that were raised by various business teams.
Monitoring list of jobs and if required kill or increase the priority of job.
Removing corrupted blocks and maintaining HDFS health.
Extensive knowledge of Business Intelligence / Data Warehousing solutions.
Running Hive jobs in Hadoop environment to process the data as per the client requirement and keeping it in Hadoop distributed file system and transferring data from HDFS to RDBMS for analytics and Adhoc reporting purpose through Hive Querying language.
Analyzing and providing support to different data sets like Orders, Quotes, Affinity, Trips, Call data, Zyme, Targets and SFDC(Salesforce).
Prevent reoccurring data quality issues and production failures and proposing mitigation plan to avoid reoccurring issues.
Analyze business and technical requirements, as well as data warehouse data models, to understand global detailed data management.
ETL design and development with technologies such as Informatica, SSIS
Clear communications to regional teams on the data quality issues and impact on daily data loads due to production environments downtime(if any).
Coordinates with multiple teams to resolve issues though incidents/tasks/bridge calls and invokes MIM (Major Incident Management) as needed.
Project # 2:
Project : US_Americre-Virt BIDAT_120101
Customer : GENERAL MOTORS FINANCIAL COMP (GMFC)
Technologies : Informatica 9X, Informatica Data Quality
Role : Developer
GM Financial is the captive finance company and a wholly-owned subsidiary of General Motors Company and is headquartered in Fort Worth; Texas.GM Financial is a global provider of automobile finance solutions, with operations in the U.S., Canada, Europe and Latin America. Through there long-standing relationships with auto dealers, offer attractive retail loan and lease programs to meet the Unique needs of each customer and offer commercial lending products to dealers to help them finance and grow their businesses. GM Financial having operations in 19 countries, providing auto finance solutions to approximately 16,000 dealers worldwide.
Roles and Responsibility:
Extensively used Informatica data quality to design multiple mappings based on business logic provided in the Design level document
Involved in code improving methods like Data cleansing and Data merging
Worked with various developer modules like profiling, standardization and matching.
Populated the data into target tables in an efficient manner.
Designed various mappings and Mapplets using different transformations such as key generator, match, labeler, case converter, standardize, parser, consolidation transformation.
Worked extensively with address validator to cleanse the address elements for country specified data.
Created data quality profiles and scorecards to identify and measure data quality using Informatica Analyst tool.
Creating Physical data objects for both flat files and oracle tables using informatica developer client.
Sending exception files back to source system for correction.
Creating mapping/Mapplets using Informatica developer to identify Data quality issues in the data
Involved in analysis, design, coding and testing.
Involved into source code review, the mappings created by my team members before sending to external review.
Prepared deployment document and assisted to deployment team while migration.
Used Workflow Manager for Creating, Validating, Testing and running the sequential and concurrent Sessions and scheduling them to run at specified time.
Creating Unit test cases for the Informatica mappings developed by other (In some cases).
Involved in Unit testing for the mappings developed by myself.
Developed different mappings by using different Transformations like Aggregator, Lookup, Expression, update Strategy, Joiner, Router etc to load the data into staging tables and then to target.
Involved in performance tuning by optimizing the sources, targets, mappings and sessions.
Used tools like Rapid SQL to run SQL queries and validate the data loaded into the target tables.
Involved in creating data quality verification and documented all ETL Development work done and Handled the conflicts raised.
Handle the new development CR from the onsite team. Clarify queries and complete the development. CR includes the specs and the test data will be provided once the development is complete. Once the QC is done on the present development, it is deployed to production.
Involved Actively in Table Load Delay Notifications to Business users.
Actively Involved In CAB Meeting.