Sign in

Data Developer

Vasant Nagar, Karnataka, India
September 26, 2019

Contact this candidate


Hari Babu Daddala



In pursuit of challenging assignments that would facilitate the maximum utilization and application of my broad skills and expertise in making a positive difference to the organization. Summary

Professional Summary

Result oriented professional with 4+ years of experience on Business Intelligence and Data Warehousing using ETL tool Informatica, SSIS, SAP DS (Data Services) and Hadoop Administration on Hortonworks Data Platform (HDP).

Having 3 years of experience on Hadoop administration and its eco system components YARN, MapReduce, Sqoop, Pig, Hive, HBase, Oozie, Zookeeper.

Responsible for syncing data to multiple reporting mini clusters (ponds) like sales, Inventory, SCP and Partner information for analytics.

Configured various property files like core-site.xml, hdfs-site.xml, mapred- site.xml and hadoop-env.xml based upon the job requirement.

Responsible for subscribing data to multiple reporting platforms like Teradata, Oracle, SQL Server and HANA DB.

Loading summarized data back to RDBMS based on the data sets and granularity of data for real-time reports.

Monitoring the Oozie jobs to make sure the jobs are getting completed within SLA response and resolution times and ensure cases are regularly updated.

Managing the MIM calls when the major issues are happened and sharing the MOM notes to leadership teams.

Diligently teaming with the infrastructure, network, database, application and business intelligence teams to guarantee high data quality and availability.

Involving with various reporting teams for migrating data from the RDBMS/respective reporting DB’s data source to Hadoop distributed File System Environment and vice versa for analytics and reporting purpose based on retention policy.

Experience in commissioning and decommissioning of nodes, keeping track of jobs, monitoring critical alerts, configuring high availability, configuring data backups and data purge.

Handling Ganglia & Nagios alerts to handle CPU, Memory, Storage and Network.

Developing Automation scripts for process improvements as a part of automation process.

Creating the snapshot and increasing the Quota setting.

Data copy from one cluster to another or cluster using distcp utility.

Maintaining the root cause analysis & issues tracker sheet.

Worked in 24x7 environment for production support in an on-call rotation.

Excellent problem-solving skills with strong technical background and good interpersonal skills. Quick learner and excellent team player, ability to meet deadlines and work under pressure.

Professional Expérience



July 2017 - Present

Dell International Services Pvt Ltd, Bangalore

Nov 2014 – July 2017

NTT Data, Bangalore

Academic Profile

Completed B.Tech from Jawaharlal Nehru Technological University, Anantapur.

Skills Set

Big Data

Apache Hadoop, YARN,Hive, Sqoop, Pig, HBase, Horton Works

ETL Tools

Informatica Power Center 8.X/9.X,SSIS

Operating Systems

Windows, Centos


MySQL, Oracle, Teradata, HANA, SQL Server, Greenplum

Reporting Tools

Tableau, HANA BOBJ, Power BI, TOLAP Cube, SSRS

Monitoring Tools

Ambari, Nagios, Informatica Monitor

Supporting Tools

Toad, Web I, Service Now, Intelligence Center

Scheduling Tools

Oozie and Control M

Key Carrier Achievements

On the Spot Award in FY19Q1

Team award in FY19 Q3

Silver award in FY20 Q1

Projects Undertaken

Project # 1:

Project : Business Management System (BMS)

Client : Dell Technologies Inc.

Environment : HDFS, Map Reduce HIVE, PIG, Sqoop,SAP DS, Informatica,TOLAP Cube,BOBJ, Oozie and Control M

Database : My Sql, HANA, Teradata, Oracle, SQL Server,

Duration : July-2017 to till date

Project Description:

BMS Reporting solution which enables a more flexible, faster global standard reporting. Simplifying navigation, analysis and drill down from GM to Order to Rep Level. Focuses on Reporting, Analytics, Data Stewardship, Business Analytics, and SFDC optimization to enable faster, more proactive and predictable decision making for CSO leadership and sales organizations.

Roles and Responsibility:

Hands on experience in installing, configuring Hadoop ecosystem components like Hadoop Map Reduce, HDFS, HBase, Oozie, Hive, Sqoop.

Expert in dealing with structured, semi-structured and unstructured data in Hadoop eco system.

Installation and configuration, Hadoop Cluster and Maintenance, Cluster Monitoring and Troubleshooting and Transform data from RDBMS to HDFS

Work along with the Service Providers to resolve the tickets that were raised by various business teams.

Monitoring list of jobs and if required kill or increase the priority of job.

Removing corrupted blocks and maintaining HDFS health.

Extensive knowledge of Business Intelligence / Data Warehousing solutions.

Running Hive jobs in Hadoop environment to process the data as per the client requirement and keeping it in Hadoop distributed file system and transferring data from HDFS to RDBMS for analytics and Adhoc reporting purpose through Hive Querying language.

Analyzing and providing support to different data sets like Orders, Quotes, Affinity, Trips, Call data, Zyme, Targets and SFDC(Salesforce).

Prevent reoccurring data quality issues and production failures and proposing mitigation plan to avoid reoccurring issues.

Analyze business and technical requirements, as well as data warehouse data models, to understand global detailed data management.

ETL design and development with technologies such as Informatica, SSIS

Clear communications to regional teams on the data quality issues and impact on daily data loads due to production environments downtime(if any).

Coordinates with multiple teams to resolve issues though incidents/tasks/bridge calls and invokes MIM (Major Incident Management) as needed.

Project # 2:

Project #1

Project : US_Americre-Virt BIDAT_120101


Technologies : Informatica 9X, Informatica Data Quality

Role : Developer

Project Description:

GM Financial is the captive finance company and a wholly-owned subsidiary of General Motors Company and is headquartered in Fort Worth; Texas.GM Financial is a global provider of automobile finance solutions, with operations in the U.S., Canada, Europe and Latin America. Through there long-standing relationships with auto dealers, offer attractive retail loan and lease programs to meet the Unique needs of each customer and offer commercial lending products to dealers to help them finance and grow their businesses. GM Financial having operations in 19 countries, providing auto finance solutions to approximately 16,000 dealers worldwide.

Roles and Responsibility:

Extensively used Informatica data quality to design multiple mappings based on business logic provided in the Design level document

Involved in code improving methods like Data cleansing and Data merging

Worked with various developer modules like profiling, standardization and matching.

Populated the data into target tables in an efficient manner.

Designed various mappings and Mapplets using different transformations such as key generator, match, labeler, case converter, standardize, parser, consolidation transformation.

Worked extensively with address validator to cleanse the address elements for country specified data.

Created data quality profiles and scorecards to identify and measure data quality using Informatica Analyst tool.

Creating Physical data objects for both flat files and oracle tables using informatica developer client.

Sending exception files back to source system for correction.

Creating mapping/Mapplets using Informatica developer to identify Data quality issues in the data

Involved in analysis, design, coding and testing.

Involved into source code review, the mappings created by my team members before sending to external review.

Prepared deployment document and assisted to deployment team while migration.

Used Workflow Manager for Creating, Validating, Testing and running the sequential and concurrent Sessions and scheduling them to run at specified time.

Creating Unit test cases for the Informatica mappings developed by other (In some cases).

Involved in Unit testing for the mappings developed by myself.

Developed different mappings by using different Transformations like Aggregator, Lookup, Expression, update Strategy, Joiner, Router etc to load the data into staging tables and then to target.

Involved in performance tuning by optimizing the sources, targets, mappings and sessions.

Used tools like Rapid SQL to run SQL queries and validate the data loaded into the target tables.

Involved in creating data quality verification and documented all ETL Development work done and Handled the conflicts raised.

Handle the new development CR from the onsite team. Clarify queries and complete the development. CR includes the specs and the test data will be provided once the development is complete. Once the QC is done on the present development, it is deployed to production.

Involved Actively in Table Load Delay Notifications to Business users.

Actively Involved In CAB Meeting.

Contact this candidate