Post Job Free

Resume

Sign in

Senior big data engineer

Location:
Chennai, Tamil Nadu, India
Posted:
April 17, 2021

Contact this candidate

Resume:

Sathish Loganathan

adlri3@r.postjobfree.com 984**-***** Chennai

Summary

Experienced Big Data Engineer having knowledge on Financial services and Telecom domain.

8+ years of professional IT experience, which includes 3+ years as Spark/Hadoop developer and 6 years as Teradata ETL developer.

Efficient in building ETL pipelines with Apache Spark.

Hands on experience in creating streaming jobs with Kafka and Spark.

Hands on experience in Hive and Sqoop for batch processing.

Experience in AWS Cloud platform (EMR, Lambda and S3).

Experience in Spark and Hive performance tuning.

Hands on experience in Delta Lake.

Good knowledge on the various file formats and compression techniques.

Ability to adapt to evolving technology, strong sense of responsibility and accomplishment. Technical Skills

Big Data Ecosystem : Apache Spark, Hive, Sqoop, Phoenix, HDFS, YARN and Delta Lake Streaming : Kafka and Spark Streaming

Programming Language : Scala and Python

Scripting Language : Shell Scripting

Querying Language : Teradata SQL and Oracle SQL

Database : Teradata and Oracle

NoSQL Database : Apache HBase

Hadoop Distribution : Cloudera

Change Management : JIRA and Azure Devops

IDE : Pycharm, Intellij Idea and Eclipse Scala IDE Build : Jenkins

Version Control : GitHub, Bitbucket and Tortoise CVS AWS Micro Service : Lambda, ElasticMapReduce (EMR) and S3

(Learning knowledge on Athena and Dynamo DB)

Work Experience

Employer Role Experience

Cognizant Senior Associate - Big Data Engineer Jan 2019 - Present Verizon India Senior Analyst - Data Engineer Aug 2012 - Dec 2018 August 2020 – Present

Role: Senior Associate Senior Big Data Engineer

Client and Employer : Charter Communications Cognizant Tech Stack : Spark Scala Kafka Sqoop Hive HBase Phoenix Oracle Unix Project : Voice Of Customer(VOC)

Project Summary: The purpose of this project is to process the customer feedback about the products. And also receive the billing information from different billing sources. Based on the billing and customer feedback data, we generate the report for the business team. Based on this report, the business team will take the key decisions on understanding and analyzing the customer purchase, usage patterns and make improvements to the process. Roles & Responsibilities

Creating Spark-Kafka consumer streaming jobs and store the results in Hbase.

Creating Sqoop jobs to import data from Oracle source and load the results to hive external table.

Create Spark jobs to do the transformations on hive tables and store the results in Hive.

Perform performance tuning on both Spark jobs and Hive.

Automated the Hive table creation process with shell script which resulted in creating 1000+ Hive tables and saved lot of effort.

Preparing the design document for the project.

Implementing the data pipeline through Tidal scheduler by creating Tidal jobs.

Leading the team into Big data domain and giving technical support to the team.

Participate in code review and bug fixing for improving the performance.

Perform Unit testing and make sure it matches with the requirement. Jan 2019 – July 2020

Role: Senior Associate Senior Big Data Engineer

Client and Employer : S&P Global Cognizant

Tech Stack : Spark Python Kafka Hive AWS EMR Lambda S3 Oracle Unix Project Summary: S&P is deriving index for many stock exchanges across the world. In this project we receives all the traded stock information across the world and we are doing ETL processing. S&P has more than 100 years history data of a stock. With this data we are calculating stock returns and performing analytics. Roles & Responsibilities

Conversion of ETL jobs into Pyspark on the AWS EMR environment.

Performed performance tuning and achieved scalability with AWS auto scaling.

Performed performance tuning and made most of the transformations executing in- memory rather than disk to achieve more efficiency.

Used Spark to import and export data among HDFS, S3, Oracle database and Hive.

Used Spark+ Kafka streaming and created JSON file from Kafka message.

Made use of AWS Lambda to process the JSON objects in the S3 buckets to trigger Spark jobs.

Involved in preparing technical design/specifications for data extraction, transformation and loading.

Involved in code review and bug fixing for improving the performance.

Unit testing of the Spark job in the local environment to confirm the expected output.

Debugging the YARN/Spark logs to see the cause of issue and fix the same.

Providing technical solution to Big data team.

Created Kafka producer and consumer jobs.

Aug 2012– Dec 2018

Role: Senior Analyst - Data Engineer

Employer and Client : Verizon Verizon

Tech Stack : Teradata Unix Data stage Sqoop Hive Role summary : Started as fresher and worked 6.5 years in Verizon as Data Engineer. During this time, handled many ETL projects under Revenue Accounting domain. Project summary: The purpose of Verizon business data warehouse is to collect, harvest and distribute invoiced billing data, account receivables, SAP journal entries and supporting reference data, which is necessary for downstream communities (Finance, Sales & Service and marketing ). BODS is the single source for all revenue data. Roles & Responsibilities

Experience with development, testing, debugging, implementation and documentation.

Worked on optimizing and tuning the Teradata views and SQL’s to improve the performance of batch and response time of data for users.

Prepare Detailed Design document for the project from HLD and give walk-through of the DD document to business team, support team and to development team.

Providing technical solution to team and worked as technical lead in some deliverables.

Ensuring client deliverables within committed timelines.

Automated multiple test scripts to reduce the manual effort by using the PMF. Much interested in Automation.

Worked on Teradata to Hadoop migration in the initial stage of the project. Certifications

Teradata 12 Basics - TE0 121

Teradata 12 SQL - TE0 122

Teradata Physical design and Implementation: TE0 123. Trainings

5 days training on Apache SPARK arranged by Verizon.

5 days training on Hadoop arranged by Verizon.

AWS Solution Architect online course by Udemy.

Achievements

Received 'Trending award’ from Verizon CIO for project deliverables and automation works.

Received ‘Spot Light award’ from VP of Verizon Enterprise Solutions for automation.

Received ‘Spot Light award’ from Delivery head for automation. Academic Profile

Degree / Course Name of the institution/ University Year of passing Percentage

obtained

B.Tech – Information

Technology

Karpagam College of Engineering (Affiliated

to Anna University), Coimbatore.

2008 – 2012,

86.9%

Class XII St. Antony’s HSS, Krishnagiri. March 2008 81.75% Class X St. Antony’s HSS, Krishnagiri. March 2006 76.4% Personal Information

Name : Sathish Kumar.L

Date of birth : 10 - 06 - 1991

Gender : Male

Father’s name : Loganathan.S

Language Known : English, Tamil, Telugu

Residential address : No 12, Elango Street, Adambakkam, Chennai. Declaration:

I hereby declare that the above stated information is true and complete to the best of my knowledge and belief.

Place : Chennai Yours Sincerely,

Date : Sathish Kumar.L



Contact this candidate