Resume

Sign in

Engineer Data

Location:
Hyderabad, Telangana, India
Posted:
September 25, 2020

Contact this candidate

Resume:

Sekhar Talupula Email:adgeo4@r.postjobfree.com Phone: +919*********

PROFESSIONAL SUMMARY:-

Software Engineer with 5 year 8 months of experience in IT industry, having hands-on knowledge in all aspects of the software development life cycle like requirements analysis, design, development and testing.

Worked as a big data developer for around 4 years in Banking COE (Centre of Excellence), Health domains and having good expertise in Big data tools like Hadoop, Hive, Sqoop, Yarn, Spark Core and Data Frames, Datasets, Spark Streaming, Kafka, Scala, Java. Worked as ETL developer for 1.6 years in the beginning of career and have experience with tool like Informatica, Pentaho, Oracle Sql .

TECHNICAL SKILLS:-

OS

WINDOWS & UNIX

Programming

Java, Scala, Python, HTML, CSS, JavaScript

Database

Oracle 11g

Big Data tools

Hdfs, MapReduce, Scala, oozie

Hive, Hbase, Sqoop, Spark Core

Apache Spark, Apache Sql (DataFrames, DataSets)

Spark Streaming, Kafka, Shell scripting

ETL Tools

Informatica PowerCenter 9.6.1,

Pentaho 6.0

Reporting tool

Business Object

Core Competencies: -

●Developed solutions using Bigdata tools like HDFS, HIVE, Sqoop, HBase, Scala, Java,Apache Spark Core, Spark SQL (Data Frames, DataSets),Spark Streaming, Kafka.

●Developed solutions using ETL tools Informatica PowerCenter 9.5.1, Pentaho 6.0, Informatica Developer 9.5.1, for various requirements from BANKING DOMAIN TEAM.

●Good knowledge in writing different types of command in SQL (DDL, DML Commands), PLSQL.

●Have worked in the Oracle 11G using SQL developer.

●Worked in Agile team and having knowledge in git, jenkins, sonar.

●Knowledge on Docker, Kubernetes, AWS

●Have been trained in Business Object B1 tool to develop various reports and dashboards for analysis purpose.

●Testing: Have sound knowledge to perform Unit Testing, System Testing.

PROFESSIONAL EXPERIENCE: -

Accenture (May 2019 – June2020)

Tata Consultancy Services, Chennai (NOV 2014 – April 2019)

PROJECT

Datalake Team

CLIENT

ESI

EXPERIENCE

12 months

DESIGNATION

Systems Engineer

ROLE

Developer

TOOLS USED

Hadoop, Hive, Spark core, Scala, Data frame, Spark sql,Sqoop, Java, Hbase, Shell scripting, Spark Streaming, Kafka

SUMMARY: -

This project contains claims, entity business data. The size of data is in terabytes. Claims involve moving data from Teradata to Hive transactional tables using Teradata import tool and applying SCD on those tables and entities involve moving json data from Kafka to HDFS, HBase using spark batch, spark streaming jobs respectively.

Entity data pulled from Kafka topic is having both commercial and non-commercial data, to separate Json data to corresponding tables the Json messages are parsed with spark UDF and after performing necessary operations they are written to corresponding HDFS Location an HBase tables.

Roles & Responsibilities: -

●Developed data ingestion jobs to load from Teradata to hdfs.

●Designed Hive repository with external tables, buckets, partitions and ORC compressions

●Performed data processing, quality analysis using spark core, dataframes,scala and stored the data into storage layer Hive.

●Loading data from kafka to hdfs as well as Hbase tables using spark streaming.

●Implemented kafka offset management in Hbase

●Performed data processing using Spark

●Implemented UDF for data processing

●Involved in performance Tuning

●Used cron jobs and oozie jobs for job scheduling.

PROJECT

Bigdata Team

CLIENT

Westpac

EXPERIENCE

2 years

DESIGNATION

Systems Engineer

ROLE

Developer

TOOLS USED

Hadoop, Hive, Spark core, Scala,

Data frame, Spark sql, Sqoop, Java,

Hbase,Shell scripting

SUMMARY: -

This project involves copying data from oracle to hive tables using Sqoop, performing operations on those tables using spark and exposing them as hive those tables to ML team. It also involves extracting zipped csv files in hdfs, performing operations using spark and exposing them as hive tables to AI team

Roles & Responsibilities: -

●Developed Sqoop jobs with incremental load from heterogeneous RDBMS using native dB connectors into HDFS.

●Performed change data capture in hive using sqoop jobs

●Unzipping csv files using java code

●Designed Hive repository with external tables, buckets, partitions and ORC compressions

●Performed data processing, quality analysis using spark core, spark dataframes, scala and stored the data into storage layer Hive which is used by BI team for reporting.

PROJECTS

CounterParty Credit Risk Management (CCR),

Interest Rate Risk in Banking Book (IRRBB),

Fundamental Review of Trading Book (FRTB)

CLIENT

Banking Domain Team

EXPERIENCE

2 years

DESIGNATION

Systems Engineer

ROLE

Developer

TOOLS USED

Hadoop, Hive, Spark core, Scala, Data frame, Spark sql, Sqoop, Hbase, Pentaho ETL tool

SUMMARY:

Base Version of these projects involve moving data present in CSV files and RDBMS tables to Hive staging tables, performing calculations using Hive sql and then loading data to Hive tables. Advanced verion of these projects involve reading CSV files using Spark core and Scala, separating detail records and creating data frame out of it and performing data frame and spark sql operations and loading data to hive tables, which are used by reporting tool for creating dashboards. Used Pentaho ETL tool to interact with HDFS and Hive in base version. Have good knowledge in testing developed application.

Roles & Responsibilities: -

●Developed Sqoop jobs with incremental load from heterogeneous RDBMS using native dB connectors into HDFS.

●Designed Hive repository with external tables, buckets, partitions and ORC compressions

●Created dataframes on top of csv files, hive tables using spark dataframes

●Performed data processing and quality analysis using spark and storing the data into storage layer Hive which is used by BI team for reporting.

PROJECTS

Generating data to project (RVP),

Extracting Email content from sequence of conversations,

Small Medium Business Advisory (SMB)

CLIENT

Banking Domain Team

EXPERIENCE

1 year

DESIGNATION

Systems Engineer

ROLE

Developer

SUMMARY: -

Developed above solutions, SMB using informatica power centre and informatica developer, data generation for RVP using Pentaho ETL tool and Email data extraction using Perl programming language.

Roles & Responsibilities: -

●Implemented SMB using informatica data quality which determines the feasibility of business using provided data.

●Developed data using ETL tool pentaho for RVP which identifies potential customers for banking.

●Extracted email body from mail chain using Perl code.

ACADEMIC QUALIFICATIONS :-

Course

Name of the Institution

Board/ university

Year of completion

Score

BTECH (ECM)

Prasad V Potluri Sidhartha Institute of Technoloy

JNTUK, AP

2014

67%

Intermediate (10+2) (MPC)

Siddhartha Jr.College

Board of Intermediate Education

2009

91%

SSC

Loyola English Medium School

Board of Secondary Education

2007

89%

PERSONAL INFORMATION:-

Date of Birth : 06-09-1991

Gender : Male

Nationality : Indian

Languages : Telugu, English

Address : 2-36, Srikakulam,Gantasala Mandal, Krishna Dist, AP,521132



Contact this candidate