Sekhar Talupula Email:******.********@*****.*** Phone: +919*********
PROFESSIONAL SUMMARY:-
Software Engineer with 5 year 8 months of experience in IT industry, having hands-on knowledge in all aspects of the software development life cycle like requirements analysis, design, development and testing.
Worked as a big data developer for around 4 years in Banking COE (Centre of Excellence), Health domains and having good expertise in Big data tools like Hadoop, Hive, Sqoop, Yarn, Spark Core and Data Frames, Datasets, Spark Streaming, Kafka, Scala, Java. Worked as ETL developer for 1.6 years in the beginning of career and have experience with tool like Informatica, Pentaho, Oracle Sql .
TECHNICAL SKILLS:-
OS
WINDOWS & UNIX
Programming
Java, Scala, Python, HTML, CSS, JavaScript
Database
Oracle 11g
Big Data tools
Hdfs, MapReduce, Scala, oozie
Hive, Hbase, Sqoop, Spark Core
Apache Spark, Apache Sql (DataFrames, DataSets)
Spark Streaming, Kafka, Shell scripting
ETL Tools
Informatica PowerCenter 9.6.1,
Pentaho 6.0
Reporting tool
Business Object
Core Competencies: -
●Developed solutions using Bigdata tools like HDFS, HIVE, Sqoop, HBase, Scala, Java,Apache Spark Core, Spark SQL (Data Frames, DataSets),Spark Streaming, Kafka.
●Developed solutions using ETL tools Informatica PowerCenter 9.5.1, Pentaho 6.0, Informatica Developer 9.5.1, for various requirements from BANKING DOMAIN TEAM.
●Good knowledge in writing different types of command in SQL (DDL, DML Commands), PLSQL.
●Have worked in the Oracle 11G using SQL developer.
●Worked in Agile team and having knowledge in git, jenkins, sonar.
●Knowledge on Docker, Kubernetes, AWS
●Have been trained in Business Object B1 tool to develop various reports and dashboards for analysis purpose.
●Testing: Have sound knowledge to perform Unit Testing, System Testing.
PROFESSIONAL EXPERIENCE: -
Accenture (May 2019 – June2020)
Tata Consultancy Services, Chennai (NOV 2014 – April 2019)
PROJECT
Datalake Team
CLIENT
ESI
EXPERIENCE
12 months
DESIGNATION
Systems Engineer
ROLE
Developer
TOOLS USED
Hadoop, Hive, Spark core, Scala, Data frame, Spark sql,Sqoop, Java, Hbase, Shell scripting, Spark Streaming, Kafka
SUMMARY: -
This project contains claims, entity business data. The size of data is in terabytes. Claims involve moving data from Teradata to Hive transactional tables using Teradata import tool and applying SCD on those tables and entities involve moving json data from Kafka to HDFS, HBase using spark batch, spark streaming jobs respectively.
Entity data pulled from Kafka topic is having both commercial and non-commercial data, to separate Json data to corresponding tables the Json messages are parsed with spark UDF and after performing necessary operations they are written to corresponding HDFS Location an HBase tables.
Roles & Responsibilities: -
●Developed data ingestion jobs to load from Teradata to hdfs.
●Designed Hive repository with external tables, buckets, partitions and ORC compressions
●Performed data processing, quality analysis using spark core, dataframes,scala and stored the data into storage layer Hive.
●Loading data from kafka to hdfs as well as Hbase tables using spark streaming.
●Implemented kafka offset management in Hbase
●Performed data processing using Spark
●Implemented UDF for data processing
●Involved in performance Tuning
●Used cron jobs and oozie jobs for job scheduling.
PROJECT
Bigdata Team
CLIENT
Westpac
EXPERIENCE
2 years
DESIGNATION
Systems Engineer
ROLE
Developer
TOOLS USED
Hadoop, Hive, Spark core, Scala,
Data frame, Spark sql, Sqoop, Java,
Hbase,Shell scripting
SUMMARY: -
This project involves copying data from oracle to hive tables using Sqoop, performing operations on those tables using spark and exposing them as hive those tables to ML team. It also involves extracting zipped csv files in hdfs, performing operations using spark and exposing them as hive tables to AI team
Roles & Responsibilities: -
●Developed Sqoop jobs with incremental load from heterogeneous RDBMS using native dB connectors into HDFS.
●Performed change data capture in hive using sqoop jobs
●Unzipping csv files using java code
●Designed Hive repository with external tables, buckets, partitions and ORC compressions
●Performed data processing, quality analysis using spark core, spark dataframes, scala and stored the data into storage layer Hive which is used by BI team for reporting.
PROJECTS
CounterParty Credit Risk Management (CCR),
Interest Rate Risk in Banking Book (IRRBB),
Fundamental Review of Trading Book (FRTB)
CLIENT
Banking Domain Team
EXPERIENCE
2 years
DESIGNATION
Systems Engineer
ROLE
Developer
TOOLS USED
Hadoop, Hive, Spark core, Scala, Data frame, Spark sql, Sqoop, Hbase, Pentaho ETL tool
SUMMARY:
Base Version of these projects involve moving data present in CSV files and RDBMS tables to Hive staging tables, performing calculations using Hive sql and then loading data to Hive tables. Advanced verion of these projects involve reading CSV files using Spark core and Scala, separating detail records and creating data frame out of it and performing data frame and spark sql operations and loading data to hive tables, which are used by reporting tool for creating dashboards. Used Pentaho ETL tool to interact with HDFS and Hive in base version. Have good knowledge in testing developed application.
Roles & Responsibilities: -
●Developed Sqoop jobs with incremental load from heterogeneous RDBMS using native dB connectors into HDFS.
●Designed Hive repository with external tables, buckets, partitions and ORC compressions
●Created dataframes on top of csv files, hive tables using spark dataframes
●Performed data processing and quality analysis using spark and storing the data into storage layer Hive which is used by BI team for reporting.
PROJECTS
Generating data to project (RVP),
Extracting Email content from sequence of conversations,
Small Medium Business Advisory (SMB)
CLIENT
Banking Domain Team
EXPERIENCE
1 year
DESIGNATION
Systems Engineer
ROLE
Developer
SUMMARY: -
Developed above solutions, SMB using informatica power centre and informatica developer, data generation for RVP using Pentaho ETL tool and Email data extraction using Perl programming language.
Roles & Responsibilities: -
●Implemented SMB using informatica data quality which determines the feasibility of business using provided data.
●Developed data using ETL tool pentaho for RVP which identifies potential customers for banking.
●Extracted email body from mail chain using Perl code.
ACADEMIC QUALIFICATIONS :-
Course
Name of the Institution
Board/ university
Year of completion
Score
BTECH (ECM)
Prasad V Potluri Sidhartha Institute of Technoloy
JNTUK, AP
2014
67%
Intermediate (10+2) (MPC)
Siddhartha Jr.College
Board of Intermediate Education
2009
91%
SSC
Loyola English Medium School
Board of Secondary Education
2007
89%
PERSONAL INFORMATION:-
Date of Birth : 06-09-1991
Gender : Male
Nationality : Indian
Languages : Telugu, English
Address : 2-36, Srikakulam,Gantasala Mandal, Krishna Dist, AP,521132