Post Job Free

Resume

Sign in

Data Engineer

Location:
Plano, TX
Posted:
February 22, 2020

Contact this candidate

Resume:

Biswajit Pan

Big Data Engineer

Mobile: +1-201-***-**** E-Mail: adbw08@r.postjobfree.com

Professional Summary

Over 8+ years of development experience applications using AWS Cloud, Java/J2EE, Hadoop and Big Data technologies in Banking, Insurance domains.

Proficient in Java, J2EE, Spring, Kafka, Spark, Scala, Python, Hive, Snowflake, Sqoop, REST, PL/SQL, Jenkins, jQuery, GitHub, Snowflake, Shell, Hibernate, Oracle, Ant, Maven, JIRA, Log4J, GIT, HTML, JavaScript, JSON, JMS, XML, JNDI, XSLT.

Software development experience using Big Data technologies to efficiently solve Big Data processing requirement.

Application Development using AWS Suite like S3, EMR, EC2, RDS, Lambda’s, SNS etc.

Good understanding in Hadoop eco system components HDFS, Map Reduce, Yarn, Sqoop, Spark, Spark SQL, Spark Streaming, and Hive for scalability, distributed computing, and high- performance computing.

Experience in using Hive/SparkSql Query Language for data Analytics.

Active contributor in Apache Spark/Hadoop user forums and proficient in developing custom frameworks.

Implemented Spark using Scala and utilizing Spark Core, Spark SQL API for faster processing of data.

Used Spark-SQL to Load various kind of data and create Schema RDD and loaded it into Hive Tables and handled Structured data using Spark SQL.

Developing Spark programs using Scala and python API's to analysis of data

Used Spark API over Cloudera Hadoop YARN to perform analytics on data in Hive.

Collected and aggregated large amounts of data from RDBMS and stored the data into HDFS for analysis.

Generated the datasets and loaded to HADOOP Ecosystem.

Involved in creating Hive tables, loading with data and writing hive queries that will run internally in MapReduce way.

Worked with various HDFS file formats like Avro, parquet File and various compression formats like Snappy, gzip.

Involved in converting Hive/SQL queries into Spark transformations using Spark RDDs, Scala and Python.

Worked in all phases of SDLC, DevOps and Agile models - Played Scrum Lead and shadow Scrum Master roles.

Expertise in technically leading Migrations to Open Source Technologies

Ability in technically leading Java/J2EE, Spark, Bigdata developers and teams.

5+ years of Onsite experience working directly with clients and handling onsite & offshore teams effectively.

Capable in initiating POCs, Project estimations, Hardware provisioning, release and configuration management

Highly motivated, results oriented and leadership skills with great team ethics.

Good in mentoring novice programmers and evaluating team performance.

Have the ability to take corrective and preventive actions in delivering desired quality.

Very strong analytical and reverse engineering skills part of technology migrations.

Experienced in working with geographically distributed development teams.

Hands on Experience in developing web applications by using Hibernate and Spring Frameworks with Apache Tomcat.

Expertise in developing data models using Hibernate Object relational mapping (ORM) framework and used Hibernate Annotations to reduce time at the configuration level.

Worked with design and development of robust and highly scalable Web Service platforms and API’s based on REST and SOAP for enterprise applications using AXIS, WSDL, UDDI, XML, JAX- WS, JAX-RS, JAXB, JAX-RPC technologies and Jersey frameworks. Technical Skills

Big Data Ecosystems : Spark streaming, Spark, Sqoop, Hive, Flume, Hadoop, MapReduce, HDFS, Pig, Kafka

Web Services : RESTful Web Services & SOAP

Languages : Java/j2ee, Scala, Python, JavaScript, Servlet, PL/SQL, Unix Shell

Web Technologies : HTML, JSP, JSF, CSS, JavaScript, Python, AJAX Server-Side Frameworks : Spring, Hibernate, Struts 2 IDEs : IntelliJ IDEA, Eclipse, RAD

Build Tools : Maven, Ant

Web Servers /App Servers : Apache Tomcat 6.0/7.0, IBM WebSphere 6.0/7.0, JBoss 4.3 Static Code Analysis : Find Bug/Check Style/PMD

Configuration Tool : GIT-Bit bucket, SVN, CVS, SOAPUI Database : Snowflakes, MongoDB, DynamoDB, Cassandra, HDFS, Oracle 11g,

Defects Triage : Quality Center 9.2.

Cloud Environment : AWS

Academic And Certificate Information

Education University

Degree : Bachelor of Technology (Computer

Science) from WBUT University, India

Certifications : - AWS Certified Solutions Architect.

- Big data expert

- Sun Certified Java Programmer.

Employment History

I am currently working in CapitalOne as Data Engineer from 2018.

I previously worked with Liberty Mutual, AAA Insurance. Company Projects Undertaken

Project: Data Express

Details: DataExpress is a framework to build Streaming applications. The tool helps fetching data from Kafka cluster, save it to AWS S3 for audit purpose and later validates and writes it to Snowflake database. To setup the configuration, one can create this configuration file through a self- service portal and pass on this configuration file provide the same during the deployment. Technology: Kafka, Spark Streaming, Scala, AWS, Java, Rest webservice Client: CapitalOne

Tools: EMR, AWS

Role: Developer and Team Lead

Team Size: 10

Duration: Jan, 2018 – Till date

Accountabilities:

Involved Design of the framework

Involved reading data from Kafka using Spark streaming

Writing data in different format in Amazon - S3(Avro, Parquet)

Writing in Snowflake Database

Design Exception handling for this frame work

Design to create auto Incident Request and Amazon SNS

Design Kafka Failover

Validation of the Schema for incoming data.

Title Customer first(CF)

Technology: Spark, Scala, Hive, Sqoop, AWS, Java, Rest webservice Client Liberty Mutual

Team Size 12

Role Design, Analysis, Detail Design & Developer

Period April 2017 – December 2017

Project Description

Customer first(CF) applications comprised of 3 different application which includes Customer first property(CFP), Customer first Auto (CFA) & CF Personal Liability Protection(CP). While individual modules can be loaded independently of each other, every module shares a set of core code which must present to successfully run CF from a development slice. These applications collectively referred as CF and these applications are used by Liberty Mutual sales staff to get insurance quotes, policy and Policy Endorsement etc. All the applications are in production and the business is rolling out new functions to the application state wise.

Team is working in agile methodology and the new functionalities will be described in Epics and assigned to the team.

Role and Responsibilities

Involved in architecture discussions and requirement analysis.

Involved in discussions with business line.

Manual Code Review

Developed the CF related service.

Title AAA Insurance Home Products

Technology: Java, J2EE, Spring, Hibernate, Soap, Rest Client AAA

Team Size 55

Role Design, Analysis, Detail Design & Developer

Period April 2012 – April 2017

Project Description

This project is a product enhancement and customization based on client specifications and requirements for rolling out a system to support foot print and partner club states. The core base service functionality is provided by Exigen product factory runtime and it is based on policy core common requirements. Creation of New custom products on the top of the policy core system by extending baseline services which includes 3 steps process (1) Product Creation with the help of independent loosely coupled components and respective service configurations, (2) Product Testing and (3) Policy Creation.

Role and Responsibilities

Interacting with the Product Managers and stakeholders at AAA NCNU to better understand their business needs and provide good solutions to them on the same.

Understanding the Business requirements and configure, customize the product as per the requirements document.

Preparing custom component lifecycle diagrams, service design documents for each custom components, UML class diagrams and sequence diagrams as per as the client requirements and specifications

Development of components/artifacts that implements Component interface contract, versioning of the components, and creation of the new business services by extending core policy core services provided by Exigen product factory

Responsible for packaging of the components into libraries, adding/removing attributes, changing attribute labels, configuration of component constraints, component deployments

Personal Details

Present City : Plano, Texas

Languages Known : English.

Visa : Approved (I-140 approved)



Contact this candidate