Post Job Free

Resume

Sign in

Data Tableau

Location:
Englewood, CO, 80112
Posted:
March 29, 2020

Contact this candidate

Resume:

Janani Harikrishnan

Mail: adci24@r.postjobfree.com

Ph: 720-***-****

Denver, Colorado

PROFESSIONAL SUMMARY

Accomplished Module Lead, skilled in reengineering, developing and implementing digital strategies in alignment with business objectives. Particularly interested in Hadoop and challenges around it. Has customer relationship experience and managing offshore resources. Excellent oral and written communication skills.

• Exposure in design and development of solutions for Big Data using the Hadoop eco system technologies (HDFS, Hive, Scala, Impala, Sqoop, Map Reduce, Apache Spark) Worked for reputed financial service companies The Bank of New York Mellon and Financial Conduct Authority and rewarded with outstanding performer during the year 2014, 2018 and 2019. CORE COMPETENCIES:

Hadoop Developer with 5+ years of IT experience and enthusiast to work in an environment which hone my skills and knowledge.

Around 3.5 years in Big Data Ecosystem capable of processing large sets of structured, semi structured and unstructured data.

Hands on experience in working with Ecosystems consisting of HDFS, Map Reduce, Hive, Sqoop, Oozie, Pig, Spark, Scala.

Good knowledge of Financial Services/Banking domain

Good understanding of distributed systems, HDFS architecture, Internal working details of MapReduce and Spark processing frameworks.

Experienced in Big data solutions and Hadoop ecosystem related technologies. Well versed with Big Data solution planning, designing, development and POC's.

Involved in converting Hive/SQL queries into Spark transformations using Spark Data frames and Scala.

Used Spark-SQL to perform transformations and actions on data residing in Hive.

Expertise in Inbound and Outbound (importing/exporting) data form/to traditional RDBMS using Apache SQOOP.

Worked with different File Formats like Textfile, Sequence File, Avro File, ORC and Parquet for Hive querying and processing.

Created and worked Sqoop jobs with incremental load to populate Hive External tables.

Hands-on experience in using Hive partitioning, bucketing and execute different types of joins on Hive tables.

Good exposure to performance tuning hive queries and map-reduce jobs in spark framework.

Great experience in Spark framework by creating RDD’s, DataFrames, Datasets and mapping with key/value pairs using Scala.

Knowledge in job workflow scheduling and monitoring tools like Oozie and Zookeeper.

Delivering report on occupancy using Excel Pivot Table, VLOOKUP & Slicer

Hands on experience in Java, Scala and python while working in production environment and some academic projects

Tuned PIG and HIVE scripts by understanding the joins, group and aggregation between them.

Hands on Experience in database Oracle 9i/10g, SQL Server 2005/2008, Sybase IQ, DB2 and Hadoop.

Hands on experience in Hadoop with Tableau using kerbose mechanism, Rserve integration.

Involved in Architecture, installation, configuration, and deployment of Tableau and troubleshooting.

Extensive experience in Tableau Desktop, Tableau Server and Tableau Reader in various versions of Tableau 8.2,9.3 & 10.1, and in various reporting objects like Facts, Attributes, Transformations, filters, prompts, Calculated fields, Sets, Groups, Parameters in Tableau.

Created Tableau scorecards, dashboards using stack bars, bar graphs, scattered plots, geographical maps, Gantt charts using the functionality.

Involved in creating dashboards and reports in Tableau and maintaining user and server activities.

Experience in using tools / IDEs spanning across Eclipse IDE, SVN, GIT, Jenkins, Remedy and JIRA

Knowledge of Core Java, SDKs and scripting (PowerShell, Batch and Shell Script).

Used various Project Management services like JIRA for tracking issues, bugs related to code and GitHub for various code reviews and worked on various version control tools like GIT, SVN

Scholastics include Bachelor of Technology in Information technology SKILLSET

Big Data Ecosystems Hadoop, HDFS, Hive, Pig, Scala, Sqoop, Oozie, Map Reduce Databases Oracle 9i/10g, SQL Server 2005/2008, Sybase IQ, DB2, PostgreSQL, HBase Hadoop Distributions Cloudera CDH

Methodologies Agile (Scrum), Waterfall

Continuous Integration Jenkins, Jira

Version Control Tortoise SVN, GIT

Programming JDK 1.8, SQL

Scripting PowerShell, Batch and Shell Script

Operating System Windows/Linux

BI Tools Tableau

EXPERIENCE DETAILS

• Worked as Module Lead in Sopra Steria from May 2019 to August 2019

• Worked as Senior Application analyst in The Bank Of New York Mellon since July 2013 to June 2018 EDUCATION DETAILS

• Bachelors in Information Technology, from Anna University, India, 2013. Project #1

Role : Hadoop Developer

Client : Financial Conduct Authority

Duration : May 2019 to August 2019

Environment : HDFS, Hive, HBase, Java, Spark SQL, Oozie, Yarn, GitHub, Unix, Cloudera, Sqoop, HDFS, Scala. Roles and responsibility:

Involved in extracting customer's Big data from various data sources into Hadoop HDFS.

Involved in importing data from Microsoft SQL Server, MySQL, Oracle and Teradata into HDFS for implementing an incremental load using Sqoop.

Load and transform large sets of structured, semi structured and unstructured data.

Created HBase tables to store huge volumes of data in rows and columns of variable data formats of input data coming from different portfolios.

Exported the analyzed data to the relational databases using Sqoop.

Created managed or external tables in Hive as per business requirement.

Involved in creating Hive tables, and loading and analyzing data using hive queries for staging and permanent tables.

Implemented Partitioning, Dynamic Partitions, Buckets in HIVE.

Created partitioned tables in Hive. Used Hive to analyze the partitioned and bucketed data and compute various metrics for reporting.

Developed custom MapReduce programs and User Defined Functions (UDFs) in Hive to transform the large volumes of data with respect to business requirement.

Data Modelled for HBase for large transaction sales data.

Analyzed large data sets by running Hive queries.

Experienced in handling different types of joins in Hive like Map-side joins, Bucket-Map join and Sort Merge Bucket(SMB) Map Join

Scheduled automated tasks with Oozie for loading data into HDFS through Sqoop.

Developed PIG scripts to analyze data.

Load and Query hive table into spark using Scala

Implemented the data processing framework using Scala and Spark SQL.

To load spark DataFrames to hive tables

Implemented Spark using Scala and Spark SQL for faster testing and processing of data. Was rewarded with the Star Performer for the second quarter of 2019 for this delivery. Project #2

Role : Hadoop Developer

Client : The Bank Of New York Mellon

Duration : June 2015 to June 2018

Environment : Hive, HBase, Spark, Oozie, Yarn, GitHub, Unix, Cloudera, Sqoop, HDFS, Scala. Roles and responsibility:

Involved in discussions with data scientists and analysts to understand the requirements and propose optimal solutions to the business problems.

Handled importing data from different data sources into HDFS using Sqoop and performing transformations using Hive, MapReduce and then loading data into HDFS.

Exporting of a result set from HIVE to MySQL using Sqoop export tool for further processing.

Created Partitions, Buckets based on State to further process using Bucket based Hive joins.

Worked on Data serialization formats for converting complex objects into sequence bits by using Avro, Parquet, JSON, CSV formats.

Created HBase tables to store various data formats of data coming from different sources.

Scheduled Oozie workflow engine to run multiple Hive and Pig jobs, which independently run with time and data availability.

Created Spark applications using Spark Data frames and Spark SQL API extensively.

Created reports in TABLEAU for visualization of the data sets created and tested native Drill, Impala and Spark connectors.

Good hands on experience with git and GitHub.

Involved in Agile methodologies, daily scrum meetings, spring planning. Was rewarded with the Star Performer for the second quarter of 2018 for this delivery. Project #3

Role : Senior application support analyst

Client : The Bank Of New York Mellon

Duration : July 2013 to June 2015

Environment : Tableau Desktop/Server 8.2, 9.3 & 10.1, R-serve integration on windows server Roles and responsibility:

Responsible for interaction with business stakeholders, gathering requirements and managing the delivery.

Created Rich dashboards using Tableau visualization tool and prepared user stories to create compelling dashboards to deliver actionable insights.

Developed reports, dashboards, and metadata layers sourced from information in Teradata, SQL Server, and Hadoop.

Worked extensively with advanced analysis Actions, Calculations, Parameters, Background images, Maps.

Created Tableau scorecards, dashboards using stack bars, bar graphs, scattered plots, geographical maps, Gantt charts using the functionality.

Extensively used data blending, embedded functionalities in Tableau and defined best practices for Tableau report development and effectively used data blending feature in tableau.

Worked creating Aggregations, calculated Fields, Table calculations, Totals, percentages using Key Performance Measures (KPI) and Measure.

Create, customize & share interactive web dashboards with simple drag & drop method and access them from any browser.

Restricted data for users using Row level security and User filters.

Designed and Optimized Connections, Data Extracts, Schedules for Background Tasks and Data Refreshes for corporate Tableau Server

Integration of Dashboard with BNY’s proprietary applications (Digital pulse, Eagle, Oncore etc.).

Involved in Migration of legacy system into Tableau shared infrastructure. Was rewarded with the Star Performer for the second quarter of 2014 for this delivery.



Contact this candidate