Post Job Free
Sign in

Data Python

Location:
Pasadena, CA
Posted:
April 02, 2020

Contact this candidate

Resume:

Sushant Burde

Palo Alto, California, United States

linkedin.com/in/sushant-burde-515a5a5a

********@*****.***

Summary

Data Enthusiast working in the tech industries for more than 5 years specializing in data engineering, visualization and analysis. Deeply expertise in SQL, Tableau, Python for Data Analysis, Hadoop echo systems HDFS, Map reduce (Non-java), Pig, Sqoop, Kafka, YARN and Hive for scaled and distributed computing and High-performance computing.

• Knowledge on advanced SQL functions.

• Developed machine learning skills (Supervised and Unsupervised learning, Linear Regression, Decision Tree Classifier, SVM, Gaussian and others) to train and test models.

• Experienced in working with various Python Integrated Development Environments like IDLE, PyCharm, PyDev, JupyterLab and Anaconda

• Built-up skills in NumPy, SciPy, Matplotlib, Scikit-learn (Machine Learning) and Pandas python libraries

• Created user friendly dashboards in Tableau for data reporting. Extracted, interpreted and analyzed data to identify key metrics and transform raw data into meaningful, actionable information.

• Good Knowledge of using Web Services and API's - AWS (Redshift, Kinesis, EC2, S3)

• Familiar with Airflow directed acyclic graphs (DAGs) of tasks

• Knowledge on Docker Containers and Images

• Familiar with Druid database - Loading native file, from hadoop, Kafka and querying data

• Extensive experience in Software Development Life Cycle (SDLC)

• Familiar with single node and multi node cluster configuration

• Good knowledge on Unix/Linux commands

Experience

Instructional Student Assistant

California State University, Los Angeles - College of Business and Economics Jan 2019 - Present (1 year 3 months +)

Data Engineer Intern

VMware

Jun 2019 - Aug 2019 (3 months)

• Researched BI solutions for Data Initiative Data Pipeline model for increasing productivity with various feature & functionalities for user to create reports and dashboards.

• Expressed the initiative of BI tool research through poster session.

• Involved in DI data pipeline ecosystem R&D research activity for enhancing data source consistency.

• Gathered information from R&D functional areas to identify primary & secondary sources, owners & documentations for DI Data Pipeline Ecosystem.

• Published mission deck documentation for R&D central unified data DI ecosystem and uploaded on Confluence and JIRA page.

• Involved in ongoing business need for expanding source/portal "search" index Confluence page to increase the accuracy of search option and help VM users to find document conveniently. Data Engineer - Google Cloud Platform, Python Data Analysis libraries - Retail Project Tata Consultancy Services

Sushant Burde - page 1

Jan 2018 - Jun 2018 (6 months)

• Responsible for migrating Teradata tables (Millions of records) into Google Cloud using Google Big Query.

• Familiar with python script and SQL for querying and analyzing data, schema matching and creating replicas in Google Big Table.

• Worked with Jenkins (Job scheduling tool) for running batches migrating data from Teradata to Big tables.

• Frequently writing codes on GitHub for running the jobs and batch scheduling.

• Responsible for creating daily reports on migrated data and client interaction for daily updates.

• Worked for migrating java codes in python (jython) using anaconda and java eclipse.

• Analyzed different table structures and implemented managed tables using Hive.

• Involved in moving all scripts and log files to new location in Google Big Table. Data Analyst - Hadoop (Big Data) Developer - Insurance Project Tata Consultancy Services

Oct 2015 - May 2017 (1 year 8 months)

• Responsible for building scalable distributed data solutions using Hadoop and migrate Customer, Agency and Policy data from Bancs interface to ACID interface.

• Developed and Designed ETL Applications and Automated using Oozie workflows for the Jobs which are running Daily, Weekly and Monthly.

• Hands on experience in moving data from databases and DWH to HDFS file system-using Sqoop and used various compression techniques to optimize the data storage in ACID (AXA Customer Insight Database) OLAP database. Developed Oozie workflows for transforming huge set of data using Pig from Capita(Steria) and loaded in Hive external tables of ACID OLAP database.

• Performed cleaning data (data wrangling) from a variety of different sources; assembling disparate datasets into a tidy format that can be easily used for analytics and visualizations.

• Implemented performance-tuning techniques along various stages of the ETL process.

• Analysed the data by performing Hive queries and running Pig scripts to know user behaviour.

• Automated all the jobs for pulling data from Bancs FTP server to load data into Hive tables using Oozie workflows.

• Involved in moving all log files generated from Bancs and CKB (Customer Knowledge base) services to HDFS for further processing through Flume.

• Involved in loading data from UNIX file system to HDFS. Tableau - Support Analyst - Banking Project

Tata Consultancy Services

Jul 2015 - Oct 2015 (4 months)

• Involved in creating database objects like tables, views, procedures, triggers, and functions using T-SQL to provide definition, structure and to maintain data efficiently.

• Building, publishing customized interactive reports and dashboards, report scheduling using Tableau server.

• Created action filters, parameters and calculated sets for preparing dashboards and worksheets in Tableau.

• Generated Dashboards with Quick filters, Parameters and sets to handle views more efficiently.

• Restricted data for particular users using Row-level security and User filters.

• Developed Tableau Visualizations and dashboards using Tableau Desktop.

• Interacted with the existing database developers and DBA to understand the existing schema.

• Participated in daily stand-ups with clients on daily design and development effort Sharpoint Administrator

Sushant Burde - page 2

Tata Consultancy Services

Oct 2014 - Jun 2015 (9 months)

• Responsible for Site creation, Deletion, Updating for various teams held in an organization.

• Provides access (Read/Read+write/Read+Write+delete) to various clients for different sites.

• Manages details about the clients and change as per the requirement.

• Involved in migration of one client to different client held in an Organization.

• Deployed and created solution packages of InfoPath forms, Lists, Web Parts.

• Created, configured and administered SharePoint sites and user permission rights.

• Collaborated with multiple departments and application leads to successfully integrate SharePoint applications with non-SharePoint applications and improve business processes Business Object Administrator - Support Analyst

Tata Consultancy Services

Oct 2014 - Jun 2015 (9 months)

• Responsible for monitoring reports, discovering data, perform analysis to derive insights and create reports that visualize the insights and updating it as per the client requirement.

• Enhance/modify existing reports by creating new Parameters, Formulas etc.

• Interact directly with the business users in gathering requirements.

• Optimize various SQL queries to make the reports run faster.

• Used SQL extensively in command objects of Crystal reports.

• Export reports in various formats like MS Excel, PDF, MS Word, Text and CSV.

• Generate reports with parameters, sub reports, cross tabs, charts using Crystal Reports. SQL Developer

Tata Consultancy Services

Oct 2014 - Mar 2015 (6 months)

• Responsible for providing customer and policy details to clients for business requirements using SQL server management studio.

• Worked on 3 batches that supplies the one scheduler to another job scheduler.

• Involved in Decommissioning of interfaces connected with one organization by running multiple SQL queries to fetch records from the Databases

• Automated jobs which run daily and weekly batches.

• Developed, implemented and maintained enterprise business information systems.

• Implementation of data access from File system, MS SQL Server etc.

• Gathered software requirements from clients and end users.

• Derived and documented process application business rules. Education

California State University-Los Angeles

California State University, Los Angeles - College of Business and Economics, Management Information Systems

2018 - 2020

Microsoft AZURE for Machine Learning (Data Science) 1. Learned to employ a range of techniques to prepare data for modeling, build effective models, and evaluate model performance to create a suitable accurate predictive solution. 2. Familiar with the basic process of creating and scoring a model. Bike rental dataset used to iterate work for further cleansing the data on range of datasets. 3. Used Apply SQL transformations and Python scripting to train and test Bike rental models dataset. SAS Visual Analytics

Sushant Burde - page 3

1. Create global and local filters, Graphs and exported to PDF/VA Reports on Healthcare Data Objects in SAS Visual Analytics.

2. Create Hierarchies to drill-down on desired category items in SAS Visual Analytics. 3. Created SAS OLAP cubes with measures, hierarchies, aggregations, etc. 4. Performed Query Optimization to improve performance of ETL process to load data. 5. Created SAS information Maps with Healthcare data for Reporting and analysis. G.H.Raisoni College of Engineering,nagpur

Bachelor of Engineering (B.E.), Electrical, Electronic and Communications Engineering Technology/Technician

2010 - 2014

Electronics Engineering

Licenses & Certifications

Hadoop Data Access - Level 1 - IBM

Learning Python - LinkedIn

78433424F67A41B2B050C516D8963E8F

Tableau 10 for Data Scientists - Lynda.com

Spark for Machine Learning & AI - Lynda.com

Mendix Rapid Developer - Mendix

14913

Honors & Awards

Best Team Award - TCS - Zimmer Biomet (Medical Device Manufacturer) - Pune Offshore

On-time Deliveries to customer with good quality/ Accuracy within given timelines and met customer expectations.

Gold Medalist - G. H. Raisoni College of Engineering, Nagpur Achieved Highest Grade within all streams

Service and Commitment Award - Tata Consultancy Services Awarded for completing more than 3 years of dedicated services On the Spot Award - Tata Consultancy Services

Crafted miniature model to demonstrate the support functionality to client Skills

SQL • Data Analysis • Python (Programming Language) • Unix • Microsoft Office • Hadoop • Software Development Life Cycle (SDLC) • Big Data • Tableau • Hive Sushant Burde - page 4



Contact this candidate