Post Job Free

Resume

Sign in

Data Engineer

Location:
Tempe, AZ
Posted:
July 28, 2016

Contact this candidate

Resume:

Yogesh Kumar Mohan www.linkedin.com/in/YogeshKumarMohan

** *** ******, ****** ****, NJ 07306 424-***-**** acvw9g@r.postjobfree.com

Education

Dual Degree: W. P. Carey School of Business at Arizona State University, Tempe, AZ August 2015 - May 2016 Master of Science in Business Analytics (MSBA), 3.97 GPA

Institute of Management and Technology Dubai, UAE August 2014 - May 2015

Master of Business Administration, 3.7 GPA

Anna University, Chennai, India May 2011

Bachelor in Engineering, Electrical and Electronics Engineering

Professional Qualifications

Data Mining & Machine Learning Algorithms

Text Mining & Web Scraping

Layer2 & Layer3 Protocols

System testing,codedebugging

Multi-threading & MPP

ETL Automation & Scripting

Python automation scripting

Functional, Integration Testing

Linux and Distributed sys.

Data Visualization(Tableau)

No SQL: MongoDB, Neo4j

NaturalLanguageProcessing

LAN & WAN switching

Jenkins, Selenium, REST

OLAP & OLTP SQL Querying

JSON/XML Parsing- Python

Agile Methodology(Scrum)

Python Packages: Pandas, Scipy, Scikit-Learn, Theano

Software: Python, R, SAS Enterprise, SQL: MySQL, Oracle, Scala, Cognos, SPSS, Django, C, Matlab. C++,Ruby, Java, Minitab, Google API, Graph API

Apache Hadoop, Pig, Spark, Hive, HBase, Kafka, Storm, KNIME, Disco, Informatica, Cuint, Junit, Json, JIRA

Data Manipulation & Pipeline Building, ProcessMining

Scripting: Python, Perl, TCL, Java Script, shell, bash

Test Plan development, test scheduling and reporting

Professional Experience

Research Associate, Department of Information Systems- WP Carey School of Business, ASU Jan 2016 – Till date

Built an automated framework for web scraping (Python: Requests, BeatifulSoup, lxml) from various web sources & social media and for data cleaning, data visualization, text mining (NLTK) & Machine Learning for a doctorate research.

Developed scripts using sqoop tool to import data into the HDFS on AWS and used PIG scrips for Data Manipulation and HIVE for Querying.

Built Map Reduce scripts in Python to perform Map Reduce tasks for performing Data Manipulation and Munging purposes on AWS EMR distributed Hadoop platform.

Used Facebooks and Twitter’s REST based API to build framework for scraping data in JSON format and processing it using Python library (json) for extracting the data like tweets, comments etc.

Senior Software Engineer, Testing & QA, Aricent Technologies, Chennai, India January 2014 – August 2014

Headed the L2 and L3 QoS Module in Test planning and implementation for the US based start-up Telecom client for their Enterprise Gateway device. It involved Rate Limiting, Scheduling, Metering & Policing of Network Traffic.

Designed Python based automation framework, developed automation test scripts using Python & TCL for system and integration testing, designed system and integration test plans for testing L2 Protocols(VLAN, MSTP, DHCP, LACP)

Developed unit test scripts in Cunit for testing the code before integration which resulted in 40% increase in efficiency

Built prototypes to test the design of advanced QoS feature, it involved porting codebase to Broadcom, Marvel & Qualcomm Chipset through APIs using C & C++ and integrated it to be controlled by LSI’s AXM network processor.

Worked on building a SOAP based web API for the Enterprise Gateway device with XML based communication through HTTP for configuration and debugging of the device from Remote client.

Developed python based scripts to test UI on the device using Selenium web drivers’ python API which improved test efficiency significantly; tested various functionalities on the device through the User Interface.

Software Test Engineer, Aricent Technologies, Chennai, India April 2012 – December 2013

Worked for Japanese Client NEC on their mobile backhaul device i-Pasolink on various Layer2 & Layer3 protocols

Developed automation scripts in Python & TCL mainly involved socket programming & text mining for automating monitoring, reporting of various L2, L3 protocols, DHCP & ELOAM, for fault prediction & detection in the network

Created Dashboards consisting of various reports on Tableau using Python data extracts from the i-Pasolink device.

Automated the ETL process on the Oracle Database using Informatica & SSIS, which improved efficiency by 30%.

Developed automated test suites using Selenium web drivers’ python interfaces to test web interface on the i-pasolink device for testing various protocol configurations, logging and alarm functionalities.

Performed Web Scraping using Python, on web servers of remote routers to monitor historical logs for fault detection which involved scraping through HTML and JSON pages and used Javascripts for updating the webpage.

Worked on multicore servers with Ubuntu and CentOS based Linux platforms for developing, running, maintaining and for version control of the automation scripts and code bases, as a configuration manager using tools like GIT & CVS.

Systems Engineer, Infosys Technologies, Mysore, India

August 2011 – February 2012

Completed extensive training on Microsoft .Net, SQL, Java, C# & C++ and worked for Bank of America in a project involving Data Warehousing and front end development using Java

Projects

Dignity Health: Social Media Analytics: (Capstone Project)

Built predictive model to measure KPIs for gauging Customer engagement on Social Media Platforms namely Facebook & Twitter for DH, extracted Facebook & Twitter Comments using python for each posts to perform sentiment analysis

Extracted data from YouTube using python APIs & performed comparative analysis on YouTube channels of hospitals.

Built item based & content based collaborative filtering recommender system using Python using user ratings.

Built a predictive model in R to gauge the helpfulness factor of the food reviews, used Amazon Food Reviews data. Involved text mining & prediction using XGBoost algorithm, performed sentiment analysis on reviews.

Developed Map Reduce for document tokenization over Hadoop using Python, and extended it to Apache Spark

Built and simulated Excel based (using Excel Add-Ons @Risk) supply chain model for optimizing the supply chain of RUTF in Africa for a UN based study. Analyzed various factors in the supply chain and simulated it for optimal solution.

Starbucks: Waiting Line Queue & Process Capability Analysis: Performed Waiting Line Analysis and incorporated Markov chain techniques in queueing theory & process capability analysis(Minitab) to analyze the process efficiency at ASU’s Starbucks

Leadership Experience

Led a team of four strategically, helped NEC Japan in technology risk management and to design and implement a stable framework and architecture for the Quality of Service module in their state of art Product.

Delivered Technology Feasibility Report covering the operational analysis and presented RFPs to the US based Client determining client’s needs and their requirements for the project.

Pioneered in Python Automation Test Framework for ETL automation, for analysis of streaming data and derive insights

Availability

Available: Immediately

Best Time to issue: Any time after 11 am EST(all days)



Contact this candidate