Yogesh Kumar Mohan www.linkedin.com/in/YogeshKumarMohan
** *** ******, ****** ****, NJ 07306 424-***-**** acvw9g@r.postjobfree.com
Education
Dual Degree: W. P. Carey School of Business at Arizona State University, Tempe, AZ August 2015 - May 2016 Master of Science in Business Analytics (MSBA), 3.97 GPA
Institute of Management and Technology Dubai, UAE August 2014 - May 2015
Master of Business Administration, 3.7 GPA
Anna University, Chennai, India May 2011
Bachelor in Engineering, Electrical and Electronics Engineering
Professional Qualifications
Data Mining & Machine Learning Algorithms
Text Mining & Web Scraping
Layer2 & Layer3 Protocols
System testing,codedebugging
Multi-threading & MPP
ETL Automation & Scripting
Python automation scripting
Functional, Integration Testing
Linux and Distributed sys.
Data Visualization(Tableau)
No SQL: MongoDB, Neo4j
NaturalLanguageProcessing
LAN & WAN switching
Jenkins, Selenium, REST
OLAP & OLTP SQL Querying
JSON/XML Parsing- Python
Agile Methodology(Scrum)
Python Packages: Pandas, Scipy, Scikit-Learn, Theano
Software: Python, R, SAS Enterprise, SQL: MySQL, Oracle, Scala, Cognos, SPSS, Django, C, Matlab. C++,Ruby, Java, Minitab, Google API, Graph API
Apache Hadoop, Pig, Spark, Hive, HBase, Kafka, Storm, KNIME, Disco, Informatica, Cuint, Junit, Json, JIRA
Data Manipulation & Pipeline Building, ProcessMining
Scripting: Python, Perl, TCL, Java Script, shell, bash
Test Plan development, test scheduling and reporting
Professional Experience
Research Associate, Department of Information Systems- WP Carey School of Business, ASU Jan 2016 – Till date
Built an automated framework for web scraping (Python: Requests, BeatifulSoup, lxml) from various web sources & social media and for data cleaning, data visualization, text mining (NLTK) & Machine Learning for a doctorate research.
Developed scripts using sqoop tool to import data into the HDFS on AWS and used PIG scrips for Data Manipulation and HIVE for Querying.
Built Map Reduce scripts in Python to perform Map Reduce tasks for performing Data Manipulation and Munging purposes on AWS EMR distributed Hadoop platform.
Used Facebooks and Twitter’s REST based API to build framework for scraping data in JSON format and processing it using Python library (json) for extracting the data like tweets, comments etc.
Senior Software Engineer, Testing & QA, Aricent Technologies, Chennai, India January 2014 – August 2014
Headed the L2 and L3 QoS Module in Test planning and implementation for the US based start-up Telecom client for their Enterprise Gateway device. It involved Rate Limiting, Scheduling, Metering & Policing of Network Traffic.
Designed Python based automation framework, developed automation test scripts using Python & TCL for system and integration testing, designed system and integration test plans for testing L2 Protocols(VLAN, MSTP, DHCP, LACP)
Developed unit test scripts in Cunit for testing the code before integration which resulted in 40% increase in efficiency
Built prototypes to test the design of advanced QoS feature, it involved porting codebase to Broadcom, Marvel & Qualcomm Chipset through APIs using C & C++ and integrated it to be controlled by LSI’s AXM network processor.
Worked on building a SOAP based web API for the Enterprise Gateway device with XML based communication through HTTP for configuration and debugging of the device from Remote client.
Developed python based scripts to test UI on the device using Selenium web drivers’ python API which improved test efficiency significantly; tested various functionalities on the device through the User Interface.
Software Test Engineer, Aricent Technologies, Chennai, India April 2012 – December 2013
Worked for Japanese Client NEC on their mobile backhaul device i-Pasolink on various Layer2 & Layer3 protocols
Developed automation scripts in Python & TCL mainly involved socket programming & text mining for automating monitoring, reporting of various L2, L3 protocols, DHCP & ELOAM, for fault prediction & detection in the network
Created Dashboards consisting of various reports on Tableau using Python data extracts from the i-Pasolink device.
Automated the ETL process on the Oracle Database using Informatica & SSIS, which improved efficiency by 30%.
Developed automated test suites using Selenium web drivers’ python interfaces to test web interface on the i-pasolink device for testing various protocol configurations, logging and alarm functionalities.
Performed Web Scraping using Python, on web servers of remote routers to monitor historical logs for fault detection which involved scraping through HTML and JSON pages and used Javascripts for updating the webpage.
Worked on multicore servers with Ubuntu and CentOS based Linux platforms for developing, running, maintaining and for version control of the automation scripts and code bases, as a configuration manager using tools like GIT & CVS.
Systems Engineer, Infosys Technologies, Mysore, India
August 2011 – February 2012
Completed extensive training on Microsoft .Net, SQL, Java, C# & C++ and worked for Bank of America in a project involving Data Warehousing and front end development using Java
Projects
Dignity Health: Social Media Analytics: (Capstone Project)
Built predictive model to measure KPIs for gauging Customer engagement on Social Media Platforms namely Facebook & Twitter for DH, extracted Facebook & Twitter Comments using python for each posts to perform sentiment analysis
Extracted data from YouTube using python APIs & performed comparative analysis on YouTube channels of hospitals.
Built item based & content based collaborative filtering recommender system using Python using user ratings.
Built a predictive model in R to gauge the helpfulness factor of the food reviews, used Amazon Food Reviews data. Involved text mining & prediction using XGBoost algorithm, performed sentiment analysis on reviews.
Developed Map Reduce for document tokenization over Hadoop using Python, and extended it to Apache Spark
Built and simulated Excel based (using Excel Add-Ons @Risk) supply chain model for optimizing the supply chain of RUTF in Africa for a UN based study. Analyzed various factors in the supply chain and simulated it for optimal solution.
Starbucks: Waiting Line Queue & Process Capability Analysis: Performed Waiting Line Analysis and incorporated Markov chain techniques in queueing theory & process capability analysis(Minitab) to analyze the process efficiency at ASU’s Starbucks
Leadership Experience
Led a team of four strategically, helped NEC Japan in technology risk management and to design and implement a stable framework and architecture for the Quality of Service module in their state of art Product.
Delivered Technology Feasibility Report covering the operational analysis and presented RFPs to the US based Client determining client’s needs and their requirements for the project.
Pioneered in Python Automation Test Framework for ETL automation, for analysis of streaming data and derive insights
Availability
Available: Immediately
Best Time to issue: Any time after 11 am EST(all days)