Post Job Free
Sign in

Data Engineer

Location:
Fremont, CA
Posted:
August 19, 2020

Contact this candidate

Resume:

Ramyakrishna Vairagyam

Fremont,CA, ***** Mob- 650-***-**** •***.***.***@*****.***

LinkedIn - www.linkedin.com/in/RamyaSCU

Visa Status: Green Card

TECHNICAL SKILLS

Applications : Tableau, Dataswarm(Similar to AirFlow)

Programming : SQL, Postgres, R, Python,Perl C++, NLTK, Shell Scripting, Perl Scripting

Databases : Presto, Hive, MySQL, Oracle, MS SQL Server

Operating Systems : UNIX, Linux, Windows

Cloud : AWS

SUMMARY

8 Years Experienced IT Professional with Master Degree specialized in Data Science and Analytics. Extensive hands-on experience in Data Analysis, Data Engineering, Python, R, SQL, Tableau and Programming (C++) expertise

• Designed, developed, documented, and tested advanced data systems that brought together data from disparate sources, made it available to data scientists, analysts, and other users using scripting and/or programming languages (Python,R, C++)

• Experience with machine learning algorithms.

• Experience with Scrum/Agile development model

• Knowledge of RDBMS, SAS/R and Data Warehousing.

• Experience with Unix/Linux Environment

• Experience with SQL and experience working with complex data sets

•Extensive experience in working with Tableau Desktop, Tableau Server and Tableau Online in various versions of Tableau

•Worked on complex KPI scorecards, heat maps, tree views, circle views, bar charts etc. Data visualizations and interactive dashboard which helped business get a good handle on the team's productivity

• Writing of advanced multi-dimensional queries and reports against multiple SQL databases

• Excellent verbal and written skills and experience interacting with executives, stakeholders and managers

• Familiar with AWS

• Experience with Business Analysis. Excellent communication skills.

• Evaluated structured and unstructured datasets utilizing data mining, and predictive analytics to gain additional business insights

• Designed, developed, and implemented data processing pipelines at scale

• Ability to translate ambiguous business requirements into analytical insights

• Critically evaluate information gathered from multiple sources, reconcile conflicts, extract and outline details, conduct needs assessment, and compose into an effective format for decision making.

• Working experience with REST, SOAP, and other web API technology

• Interacted with Salesforce for Data Collection

• Extracted data from multiple sources, integrated data into a common data model, and integrated data into a target database, application, or file using efficient programming processes.

• Wrote and refined code to ensure performance and reliability of data extraction and processing.

• Communicated with all levels of stakeholders as appropriate, including executives, data modelers, application developers, business users, and customers

• Participated in gathering requirements with business and technical staff to distill technical requirements from business requests

• Experience with Relational databases

• Conduct detail-oriented rigorous data examination of source systems using SQL and Excel, including data validation to check for inconsistencies

• Demonstrated development and application of predictive statistical procedures in R

• Experience data modeling and transformation of large scale data sources using SQL

• Knowledge of scripting for automation Python, Perl

• Strong active listener with solid written and verbal communication skills

Design, build and launch extremely efficient and reliable data pipelines to move data across a number of platforms including Excel, Data Warehouse, Scuba and real-time systems

WORK EXPERIENCE

Facebook (With TekSystems)

Data Engineer/BI Engineer October 2019 – Present

Project: Data Analytics in Infrastructure Data Centers

Collaborate with the Data Center SMEs, Data Scientists, Data Engineers and Program Managers to procure data needs and define metrics

Communicate progress, risks, and completion of projects to stakeholders

Built high-performance, scalable data warehouse application

Securely source external data from numerous global partners

Ownership of the end-to-end data engineering component of the solution

Design, build and launch extremely efficient and reliable data pipelines to move data across a number of platforms including Excel, Data Warehouse, Scuba and real-time systems.

Communicate, at scale, through multiple mediums: Presentations, dashboards, company-wide datasets, bots and more.

Built data expertise and own data quality for your areas.

Designed Intelligent data models for optimal storage and retrieval

Built business intelligence solutions that balance high-performance and usability

Designed, built, and launched reliable dashboards that provide insights across cross-functional global teams

Designed visualizations that focus on usability and ease of interpretation

Deployed inclusive data quality checks to ensure high quality of dashboards

Enhance existing dashboards and maintain of all domain-related dashboardsoure

Educated business partners: Use your data and analytics experience to ‘see what’s missing’, identifying and addressing gaps in their existing logging and processes.

Leverage data and business principles to solve large scale web, mobile and data infrastructure problems.

AllScripts May 2019 - Sept 2019

Data Engineer/BI Engineer

Project: Business Analytics in Health Care

Skilled at defining and asking key questions that help clients discover their needs, communicating effectively with clients about status and issues, developing and presenting convincing recommendations that address key client objectives, setting appropriate expectations with client and managing them throughout the project life cycle

Analyzed industry trends, conducted discussions for insights, interpreted data from Allscripts’s healthcare repository, examined models and tools, formulated studies and made recommendations to stakeholders. Designed, built and deployed BI solutions in Tableau and created ad-hoc reports as needed.

Supported all aspects of projects; including formulating research and analysis plans, developing strategies instrumental in product success, managing data collection, executing data analysis, writing reports, and presenting findings while ensuring quality and accuracy. Provided on-going tracking and monitoring of performance decision systems and statistical models

Owned projects end-to-end and working in a multidisciplinary team that is agile in their approach to analytics, ensuring that everything they do is applied to the wider business.

Developed effective, quality analysis and visual insights that meet business requirements and communicated findings.

Wrote Advanced SQL queries to create reports for Finance and Medical Teams

Build Tableau Reports to better understand Data and answer the Business Questions

Brought Insights from Data

Responsible for asking Key Questions that help client discover their needs

Helped the client move their 500 reports efficiently with 2 weeks from Netezza to Yellowbrick using Tableau Backend DataSource using Postgres

Developed Dashbaords on AWS

Verizon Apr 2017- Apr 2019

Data/BI Engineer Sep 2011 – Aug 2015

Project: Fault Isolation

Responsible for extracting network surveillance and customer datasets across the organization to support business initiatives

Provided multiple demonstrations of Tableau functionalities and efficient data visualizations using Tableau to the senior management and executives

Analyzed Trouble Tickets created by VoicePortal to make more informed decision on Reducing Technician Dispatch

Visualized Delphi Machine Learning Models Performance according to Business KPI and Brought into light, the Insights on Dispatches

Designed, developed, tested, and maintained Tableau functional reports based on user requirements.

Experience designing and deploying Graphic and informative visualizations with Drill Down and Drop down menu option and Parameters using Tableau.

Worked on Technician Dispatch Reduction which created huge savings to Verizon

Built large datasets from log files

Exploratory data analysis in Tableau, Python and R for various functional

groups.

Member of the Data warehousing team involved in understanding business requirements, analysis and design of the ETL process

Actively involved in understanding business requirements, analysis and design of the ETL process

Experience studying and mining data sets in partnership with IT, cross functional teams, derive actionable insights and drive change.

Scrutinized and track trouble ticket data behavior to identify trends and unmet needs.

Provided in-depth analytical insights by developing or analyzing data models, assess the impacts, and partner with cross functional teams to safeguard organizational impacts.

Lead efforts to drive teamwork by engaging and partnering on strategic objectives.

Assisted in developing internal tools for data analysis.

Performed Advanced Analytics and showed Trend lines

Data analysis duties involve turning large volumes of data into actionable insight for the business by:

Experience asking relevant and meaningful business questions.

Collected the best data that is relevant for the decision from multiple large volume data repositories.

Reviewed and analyzed the data in context to the business and how it changes.

Cleaned and transformed the data to prepare for data analysis models.

Analyzed the data, creating data visualizations, and studied visualizations to answer the business question.

Communicated results to stakeholders using applicable metrics, visuals, actionable insights, recommendations and decisions.

Researched Networks - Service, Network elements, capabilities and diagnostics

Created framework, using data inferences and automation, to develop reproducible data driven recommendations optimized stated business objectives.

Created Advanced visualizations in Tableau, ad-hoc dashboards, traditional BI analytics

Published weekly reports that display network component metrics such as percentage utilization, migration plans, quantity of customers

Wrote Advanced SQL Queries to Integrate data from different sources of various Network Layers

Assisted in the creation and continued support of internal analytics databases to ensure actionable outcomes for analysis and reporting.

Wipro Technologies Aug 2009 – Aug 2011

Project Engineer

Project: EBX-PLATFORM

oCoding in C++ to implement use cases in Service Layer of MFP Application from scratch

oDevelopment of Ramdisk application to support various features such as user-authentication, platform component update, digital signature verification, file-system integrity check

oUpgradation of MFP and system boards with latest released platform (Kernel image, boot loader and root file system)

Project: NSN-FlexiPlatform(MiddleWare)

oDevelopment of userStory in C++

oDevelopment of SNMP Mediator component which converts SNMP traps to alarms.

oUserStory development of IF-MIB,Pronto Correction

oPatching of windriver net-snmp package with open-source netsnmp Package.

oPart of TestCase Automation framework.

oPart of Agile Methodology-Sprint Planning,DSM,Sprint Retro.

EDUCATION

Santa Clara University, Santa Clara, CA Aug 2015 - March 2017 Master of Science, Computer Science

Sri Venkateswara University, India Sep 2004 – April 2008

Bachelor of Science, Computer Science and Engineering

PROJECTS Sept 2015 – March 2017

NCDC Weather Data Analysis using Hadoop, Spark and Weka – Classified 150 GB Weather data by NCDC using K-Means Clustering by reducing the data using MapReduce(Python) across the world. Visualized the temperature clusters on world Map using Tableau across different quarters. Compared the performance of Machine Learning in Hadoop, Weka and Spark

Personality Prediction basing on Sentiment Analysis of Facebook Posts – Performed text Analysis on 10k Facebook posts of 250 Users to understand personalities of users and classified the user’s personality by building Classification Data Model in Naïve Bayes, Support Vector Machine, Random Forest and KNN

Predicting Bike Sharing Demand – Achieved higher accuracy of 80% on Predicting Bike Sharing Demand Data comprises of 10K Fields using Machine Learning Techniques – Linear Regression, Support Vector Regression and Random Forest. Performed preprocessing, exploratory data Analysis, Model Building in R

Object Oriented Analysis and Design

Build an Application using OOPS concepts to Automate Campus Cafes and Vending Machine by building Front End UI using Java Swings, with backend Code in Java and designed database in oracle



Contact this candidate