Ramyakrishna Vairagyam
Fremont,CA, ***** Mob- 650-***-**** •***.***.***@*****.***
LinkedIn - www.linkedin.com/in/RamyaSCU
Visa Status: Green Card
TECHNICAL SKILLS
Applications : Tableau, Dataswarm(Similar to AirFlow)
Programming : SQL, Postgres, R, Python,Perl C++, NLTK, Shell Scripting, Perl Scripting
Databases : Presto, Hive, MySQL, Oracle, MS SQL Server
Operating Systems : UNIX, Linux, Windows
Cloud : AWS
SUMMARY
8 Years Experienced IT Professional with Master Degree specialized in Data Science and Analytics. Extensive hands-on experience in Data Analysis, Data Engineering, Python, R, SQL, Tableau and Programming (C++) expertise
• Designed, developed, documented, and tested advanced data systems that brought together data from disparate sources, made it available to data scientists, analysts, and other users using scripting and/or programming languages (Python,R, C++)
• Experience with machine learning algorithms.
• Experience with Scrum/Agile development model
• Knowledge of RDBMS, SAS/R and Data Warehousing.
• Experience with Unix/Linux Environment
• Experience with SQL and experience working with complex data sets
•Extensive experience in working with Tableau Desktop, Tableau Server and Tableau Online in various versions of Tableau
•Worked on complex KPI scorecards, heat maps, tree views, circle views, bar charts etc. Data visualizations and interactive dashboard which helped business get a good handle on the team's productivity
• Writing of advanced multi-dimensional queries and reports against multiple SQL databases
• Excellent verbal and written skills and experience interacting with executives, stakeholders and managers
• Familiar with AWS
• Experience with Business Analysis. Excellent communication skills.
• Evaluated structured and unstructured datasets utilizing data mining, and predictive analytics to gain additional business insights
• Designed, developed, and implemented data processing pipelines at scale
• Ability to translate ambiguous business requirements into analytical insights
• Critically evaluate information gathered from multiple sources, reconcile conflicts, extract and outline details, conduct needs assessment, and compose into an effective format for decision making.
• Working experience with REST, SOAP, and other web API technology
• Interacted with Salesforce for Data Collection
• Extracted data from multiple sources, integrated data into a common data model, and integrated data into a target database, application, or file using efficient programming processes.
• Wrote and refined code to ensure performance and reliability of data extraction and processing.
• Communicated with all levels of stakeholders as appropriate, including executives, data modelers, application developers, business users, and customers
• Participated in gathering requirements with business and technical staff to distill technical requirements from business requests
• Experience with Relational databases
• Conduct detail-oriented rigorous data examination of source systems using SQL and Excel, including data validation to check for inconsistencies
• Demonstrated development and application of predictive statistical procedures in R
• Experience data modeling and transformation of large scale data sources using SQL
• Knowledge of scripting for automation Python, Perl
• Strong active listener with solid written and verbal communication skills
Design, build and launch extremely efficient and reliable data pipelines to move data across a number of platforms including Excel, Data Warehouse, Scuba and real-time systems
WORK EXPERIENCE
Facebook (With TekSystems)
Data Engineer/BI Engineer October 2019 – Present
Project: Data Analytics in Infrastructure Data Centers
Collaborate with the Data Center SMEs, Data Scientists, Data Engineers and Program Managers to procure data needs and define metrics
Communicate progress, risks, and completion of projects to stakeholders
Built high-performance, scalable data warehouse application
Securely source external data from numerous global partners
Ownership of the end-to-end data engineering component of the solution
Design, build and launch extremely efficient and reliable data pipelines to move data across a number of platforms including Excel, Data Warehouse, Scuba and real-time systems.
Communicate, at scale, through multiple mediums: Presentations, dashboards, company-wide datasets, bots and more.
Built data expertise and own data quality for your areas.
Designed Intelligent data models for optimal storage and retrieval
Built business intelligence solutions that balance high-performance and usability
Designed, built, and launched reliable dashboards that provide insights across cross-functional global teams
Designed visualizations that focus on usability and ease of interpretation
Deployed inclusive data quality checks to ensure high quality of dashboards
Enhance existing dashboards and maintain of all domain-related dashboardsoure
Educated business partners: Use your data and analytics experience to ‘see what’s missing’, identifying and addressing gaps in their existing logging and processes.
Leverage data and business principles to solve large scale web, mobile and data infrastructure problems.
AllScripts May 2019 - Sept 2019
Data Engineer/BI Engineer
Project: Business Analytics in Health Care
Skilled at defining and asking key questions that help clients discover their needs, communicating effectively with clients about status and issues, developing and presenting convincing recommendations that address key client objectives, setting appropriate expectations with client and managing them throughout the project life cycle
Analyzed industry trends, conducted discussions for insights, interpreted data from Allscripts’s healthcare repository, examined models and tools, formulated studies and made recommendations to stakeholders. Designed, built and deployed BI solutions in Tableau and created ad-hoc reports as needed.
Supported all aspects of projects; including formulating research and analysis plans, developing strategies instrumental in product success, managing data collection, executing data analysis, writing reports, and presenting findings while ensuring quality and accuracy. Provided on-going tracking and monitoring of performance decision systems and statistical models
Owned projects end-to-end and working in a multidisciplinary team that is agile in their approach to analytics, ensuring that everything they do is applied to the wider business.
Developed effective, quality analysis and visual insights that meet business requirements and communicated findings.
Wrote Advanced SQL queries to create reports for Finance and Medical Teams
Build Tableau Reports to better understand Data and answer the Business Questions
Brought Insights from Data
Responsible for asking Key Questions that help client discover their needs
Helped the client move their 500 reports efficiently with 2 weeks from Netezza to Yellowbrick using Tableau Backend DataSource using Postgres
Developed Dashbaords on AWS
Verizon Apr 2017- Apr 2019
Data/BI Engineer Sep 2011 – Aug 2015
Project: Fault Isolation
Responsible for extracting network surveillance and customer datasets across the organization to support business initiatives
Provided multiple demonstrations of Tableau functionalities and efficient data visualizations using Tableau to the senior management and executives
Analyzed Trouble Tickets created by VoicePortal to make more informed decision on Reducing Technician Dispatch
Visualized Delphi Machine Learning Models Performance according to Business KPI and Brought into light, the Insights on Dispatches
Designed, developed, tested, and maintained Tableau functional reports based on user requirements.
Experience designing and deploying Graphic and informative visualizations with Drill Down and Drop down menu option and Parameters using Tableau.
Worked on Technician Dispatch Reduction which created huge savings to Verizon
Built large datasets from log files
Exploratory data analysis in Tableau, Python and R for various functional
groups.
Member of the Data warehousing team involved in understanding business requirements, analysis and design of the ETL process
Actively involved in understanding business requirements, analysis and design of the ETL process
Experience studying and mining data sets in partnership with IT, cross functional teams, derive actionable insights and drive change.
Scrutinized and track trouble ticket data behavior to identify trends and unmet needs.
Provided in-depth analytical insights by developing or analyzing data models, assess the impacts, and partner with cross functional teams to safeguard organizational impacts.
Lead efforts to drive teamwork by engaging and partnering on strategic objectives.
Assisted in developing internal tools for data analysis.
Performed Advanced Analytics and showed Trend lines
Data analysis duties involve turning large volumes of data into actionable insight for the business by:
Experience asking relevant and meaningful business questions.
Collected the best data that is relevant for the decision from multiple large volume data repositories.
Reviewed and analyzed the data in context to the business and how it changes.
Cleaned and transformed the data to prepare for data analysis models.
Analyzed the data, creating data visualizations, and studied visualizations to answer the business question.
Communicated results to stakeholders using applicable metrics, visuals, actionable insights, recommendations and decisions.
Researched Networks - Service, Network elements, capabilities and diagnostics
Created framework, using data inferences and automation, to develop reproducible data driven recommendations optimized stated business objectives.
Created Advanced visualizations in Tableau, ad-hoc dashboards, traditional BI analytics
Published weekly reports that display network component metrics such as percentage utilization, migration plans, quantity of customers
Wrote Advanced SQL Queries to Integrate data from different sources of various Network Layers
Assisted in the creation and continued support of internal analytics databases to ensure actionable outcomes for analysis and reporting.
Wipro Technologies Aug 2009 – Aug 2011
Project Engineer
Project: EBX-PLATFORM
oCoding in C++ to implement use cases in Service Layer of MFP Application from scratch
oDevelopment of Ramdisk application to support various features such as user-authentication, platform component update, digital signature verification, file-system integrity check
oUpgradation of MFP and system boards with latest released platform (Kernel image, boot loader and root file system)
Project: NSN-FlexiPlatform(MiddleWare)
oDevelopment of userStory in C++
oDevelopment of SNMP Mediator component which converts SNMP traps to alarms.
oUserStory development of IF-MIB,Pronto Correction
oPatching of windriver net-snmp package with open-source netsnmp Package.
oPart of TestCase Automation framework.
oPart of Agile Methodology-Sprint Planning,DSM,Sprint Retro.
EDUCATION
Santa Clara University, Santa Clara, CA Aug 2015 - March 2017 Master of Science, Computer Science
Sri Venkateswara University, India Sep 2004 – April 2008
Bachelor of Science, Computer Science and Engineering
PROJECTS Sept 2015 – March 2017
NCDC Weather Data Analysis using Hadoop, Spark and Weka – Classified 150 GB Weather data by NCDC using K-Means Clustering by reducing the data using MapReduce(Python) across the world. Visualized the temperature clusters on world Map using Tableau across different quarters. Compared the performance of Machine Learning in Hadoop, Weka and Spark
Personality Prediction basing on Sentiment Analysis of Facebook Posts – Performed text Analysis on 10k Facebook posts of 250 Users to understand personalities of users and classified the user’s personality by building Classification Data Model in Naïve Bayes, Support Vector Machine, Random Forest and KNN
Predicting Bike Sharing Demand – Achieved higher accuracy of 80% on Predicting Bike Sharing Demand Data comprises of 10K Fields using Machine Learning Techniques – Linear Regression, Support Vector Regression and Random Forest. Performed preprocessing, exploratory data Analysis, Model Building in R
Object Oriented Analysis and Design
Build an Application using OOPS concepts to Automate Campus Cafes and Vending Machine by building Front End UI using Java Swings, with backend Code in Java and designed database in oracle