Post Job Free

Resume

Sign in

Graduate Assistant Prophet

Location:
Utica, MI
Salary:
80000
Posted:
June 16, 2021

Contact this candidate

Resume:

BINDU BOPPUDI

201-***-**** adm6ee@r.postjobfree.com LinkedIn

EDUCATION

Master’s in computer science Aug 2019 - May 2021

Indiana State University, Terre Haute, IN GPA: 3.85/4.0 Relevant Coursework: Programming fundamentals, Programming in Data Science and Analytics, Database Systems, Concrete Mathematics, Web technologies, Big Data and Data Mining, Statistics for Data Science, Cloud Computing Online Coursework: Python for Data Science and Machine Learning, Big Data-Hadoop and Apache Spark, Tableau Hands-on, Data Engineering, Python with Data Analysis

Bachelor of Technology, Computer Science July 2015 – May 2019 Jawaharlal Nehru Technological University Hyderabad, India GPA: 9.7/10 TECHNICAL SKILLS & TOOLS

Programming Languages Python, R, Linux, JAVA, C ++, Scala Databases SQL, MySQL, Sqlite3, PostgreSQL, Oracle PL/SQL, NoSQL- MongoDB, Redis, Cosmos DB, T-SQL

Big Data Technologies Hadoop- HDFS, MapReduce, YARN, Node Manager, PIG, SQOOP, Hive, Flume, PySpark, Storm, Air Flow, Oozie, Kafka

Python Libraries NumPy, Pandas, NLTK, Scikit Learn Data Analysis Excel, Pivot tables, VLOOKUP

Visualization tools Tableau, Power BI, Matplotlib, Seaborn, Plotly Cloud Technologies Azure Databricks, AWS ML, Amazon S3, Amazon EC2 Web Technologies HTML, CSS, XML, JavaScript, JQuery, PHP, JSP, JSON, Django, Flask Platforms Jupyter Notebook, PyCharm, Putty, TensorFlow, Eclipse Windows Microsoft Office, Ubuntu, Kali Linux

WORK EXPERIENCE

IT Data Analyst Graduate Assistant, Aug 2020-May 2020 Residential Life, Indiana State University, IN, USA Roles & Responsibilities:

As a Team we provide full IT support to all the residence halls.

Handling the databases and maintaining a Starrez housing application.

Generating Data Subs for executing different queries with StarQL(Automated from SQL) language.

Managing and analyzing of the inventory data using Microsoft Excel.

Reporting from SQL Server to Excel for creating Pivot Tables and V-Lookups.

Working closely with the CBORD card access system.

Generating reports & creating the insights reports of Student Occupancy rate with data analysis tools like Tableau and Excel.

Managing of devices with cost budgeting and weekly reporting review meeting with Business operations Team (Bi-ops)

Converted data from PDF to XML using python script in two ways i.e. from raw XML to processed XML and from processed them too .CSV files.

Data Engineer/ Analyst Jan 2020 – Aug 2020

Cummins Inc., IN USA

Roles & Responsibilities:

Extract Transform and Load (ETL) data from Sources Systems to Azure Data Storage services using a combination of Azure Data Factory, T SQL, Spark SQL and Azure Data Lake Analytics and processing the data in Azure Databricks.

Worked with Cloud product team to support daily data requests using Python, SQL and Tableau.

Installed and configured Hadoop MapReduce, HDFS with developing multiple MapReduce jobs in python for data cleaning and preprocessing. Utilized Apache Hadoop by Sandbox Hortonworks to monitor and manage the Hadoop Cluster.

Creating Hive tables and written Hive queries for data analysis.

Installed and configured Apache Pig and written Pig Latin scripts.

Involved in scheduling Oozie workflow engine to run multiple Hive and pig jobs.

Worked with writing SQL queries and stored procedures with PL/SQL code.

Gathered a brief knowledge of Snowflake data warehouse, by extracting the data from data Lake and sending it to other stages of integration Also used Snowflake to Maintain and develop complex SQL queries, views, functions, and reports. ETL pipelines were used with SQL, No SQL.

Also exported data using Sqoop from HDFS to Teradata on regular basis.

Coordinate development work with team members, review ETL jobs, and create scripts for scheduling jobs and implementation. We used Airflow for scheduling.

Optimizing of existing algorithms in Hadoop using Spark Context, Spark SQL, Data Frames and Pair RDD’s. Also analyzed SQL scripts and designed the solutions to implement using PySpark.

Imported required tables from RDBMS to HDFS using Sqoop and used Apache Spark streaming and Kafka to get real time streaming of data into HBase.

Machine Learning Analyst May 2018 – June 2019

CODE BEAT, Hyderabad, India

Roles & Responsibilities:

Analyzed the existing real-world applications.

Extensively used Pandas, NumPy, Seaborn, Matplotlib, Scikit learn, SciPy and NLTK with R as working with various machine learning algorithms.

Learned and worked with applying the Supervised & Unsupervised learning algorithms like Classification, Regressions, Clustering to problems by choosing the appropriate one.

Gathered, analyzed, documented, and translated application requirements into data models and Supports standardization of documentation and the adoption of standards and practices related to data and applications.

Executed SQL queries from R/Python on complex table configurations.

Data which is stored in sqlite3 datafile (DB.) were accessed using the python and extracted the metadata, tables, and data from tables and converted the tables to respective CSV tables.

Implementations of Regressions data modeling plotting in scatter plots using R language.

Used Natural Language Processing (NLP) to pre-process the data, determine the number of words and topics in the emails and form cluster of words.

PROJECTS

Calorie Intake Calculator using Python Django March 2021 – May 2021 Libraries & technologies: Python, Django Framework, HTML, CSS, Bootstrap Building a Calorie Calculator for everyday meal starting from Breakfast through Dinner. We can record and estimate the calorie intake every day. Tracking calories weakly and workout-based meal option also available. Used Django filters and libraries for GUI development.

Python CHATBOT using Deep Learning Aug 2020 – Nov 2020 Libraries & Technologies: Python, NLP, Keras, TensorFlow, Pickle, Tkinter A retrieval based chatbot which used for customer support, marketing, and instant messaging with the client. The chatbot will be trained on the dataset which contains categories, patterns, and responses. The recurrent neural network (LSTM) is used classify which category the user’s message belongs to and then we will give a random response from the list of responses. Tkinter is used to build the Graphical User Interface. Pickle files were used to handle large datasets. Covid Data Analysis with Python May 2020 – June 2020 Libraries & technologies: Python, Pandas, Matplotlib, Tableau, Facebook Prophet, Scikit Learn Gathering the public covid-19 datasets from popular open sources. Developed a predictive model of increasing covid cases and future trends using pandas, Facebook prophet & Scikit Learn for US states. Managed datasets using Panda data frames and MySQL, queried MYSQL relational database (RDBMS) queries from python using Python-MySQL connector MySQL dB package to retrieve information. Also, employed tableau for creating more interactive visuals. Market Basket & Customer Analysis – Associative learning Feb 2020 – Apr 2020 E-Grocery products analysis with association with one product and then most likely to buy the other products too. Libraries & technologies: Python, Numpy, Pandas, SQL, Scikit Learn, Matplotlib, Power BI, Flask Roles & Responsibilities:

Data collection & simulation using google forms.

Performing EDA analysis on the data for data visualization

Developed clustering algorithm for customer segmentation by demographics using Python and Sklearn.

Model building using Apriori algorithm & Eclat algorithm with generating the association rules.

Used flask to deploy the Model, creating and connecting with the webpage.

Dashboards Reports are created by using Power BI.



Contact this candidate