Sushmitha Jarugula
Email: ********.*********@*****.*** Phone: 815-***-****
Address: **** ******* *****, **** ***, Naperville, IL 60563 SUMMARY
Around 3 years of progressive experience in software development using JAVA/J2EE and Hadoop ecosystem tools.
Experience in implementation of various solutions arising out of the large data processing (TB's / PB's) over various NoSQL / Hadoop and MPP based products
Experience in building scalable, cost effective and concrete technology solutions in Big data area.
Experience in coding Java Map Reduce, Pig programming, HiveQL.
Experience with NoSQL databases such as HBASE and Experience in designing data models in Hadoop.
Experience on Data Access and Analytics through PIG, HIVE and NoSQL
Experienced on Hadoop, Hbase, Sqoop, Flume and MapReduce programming with Java.
Good experience on Bigdata integration Unstructured, Social media and moving data, stream data with HADOOP, Hbase and Sqoop.
Experience in Spark Core, Spark SQL, Data Frames and RDD API using Python and Scala.
Strong debugging and problem solving skills with excellent understanding of system development methodologies, techniques and tools.
Good analytical abilities, quick grasping power and zeal for learning new things. Technical Skills:
Programming Languages: C, C#, Java, Scala, Python, R Big Data Technologies: HDFS, MapReduce, Hive, Pig, Apache Spark, Sqoop, Flume, Kafka Oozie, Impala, HBase
Scripting: PowerShell, Linux
Web Technologies: HTML, CSS, ASP.Net
Database: MySQL, SQL (including SQL data tool SSAS: OLAP concepts) Data Visualization Tools: Excel, Tableau, Power BI Professional Experience:
Asian American Resource Center Database Developer Sept’2017-May’2018 Northern Illinois University
Parent Petroleum, Inc. Junior Consultant Aug’2017-Dec’2017 Tata Consultancy Services Internship Jan’2016 – May’2016 Academic Work Experience
Graduate Research Assistant Sep’2017 – May’2018
Asian American Resource Center, Northern Illinois University, DeKalb, Illinois Marketing:
• Designed Posters, Flyers, and templates for digital signage by following the university and departmental media protocols.
• Developed and implemented the departmental marketing strategies by utilizing webpage, social media, e-mail announcements.
• Responsible for planning, monitoring, and controlling Public Relation activities. Peer-mentoring:
• Supervised the mentors and maintained regular contact with their proteges.
• Developed and implemented the weekly mentor training. Supplemental Instruction Leader Aug’2016- Sept’ 2017 ACCESS Tutoring and Support Services, Northern Illinois University, DeKalb, Illinois
• Facilitated undergraduate students in math and statistics by conducting weekly sessions to 100 students throughout the term using various learning and study strategies. Academics:
Master’s in Management Information Systems Aug’2016 – May’2018 Northern Illinois University, DeKalb, Illinois(GPA-3.9) Bachelor of Technology Aug’2012 – May’2016
Jawaharlal Nehru Technological University, Hyderabad, India (GPA-3.67) Projects:
Project Name : Opinion Mining
Software : Java, HDFS, MapReduce, R, Cloudera CDH4, Jfree Charts, Primefaces, MySQL Client : Financial Services customer
Role : Hadoop Developer
This project was done for a financial customer that deals with the analytics on the data taken from Social Media and different Blogs. The analytics shows benefits for business development for the customer based on the end user responses in the Social media sites. Extracted data from different sources like facebook, twitter, etc and stored into HDFS for performing Opinion/Sentiment Analysis using text mining algorithms to find out the sentiment/emotions & opinion of the company/product in the social circle. Based on these analytics the customer can take business decisions on the Opinion/sentiment of the end users for the business improvement.
Roles / Responsibilities:
• Analyzing the requirements
• Develop Map/reduce program
• Extract information from Social WebPages using Social media API and Web Crawler Developed components to interact web, HDFS and reports Project Name : Automated Communication Surveillance Software : Hadoop, HDFS, MapReduce, Hive, R and Eclipse 3.7 Client : Banking and Financial Services
Role : Hadoop Developer
The purpose of this project is to identify the lexicons (words) used in the mails being sent or in the chats of the traders of the organization. If any trader uses a word in the predefined list of lexicons, then the alert will be generated for a user and will be send the details to the business users. Roles / Responsibilities:
• Data model design for storing the data in HIVE.
• Responsible for technical architecture design.
• Written the Hive Queries to parse the xmls and store in Hive tables.
• Written UDF’s to get the trader details using lookup operations.
• Involved in writing shell script to split the source xml files.
• Extracted the data from Twitter, Facebook, and Product Reviews from the E-commerce website.
• Cleaned the unstructured data to get useful insights.
• Created Word cloud, topic models.
• Conducted Lexicon sentiment analysis.
• Performed data visualization and developing insights on the sentiment and accompanying reasons for those sentiments using Tableau
Project Name : Risk Data Rollup Engine
Software : Hadoop, HDFS, MapReduce, HBASE, Java, Talend and Eclipse 3.7 Client : Banking and Financial Services
Role : Hadoop Developer
This project is a migration project from Main Frame to Hadoop. Currently the data is being stored in IMS DB in Main Frame. This has to be migrated to Hadoop and the data should be stored in HBASE. This application will run the program on the entire data for customer/account level assessments based on the different segments.
Roles / Responsibilities:
• Data model design for storing the data in HBASE.
• Responsible for technical architecture design.
• Written the Map Reduce programs for the data bulk load to HBASE.
• Written the Map Reduce programs for the data delta load to HBASE.
• Written oozie configuration scripts for scheduling the map reduce programs for delta load on daily basis.
• Involved in writing the code for data transformation using Talend tool.
• Understanding the business impact of customer churn and developing a predictive model to return a probability score of securing a loan based on input data by customers.
• Performed data exploration and cleansing using R and visualizations in Tableau.
• Developed Logistic regression models, decision trees to provide probability score of approval depending on customers input.
• Data Visualizations used to communicate analytical findings to project sponsors.