Post Job Free
Sign in

Machine Data

Location:
Manchester, NH
Salary:
100000 USD
Posted:
December 06, 2020

Contact this candidate

Resume:

Reddy Vishnuvardhan Reddy

Email: ********@****.***,*****************@*****.***

Linked in: https://www.linkedin.com/in/reddy-vishnuvardhan-challapalli-bb1b99110/ Website: https://vishnuvardhan1028.wixsite.com/website Git hub: https://github.com/vishnuvardhanreddychallapalli SKILLS

PROGRAMMING:

Comfortable:

Python• C • C++• Java• STL(C++)

•HTML•CSS•JavaScript•NodeJS

Advanced Learner:

Jupyter Notebooks • Pandas • Numpy

• Scikit Learn

Intermediate:

TensorFlow • Kera’s • Spark •

Hadoop•Pytorch•OpenCV•A/B Testing

EDUCATION

University of North Carolina

M.S COMPUTER SCIENCE

Fall 2017 to December spring 2018.

Grade Point Average:3.5/4

OUTSIDE WORK

Algorithms implemented

from scratch in python and

C++:

•Decision Trees

•K-NN(locality sensitive Hashing)

•Naïve Bayes (Multinomial)

• Linear and Logistic Regression.

COURSEWORK

MOOCS

•Deep Learning by Stanford University

•Build Self Driving Cars by MIT

University of

North Carolina

at Charlotte:

•Machine Learning

•Algorithms

•Data Mining

•Intelligent Systems

•Parallel Computing

•Cloud computing

•Database Systems

Achievements

•top 2% in the class of Bachelors

Degree.

•Two times Hackathon winner in

Anna University Digital Design

Competition using VHDL

EXPERIENCE

Deka Research and Development Machine learning Software Engineer

February 2019 Manchester New Hampshire

• Worked on applying machine learning techniques and Sensor Fusion Approaches to build Machine Learning models using Robot Operating System (ROS).

• Deployment Object Detection algorithms like SSD and Faster RCNN in C++ using NVIDIA Tensor RT.

• Implemented Free Space Estimation that Generates Occupancy Grid using Ultrasonic sensors that runs at 5 frames per second using Open MP in c++.

•Implemented Hungarian Algorithm for 2X Faster data association in Object Tracking.

•Worked on Semantic Segmentation with focal loss and improved the IOU metric by 10% and Used Knowledge Distillation to bring the model size by 5 times smaller.

•Deployed Deep Neural Network Architectures for Semantic Segmentation using Tensor RT NVIDIA framework to achieve 30Hz frequency.

•Worked with a Proof of Concept for Car Blinker Detection using Conv-LSTM and Encoder Decoder Models with Attention Mechanism.

•Implemented Cuda kernels for SoftMax and Transpose Convolutions using CUDA on GTX1060-T that runs 30 times that runs faster than CPU Parallel Algorithms.

PROJECTS

Implemented a Self-Driving car model using Deep Learning with Object Detection

USING CONVOLUTIONAL NEURAL NETWORKS PYTHON and

TENSORFLOW and OPENCV

• Implemented the NVIDIA Self-Driving neural network using Keras to train a video streaming recorded from the front dash cam of the car.

• Framed the problem to predict the steering angle of the car given an image of the car by converting a video to sequence of images.

• Observed the accuracy to be 93% when I applied the neural architecture similar in the paper.

•Improved the accuracy of the model by adding drop outs and Batch Normalization techniques and proper weight initializations like Xavier-glorot techniques to 98%.

•Applied Tensor Flow object detection model with Faster R-CNN on the video to detect cars in the video.

Recommendation Systems on Movie dataset Tensor

flow and scikit learn and spark MLLIB Spark Graph X

•This is Graph based data used spark Graph X module to process this data.

•Used several traversal techniques like Breadth first search and depth first search to store relative meta information by graph traversing for computing similarities between movies and users.

•Applied Classical Machine learning techniques like collaborative filtering techniques like item-item similarity and user-user similarity and Singular Value

Decompositions or Matrix

Factorization techniques.

•Applied Machine learning

techniques like Gradient

boosting decision trees on the

dataset and observed a

dramatic improvement in the

model by 2 times as that of

Collaborative techniques to

see that content based

recommendations works

better than collaborative

filtering techniques.

•Tried both LSTM and

Gradient recurrent units and

found GRU’s work better than

LSTM.

Amazon Fine Food

reviews(Given a text find

the polarity of the review)

Tensor flow and scikit learn

•Applied Data cleaning

techniques like Bag of words,

TF-IDF, Word2vec

•Used naïve bayes and

Gradient boosting decision

trees but accuracies are not

higher because of the

improper word embeddings

•Applied LSTM’s on the

dataset but trained word

embeddings with respect to

data instead of using

word2Vec and Glove

Techniques observed an

accuracy of 95%for 3 epochs.

Implemented the

following algorithms in

Mapreduce using Spark

without MLLIB Used Scikit

learn and Numpy and pandas

•K-means

•K-nearest Neighbors

•Logistic Regression and

Support Vector machines

•Naïve Bayes

•Linear Regression in both

ordinary least squares and

Gradient Descent approaches.

• Random Forests.

PROJECTS on Distributed and parallel computing.

Implemented a Scheduler in Pthreads c++ Library for High intensity Numerical Function:

•High intensity functions take a lot of time to compute so there is a severe necessity to reduce the computation time. One of the classical approaches to perform work decomposition is through multi-threading.

•Computed the Dependency Graph and look at some of critical Metrics in the systems like Width and critical path because they are essential to compute speed ups and no of processors to be used when performing parallel computing.

•Used techniques like Static scheduling, Dynamic scheduling with different chunk sizes and computed speed ups.

Implemented Massively Serializable Algorithms Like Prefix sum, Reduce, Merge Sort, Longest Common

Subsequence in parallel using OPENMP in c++.

•Implemented the classical Serial algorithms in c++ and computes how much time it is taking for different values of n from 10 to 1 billion elements and observed that there is a necessity to bring speed ups in the code.

•Implemented parallel Prefix sum, Parallel reduce, Parallel Merge sort, Parallel Longest Common Subsequence with a Strong Scaling experiment and observed dramatic rise in the speed ups with static scheduling and Dynamic scheduling for different chunk sizes. Implemented Search Engine for information retrieval course using TFIDF and page rank scores Used Hadoop Map-reduce in java.

•Implemented the Page rank Algorithm by google in distributed Fashion and performed 10 iterations to converge the algorithm on the whole Wikipedia data using Hadoop Map reduce.

•Implemented Term Frequency and Inverse Document Frequencies for Giga bytes of Data and sort all the words in Ascending order.

•Target is to retrieve the pages when a query is typed based on the high page rank and high tfidf scores.

•Used Machine learning algorithms like Logistic Regression combinedly with page ranks and perform classical tests on how fast the systems is responding to a query and Used L1-regularization for Faster querying.

Implemented Convolution operation on a very High

Dimensional image of 200 GB in a distributed Fashion in MPI (Message passing Interface in c++) and Matrix multiplication in CUDA.

•Implemented the Convolution operation in a distributed Fashion and handled all the communication on multiple computers in c++.

•Implemented the Classical Matrix- vector multiplication and matrix- matrix multiplication on Both MPI and CUDA to visualize the speed ups differences when using CPU and GPU.



Contact this candidate