Post Job Free
Sign in

Data Engineer

Location:
Powder Springs, GA
Posted:
March 29, 2020

Contact this candidate

Resume:

SAI BHARATH RAVULA

Big Data Engineer

Atlanta, Georgia, 30127

408-***-****

adcity@r.postjobfree.com

Ø Experienced Big data Engineer with close to 7 years of experience. Ø Excellent reputation for resolving problems, improving customer satisfaction, and driving overall operational improvements. Ø Involved in architecture, design and deployment of Big data Technologies on both on premise and on Cloud.

Ø Conversant in Support and prototyping of ingenious Technologies in Big data and Dev-ops stacks on AWS, GCP and Azure native technologies. Skills

HDFS, Hive, Sqoop, Yarn, Kafka, Spark,

Redshift, RDS, EMR, Glue, CDH, Cloudera

Stack, Amazon S3, Amazon Athena,Aurora

DB.

Very Good

NoSQL: MongoDB, HBase

Very Good

RDBMS: Mysql and SQL.

Very Good

Security: Kerberos, Knox, Sentry, Ranger

Very Good

CICD: Jenkins,git,Bamboo

Very Good

Operations management

Very Good

Container: Docker, EKS

Good

Automation: Ansible puppet and chef

Good

Yaml and Unix Scripting

Good

Python, scala with Spark API

Good

Work History

Big Data Engineer

Deloitte, Innovation and Platforms, ConvergeHealth, Atlanta, Georgia

• Designed and build data processing pipelines using tools and frameworks in the Hadoop ecosystem

• Built frameworks/prototypes that integrate Big Data and advanced analytics to make business decisions

• Involved in creating Hive tables loading with data and writing hive queries which will run internally in map reduce way

• Responsible for analyzing health data and producing, verifying and interpreting client reports with very little oversight. Performing the analyses of health care data, including medical and pharmacy claims, membership files and health advisory/coaching interaction

• Enabled Data Encryption On-wire and at Rest for security of PHI/PII data stored and analyzed in/from HDFS

• Experience in creating dashboards, SPL searches, reports and views, administration, upgrading, alert scheduling, Add - Ons and Splunk infrastructure

• Worked on SDLC phases for various Big data technologies along with connectivity of external tools

• Actively involved in Datacenter migration for all applications across the platforms from on Premise to Cloud

• MongoDB installations, version upgradations, database maintenance activities

• Spark query tuning and optimization to balance executors

• Experienced in Spark API in python (pyspark) and Scala.

• Worked on importing and exporting data from Oracle and DB2 into HDFS and HIVE using Sqoop

• Collected the logs data from web servers and integrated in to HDFS using Flume on integration Splunk

• Implemented Visualization of Dashboard from R analytical data to Tableau

• Setup Active batch job scheduling process and data transfer between SFTP server and other datastore servers

• Implemented connectivity of R server to Hadoop ecosystem to analyze the data sets and machine learning using R

• Built Databricks Analytical platform integration with AWS Stack

• Worked on live data streaming from multi data collectors into Kafka clusters and into spark streaming and Storage

2016-02 - Current

• Strong experience in Planning and Implementation of Continuous Integration (CI) and Continuous Deployment (CD) using GIT, Maven, Jenkins and Urban Code Deploy, Chef deploy

• Deployed and administered EKS clusters on AWS native services

• Deployed Splunk in clustered and non-clustered environments

(based on customer needs; includes indexer clusters, multi-site indexer, and search heads)

• Implemented ADFS implementation with Microsoft and Redshift for LDAP authorization

• Guided the customer and support staff to apply best practices to management of Splunk knowledge objects (Apps, Dashboards, Saved Searches, Scheduled Searches, Alerts)

• Environment: Cloudera, Hadoop, Yarn, RHEL, Hive, Spark, Zookeeper, Kafka, Oozie, Active batch, Yarn, MySQl, Mongodb, ansible, R programming, AWS, Kerberos, Splunk, EKS, Route 53, ELB, S3, Bitbucket and Jenkins

Hadoop Admin/ Platform Engineer

Cisco, Austin, Texas

• Substantiated on Micro analytics with Mesos and Marathon integration to Hadoop cluster

• Experience in security for Hadoop and integrating with Kerberos authentication

• Worked on Rack aware configuration for quick availability and processing of data

• Implemented Job scheduling using Oozie and monitoring using Ganglia

• Worked with Terraform and ansible to create init and fill in pre- requisites for linux servers. Managed different infrastructure resources, like physical machines, VMs and Docker containers.

• Generated certificates with key tool, OpenSSL and get certificate signed with CA signing authority and imported to key store in respective environment

• Installed and configured various components of Hadoop ecosystem and maintained their integrity

• Planning for production cluster hardware and software installation on production cluster and communicating with multiple teams to get it done

• Involved in designing and modeling access controls for various applications residing on the cluster to ensure security and data integrity

2015-02 - 2016-01

• Implemented and created jobs to perform CI/CD from version control systems like Bitbucket,Stash and Bamboo

• Capacity planning and designing network segments, security groups, etc for new projects owned in the environment

• Installed and configured major components in Hadoop ecosystems including Kafka, Storm, Hive, Pig, HBase, Sqoop and Oozie on the cluster

• Involved in implementing High Availability and automatic failover infrastructure to overcome single point of failure for Name node utilizing zookeeper services

• Worked with big data developers, architects and data scientists in troubleshooting map reduce job failures and issues with Hive, Phoenix Query, Pig and Flume

• Configured custom interceptors in Flume agents for replicating and multiplexing data into multiple sinks

• Responsible for implementation and ongoing administration of Big Data Platform- Hadoop infrastructure accommodating multiple tenants

• Implemented Kerberos Security in all environments

• Installed and configured Apache Ranger for enabling more granular level security over HDFS and Hive

• Implemented encryption on database and cluster level on On-wire and data at REST

• Environement:Mesos, marathon, ansible, openssl, HDFS, YARN, Zookeeper, HBase, Hortonworks, HDP 2.x, Ambari 2.x, MapReduce, HIVE, MySQL,AWS REST, Sqoop, Kafka, Kerberos, Ranger System Engineer

Adroix Corp, Alpharetta, GA

• Forecast storage needs

• Work with the site management to determine future disk requirements

• Conducted daily system health checks using Solarwinds and generated reports

• Applied daily OS patches and upgrades to counter potential threats to system

• Implemented various security mechanisms, including firewalls, two- factor authentication and antivirus software to maintain network integrity

• Generated CSR and configured SSL/TLS certificates on Apache HTTP Server. And Configured SSL/TLS trust store of Application server and 2014-04 - 2015-01

imported third-party application signer's certificates to the trust store for SSL/TLS communication

• Configured Jenkins as a common build engine to build and deploy applications to DEV, QA, and UAT and PROD environments and monitoring and Troubleshooting the build issue during the Jenkins build process

• Strong understanding of NoSQL databases like HBase, MongoDB & Cassandra

• Installation and configuration of Red Hat Linux and CentOS on new server builds as well as during the upgrade situations

• Created Docker images using a Docker file, worked on Docker container snapshots, removing images and managing Docker volumes also deployed Docker Swarm using Ansibleon MVP and POC environments

• Log management like monitoring and cleaning the old log files

• System audit report like no of logins, success & failures, running cron jobs

• Assembled and Deployed the application in staging, in productions, following change management practices and 24/7 support, worked in shifts

• Created user roles and groups for securing the resources using local operating. System authentication

• Experienced in tasks like managing User Accounts and Groups, managing Disks and Filesystems

• Install and configure Instruction Detection System (IDS) like Tripwire, Snort, and Lids

• Configuring & monitoring DHCP server

• Created and Managed User roles and Groups for Securing the Resources using local operating system authentication in LDAP Servers and mapped them to the enterprise applications running on the Application Servers

• Created and Implemented CRON JOBS for BPM Batch jobs and automation of BPM/WebSphere Administration

Systems Analyst

Tekk Force Inc, Dallas, Texas

• User account management and support for NFS, FTP, SSH, LDAP integration

• Performed standard IT functions including but not limited to design, analysis, evaluation, testing, debugging, and implementation of all system programs that provide support to Accounting, IT, Human Resources, and Operations

2013-08 - 2014-04

• Recommended test and evaluation or T&E strategies including lifecycle integration planning on current systems and networks to be used as a baseline reference for future infrastructure

• Resolved issues effectively by providing supports in System Integration Testing, Business Integration Testing and Production Responsible for reviewing all open tickets, resolve and close any existing tickets using zendesk

• Installation, Configuration, upgradation and administration of Sun Solaris, Red-Hat Linux

• Experienced in tasks like managing User Accounts and Groups, managing Disks and Filesystems

• Organized system infrastructure documentation and operating procedures, strengthening controls and enhancing overall performance

Transaction Risk Investigator

Amazon, Hyderabad, Telangana

• Take appropriate action to identify and help minimize the risk posed by fraud patterns and trends

• Consistently achieve/exceed weekly productivity and standards

• Participate in ad hoc projects/assignments as necessary

• Specific expertise of PQ/MRI/BRI/Sanctions/Payment/AWS

/Chargebacks within TRMS

Education

Master of Science: Engineering Technology

Pittsburg State University - Pittsburg, KS

Bachelor of Science: Engineering Technology

Mahatma Gandhi Institute of Technology - Hyderabad Certifications

AWS Solution Architect Associate

PMBOK Certified Associate

Splunk CLusterAdmin( Splunk)

Splunk Administration (Splunk)

Python Data Structures(Coursera)

Programming python (Coursera)

2011-07 - 2011-12

2012-01 - 2013-06

2007-09 - 2011-06



Contact this candidate