Big Data Hadoop Administrator
Defacto Veritas Pvt.Ltd from Dec-2015 to Present
Current Location: Pune
Designation: Hadoop Administrator.
Total 3.9 years of work experience in Big Data Hadoop Administration
Roles and Responsibilities:
Experience in Planning, Installing, Configuring, Deploying & Securing the Hadoop clusters on AWS cloud.
Hands on experience in deployment, configuration, supporting and managing Apache Hadoop Clusters using Cloudera Enterprise, Hortonworks and Apache Distributions.
Capable of planning and estimating cluster capacity and creating roadmaps for Hadoop cluster deployment.
Configuring and Managing High Availability (HA) for various services like HDFS HA, Resource Manager HA, in a Hadoop Cluster.
Followed Best practices for preparing and maintaining Apache Hadoop in production
Monitoring and Configuring Hadoop Services using Cloudera Manager.
Deploying Test and Production cluster on AWS Cloud.
Responsible and involve working with Data Ingestion Team.
Decommissioning and Commissioning of nodes in the cluster.
Experience in Rebalancing the data in the cluster.
Hands on performance management on cluster.
Installing, Configuring, Maintaining and Troubleshooting the Hadoop Ecosystem: HDFS, Yarn, MapReduce, Hive, Hue, Flume, Sqoop, Sentry, Kerberos, Kafka, StreamSets, Oozie and Zookeeper.
Loading data into the hadoop cluster from dynamically-generated files using Flume and from RDBMS using Sqoop, also from the local file system
Setting Alerts configuration from Cloudera Manager
Setting disk and space quota on HDFS
Troubleshooting Hadoop cluster and services configuration related issues
Loading Data from S3 to HDFS and vice versa, assisting in creation of data pipeline
Experienced in creating PAAS EMR cluster for running transient jobs using AWS console.
Troubleshoot and debug Hadoop Ecosystem runtime issues.
Experience in upgrading the existing Cloudera Distributed Hadoop from 5.7.6 to 5.9.3.
Troubleshooting, diagnosing, tuning and solving Hadoop issues & Job issues.
Adding users into the group and availing set of permissions to the specified group.
Experienced in Securing the complete access to the Cluster and its Components using Hadoop Security Authentication and Authorization Protocols.(Kerberos,ACL's and Apache Sentry)
Expertise in planning and implementing Backup & Disaster Recovery for Hadoop Cluster
To recognize and kill the jobs which are stuck and not progress to maintain the clusters performance.
Assigning permission on Roles using sentry in the cluster and users and group creation.
Responsible for adding users principles in Kerberos for security.
BE (Information Technology) from P.E.S College of Engineering,
Aurangabad in 2015.
Name : Syed Saleem
Date of Birth : 06 Dec 1988
Nationality : Indian.
Passport : Yes
Languages : English, Hindi, Urdu and Marathi.