PANKAJ GANESH WANI PERSONAL
HADOOP ADMINISTRATOR DETAILS
PROFILE Ganga Darpan 3 Socy
Overall IT experience of 4+ years in Hadoop Administration. Pune-15.
Currently working as a Hadoop Administrator with Afsa Infosystems Pvt.Ltd, Pune.
Deployed Production Ready Cluster on AWS Datacenter.
Cluster and Capacity Planning including Hardware setup, Diagnosing, Troubleshooting, Mobile NO-866-***-****
solving runtime Issues and Securing cluster.
Exposure of Cluster Installation, Monitoring, Configuration, Node Commissioning, Email id-
Decommissioning, Load Balancing, High Availability, Kerberos Security, Authentication
and Backup (BDR- Replication Schedule).
Roles & Responsibilities
Responsible for day-to-day activities which includes HDFS support and maintenance, LANGUAGES
Deployed production grade cluster UAT(User Acceptance Testing ) and deployment cluster English
in Linux flavour of RHEL / Centos / Ubuntu. Specialist in deployment of Hadoop in AWS Hindi
Managing and Supporting large scale Production Hadoop environments. Secure the cluster
using centralized mechanism Kerberos Authentication with AD to achieve one way cross
realm trust and Sentry for Authorization Key Management Server (KMS) for HDFS
Responsible for implementation and support of the Enterprise Hadoop environment
Hadoop Cluster Tuning, monitoring performance of ecosystem.
Setting up new Hadoop users, Linux users, and Kerberos principals.
Deployed production grade Hadoop cluster using local repository. Configured
prerequisites for Hadoop on Linux based servers.
Upgraded CM & CDH. Manage and review Hadoop log files in accordance with
troubleshooting real-time issues.
Enabled HDFS Namenode HA and YARN (MRv2) Resource Manager HA using Cloudera
Experience in Backup and Disaster recovery process.
Sqoop configuration to import/export data to/from MySQL databases.
Decommissioning & Commissioning nodes on running cluster including Balancing HDFS
Managed services on AWS Cloud such as VPC, EC2.
Configured and Deployed Cloudera Director.
Troubleshoot, diagnose and solve the Hadoop issues and making sure that they do not
Installation of various Hadoop Ecosystems and Hadoop Daemons.
Secured the Cloudera Hadoop Cluster with Kerberos in Active Directory. Configured LDAP,
Sentry for Authorization, KMS for Data Encryption, Extended ACL for HDFS.
Implemented Data pipeline and performed DISTCP between development and production
Collaborated with the Network, Deployment, BI, Infrastructure and Database teams.
Addressing & resolving dynamic production cluster issues and providing support to Data
scientists, Data Engineers and Big Data Developers.
Working as Hadoop Administrator in Afsa Infosystems Pvt.Ltd. Pune,
Bachelor of Engineering from Pune University.
Administration from Big Data University with IBM.
Hadoop Fundamentals and foundation from Big Data University with IBM.
Hadoop Data Access from Big Data University with IBM.
Monitoring Cluster by using Cloudera Manager.
Managing Hadoop cluster and connectivity.
Commission and Decommission of nodes in cluster.
Taking backup and making sure no accidental file is deleted.
Perform tuning on cluster.
Secured Cluster by using Kerberos, Sentry.
- Reduces query times from hours and days to seconds and minutes.
- Provides insights to deliver precision medicine and improve patient health.
Operating systems: Windows, Linux, RHEL, Centos & Ubuntu.
Networking: DNS, TCP/IP.
Big Data Eco-System: HDFS, Yarn, Hive, Sqoop, Zookeeper, Spark,
Security: Kerberos, Active Directory & Sentry.
Cluster Monitoring Tool: Cloudera.
Cloud Platform: Amazon AWS.
Cloud Services: EC2.