sateesh kumar chillakuru
● Mobile: 647-***-**** ● Email: *************@*****.***
PROFILE
12+ Years of varied experience in Big Data Hadoop, Data Warehouse and Mainframes systems covering all the stages of Software Development Life Cycles (SDLC) & STLC.
Around 3 Years of exclusive experience in Big Data Hadoop components like HDFS, Spark, Hive, Map Reduce, Apache Pig, Sqoop, HBase, Oozie and R Tool.
Played different roles as Developer, Development Lead, Test Lead & Solution Architect and involved in all the phases of implementation, application support and maintenance.
Experience designing solutions to process high volume of data, stream ingestion, DQ processing and low latency data provisioning by using the in-memory computing capabilities of Spark.
Basic knowledge in scripting languages like Scala and Python and streaming component Kafka and MLlib.
Coordinating with business customers to gather business requirements, and also interacting with other technical peers to derive Technical and System Requirements.
Preparation of Business Requirements Document (BRD), Use Case Specifications, Functional Specifications (FSD), Systems Design Specification (SDS), Technical Design Document (TDD), Requirement Traceability Matrix (RTM) and Final Guideline Document for users instruction.
Experience in setting small size Hadoop Cluster.
Experience with Agile, Scrum methodology of development process.
Highly proficient in handling the activities from Business Requirement/Issue Analysis, Solution Design and Implementation through Coding, Unit & System Testing and Delivery of medium to large scale projects with Citi, Travelers, Wal-Mart, SIS Canada, TD Bank, IBM Warehouse, Sprint Nextel and AMEX through L&T Infotech, IBM Canada & IBM India.
CERTIFICATIONS
Certified in Cloudera Certified Developer Hadoop - CCDH 410 (License# 100-011-235).
Certified in ISTQB Software Tester Foundation Level – BH0-010.
Certified in IBM DB2 Certified Database Associates/DB2 UDB V8.1 Family – Exam 700.
highlights of qualifications
Excellent Analytical and Problem Solving Skills
Results Oriented and Focused Leader
Effective Time Management
Exceptional Team Motivator
Strong Written and Verbal Communication Skills
Strong Interpersonal Skills
Education qualification
M.Sc. (Master Of Science- Statistics) from S.V University, Tirupati, A.P, India- 1997-99
TECHNICAL SKILLS
Programming Languages & Web Technologies: Java, UNIX, MapReduce, YARN-Mapred, Apache Pig, Python, Scala, COBOL, PL/1, JCL, CICS, HMTL and VBScript.
Database & Tools: HBase, IBM DB2, IMS DB, Oracle and MS SQL Server, Sqoop & Hive.
Operating environments: UNIX, Linux, ZOS/390, IBM/MVS, Windows 2000 and Windows 7.
Framework: Hadoop, SPARK–HDFS and YARN.
Utilities and Softwares: Eclipse, Flume, Kafka, OOZIE, R, ETL/Informatica, TSO/ISPF, VSAM, CA7, File Aid, Expeditor, OPC/TWS, SAR, SCLM, NDM, Infoman, Change man, Endeavor, QMF, Abend aid, Autosys, RPM, QC and Agile methodology.
RESIDENCY: Holding Canadian Permanent Residency.
EXPERIENCE / ACCOMPLISHMENTS
PROJECT : VALLAR NG – Internal Fraud Detection (Asia)
Hadoop Technical Solution Architect
Client : Citi Bank, USA.
Work Location : L&T Infotech, Chennai, India.
Team Size : 5
Duration : May 2016 to Jan 2017
Domain : Banking – Fraud Detection System (AML)
Environment : Hadoop- CDH, Spark, Java, Hive, AVRO, Auto Sys, Auto Watch & UNIX
Project Type : Development
Description :
The Vallar Next Generation program was designed and deployed in Asia as a result of significant Internal Fraud identified in late 2010, causing some USD 100MM in losses. Today the deployment proactively detects suspicious employee activity as a catalyst for investigation.
Entity Resolution is a generic component on Enterprise Analytical platform (EAP 2.0), which will leverage Big Data Technologies for satisfying future analytical reporting needs in the AML area.
A cost effective data platform for storage, processing and high performance solution for entity matching.
Entity Resolution component is highly configurable and able to process the large amount of data of any two entities.
For this Vallar Next Generation project, Citi Employee and Party were two entities for matching with 250 K Vs 55 MM of records respectively.
A Spark-SQL-Java based component for efficient processing of large volumes of data as it is a memory based computation.
Responsibilities:
Extensively involved in analysis of the requirement and prepared architect profiles.
Daily interaction with client partner on work status/reviews and requirement (FRD) understanding and also with business user/architect for test results validations.
Prepared high & low level solution design documents, provided estimations and allocated the tasks to the team and follow up on daily basis.
Prepared Technical Design Document and Testing approach documents.
Involved in reviews of code, test cases and results and preparation of test scenarios and data.
Perform application optimization on Spark Jobs by working with developers, customers and system support personnel.
Implemented different read optimized data storage formats like HIVE-Parquet & AVRO to achieve efficiency.
Implemented different file formats like CSV, Sequence, AVRO, RC, ORC, JSON and PARQUET files and different compression techniques like LZO, Gzip, Bzip2 and Snappy.
Implemented Partitioning, Dynamic Partitions and Buckets in HIVE for efficient data access.
Added many runtime configuration parameters to the Spark Job as it requires as input parameters in the cluster, based on the cluster size and resources availability.
Performed all the testing activities from Unit, SIT, Stress Test and UAT and reported.
Involved in the setup of Autosys and Auto Watch components in SIT, UAT & PROD regions.
Involved in the promotion of code into SIT, UAT & Prod environments and provided support during application runs on daily.
PROJECT : EMS - Data Profiling
Hadoop Technical Architect
Client : Travelers, Hartford, USA.
Work Location : L&T Infotech, Chennai, India.
Team Size : 4
Duration : Oct 2015 to April 2016
Domain : Insurance- Evidence Management System (EMS)
Environment : Hadoop- CDH, MapReduce, Hive, Sqoop, Java, Oozie, UNIX Shell Script, R & MySQL
Project Type : Development
Description :
This Project is all about the rehousting of their (Travelers) current existing on Ab Initio process to get Data Profiling for their Data Quality into Big Data Hadoop environment. Data Profiling (DP) leveraging Big Data Technologies for satisfying present & future Data Profiling reporting needs in the Evidence Management System.
DP as single source of truth for all Data Profiling.
A scalable storage & Data Profiling platform for working with large datasets to generate meaningful insights.
A cost effective data platform for storage, processing and reporting.
Responsibilities :
Completely involved in analysis of the requirement and prepared architect profiles.
Involved in designing SAS Conversion and Mainframe FTP jobs to transfer DDLs and host DB2 data into Hadoop environment.
Prepared Dynamic Scripts to generate External Hive tables and Loading with corresponding data.
Designed, code and tested the Data Profiling scripts, which will generates the Table Statistics for any number of tables and any number of columns that given in the Evidence Management System.
Focused on performance of the process by implementing Process ID concept to run parallel for all the tables and columns at the same time.
Involved in preparing R Scripts to generate simple dashboards like Pie & Bar graphs for Table statistics, which will give more visualization for business decisions.
PROJECT : Web Intelligence
Senior Hadoop Developer
Client : Wal-Mart, Bentonville, Arkansas, USA.
Work Location : IBM India Pvt Ltd, Hyderabad, India.
Team Size : 12
Duration : Jan 2013 to Jul 2014
Domain : Retailer
Environment : Hadoop-CDH, MapReduce, Apache Pig, Hive, Sqoop, Java, Oozie, UNIX, MySQL
Project Type : Development and Enhancements
Description :
This Project is all about the rehousting of their (Wal-Mart) current existing web project into Hadoop platform. Previously Wal-Mart was using WebHarvesting a third party JAR and MySQL Database to keep their competitor’s retailer’s information [The Crawled web data]. Earlier, Wal-Mart use to have only 4 competitor retailers namely Amazon.com, Target.com, Best Buy & Sears.
But as and when the competitor retailers are increasing, then the data generated out of their web crawling is also increased massively and which cannot be accommodate in MySQL database. Hence with the same reason, Wal-Mart wants to move it into Hadoop, where exactly we can handle massive amount of data by means of its cluster nodes and also to satisfy the scaling needs of the Wal-Mart business operations.
Responsibilities:
Completely involved in analysis of the requirement, designing, development, testing & implement & setup of Hadoop cluster.
Created two different users (hduser for performing HDFS operations and map red user for performing Map Reduce operations).
Setting up cron job to delete Hadoop logs/local old job files/cluster temp files.
Configured Hive Metastore to use Oracle database to establish multiple user connections to hive tables.
Moved all the crawled data flat files, which are generated from various retailers into HDFS for further processing by writing the Apache Pig scripts.
Created External & Managed Hive tables to store the processed results in a tabular format based on ad-hoc business request.
Developed the UNIX shell scripts for creating the simple reports from Hive data.
Developed the Sqoop scripts in order to make the interaction between Pig and MySQL Database.
Written Map Reduce code that will take input as log files and parse the logs and structure them in HIVE tabular format to facilitate effective querying on the log data and writing the Hive-Queries and UDF's in order to do advance analytics on top of the table data.
Good experience in optimizing MapReduce algorithms using Mappers, Reducers, Combiners and Partitioners.
Has extensive experience in resource allocation, scheduling project tasks, tracking and reporting and also risk analysis activities.
Have good experience in the creation and maintenance of Master test strategies/plans, test cases and test scripts.
PROJECT : Securities Information System (SIS) Canada, Test Lead
Client : SIS Canada- PL/1 Enterprises and ETL Migrations - Canada
Work Location : IBM Canada. Toronto, Canada
Team Size : 11
Duration : Dec 2009 to Dec 2012
Domain : Financial Services
Environment : Mainframes- PL/1, COBOL, JCL, DB2, IMS–DB, Informatica ETL, QTP & Oracle
Project Type : Development and Testing
Responsibilities:
Client facing role and also manage the resources both at onshore and offshore.
Preparing the Project Migration Strategy and Testing Plan as an Architect.
Compiling the mains with dual compiler and fixing errors, if any and getting reviewed by concern development leads to promote next level of the project.
Running the jobs for new and old loads, comparing the results and getting reviewed by Dev Leads for final migration.
Handled Data Center Migration Projects
Designing and optimizing the mappings, sessions to load the data into Data warehouse.
Involving in preparation of RTM and various levels of test execution like Smoke test, Real test, Retest, Sanity test, Regression test and Final Regression test
Create, test and maintain automation scripts to eliminate manual intervention
Have good experience in the creation and maintenance of Master test strategies/plans, test cases and test scripts.
PROJECT : SCNR-Visa Integration Layer, Host Senior Developer
Client : TD Bank- Canada
Work Location : IBM Canada, Toronto, Canada.
Team Size : 8
Duration : Aug 2009 – Nov 2009
Domain : Banking- Credit Cards
Project Type : Development and Support
Environment : Mainframes- PL/1, COBOL, JCL, DB2, IMS–DB, VSAM
Responsibilities:
Involving in preparation of System Specification Documents (SDS) as per the requirements.
Coding, once the SDSs were reviewed and approved and Preparing Unit Test Plan (UTP) along with Test Data and doing Unit Testing.
Preparing CMRs for UAT (K3) promotions and instruction on scheduling and supporting UAT (K3) runs and instructing the support team for failures runs.
Working on Change Requests (CRs) for business improvement.
Analyzing production abends and fixing them, once fixed and reviewed them, requesting to run in UAT (K3) environment by creating CMRs.
Be pro-active and responsive by having no late surprises, avoiding escalations and negative feedback from customers.
PROJECT : RMS- IBM Canada, Host Senior Developer
Client : IBM Global Service – Canada
Work Location : IBM Canada, Toronto, Canada.
Team Size : 9
Duration : Oct 2008 – Jul 2009
Domain : Ware house/Inventory
Project Type : Development & Application Support
Environment : Mainframes- PL/1, COBOL, JCL, DB2, IMS-DB, CICS and SCM
Responsibilities:
Preparing Internal Design Documents (IDD) as per the approved EDD.
Coding, once the IDDs were reviewed and approved.
Preparing Unit Test Plan (UTP) along with Test Data and doing Unit Testing.
Performing the System Testing activities and Integration Testing.
Working on Change Requests (CRs) for business improvement.
Supporting UAT runs and instructing the build team for failures runs.
Requesting the Build team to UAT promotion & XFR promotions and following OPS team to move modules into production.
Preparing for Configuration Audits, Access Validations, and Asset Protection Plans
Recovering the system when notified of any unscheduled maintenance within 2 hours.
Notifying the User community about the outages and updating the status in a timely manner until the system is restored.
Coordinating with the onshore team to understand and accept that task can be managed from offshore.
PROJECT : CEDS (Customer Engineering Data Server), Dev Team Lead
Client : IBM Global Services, Netherland
Work Location : IBM India, Hyderabad, India.
Team Size : 18
Duration : Jun 2006 – Sep 2008
Domain : Inventory
Project Type : Development and Enhancements
Responsibilities:
Gathering of requirements from client and clarification through calls. The aim is to freeze and accept the requirements after architect clarification.
Providing ROM sizing for the accepted requirements.
Preparing External Design Documents (EDD) and OPP for the accepted requirements and distributing them for external approvals.
Preparing Internal Design Documents (IDD) as per the approved EDD.
Coding, once the IDDs were reviewed and approved.
Preparing Unit Test Plan (UTP) along with Test Data and doing Unit Testing.
Initiating and performing all Project Quality activities such as submitting the metrics, maintaining project quality documents, participating in internal Process Reviews (PRs) and Quality Audits.
Conducting trainings on Application Knowledge and PL/I to the new members
Working on Change requests for business improvement.
Preparing for Configuration Audits, Access Validations, and Asset Protection Plans
Supporting PM in terms of project deliverables to achieve a CSAT of > 7
Preparing Project Metrics, Defect Reports and other documents related to Delivery.
Achieved Delivery Excellence and received Bravo award from IBM.
PROJECT : CRP-CVP, Dev Team Lead
Client : SPRINT NEXTEL - USA
Work Location : IBM India, Bangalore, India.
Team Size : 12
Duration : Apr 2004 – May 2006
Domain : Telecom
Project Type : Maintenance, Enhancements and Development
Responsibilities:
Along with the role of Team Lead, developed and debugged programs, jobs, etc.,
Anticipated potential problems and reacted pro-actively to ensure smooth running of the application.
Documented technical, application-specific information and also prepared test cases, plans and strategies.
Creating RFC, Work Orders and CSMR tickets for migration of changed modules
Evaluated and estimated project-tasks at offshore and planned for their timely delivery.
Applied CMM Quality methodology in the Project to ensure minimum-defects
Also handling the other team members working on UNIX platform for the same project.
PROJECT : MYSTIC - AMEX, Application Developer
Client : American Express (AMEX), Florida – USA
Work Location : IBM India, Pune, India-Service Provider- Melstar Information Technology Pvt Ltd
Team Size : 6
Duration : Aug 2003 – Mar 2004
Domain : Banking- Credit Cards- AIRLINE
Project Type : Maintenance, Enhancements and Development
Responsibilities:
Involving in Analysis of requirement and preparing document.
Coding will be done according to analysis document.
Preparing test plans/cases and data and performing Unit testing for those modules.
Finding bugs by debugging modules through expeditor tool and fixing those bugs.
Anticipated potential problems and reacted pro-actively to ensure smooth running of the application.