Post Job Free

Resume

Sign in

Data Developer

Location:
Vienna, VA
Posted:
May 24, 2020

Contact this candidate

Resume:

SUMMARY

**+ years as a Software Professional including Hadoop QA Lead, Hadoop ETL Lead, ETL Quality analyst, Business System Analyst, Java and Mainframe Tester /Developer and Onsite Coordinator

TECHNICALSKILLS SUMMARY

Big Data Engineering, Analytics

•Designing and building large-scale real time distributed data processing and analysis systems with Data warehouse, ETL, ELT, Hadoop, NoSQL and Analytical databases.

•NoSQL databases – HBase, MongoDB

•Hadoop – HDFS, Hive, Pig, Sqoop, Oozie, Flume, Spark, Ranger

•Cloud Platforms – Amazon Web Services(AWS),Azure Cloud, BLOB, ADW, AAS, Power BI and Google Analytics

•BI Reports - – MicroStrategy and Tableau

•Automation testing – Selenium WebDriver, TestNG, GitHub, Alteryx, Python and Unix

•Real Time Analytics – Spark, JITP, Tableau

•Data Science Languages - R, R Studio, R Shiny and Python

•Machine Learning Algorithms - Statistics and Machine Learning - Descriptive Statistics, Inferential Statistics, Prediction, Logistics Regression, K-Means Clustering, Na ıve Bayes and Decision Trees, CNN and RNN

Java, Oracle and Mainframe Technologies

•Java Developer – MS Windows, Java, JSP, Servlets, HTML, JavaScript, EJB, WebSphere, Oracle, Eclipse

•Architecture - SOA Architecture, J2EE Architecture(MVC), Client Server

•Web Services – XML RPC, SOAPUI, REST API

•Mainframe Developer – MVS, MVS/ESA, Cobol, CICS,DB2, Visio, MS Project, VSS, PVCS, MS-Word, MS Excel, PowerPoint, Crystal Reports, CA-Interest, Easytrieve, XPEDITER

•Data Modeling – Physical and logical modeling, Normalization, ERDs.

•Relational Databases – Oracle, DB2, MySQL, Informatica, Oracle Developer

•Quality Tools– Hive Query Automation, Quality Centre(ALM), Selenium,QTP,LoadRunner, Rational ClearCase, ClearQuest, Robot, Talend, JUnit, MRUnit, MS Office, MS Project

BUSINESS MANAGEMENTSKILLS SUMMARY

Project and People Management

•15+ years of managing teams and clients.

•Managed teams with 5- 25 employees in multi- cultural environment in developed (US) and Emerging (India) markets at offshore and onsite locations.

•Enjoys mentoring, building effective teams and retaining them.

•Agile project management, Performance management.

•Known for being process oriented, project planning, delivering on time with quality.

Business Development

•Entrepreneurially inclined in finding and solving problems.

•Known for customer relationship management and retaining many for 10+ years.

Personal Skills

•Active in Business Intelligence and Big Data community in US and India

•Enjoys writing blogs, articles in newspapers and magazines.

Enjoys learning what I don’t know and teaching what I know.

SELECTIVE ACCOMPLISHMENTS

Data aNd Analytics (DNA), Revenue Management, Hilton Inc., McLean, VA Jan’18 – Till Now Lead Test Engineer

Responsible for gathering requirements from the client and translate the business requirements into technical design and perform various transformations on the incoming data.

Validate Data Ingestion from external Source to Target Hive

Validate data in Hive data after Spark jobs are run.

Validate data between Hive and Amazon Redshift warehouse by using Alteryx workflow for a full dataset

Validate data between AWS Redshift and MSTR Cubes/Dashboards/Views/Dossiers by using Alteryx workflow for a full dataset

Validate Metrics and Derived Metrics between AWS Redshift Vs. Dashboards by using SQL

Validate Mapping transformations between file to Hive and Hive to AWS Redshift and Redshift to MicroStrategy Cubes/Dashboards

Validate ETL Transformations between Hive to AWS Redshift and Redshift to Tableau/MicroStrategy Cubes/MSTR Dashboards/Dossiers/Views

Automated functional tests to be added to regression using Selenium WebDriver

Developed automation scripts using Selenium and TestNG for common and repeatable tasks for Login, create customer/account, balance enquiry and certain service changes

Created and designed Behavior driven tests cases using Cucumber and Selenium WebDriver

Implemented the open source approach with Selenium

Involved in automation infrastructure setup using Selenium with Java

Validate BI Dashboards functionally, data accuracy wise and compatibility wise

Worked in Agile methodology, test cases were executed in sprints and stories were updated on Jira board.

Lead the change management processes and facilitate the transition of the project to the Operations teams.

Participate Agile Sprint Planning, come up with user stories, provide estimations for stories

Participate in discovery meetings while brain storm on business requirements, data modeling and ETL code reviews

Participate in Story grooming sessions and retrospective meetings, provide feedback

Prepare test data by running Spark jobs

Validate Data science apps using R, Python for different algorithms like random forest classifier and linear regression

Provide status report weekly basis on Testing status, resource status and defect status in Jira, service-now

Coordinate offshore onsite tasks

Prepare test approach, strategy for E2E IMT project and Power BI Reports

Technologies: Hortonworks, HDP, Ranger, Hadoop, HDFS, Hive, Yarn, AWS, Redshift Warehouse, S3, MicroStrategy Cubes, Tableau, Cubes, MSTR BI Reports, Unix, Jira, MS Excel, MS Word

Inventory Management Tool, Gap Inc., San Francisco, CA Oct’17 – Dec’17 Business Intelligence QE/Onsite Lead/Azure Consultant

Responsible for gathering requirements from the client and translate the business requirements into technical design and perform various transformations on the incoming data.

Validated the data in Hadoop by executing ETL jobs in Data Stage.

Perform data validation to test the ingestion framework which populates the data from Big data and Teradata to Azure Blob and Azure Data Warehouse.

Performed data validations in SSAS cubes by executing MDX queries.

Worked in Agile methodology, test cases were executed in sprints and stories were updated on Jira board.

Lead the change management processes and facilitate the transition of the project to the Operations teams.

Validate Migrating Data from various data sources (Big Data, TeraData, Oracle Venus and Essbase) to Azure Data Lake Service(BLOB) Cloud Platform

Validate data lake after data ingestion by using talend

Validate Implementation of standardized metric definitions in ADW and AAS

Validate Power BI Reports on top of Azure Analysis Services(AAS) Layer

Validate Batch streaming data into Power BI Reports

Validation of Power BI reports including Report or Dashboard or Layout Design Check, Prompts Check, Report data accuracy Check and Browser Compatibility Check

Validate user access security, Power BI Reports Load Testing

Coordinate offshore onsite tasks

Prepare test approach, strategy for E2E IMT project and Power BI Reports

Technologies: Hortonworks, Cygwin, Win SCP/Putty, Ultra-Edit, Chrome Browser, HDP, Ranger, Hadoop, HDFS, HUE, Hive, Yarn Mapreduce 2), PIG, HCat, Azure Cloud, BLOB, ADW, AAS, Power BI Reports, Unix, Jira, Polybase, MS Excel, MS Word

Home Quotes, Progressive, Mayfield Village, OH Feb’16 – Sept’17

Big Data/Hadoop QA Lead

Participate in Project planning and understand requirements

Provide test strategy, test plan and testing approach for a Data Ingestion and Incremental approach

Validate Hive table schema, ETL transformation rules and aggregations

Execute test cases and Interact with developers on defect resolution

Provide root cause analysis for defects

Develop Functional, System Integration, End to End, Regression Testing test cases

Validate data lake and data access layer after data ingestion by using Java framework

Perform Unified Functional Testing by using nightly jobs

Validate upgrade of HDP components stack such as HDFS, Hive, Hue, PIG, HBase, Spark and TEZ

Provide mockup data for different testing data driven scenarios

Developed new and updated existing Automation test cases using Selenium Web Driver and TestNG, using SVN as VCS for a Java based Web application, using Maven as build tool and Jenkins to create and run deployment jobs

Worked on Continuous Integration using Jenkins to execute automation scripts using Selenium and SOAP UI

Version controlling of scripts using GIT, STASH

Provide overview on project testing approach, setting up testing standards and processes

Build Data Science testing competency, mentor Testers, review performance and train testers on Data Science and Hadoop technologies

Provide framework for Automation/Regression testing by suing Unix and Python

Validate R shiny web applications functionality, dashboards, graphs using R Studio

Training and mentoring resources to ensure they are up to date with the changing environment, technology, and best practices for Hadoop testing.

Planning for daily releases, resource allocation for large number of initiatives and adapting to aggressive timelines.

Utilization of HP Quality Center for Test Case storage, allocation and defect logging and management

Responsible for providing analysis and metrics for projects and overall release on an ongoing basis

Technologies: Hortonworks, Cygwin, Win SCP/Putty,Ultra-Edit,Chrome Browser,HDP, Ranger, Hadoop, HDFS, HUE, Hive, Yarn(Mapreduce 2),PIG, HCat,HBAse, Spark, Unix, Selenium WebDriver, SOAPUI,GIT,STASH, HP ALM 12.21,DB2, SQL Server,Polybase, SSIS, PDW, Syncsort, MS Excel, MS Word

Predictive Analytics and Data Architecture (PADA), Zurich, Schaumburg, IL Feb’15 – Jan’16

Hadoop QA Lead

Participate in Project planning and understand requirements

Provide test strategy, test plan and testing approach for a Data Ingestion and Incremental approach

Validate data landing zone, PIG location and hive tables

Validate data lake and data access layer after ingesting data by using Python framework

Validate input data streams deposited on HDFS storage(landing zone)

Validate PIG UDFs from landing zone

Validate hive tables from PIG UDFs

Validate Hive table schema, transformation rules and aggregations

Execute test cases and Interact with developers on defect resolution

Provide root cause analysis for defects

Develop Functional, System Integration, End to End, Regression Testing test cases

Review Test Plan, strategy and Test Scripts

Coordinate with developers on defects tracking on ALM 12.21

Perform validation on QA, Prod and Research environments

Create test environment, copy data from one HDFS node to another node using file zilla and winscp

Create Hive tables and partitions using HQL

Provide mockup data for different scenarios

Coordinate with offshore testing team on day to day tasks

Mentor offshore team on validation tasks and make sure deliverables delivered timely

Provide overview on project testing approach, setting up testing standards and processes

Technologies: Hortonworks, HDP 2.6.0, Ranger, Hadoop, HDFS, HUE, Hive, Yarn(Mapreduce 2),PIG, HCat, Spark, Unix, HP ALM 12.21,DB2, SQL Server, MS Excel, MS Word

network Digital Video Recording (nDVR), Arris, Beaverton, Oregon Aug ’14 – Jan’15

QA Analytics Principal Hadoop Lead

Participate in Scrum planning, user story creation, retrospective meetings

Provide test strategy, test plan and testing approach for a brand new Hadoop application

Develop Automation Test framework, prototype and implement

Validate data intake layer, processing layer and output layer

Validate JSON and XML inputs, validate Avro formatted data

Validate input data streams deposited on HDFS storage

Validate Mapreduce Uber table avro schema data

Validate Hive table schema, transformation rules, aggregations against Uber table aggregations

Validate Tableau BI Reports against Hive table aggregations and transformation rules

Validate Analytics Tableau Data Analytics Batch Reports and real time reports

Validate Tableau Reports Skeleton, Legends, aggregations against Hive data

Develop manual and automated test scripts for regression of existing functionality, bug isolation and new functionality

Execute test cases and Interact with developers on defect resolution

Research new features and technologies for analysis and implementation. Perform feasibility studies on new technology, tools and product concepts

Develop Functional, System Integration, End to End, Regression Testing test cases

Review Test Plan, strategy and Test Scripts

Technologies: Spring-XD, RabbitMQ, Scala, Uber Table, Hadoop, HDFS, Hive, Yarn(Mapreduce 2), Spark, Postgres, Docker Image, GitHub, Cloudera Manager,XML, Json, Avro, Python, Realtime Analytics, Tableau Reports, RQM, RTC Scrum and Cloud Platform

SBM –Data Clickstream, Intuit, Menlo Park, CA: Feb’14 – July‘14

QA Lead Engineer

Participate in Scrum planning, user story creation, retrospective meetings

Analyzed Business Requirements Documents,FSD,PDM and Data Mapping Document

Interact with Business User to get requirements

Write Test strategy, test plan and approach for Hadoop data application data sets

Validate Raw data against refined data

Load data into Hive tables using HQL scripts

Cleanse data with Pig scripts

Design Test Plan, Test Strategy for Adlens CD,TD, TList and Plan and Forecast Datasets

Develop Functional, Regression, End to End Test scripts using Hive Query Language

Review Test Plan, strategy and Test Scripts

Run Oozie workflow to perform pig action, hive actions and loading data into table

Execute test cases and interact with developers on defects tracking

Perform Avro Schema validation using Pig Scripts for Data quality on 3rd party Datasets

Perform Functional, Integration, regression and End to End testing using Automated Hive scripts

Validate Data Mapping between Hive, Vertica(MPP) Data Marts and Tableau BI Reports

Automate Sqooping from Informatica to HDFS using python scripts

Automate validation of data from Oracle/Informatica to HDFS/AWS using python scripts

Validated Business Intelligence Tableau reports skeleton, legends, aggregation rules for Data Analytics

Validation of reports after ETL/transformation work flows are executed for all source systems.

Validation of non-functional testing like performance and failover testing plays a key role to ensure the whole process is scalable and is happening within specified SLA.

Technologies: Hadoop, AWS, HDFS, Hive, Pig, Avro schema, Sqoop,Informatica, Vertica, Tableau,Jira, Silk Centre, Rally Scrum, CSV file, Unix, Python, Talend

Abbot Laboratories,Chicago,IL Jan’ 13 – Dec’13

Lead Quality Analyst

Designingthe architecture of a large implementation of Big Data Analytics platform – NoSQL and Analytical databases, Data Visualization tools.

Perform Functional testing activities such as pre-Hadoop processing, map reduce process, structured and unstructured validation.

Validation of pre-Hadoop processing to ensure that that data is getting processed without any errors.

Validation of data from different source systems and loading into HDFS using Sqoop

Validation of Hadoop Map Reduce process data output processing to ensure that that data is getting processed without any errors

Perform MapReduce operations involve in processing the input files and applying map and reduce operations to get desired output.

Responsible to gauge the Test progress activities and forecast the Risks involved in the testing process;

Participate in QA best practice initiatives across QA organization.

Validation of transformation rules are applied correctly.

Validation of reports after ETL/transformation work flows are executed for all source systems.

Data Storage validation is done to ensure that the data is correct and is of good quality.

Validation of non-functional testing like performance and failover testing plays a key role to ensure the whole process is scalable and is happening within specified SLA.

Manage team of 5 testers

Technologies: Informatica, Oracle Developer, Hadoop, HDFS, MapReduce, Hive, Talend, Pig, Sqoop, Oozie, MRUnit, Quality Centre(ALM), Agile, UNIX

Cloudwick Technologies Inc., Newark, CA Jan’12 – Dec’12

ETL Sr. QA Analyst

Facilitated status meetings, metrics reporting, and status reporting.

Gathered business requirements from users and created designs for enhancements and to remedy defects

Converted business requirements in use cases

Developed application using R language

Developed Test case scenarios using use cases

The project generates reports from various unstructured sources

Executed unit, integration and performance testing using R language scripts

Students performance graphs were tested using R language script and built in functions

Faculty performance graphs were tested using R language script and built functions

Find predictive analytics using students historical data

Technologies:ETL, Informatica, Oracle Developer, Hadoop, HDFS, Hive, Pig,Sqoop, Flume,Avro Schema, Quality Centre

RouteOne, Farmington, MI Jun’10 – July ‘11

Business Systems Analyst

Responsible for primary development of testing strategies.

Create test plans and scripts for the projects.

Perform moderately complex to complex test data conditioning, regression testing, and testing validation.

Provide specific guidance on defects to developers.

Give direction to less experienced analysts.

Logs, tracks, and verifies resolution of software and specification defects.

Documents all phases of the Systems QA process.

Facilitate requirements and design sessions with both business and technical team members.

Act as the go-to person for requirements questions and clarification. In many cases, become an expert on specific features and participate in changes even after implementation to production.

Analyze high level business requirements and develop lower level functional and technical requirements. Various deliverables include System Requirement Specifications, User Interface Specifications, screen mock-ups, report mockups, process and navigational flows, etc.

Determine impact of functional requirements in collaboration with other feature team members on various aspects of existing system.

Technologies: Java, JSP, Web Services, XML, XSLT, Websphere, Eclipse, EJB, Quality Centre

Blue Cross Blue Shield, Detroit, MI Jan ’06 – May ’10

Sr. QA Analyst / QA Tester

Responsible for Institutional, Professional claims (HIPAA compliance) system to validate

Reviewed Use cases, design documents and Prepared a test plan

Prepared test plans for SIT and UAT tests

Validated batch results against backend database tables

Ran Batch mainframe jobs and validated batch results

Performed SIT, Regression, Performance and UAT tests in the mainframe environment

Performed defect tracking with development team lead

Executed JCL to run batch jobs in the testing environment

Prepared test data for different claim types testing

Done research on different bugs causing problems in the batch claims system

Technologies: MVS/ESA, Cobol, DB2, VSAM, Java, Servlets, JSP, Quality Center, Java, JSP, Web Services, Websphere

Ford Motor Company, Dearborn, MI, Oct’ 04 – Dec ‘05

Quality Analyst / Tester

Reviewed requirements, design documents and Prepared a test plan

Written Use cases using Rational Rose

Developed System threads for Batch Jobs to integrate multiple systems

Prepared Test data for Batch jobs and online screens

Performed regression testing

Managed a team of eight people

Performed Auditing for test cases

Ran Batch jobs using UNIX script for ETL jobs

Validated batch results against backend database tables

Entered defects in the Test Director and communicated with corresponding Development team

Perform Quality Assurance Inspections, Audits, and Reviews

Assist in Ongoing Process Improvements to Ensure Test Planning, Execution, and Reporting is Effective and Efficient, Standardized, Coordinated, and Integrated.

Involved in RUP based SDM methodology implementation

Technologies: MVS/ESA, Cobol, DB2, VSAM, Java, Servlets, DataStage, Business Objects, Quality Center, Websphere

GE Financial Assurance, Richmond, VA Jan ’01 – Sep ‘04

Mainframe/Java Developer

Worked closely with software developers to resolve software problems and issues uncovered during testing.

Utilized client/server and PC-based action tracking software to analyze and report testing status to management and clients.

Developed test cases and test scripts to test financial aid software while maintaining a focus on quality assurance.

Developed HTML, JSP and JavaScript for front-end design.

Wrote code according to MVC architecture.

Worked with application architects and DBAs to understand the architecture of the application as well as the database design.

EDUCATION & TRAINING

• BS in Computer Science Engineering, NIT Warangal, India

• Relevant Courses:

Hadoop Administration (Cloudwick), Computing for Data Analysis (Coursera), Predictive Analytics, Statistics One (Course), Machine Learning (Coursera)



Contact this candidate