Post Job Free

Resume

Sign in

Business Systems Intelligence

Location:
Morrisville, NC
Posted:
February 26, 2024

Contact this candidate

Resume:

SANDEEP NATUVA

ad3xqf@r.postjobfree.com

Contact # 732-***-****

SUMMARY

●Around 14 years of experience in Client/Server business systems design/analysis/testing, Data Warehousing, and Business Intelligence applications

●11 + years of experience in working with ODS, Data warehouses/Datamarts with Extraction Transformation and Loading (ETL) processes tool like Talend 6.4/7.3, SSIS, Matillion, IBM DataStage Server 11.7/11.5/9.1/8.5/8.1/7.5/5.1, Parallel Extender PX, Informatica 7.1.3/5.2.

●Experience in Importing and Exporting the Data using SQOOP from HDFS to Relational Database management system.

●Designed and implemented solutions on Microsoft Azure that included aspects like compute, network, storage, and security.

●Worked on SnowSQL and Snowpipe.

●Used Azure Download/upload tasks to download and upload blobs.

●Proficient in Development methodologies like Agile, Scrum and Waterfall.

●Experience in using SSIS tools like Import and Export Wizard, Package Installation, and SSIS Package Designer.

●Deployed and designed pipelines through Azure data factory and debugged the process for errors

●Designed Server jobs, Job Sequencers, Batch Jobs and Parallel jobs.

●Worked with and extracted data from various database sources SQL Server 2008/2005/2000, Oracle 11g/10g/9i/8i, DB2UDB, DB2, Cobol, Teradata V2R5.0, Ms-Access, Sequential files/Flat files.

●Experienced in Developing and Monitoring Batch Jobs using UNIX Crontab and Shell Scripts.

●Involved in E-R modeling and Dimensional data modeling, design of Star Schema and Snow flake Schema. Used Erwin for physical and logical data modeling.

●Heavily involved in Matillion Development to load data into Snowflake/RedShift.

●Strong understanding of the principles of Data Warehouse using Fact Tables, Dimension Tables, star schema modeling, Ralph-Kimball approach and Bill-Inmon approach.

●Proficiency in Data warehousing techniques for data cleansing, Slowly Changing Dimension phenomenon, surrogate key assignment.

●Experience in writing, testing and implementation of procedures, functions, triggers at Database level using PL/SQ and T-SQL.

●Experience with Snowflake cloud data warehouse and AWS S3 bucket for integrating data from multiple source system like csv and tsv files into snowflake table.

●Knowledge of Developing ETL processes in AWS Glue to Migrate data from external soures like S3, ORC/Parquet/Text Files into AWS RedShift.

●Extensive Knowledge of Snowflake Database, Schema and Table structures.

●Loaded and unloaded data in and out of Snowflake using web ui.

●Creating Pipelines with GUI in Azure Data Factory V2

●Participated in the development improvement and maintenance of snowflake database applications

●Scheduling Pipelines and monitoring the data movement from source to destinations

●Transforming data in Azure Data Factory with the ADF Transformations

●Significant Multi-dimensional and Relational data modeling experience, Data Flow Diagrams, Process Models, ER diagrams with modeling tools like ERWIN & VISIO.

●Extensive experience in laying out technical specifications of an ETL process using Visio Tools.

●Hands on experience in Production Support and Maintenance.

●Experience in Performance tuning of ETL processes and DataStage jobs in SMP and MPP environments from both system as well as job design perspective.

●Develop stored procedures/views in Snowflake and use in DataStage for loading Dimensions and Facts.

●Built, deployed and maintained python codes for automation.

●Heavily involved in testing Snowflake to understand best possible way to use the cloud resources.

●Experience in Scheduling Jobs using various third party tools like ASG Zena, Control -M, Autosys.

●Creating interface design document for developed jobs (Detailed code description).

●Experience with building Hadoop clusters. Have a very good knowledge of architecture of Bigdata.

●Good Knowledge on Talend Big Data Studio to ingest data into HDFS.

●Excellent analytical and functional skills with strong communication and interpersonal skills. Actively involved in the full project life cycle, user requirements to implementation. Willing to learn and adapt to new challenges.

TECHNICAL SKILLS:

TECHNICAL SKILLSETL Tools:

MS SQL Server Integration (SSIS), IBM Information Server 11.5/9.1/8.5/8.1/8.01, Quality Stage, Ascential DataStage V7.5, Parallel Extender, Talend 6.4/7.3, Matillion

Operating System

IBM UNIX AIX 6.1/ 4.1, HP UNIX 10.2, Windows 2000/NT, Sun Solaris 12/10

Data And Analytics

Azure Synapse Analytics Workspace, Azure Data Factory, Azure Databricks, Spark, Azure Data Explorer, Azure Blob Storage, Azure Data Lake, Azure Logic Apps, Azure Functions, Azure Key Vault, Azure Purview

Databases

Azure SQL, Azure Synapse Analytics, SQL Server, Oracle, MySQL, MS Access, DB2

Languages

SQL, T-SQL, PL/SQL, PySpark, Spark SQL, Python

Other Software

AWS, Azure DevOps, Azure PowerApps, SqlDBM, Red gate, MS office, MS Visio, Erwin Studio, TOAD, SQL Assistant, SQL Developer, Visual Studio, Azure Data Studio

Operating Systems

Windows 10/7/XP

Version Control Tools

GIT, Subversion(SVN), Bitbucket Clear Case and Team city

Scheduling Tools

ASG Zena, Autosys, Cron-tab, Control- M, Tivoli

EDUCATIONAL QUALIFICATION

●BS in Computers Science, India

●MS in Engineering, Dayton, Ohio

EXPERIENCE:

Aug 2017 to Till Date

Client: Dearborn Group, DownersGrove, IL

Sr. ETL Developer

Dearborn National offers a broad selection of highly competitive insurance and financial products covering diverse markets including: Group Benefits (employer-paid/voluntary), Worksite, Individual and an array of value-added services. My Major responsibilities are processing Members data into our Dearborn National Data ware house systems and also migrated ETL jobs to Snowflake environment.

Environment: Snowflake, Matillion, DBT, Azure Data Factory, Azure Synapse, SQL Server 2016/2012/2008/2005, SQL Integration Services(SSIS), Linux shell scripting, Eclipse, ETL DataStage 11.7/11.5, Talend 6.4.1/7.3, Control M, ASG Zena, AWS Glue, Jira.

Responsibilities:

●Involved in understanding of Business processes and coordinated with Business analysts to get specific user requirements.

●Performed Peer review for ETL detail designs.

●Worked with data feeds from various source systems from flat files, XML files, DB2 databases.

●Prepared mapping documents for designing and developing the DataStage jobs

●Involved in preparing ETL Detail technical design document for development

●Involved in legacy historical data conversion from existing ODS & DW tables to new ODS & DW tables. ETL Design & use of Complex ETL routines to extract data from ODS (converting business functional specifications into mappings/workflows).

●Replicate operational tables into staging tables, Transform and load data into enterprise data warehouse tables using DataStage from their legacy systems and load the data into targets by ETL process through scheduling the workflows.

●Heavily involved in DataStage Migration projects from 7.5 to 8.5 to 11.5.

●Designed Server jobs, Job Sequencers, Batch Jobs and Parallel jobs.

●Developed Multi Instance (reusable) jobs.

●Worked with and extracted data from various database sources SQL Server 2008/2005/2000, Oracle 11g/10g/9i/8i, DB2UDB, DB2, Cobol, Teradata V2R5.0, Ms-Access, Sequential files/Flat files.

●Experienced in Developing and Monitoring Batch Jobs using UNIX Crontab and Shell Scripts.

●Involved in E-R modeling and Dimensional data modeling, design of Star Schema and Snow flake Schema. Used Erwin for physical and logical data modeling.

●Strong understanding of the principles of Data Warehouse using Fact Tables, Dimension Tables, star schema modeling, Ralph-Kimball approach and Bill-Inmon approach.

●Proficiency in Data warehousing techniques for data cleansing, Slowly Changing Dimension phenomenon, surrogate key assignment.

●Experience in writing, testing and implementation of procedures, functions, triggers at Database level using PL/SQ and T-SQL.

●Extract Transform and Load data from Sources Systems to Azure Data Storage services using a combination of Azure Data Factory, T-SQL, Spark SQL, and U-SQL Azure Data Lake Analytics. Data Ingestion to one or more Azure Services - (Azure Data Lake, Azure Storage, Azure SQL, Azure DW) and processing the data in In Azure Databricks.

●Created Pipelines in ADF using Linked Services/Datasets/Pipeline/ to Extract, Transform, and load data from different sources like Azure SQL, Blob storage, Azure SQL Data warehouse, write-back tool and backwards.

●Developed JSON Scripts for deploying the Pipeline in Azure Data Factory (ADF) that process the data using the Sql Activity.

●Developed Multi Instance (reusable) jobs.

●Worked on code commit on source control and versioning.

●Worked on migrations of complex ETL pipelines from on premise system to cloud databases like Azure SQL Database and Azure Synapse Analytics for migration of reports to cloud sources.

●Created Stored procedures in load data into Dimensions and Fact tables of the Snowflake warehouse.

●Facilitated Life Cycle Management activities like code migration, deployment of ETL components between Development, Quality Acceptance and Production systems along with their Version controlling.

●Extensively involved in studying the business requirements and develop the mapping specifications from the business requirements and the data model for reporting and analytics solutions.

●Extensively worked on building summary and reporting tables based on detail tables, to be used in developing reports.

●Created and supported event based and scheduled based pipeline triggers.

●Worked with cross functional teams to go over the data for their reporting and analytical needs.

●Coordinated with the product manager/product owner/scrum master/team lead on the status of development tasks and of any risks/issues/blocks.

.

Nov 2015 – Aug 2017

Client: BarclayCard, Willimgton, Delaware

Sr. Data Engineer

Responsibilities:

●Involved in Design and Development of Technical Specifications using Hadoop Technology.

●Involved in implementing POC’s that demonstrate the advantages Business gain by migrating to Hadoop.

●Performed real time Analysis on the incoming data.

●HiveQL Scripts to create, load and query tables in Hive.

●Involved in creating Hive Tables, loading with data and writing Hive Queries which will invoke and run MapReduce jobs in backend.

●Monitored Hadoop Eco System using Cloudera Manager.

●Loaded data into the cluster from dynamically generated files using Flume and from relational database management systems using Sqoop.

●Involved in installing and maintaining MS SQL Server 2000/2005.

●Created Data Model or the Marketing Information System Team using Erwin.

●Created SSIS package for loading the data coming from various interfaces like OMS, Orders, Adjustments and Objectives and also used multiple transformation in SSIS to collect data from various sources.

●Worked on SSIS Package, DTS Import/Export for transferring data from Database (Oracle and Text format data) to SQL Server.

●Worked on the data warehouse design and analyzed various approaches for maintaining different dimensions and facts in the process of building a data warehousing application.

●Involved in converting Hive/SQL queries into Spark transformations using Spark RDDs and Scala.

●Created multiple DataStage jobs to Parse and load data into existing DWH system.

●Designed Sequence jobs as per Barclays Standards and scheduled them using Control M .

●Coordinated with Testing Team for Functional Testing of Hadoop and ETL Code.

●Involved in Code and Peer review meetings.

●Worked closely with Developers to triage the defects.

●Worked on Build Tools like Jerkins, Maven.

●Used SVN, Tortoise SVN version control tools for code management (check-ins, checkouts and synchronizing the code with repository)

●Involved in Onsite & Offshore Coordination.

●Provided 24/7 On-call Support.

Aug 2010 – Nov 2015

Client: Dearborn National, DownersGrove, IL

Sr. ETL Consultant

Responsibilities:

●Involved in understanding of Business processes and coordinated with Business analysts to get specific user requirements.

●Performed Peer review for ETL detail designs.

●Involved in preparing ETL Detail technical design document for development

●Involved in writing datastage jobs with support to Audit, Balance and Control process for the aggregated amounts from the source and target.

●Replicate operational tables into staging tables, Transform and load data into enterprise data warehouse tables using DataStage from their legacy systems and load the data into targets by ETL process through scheduling the workflows.

●Building the pipelines to copy the data from source to destination in Azure Data Factory (ADF V1)

●Worked on creating dependencies of activities in Azure Data factory

●Responsible for designing and managing the Sqoop jobs that uploaded the data from Oracle to HDFS and Hive.

●Created Dimension, Fact and Aggregated Tables and referential integrity (or) Relations between them using PLSQL.

●Relations between them using TSQL.

●Involved in legacy historical data conversion from existing ODS & DW tables to new ODS & DW tables. ETL Design & use of Complex ETL routines to extract data from ODS (converting business functional specifications into mappings/workflows).

●Designed and developed ETL processes using DataStage designer to load data from Sources like MS SQL Server, Flat Files (Fixed Width) and XML files to Targets like Flat Files, MS SQL Server etc.

●Used DataStage stages namely Sequential file, Transformer, Aggregate, Sort, Datasets, Join, Lookup, Change Capture, Funnel, Remove Duplicate Stage, Peek, Row Generator stages in accomplishing the ETL Coding.

●Extensively used Looping Logic in Transformer Stage of DataStage V8.5 to avoid other complex Design using Transformers.

●Maintained warehouse metadata, naming standards and warehouse standards for future application development Parsing high-level design spec to simple ETL coding and mapping standards.

●Wrote SQL Scripts to extract data from source Database.

●Extensively used DS Director and Monitor for Monitoring Job logs to resolve issues.

●Developed Multi Instance (reusable) jobs.

●Effectively implemented Partitioning and Parallelism techniques to fully utilize the resources and enhance job performance

●Wrote Test Cases for ETL to compare EDW data with Target database systems.

●Used Shared and Local containers to create reusable stages.

HORIZON COMPUTER SERVICES,

New BRUNSWICK, NJ

Role: Application Developer

October 2009 – July 2010

Client: AT T Alpharetta, GA

Globus IT INC., Dallas

January 2009 – Sep 2009

Trainee as a developer



Contact this candidate