Responsibilities:
* Participate in Team activities, Design discussions, Stand up meetings and planning Review with team.
* Provide Snowflake database technical support in developing reliable, efficient, and scalable solutions for various projects on Snowflake.
* Ingest the existing data, framework and programs from ODM EDW IOP Big data environment to the ODM EDW Snowflake environment using the best practices.
* Design and develop Snowpark features in Python, understand the requirements and iterate.
* Interface with the open-source community and contribute to Snowflake's open-source libraries including Snowpark Python and the Snowflake Python Connector.
* Create, monitor, and maintain role-based access controls, Virtual warehouses, Tasks, Snow pipe, Streams on Snowflake databases to support different use cases.
* Performance tuning of Snowflake queries and procedures. Recommending and documenting the best practices of Snowflake.
* Explore the new capabilities of Snowflake, perform POC and implement them based on business requirements.
* Responsible for creating and maintaining the Snowflake technical documentation, ensuring compliance with data governance and security policies.
* Implement Snowflake user /query log analysis, History capture, and user email alert configuration.
* Enable data governance in Snowflake, including row/column-level data security using secure views and dynamic data masking features.
* Perform data analysis, data profiling, data quality and data ingestion in various layers using big data/Hadoop/Hive/Impala queries, PySpark programs and UNIX shell scripts.
* Follow the organization coding standard document, Create mappings, sessions and workflows as per the mapping specification document.
* Perform Gap and impact analysis of ETL and IOP jobs for the new requirement and enhancements.
* Create mockup data, perform Unit testing and capture the result sets against the jobs developed in lower environment.
* Updating the production support Run book, Control M schedule document as per the production release.
* Create and update design documents, provide detail description about workflows after every production release.
* Continuously monitor the production data loads, fix the issues, update the tracker document with the issues, Identify the performance issues.
* Performance tuning long running ETL/ELT jobs by creating partitions, enabling full load and other standard approaches.
* Perform Quality assurance check, Reconciliation post data loads and communicate to vendor for receiving fixed data.
* Participate in ETL/ELT code review and design re-usable frameworks.
* Create Change requests, workplan, Test results, BCAB checklist documents for the code deployment to production environment and perform the code validation post deployment.
* Work with Snowflake Admin, Hadoop Admin, ETL and SAS admin teams for code deployments and health checks.
* Create re-usable framework for Audit Balance Control to capture Reconciliation, mapping parameters and variables, serves as single point of reference for workflows.
* Create Snowpark and PySpark programs to ingest historical and incremental data.
* Create SQOOP scripts to ingest historical data from EDW oracle database to Hadoop IOP, created HIVE tables and Impala views creation scripts for Dimension tables.
* Participate in meetings to continuously upgrade the Functional and technical expertise.
REQUIRED Skill Sets:
* Proficiency in Data Warehousing, Data migration, and Snowflake is essential for this role.
* Strong Experience in the implementation, execution, and maintenance of Data Integration technology solutions.
* Minimum (4-6) years of hands-on experience with Cloud databases.
* Minimum (2-3) years of hands-on data migration experience from the Big data environment to Snowflake environment.