Post Job Free
Sign in

Big Data Developer

Company:
Cloudinity Inc
Location:
Clinton Township, OH, 43224
Posted:
May 04, 2025
Apply

Description:

Responsibilities:

* Participate in Team activities, Design discussions, Stand up meetings and planning Review with team.

* Provide Snowflake database technical support in developing reliable, efficient, and scalable solutions for various projects on Snowflake.

* Ingest the existing data, framework and programs from ODM EDW IOP Big data environment to the ODM EDW Snowflake environment using the best practices.

* Design and develop Snowpark features in Python, understand the requirements and iterate.

* Interface with the open-source community and contribute to Snowflake's open-source libraries including Snowpark Python and the Snowflake Python Connector.

* Create, monitor, and maintain role-based access controls, Virtual warehouses, Tasks, Snow pipe, Streams on Snowflake databases to support different use cases.

* Performance tuning of Snowflake queries and procedures. Recommending and documenting the best practices of Snowflake.

* Explore the new capabilities of Snowflake, perform POC and implement them based on business requirements.

* Responsible for creating and maintaining the Snowflake technical documentation, ensuring compliance with data governance and security policies.

* Implement Snowflake user /query log analysis, History capture, and user email alert configuration.

* Enable data governance in Snowflake, including row/column-level data security using secure views and dynamic data masking features.

* Perform data analysis, data profiling, data quality and data ingestion in various layers using big data/Hadoop/Hive/Impala queries, PySpark programs and UNIX shell scripts.

* Follow the organization coding standard document, Create mappings, sessions and workflows as per the mapping specification document.

* Perform Gap and impact analysis of ETL and IOP jobs for the new requirement and enhancements.

* Create mockup data, perform Unit testing and capture the result sets against the jobs developed in lower environment.

* Updating the production support Run book, Control M schedule document as per the production release.

* Create and update design documents, provide detail description about workflows after every production release.

* Continuously monitor the production data loads, fix the issues, update the tracker document with the issues, Identify the performance issues.

* Performance tuning long running ETL/ELT jobs by creating partitions, enabling full load and other standard approaches.

* Perform Quality assurance check, Reconciliation post data loads and communicate to vendor for receiving fixed data.

* Participate in ETL/ELT code review and design re-usable frameworks.

* Create Change requests, workplan, Test results, BCAB checklist documents for the code deployment to production environment and perform the code validation post deployment.

* Work with Snowflake Admin, Hadoop Admin, ETL and SAS admin teams for code deployments and health checks.

* Create re-usable framework for Audit Balance Control to capture Reconciliation, mapping parameters and variables, serves as single point of reference for workflows.

* Create Snowpark and PySpark programs to ingest historical and incremental data.

* Create SQOOP scripts to ingest historical data from EDW oracle database to Hadoop IOP, created HIVE tables and Impala views creation scripts for Dimension tables.

* Participate in meetings to continuously upgrade the Functional and technical expertise.

REQUIRED Skill Sets:

* Proficiency in Data Warehousing, Data migration, and Snowflake is essential for this role.

* Strong Experience in the implementation, execution, and maintenance of Data Integration technology solutions.

* Minimum (4-6) years of hands-on experience with Cloud databases.

* Minimum (2-3) years of hands-on data migration experience from the Big data environment to Snowflake environment.

Apply