Post Job Free

Resume

Sign in

Data Engineer Project Management

Location:
Miami, FL
Salary:
110000
Posted:
February 02, 2024

Contact this candidate

Resume:

309-***-****

ad3be7@r.postjobfree.com

Kent, OH

https://www.linkedin.com/in/ajit-kolluri/

Ajit Kolluri

Data Engineer

SUMMARY:

Experienced Data Engineer specialist with 5 years of experience. Enhanced data analysis capabilities through cloud technologies and reporting tools, creating insightful data-driven insights. Collaborated closely with business owners, developed critical dashboards in Power BI and Tableau, designed ER diagrams for dimensional models, and played a key role in KPI strategy. Managed end-to-end data lifecycle in projects, including migration to Microsoft tech landscape, hands-on experience with Databricks, and successful migration of ETLs to Azure Cloud. Proven expertise in optimizing Hive SQL queries, Spark Jobs, distributed transactions, high availability setups, and AWS utilization for large dataset ETL processing, system reliability. KEY SKILLS:

Platforms AWS, Linux, Windows, Azure

Languages Java, JavaScript, Python

Relational Databases PostgreSQL, MySQL, Snowflake, Oracle, MariaDB, Aurora, Azure SQL, NoSQL Databases DynamoDB, MongoDB

Version Control Git/GitHub/Azure DevOps

Project Management Rally, Jira, Agile, Wireframing, Road mapping Cloud/AWS/Azure

Compute Azure VM, Azure Functions, EC2, Lambda, ELB, Auto Scaling Azure Data Factory, Synapse Analytics, Event Hubs, Data Lake Storage, Databricks Storage ADLS, S3, EBS, ADLS, Blob storage,

Database DynamoDB, RDS, Redshift, Neptune, Azure SQL Networking VPC, ELB, Route 53, Azure VN

Analytics Kinesis, Glue, EMR, Databricks, Azure Data Factory, Azure HDInsight, Azure Synapse

Mobile SNS

Cloud Management Code Deploy, Cloud Watch, Cloud Formation, ARM Security Identity & Access Management (IAM), AAD, Managed identities, service principal

App Services SQS, SNS

Server Side

Technologies JSP, JDBC

Frameworks Spring MVC, Spring Boot, Hibernate, Spark, Pandas, Spark Application Servers Tomcat

IDE IntelliJ IDEA, Eclipse

Build / CI/CD Azure DevOps

Testing Junit

Monitoring AppDynamics, Kibana

PROFESSIONAL EXPERIENCE:

Data Engineer Oct 2022 - Current

UTIMCO – Contractor role

• Responsible for ensuring Systems, Network Security, maintaining performance & setting up monitoring using Cloud Watch

• Built S3 buckets and managed policies for S3 buckets and used S3 bucket and Glacier for storage and backup on AWS.

• Designed and setup Enterprise Data Lake to provide support for various uses cases including Analytics, processing, storing, and Reporting of voluminous, rapidly changing data.

• Responsible for maintaining quality reference data in source by performing operations such as cleaning, transformation and ensuring Integrity in a relational environment by working closely with the stakeholders &solution architect.

• Performed end- to-end Architecture & implementation assessment of various AWS services like EMR, Redshift and S3

• Implemented machine learning algorithms using python to predict the quantity a user might want to order for a specific item so we can automatically suggest using kinesis firehose and S3.

• Used AWS EMR to transform and move large amounts of data into and out of other AWS data stores and databases, such as Amazon Simple Storage Service (Amazon S3) and Amazon DynamoDB

• Creating Lambda functions with Boto3 to deregister unused AMIs in application regions to reduce the cost for EC2 resources.

• Designed Redshift distribution types, implemented Snowflake DW, and created DataMart for analytics with regular data ingestion and report generation.

Data Engineer Feb 2022 – Oct 2022

Unified Door & Hardware Group – Contractor role

• Processed Data into AWS S3 by developing solutions, analyzed the data using spark-sql.

• Import Data from various systems/sources in on-prem into EDL.

• Implemented multi-processing and multi-threading in Databricks notebooks for huge loads of data ingestion.

• Extract Transform and Load data from sources Systems to S3 using a combination of AWS DMS, T-SQL, Spark SQL.

• Data ingestion to enterprise EDL using data bricks.

• Create and maintain optimal data pipeline architectures Build the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data.

• Involved in creating Table and then applied HiveQL on those tables for Data validation.

• Involving on loading and transforming large sets of structured, semi structured and unstructured data

• Extract, Parsing, Cleaning and ingest data.

• Develop Spark applications using PySpark and spark SQL for data extraction, transformation, and aggregation from multiple file formats for analyzing and transforming the data uncover insight into the customer usage patterns.

• Provisioning Hadoop and Spark clusters to build the On-Demand Data warehouse and provide the data-to Data scientist. Data Engineer Aug 2021 – Feb 2022

Eli Lilly and Company – Contractor role

• Developed a Privacy Playbook for managing Sensitive Data, ensuring compliance with global laws and regulations.

• Contributed to Privacy customer experiences and standards, drove delivery of a Privacy capability framework, and had hands-on experience with the Privitar data privacy tool.

• Utilized Privitar for PII deidentification, configured Spark clusters on HDInsight, and maintained optimal data pipeline architecture on Azure.

• Created ADF pipeline as a data privacy template, leveraged Azure Data Factory and Databricks for complex data sets, and managed secret key-vault for PII tokenized data

• Demonstrated expertise in data ingestion, Azure storage technologies, and architecting BI solutions on Azure Data Platform. Assemble large, complex data sets that meet functional / non-functional business requirements.

• Implemented data security and governance policies, migrated data to Azure databases, and recreated application logic in Azure environments.

Information Technology GA May 2020 – Jul 2021

Northern Illinois university

• Established external stages in Snowflake for seamless data integration with Snowpipe and SQS events, dynamically scaling clusters as per business needs.

• Implemented Data Vault 2.0 data model for efficient migration to Snowflake from legacy databases.

• Managed roles and privileges in Snowflake, establishing connections with other cloud applications.

• Utilized Change Data Capture in Striim to ingest continuous data from PostgreSQL to Snowflake, monitoring data behavior. Created a metrics dashboard with Grafana for time-based monitoring.

• Converted large Parquet files to CSV using Azure Data Factory, storing in Azure Blob Storage, and imported to PostgreSQL.

• Developed interactive dashboards with Python libraries (plotly-dash, seaborn, matplotlib) and utilized tkinter for UI. Data Engineer Aug 2018 – Dec 2019

Cloud4c – Contractor role

• Managed large datasets in multi-million-row databases, meeting tight deadlines for import, export, and manipulation.

• Supported multiple ERP systems, developed and updated internal applications, and presented analyses using SQL, SSIS, MS Access, Excel, and PowerBI dashboards.

• Automated tools and scripts to enhance departmental efficiency and collaborated with cross-functional teams on project data gathering.

• Generated daily reports in Excel with pivot tables, built SSIS packages for data warehouse ETL, and automated Power BI dashboards.

• Addressed operational data issues, communicated as a technical and functional expert, and optimized processes in the Data Warehouse for data retrieval and analysis in the Data Mart System. EDUCATION:

M.S. Operations and Management Information Systems Apr 2021 Northern Illinois University, Dekalb, Il (3.84/4.00)

• College of Business Dean’s list (with > 3.75 GPA every semester)

• OM&IS esteemed student scholarship.

Bachelor degree Apr 2019

JNTU, Hyderabad, India



Contact this candidate