Post Job Free
Sign in

Data Engineer Senior

Location:
Virginia
Posted:
September 02, 2025

Contact this candidate

Resume:

Srikar Jalda

**********@*****.***

PROFESSIONAL SUMMARY:

●9 years of experience in software development which includes Design and Development of Enterprise and Web-based applications.

●Hands-on technical experience in Python, MySQL, AWS, GCP, Machine Learning modeling, DB2 SQL, Rprogramming in various domains from finance, banking, e-commerce, to healthcare.

●Experience with Amazon Web Services (AmazonEC2, AWSS3,AWSRDS,AWS Glue, AWS Kinesis, Amazon Elastic Load Balancing, Amazon SQS, AWS IAMs, Amazon SNS, AWS Cloud Watch, Amazon EBS, Amazon Cloud Front, VPC, DynamoDB, Lambda and Redshift ) .

●Experience with Google Cloud Platform for big query, cloud dataproc and Apache airflow services.

●Experience in using python integrated IDEs like PyCharm, Sublime Text, and IDLE.

●Experience in developing web applications and implementing Model View Control (MVC) architecture using server side applications Django and Flask.

●Proficient in programming languages Python, SQL, and Scala.

●Strong experience working with large datasets and designing highly scalable and optimized data modeling and data integration pipelines.

●Working knowledge on Kubernetes to deploy scale, load balance, and manage Docker containers

●Extensive experience in Data Extraction,Transforming and Loading(ETL) using various tools such as SQL Server Integration Services(SSIS),Data Transformation Services(DTS).

●Experience in Database Design and development with Business Intelligence using SQL Server Integration Services(SSIS),SQL Server Analysis Services(SSAS), OLAP Cubes, Star Schema and Snowflake Schema.

●Expertise in designing data intensive applications using Hadoop Ecosystem, Big Data Analytical, Cloud Data engineering, Data Warehouse, Data Visualization, Reporting,and Data Quality solutions

●Adept in building Data Warehouse using Star and Snowflake schemas.

●Expertise in Exploratory Data Analysis, Big Data Analytics using Spark, Predictive analysisusing Linear and Logistic Regression models and good knowledge in supervised/unsupervisedalgorithms.

●Worked on different statistical techniques like Linear/Logistic Regression, Random Forest, A/B Testing, ANOVA, Chi-Square Analysis K-means clustering.

●Hands-on experience on Visualizing the data using Power BI, Tableau, R(ggplot), Python (Pandas,matplotlib, NumPy, SciPy).

●Proficient in all phases of Software Development Life Cycle(SDLC) including Requirements,Analysis, Design, reviews,coding, Unit Testing,and Integration Testing.

●Analyzed the requirements and developed Use Cases, UML Diagrams, Class Diagrams,Sequence and Sequence and State Machine Diagrams..

●Proven leadership and people management skills with ability in resolving complex business problems.

●Direct interaction with client, offshore and onshore teams and business users across different locations from critical issues to production launches.

TECHNICALSKILLS:

BigDataTechnologies

Hadoop,HDFS, MapReduce,Hive,Sqoop,Pig,HBase,Impala,Kafka,Spark,Airflow

Cloud Technologies

AWS(S3, EC2,EMR, RedShift,Lambda,Glue, Kinesis, and more)and

GCP(Big Query,DataProc,Dataflow,Airflow), GCP (Google Cloud Platform), Snowflake

Programming Languages

Python, Scala,Java,R.

Databases

MySQL,Oracle

Development&ETLTools

Eclipse,IntelliJ,Maven,Jenkins,Tableau,ApacheAirflow,Informatica

OtherTools

Putty, WinSCP, Amazon AWS Console, Apache Ambari, PyCharm,VisualStudio,RStudio, PowerBI,SASStudio,Eclipse,Mainframes,Notebook,Databricks, Terraforms

VersionControl

GitHub,SVN,CVS

Methodologies

Agile,Waterfall

OperatingSystems

Windows,Unix,Linux

Project Management Tools

JIRA, Rally, MS Project Professional, SharePoint, Service Now, Contact Center- Genesys.

Reporting Tools

Tableau, Power BI and Advanced Excel with VBA

EDUCATION:

●Master’s degree in information technology – Southern New Hampshire University – 2024

●Bachelor’s degree in ECE – JNTUH 2015

WORKEXPERIENCE:

Bank of America, Dallas, TX June 2024 – Present

Sr. Azure DataEngineer

Responsibilities:

●Designed and implemented real-time and batch data pipelines in Azure Databricks with PySpark and Spark Structured Streaming, integrated with Kafka for AML and fraud detection.

●Built scalable ETL workflows in Azure Data Factory (ADF) to orchestrate ingestion from DB2, Oracle, PostgreSQL, and APIs into Azure Data Lake Storage.

●Developed Synapse Analytics marts and views for AML/KYC reporting, enabling fast analytical queries for compliance and business users.

●Automated CI/CD deployments using GitLab + Azure DevOps pipelines, ensuring smooth version control, testing, and release cycles.

●Implemented data quality frameworks in ADF and PySpark (not-null, uniqueness, referential integrity checks) to guarantee accuracy and reliability.

●Secured pipelines with Azure Key Vault (AKV), User-Assigned Managed Identity (UAMI), and RBAC, ensuring strict access control for sensitive PII/financial datasets.

●Integrated Azure Purview for metadata cataloging, lineage tracking, and governance of AML and trading-related data flows.

●Migrated legacy Hadoop-based AML batch jobs to Azure Databricks, achieving a 40% cost reduction and faster data processing.

●Deployed containerized ETL services with Docker + Kubernetes integrated into Databricks for elastic scaling of high-volume transactions.

●Partnered with global analysts and SMEs to validate trading/AML data requirements, applying Agile delivery for iterative enhancements.

●Built audit-ready Snowflake models for regulatory reporting, implementing RBAC, masking, and clustering for performance optimization.

●Monitored pipelines using Azure Monitor, Log Analytics, and Databricks job metrics, reducing SLA breaches and pipeline downtime by 30%.

●Integrated NoSQL (CosmosDB, MongoDB) with Snowflake to unify structured and unstructured KYC datasets for compliance reporting.

●Delivered self-service Power BI dashboards for compliance officers, powered by curated AML datasets in Synapse and Snowflake.

●Mentored junior engineers on Databricks optimization, ADF orchestration, Spark best practices, and secure Azure data architecture.

Lloyds, Hyderabad, India July 2019 – Oct 2022

Azure Data Engineer

Responsibilities:

●Architected Azure Data Factory pipelines for hybrid cloud ETL workflows

●Implemented real-time analytics using Azure Event Hubs and Stream Analytics

●Designed Azure Synapse Analytics solutions for enterprise data warehousing

●Automated data lake deployments with ARM templates and Azure DevOps

●Optimized Delta Lake performance in Azure Databricks

●Secured data pipelines using Azure Active Directory (AAD) and RBAC

●Built self-service BI platforms with Power BI Embedded and Azure Analysis Services

●Migrated on-prem SQL workloads to Azure SQL Managed Instance

●Implemented Azure Purview for end-to-end data governance and cataloging

●Scaled big data processing with HDInsight and Spark clusters

●Developed serverless APIs using Azure Functions and API Management

●Enabled ML model deployments via Azure Machine Learning Studio

●Monitored pipelines with Azure Monitor and Log Analytics

●Designed Azure Data Factory (ADF) pipelines with parameterized datasets and trigger-based execution, cutting pipeline runtime by 30%.

●Migrated SSIS packages to Azure Synapse Pipelines, implementing PolyBase for high-volume data loads into Azure SQL Data Warehouse.

●Migrated Azure Synapse workloads to Snowflake, leveraging Azure Blob Storage as a staging layer to cut query costs by 40%.

●Implemented Snowflake data clustering on Azure-hosted pipelines to optimize query performance for Power BI dashboards.

●Managed Airflow on Azure Kubernetes (AKS) to schedule ADF pipelines, using Kubernetes Pod Operator for resource-intensive tasks.

●Built idempotent Airflow DAGs with Azure Key Vault integration for secure credential management across pipelines.

ITC Infotech,Hyderabad, India Apr 2016 – June 2019

DataEngineer

Responsibilities:

●Deployed GCP Dataflow (Apache Beam) pipelines for streaming ETL, using Pub/Sub as a message queue to process 10K+ events/sec.

●Automated BigQuery ETL jobs with Cloud Composer (Airflow), integrating Dataform for SQL-based transformations and lineage tracking.

●Engineered BigQuery ML models for predictive analytics

●Connected Snowflake to BigQuery via External Tables, enabling hybrid analytics without data duplication.

●Scaled Snowflake workloads using GCP’s Cloud Composer (Airflow) to dynamically resize warehouses during peak loads.

●Orchestrated GCP Dataflow jobs with Airflow, using sensors to trigger pipelines only when source data landed in Cloud Storage.

●Monitored Airflow metrics with GCP Stack driver, setting alerts for DAG failures and latency thresholds.

●Designed Pub/Sub + Dataflow streaming pipelines for real-time data ingestion

●Automated GCP resource provisioning with Terraform and Deployment Manager

●Built multi-region data lakes using Cloud Storage and Dataproc

●Implemented Data Fusion for codeless ETL and data integration

●Optimized BigQuery performance with partitioning and clustering

●Secured data with Cloud IAM and Data Loss Prevention (DLP)

●Developed Looker dashboards for executive-level insights

●Migrated on-prem Hadoop workloads to Dataproc and BigQuery

●Orchestrated workflows with Cloud Composer (Apache Airflow)

●Implemented Data Catalog for metadata management and discovery

●Scaled AI/ML pipelines using Vertex AI and AutoML

●Monitored pipelines with Cloud Operations Suite (Stack driver)

Panamax Infotech, Hyderabad, India May 2015 – Mar2016 PythonDeveloper -

Responsibilities:

●Createdweb-basedapplicationsusingPythononDjangoframeworkfordataprocessing.

●Implementedpreprocessingproceduresalongwithdeploymentbycreating virtualmachinesusingEC2.

●Analyzed user behavior data and collaborated with program managers, business analysts, developers and other key stakeholders to develop effective productsolutions.

●Extracted data from S3 through SQL database, performed ETL using AWS Glue and utilized kinesis Apache Flink for data analytics.

●Applied SQL Queries, procedures for data manipulation, extraction, and analysis for product optimization.

●Created interactive dashboards using Power BI for data visualization, by translating complex model outputs into engaging, visual reports and enhanced business decision-making, and actions.

●Defineddataneeds,evaluateddataquality,andextracted/transformed dataforanalyticprojectsand research.

●Designedandmaintaineddatabasesusing Python and developed Python based API (RESTful Web Service) using Flask, andPostgreSQL.

●Workedonserver-sideapplicationsusingPythonprogramming.

●Employed Visual Studio Code, and Jupyter Notebook to streamline code writing and debugging processes, reducing development time and improving overall code quality.

●Conducted through software maintenance, testing, and troubleshooting, to ensure smooth operations.

●Researched and identified industry trends, providing valuable production improvement recommendations.

●Performed efficient delivery of code and continuous integration to keep in line with Agile principles.

●Experienced in Agile framework, from sprint planning and meetings to retrospectives, product backlog management and writing user stories.

●Researched and identified industry trends, providing valuable production improvement recommendations and maintained program libraries, user's manuals and technical documentation.



Contact this candidate