BHARGAVI ATTULURU
Cell # 401-***-**** Email: ********.**********@*****.***
PROFESSIONAL SUMMARY:
Around 12 years of IT experience focusing on the Data Engineering, design, development, enhancement, testing and maintenance of OLAP & OLTP applications using ETL, RDBMS, Data Warehousing, and Reporting tools.
Experience in all phases of the Data warehouse & Application development life cycles – Requirement Gathering, Analysis & Design, Coding, Testing, Maintenance and Documentation.
Designed and implemented end to end data solutions for source-to-Data Lake integration, defining data contracts, SLAs, and security boundaries with cross functional teams, internal and external stakeholders.
Designed, developed, and maintained ETL jobs using SSIS, IBM DataStage, Talend BIG Data and Informatica for large-scale OLAP and OLTP systems.
Extensive experience on Data migration, Data cleansing, Data Masking, Data Mining, Data validation and Data Analysis.
Experience using Rundeck and Autosys to orchestrate and monitor daily, weekly, and monthly ETL workflows, troubleshooting issues, and optimize job dependencies to improve runtime performance.
Optimized SSIS, DataStage and Talend ETL jobs by tuning job design and database queries using SQL Trace, AWR, Explain Plan, and DBMS_PROFILER reducing ETL processing time by up to 30%.
Designed and developed Python-based ETL/ELT using pandas to ingest data from APIs, files (SFTP), and databases; standardized storage in Parquet/Delta to enable analytics and auditing.
Extensive experience in Dimensional & E/R modeling methodologies
Experience in creating Reusable Transformations (Joiner, Sorter, Aggregator, Expression, Lookup, Router, Filter, Update Strategy, Sequence Generator, Normalizer and Rank) and Mappings using Informatica, Talend to move data from multiple sources into targets.
Implemented Slowly Changing Dimension Type1, Type2, Type 3 in ETL jobs for certain Dimension tables.
Extensively used PowerBI, Tableau, and SSRS for reports.
Implemented medallion architecture (bronze/silver/gold) with schema evolution and partitioning to improve data quality, lineage, and reuse across downstream analytics.
Mentored engineers supported testing and code reviews, and enforced best practices in orchestration, modular ETL design, naming/versioning standards, and error handling.
Expert in PL/SQL and SQL programming, including Tables, Views, Materialized Views, Functions, Procedures, Packages, Collections, and Triggers to implement business rules and security.
Led UAT and production deployments, preparing Deployment Operation Plans (DOP) and provided on call production support for ETL workflows, handling on-call issues and ensuring timely resolution with minimal downtime.
Established data governance, observability, and lineage using catalog/lineage tools Purview, Open Metadata and Glue Catalog with structured logging, metrics, and dashboards for auditability.
Strong collaboration and leadership skills in offshore–onsite model environments and responsible for status reporting, issue resolution, and acting as a point of contact (POC) for cross-team communication.
Fast learner with strong analytical thinking, effective communication, and the ability to explain technical issues to both technical and non-technical stakeholders.
CERTIFICATION
AWS Solution Architect-Associate
DOMAIN SKILLS
Financial Domain, Insurance Domain, Healthcare & Telecommunications Domain
EDUCATION
Master of Computer Applications (MCA), SV University, Tirupati, India
PROFESSIONAL EXPERIENCE
BPM Links Inc, USA Jan 2025 – Till Now
Role: Sr. Data Engineer
Responsibilities:
Working on all phases of the Data Warehouse, Data Lake, including Requirement Gathering, Analysis & Design, Coding, ETL, Testing, Reporting, Maintenance, and Documentation.
Designed and implemented a centralized data repository to consolidate and manage data from multiple disparate sources, ensuring a single source of truth and improving data accessibility.
Developed ETL pipelines using Talend and AWS Glue to ingest, transform, and load data into an AWS S3-based Data Lake, enabling scalable storage and efficient retrieval.
Leveraged AWS Redshift to create a high-performance data warehouse, optimizing query execution and enabling advanced analytics and reporting on user and product usage patterns.
Designed, created, and maintained databases in AWS RDS, implementing high availability configurations, automated backups, failover mechanisms, and security measures to ensure scalability, reliability, and compliance.
Automated workflows and job scheduling using Rundeck, reducing manual intervention and ensuring timely updates for ETL and data processing tasks.
Implemented data quality checks and validation processes, ensuring the integrity and consistency of ingested data and improving overall data reliability.
Optimized ETL pipelines and query execution plans by identifying bottlenecks, implementing indexing strategies, and leveraging Redshift's distribution and sort keys, resulting in a 30% improvement in data processing times.
Designed normalized and denormalized data models to support efficient querying and reporting for business-critical use cases, improving data accessibility and performance.
Collaborated with cross-functional teams, including BAs, DBAs, and external stakeholders, to gather requirements, define ETL specifications, integrate new data sources, and resolve critical production issues.
Implemented Slowly Changing Dimensions (SCD Type 1, Type 2, and Type 3) in data modeling and ETL jobs for data warehouse tables, ensuring accurate historical data tracking.
Created and optimized database procedures, functions, triggers, views, and materialized views using SQL, PL/SQL, and T-SQL to support complex business logic and reporting needs.
Designed and developed Power BI dashboards and reports, providing actionable insights to product owners, executives, and users for data-driven decision-making.
Migrated legacy data from on-premise systems to a centralized data platform, ensuring minimal downtime, data integrity, and seamless integration with modern cloud-based solutions.
Conducted performance tuning for Redshift clusters, optimizing query execution times and improving system performance by 20%, while reducing query execution times by up to 50% through advanced tuning techniques.
Implemented data security measures for S3 buckets and Redshift, including encryption, IAM roles, and lifecycle policies, ensuring compliance with organizational standards and regulatory requirements.
Environment: AWS, AWS Redshift, PostgreSQL, Talend, AWS Glue, Rundeck, SQL Server, MySQL 8.0, MySQL Workbench, PySpark, Python, UNIX, WinSCP, Git, Confluence, Jira, Amazon S3, PowerBI.
HTC Global Services Inc Jan 2021 – Nov 2024
Role: Sr. Data Engineer
Responsibilities:
Participated in Agile Scrum and Kanban methodologies across the SDLC, including status calls, pre-planning, backlog grooming, story sizing, sprint planning, and client meetings for technical and business reviews.
Designed and implemented ETL pipelines using Azure Data Factory (ADF) and SSIS to ingest, transform, and load policy, claims, and billing data from multiple sources into Azure Data Lake Storage (ADLS) and SQL Server, ensuring scalable and efficient data processing.
Developed a centralized data warehouse in Azure SQL and Azure Synapse Analytics to consolidate healthcare insurance data, enabling advanced analytics and reporting for claims processing and policy management.
Created and maintained databases in Azure SQL, implementing high availability configurations, automated backups, failover mechanisms, and security measures to ensure scalability, reliability, and compliance for transactional and analytical workloads.
Built dimensional data models for customers, claims, and policies with conformed dimensions to support cross-functional analytics and improve data consistency across reporting tools.
Gained hands-on experience with Azure Databricks and PySpark for building distributed data processing pipelines.
Optimized ETL performance using partitioned loads, bulk inserts, and set-based operations, reducing nightly refresh times from 3 hours to 20 minutes and improving data pipeline efficiency.
Designed and developed SSRS reports for claims cycle time, policy amendments, and agent performance, enabling stakeholders to make time-sensitive decisions and improve operational workflows.
Developed Power BI dashboards to provide actionable insights into claims processing, policy renewals, and customer trends, enabling data-driven decision-making for stakeholders and executives.
Implemented data quality checks (null, range, referential, and reconciliation rules) to ensure the accuracy and consistency of healthcare insurance data, reducing data defects by 35% and improving audit traceability.
Automated job scheduling and monitoring using Azure Automation, SQL Server Agent, and Rundeck, ensuring timely updates, reducing manual intervention, and improving SLA adherence.
Collaborated with cross-functional teams, including business analysts, product owners, and external stakeholders, to gather requirements, define ETL specifications, and deliver scalable data solutions.
Implemented data security measures for ADLS, Azure Synapse, and SQL Server, including encryption, role-based access control (RBAC), and lifecycle policies, ensuring compliance with healthcare data regulations (e.g., HIPAA).
Migrated data from on-premise systems to Azure SQL and Azure Synapse, ensuring minimal downtime, data integrity, and seamless integration with modern cloud-based solutions.
Conducted performance tuning for Azure Synapse workloads and SQL Server databases, optimizing query execution times and improving system performance by 20%.
Prepared comprehensive technical documentation, including ETL process flows, data mapping, and troubleshooting guides, to support audits, on-call rotations, and knowledge sharing across teams.
Implemented Slowly Changing Dimensions (SCD Type 1 and Type 2) in data modeling and ETL jobs for accurate historical tracking of policy and claims data, ensuring data reliability and consistency.
Provided ongoing support and maintenance for the centralized data platform, including monitoring, troubleshooting, and enhancements to ensure high availability and reliability for mission-critical applications.
Environment: Azure SQL, Azure Synapse, SQL Server 2022,2019, ADF, Azure Databricks, PySpark, SSIS, SSRS, SSMS, Oracle12C, MySQL 8.0, 5.7/5.6, SQL Developer, MySQL Workbench, T-SQL, UNIX, WinSCP, Confluence, Jira, Azure Blob, PowerBI.
HTC Global Services Inc July 2017 – Dec 2020
Role: Sr ETL/Database Developer
Responsibilities:
Involved in all phases of database development, including needs assessment, design, QA/QC, support, and documentation, ensuring high-quality deliverables.
Participated in Agile Scrum ceremonies, including backlog grooming, story sizing, sprint planning, and client meetings for technical and business requirements reviews.
Extensively used PL/SQL,T-SQL programming functions, procedures, packages, collections, and triggers to implement complex business rules and security measures.
Developed ETL jobs to load data from flat files and source systems into target tables, ensuring efficient data integration and transformation.
Designed physical and logical data models and data flow diagrams using Erwin, creating graphical representations of entity relationships to support application development.
Optimized code performance using Bulk SQL and Bulk Binding, improving query execution times and system efficiency.
Created reusable transformations (Joiner, Sorter, Aggregator, Lookup, Filter, etc.) and mappings using Informatica Designer, processing tasks with Workflow Manager to move data from multiple sources into targets.
Developed and scheduled ETL jobs using DataStage and informatica, automating data movement from source systems to target systems.
Extensively worked on Data Warehouse design, creating fact and dimension tables, performing data loads, and ensuring ongoing maintenance.
Leveraged SQL tuning utilities such as EXPLAIN PLAN, TKPROF, AWR, SQL Tuning Advisor, and SQL TRACE to fine-tune PL/SQL subprograms and batch jobs.
Performed optimization of long-running queries using advanced techniques, reducing response times and improving system performance.
Developed automated jobs using PL/SQL, UNIX, and Autosys to copy data from application main schema to history schema, ensuring data integrity and automation.
Written and scheduled shell scripts using cron jobs to automate routine tasks and improve operational efficiency.
Performed ETL processes and automated Data Warehousing tasks from staging to fact tables, ensuring data accuracy and reliability.
Extensively worked on UAT and production deployments, including database deployments, ETL pipelines, application code, and reports using CI/CD pipelines with Jenkins
Environment: AWS, AWS RDS, Oracle 12c, SQL Server 2012, DataStage, Autosys, DB2, PL/SQL, SQL Developer, PL/SQL Developer, Toad, TFS.
HTC Global Services Inc Oct 2013 - June 2017
Role: Database Developer
Responsibilities:
Involved in all phases of database development, including needs assessment, design, QA/QC, support, and documentation.
Designed and implemented database packages, procedures, functions, triggers, views, and materialized views using Oracle, PL/SQL and T-SQL.
Performed ETL processes, monitored and automated Data Warehousing tasks from staging to fact tables, and created ETL jobs using informatica.
Conducted data analysis for source and target systems, analyzed mapping documents, and updated coding logic to meet evolving requirements.
Optimized long-running SQL queries using advanced techniques such as Explain Plan and Hints, significantly reducing response times.
Developed and implemented data quality plans, making informed decisions on source and target data cleansing strategies.
Automated processes using DBMS_JOB, cron jobs, and workload automation tools, triggering events based on predefined schedules.
Participated in user acceptance testing (UAT), incorporating user feedback to refine coding structures and improve functionality.
Designed normalized and denormalized database schemas, ensuring scalability, data integrity, and optimal performance.
Integrated databases with external systems and APIs, enabling seamless data exchange and real-time updates.
Created comprehensive design documents for processes, identifying issues, providing solutions, and defining logical and physical table structures.
Worked in an Agile environment to fine-tune and optimize queries, resolving defects and improving system performance.
Utilized Oracle-supplied functions like PIVOT, UNPIVOT, and MODEL clauses to streamline data transformations.
Generated SQL reports, including comma-delimited reports, and integrated them with PL/SQL email routines for management.
Developed and monitored Tableau reports to provide actionable insights to users.
Environment: Oracle 11g, PL/SQL, T-SQL, MSSQL server 2005, Informatica, SQL Developer, PL/SQL Developer, Toad, TFS, OBIEE 11G, Tableau.
References:
Provided upon request.