CHANDANA CH
DATA ENGINEER
**************@*****.*** +1-732-***-**** LOCATION - EDISON, NJ/NY
PROFESSIONAL SUMMARY
• Over 3+ years of professional experience as a Data Engineer, delivering scalable, data-driven solutions and driving business growth through advanced analytics.
• Proficient in SQL, Python, Power BI, and Tableau for analyzing and visualizing complex datasets, enabling actionable insights and strategic decision-making.
• Expertise in data warehousing concepts, including data modeling, ETL/ELT processes, and building scalable pipelines using Azure Data Factory, Azure Databricks, and PySpark.
• Skilled in big data tools such as Apache Spark, Delta Lake, and Snowflake, optimizing performance and ensuring data integrity across large-scale systems.
• Hands-on experience with Azure Cloud Services (Data Lake Storage Gen2, Blob Storage, Synapse Analytics) and Databricks Asset Bundles, ensuring seamless deployment and governance.
• Proven ability to design dynamic dashboards in Power BI and deliver real-time data solutions using Spark Streaming, enabling efficient monitoring and reporting.
• Utilized AWS Glue to integrate and transform data from multiple sources (S3, RDS, Redshift), improving data accessibility and streamlining ETL workflows.
• Leveraged Amazon S3 and Redshift to centralize data storage, ensuring high availability, scalability, and secure access.
• Implemented AWS Lambda functions to automate data processing tasks, reducing manual effort and enhancing pipeline efficiency.
• Demonstrated success in automating workflows and data pipelines, reducing manual efforts by 40% through advanced Python scripting and ETL optimization.
• Experienced in collaborating with cross-functional teams to identify data-driven opportunities, define KPIs, and implement analytics solutions that improve operational efficiency.
• Certified in Azure Data Engineering, showcasing a commitment to continuous learning and proficiency in cutting-edge technologies.
TECHNICAL SKILLS
Databases: MySQL, MongoDB, SQL Server, Redshift
Data Warehousing: Data modeling, ETL/ELT processes, Azure Data Factory, Azure Databricks, Azure Synapse Analytics, Data Lake Storage Gen2
Big Data Tools: Apache Spark, PySpark, Databricks (Delta Lake, Unity Catalog, Lakehouse Architecture, Delta Live Pipelines), Snowflake
Cloud Services: AWS,Microsoft Azure (Active Directory, Data Factory, Synapse Analytics, Databricks, Data Lake Storage Gen2, Blob Storage), AWS S3
Programming/Scripting: Python (Pandas, NumPy, Pytest), SQL Version Control and Deployment: GitLab, Databricks Asset Bundles Data Visualization: Tableau, Power BI
Development Tools: Visual Studio Code
PROFESSIONAL EXPERIENCE
Client: Con Edison, New York, USA
Role: Data Engineer December 2023 – Present
• Leveraged Power BI, Excel, SQL, and Python to analyze data, driving decisions and generating $500K in cost savings.
• Designed visually impactful dashboards in Power BI and Excel, boosting team efficiency by 20% and reducing reporting time by 30%.
• Developed automated ETL pipelines using Azure Data Factory and PySpark to process large datasets, ensuring scalability and reducing manual efforts by 40%.
• Conducted regression analysis and market forecasting, contributing to a 15% increase in revenue through data-driven strategic decisions.
• Utilized AWS Glue to integrate and transform data from multiple sources (S3, RDS, Redshift), improving data accessibility and streamlining ETL workflows.
• Leveraged Amazon S3 and Redshift to centralize data storage, ensuring high availability, scalability, and secure access.
• Implemented AWS Lambda functions to automate data processing tasks, reducing manual effort and enhancing pipeline efficiency.
• Enhanced data quality by implementing Python validation scripts, reducing discrepancies by 25%.
• Conducted A/B testing and trend analysis to optimize marketing campaigns, resulting in a 10% improvement in conversion rates.
• Partnered with cross-functional teams to identify data-driven opportunities, influencing strategic business decisions.
• Trained junior analysts on data visualization tools and analytical methodologies, fostering team development and growth.
Client: VIRTUSA, Hyderabad, India
Role: Data Engineer March 2021 – April 2022
• Designed and optimized ETL pipelines using Azure Data Factory and Databricks, improving data processing efficiency by 30%.
• Built and maintained scalable data warehouses to support enterprise-wide analytics, ensuring data integrity and availability.
• Automated repetitive tasks using Python scripting, saving over 100 hours annually and improving workflow efficiency.
• Utilized Azure Data Factory to integrate data from multiple sources (e.g., Azure Blob Storage, ADLS), improving data accessibility and governance.
• Developed interactive dashboards with Power BI, delivering actionable insights that enhanced decision- making across departments.
• Migrated on-premises data infrastructure to Azure Cloud, enhancing scalability and reducing infrastructure costs by 25%.
• Preprocessed large datasets with PySpark, collaborating with data scientists to improve predictive model accuracy by 15%.
• Implemented data quality checks and validation frameworks, reducing discrepancies by 20%.
• Conducted root cause analysis on data anomalies, deploying fixes that improved pipeline stability by 30%.
• Integrated data from APIs, flat files, and Azure Data Lake Storage Gen2, creating unified data repositories for faster analytics.
• Established data governance policies, ensuring compliance with regulatory requirements. EDUCATION
Pace University — New York, USA
Master of Science, Information Systems GPA: 3.84/4 Dec 2023 Relevant Coursework: Information Systems Design & Development, Legal Issues in Information Systems, Database Design, Python Programming, Information Systems Organizational Strategy, Project Management Vaagdevi College of Engineering — Warangal, India
Bachelor of Technology, Computer Science July 2021 PROJECTS
Capstone Research Project AI-Powered Chatbots, Mixed-Methods Research
- Developed process models and reusable architecture, reducing testing time by 20%, and demonstrated stellar collaboration and planning skills using the Scaled Agile Framework.
- Conducted mixed-methods research analyzing surveys from 56 medical professionals on implementing AI-powered chatbots in healthcare using thematic analysis.
- Performed qualitative data analysis and coding to derive insights, providing actionable recommendations for integrating AI chatbots in healthcare facilities.
Verizon Cloud Platform Job Simulation Python, Cloud-Native Traits, Application Security
- Completed a job simulation for Verizon’s Cloud Computing team by building a hypothetical new VPN product, testing for redundancy, resiliency, and least privilege using command line Python.
- Researched approaches to achieve application security and effectively communicated insights through a comprehensive PowerPoint presentation.
CERTIFICATIONS
- Microsoft Certified – Azure Data Engineer Associate
- R for Data Science: Analysis and Visualization
- UiPath: Robotic Process Automation (RPA)