Post Job Free
Sign in

Data Analyst Power Bi

Location:
Fort Worth, TX
Posted:
August 05, 2025

Contact this candidate

Resume:

Dhanusha Punyala

Dallas, Texas, ***** +1-682-***-**** ***************@*****.***

PROFILE

Developed PySpark and Spark SQL code to process the data in Apache Spark on Databricks to perform the necessary transformations based on the STMs developed.

Good understanding of Spark Architecture including Spark Core, Spark SQL, Data Frames, Spark Streaming, Driver Node, Worker Node, Stages, Executors and Tasks.

Managed and verified data integrity and cleanliness to optimize data accuracy by 15%

Good understanding in cloud computing on GCP with various Technologies like Data flow, Big query, Cloud Data storage.

Experience in using GCP Cloud Dataproc for big data processing using Apache Hadoop, Apache Spark and Apache Hive.

Strong Experience in migrating ETLs to GCP using cloud native tools such as Big Query, Cloud Data Storage.

Working knowledge of Big Data concepts viz. Hadoop/HDFS and Map - Reduce with applications like HBase, Apache Spark and Microsoft Azure ML.

Built performant, scalable ETL process to load, cleanse and validate data.

Used Spark Data frame and Spark API to implement batch processing of Jobs.

Created complex SQL queries and used JDBC connectivity to access the database.

Implemented Partitioning, Dynamic Partitions, Buckets in HIVE.

Created several Databricks Spark jobs with Pyspark to perform several tables to table operations.

Data Analyst

Analog It Solutions, Hyderabad. Jul 2020 – July 2023

Environment: Azure Data Factory, Microsoft Fabric, Azure Data Lake, Azure SQL, Databricks, Synapse, Tableau, Power BI, Python, SQL, Terraform, Git, Docker

Queried and combined large datasets from multiple sources using SQL and Python to support analytical reporting and business questions.

Designed and maintained dashboards in Tableau and Power BI to track KPIs, performance metrics, and product adoption.

Investigated data anomalies, collaborated with engineering and QA teams, and provided feedback to upstream systems to improve data quality.

Gathered and translated reporting requirements into user stories, documentation, and actionable tasks.

Conducted pre- and post-launch data validation to ensure data integrity during feature releases.

Led roadmap and sprint reviews, contributing to strategy development and performance optimization.

Created technical documentation on reporting metrics, data transformations, and logic definitions.

Balanced multiple project deadlines, ensuring timely and accurate data delivery to business users.

Data Analyst Intern

Analog It Solutions, Hyderabad. Jan 2020 – Jul 2020

Environment: Azure SQL, Tableau, Power BI, SSIS, Talend, AWS Glue, Redshift, Python, SQL, Excel, MongoDB, C#

• Built and maintained real-time data pipelines using Kafka, Flink, and Scala to stream millions of records daily for downstream analytics and dashboards.

• Designed modular, reusable object-oriented (OOP) Python and Java scripts to automate transformation logic across multiple data sources.

• Worked with NoSQL databases like MongoDB to extract unstructured data, enabling deeper insights for business units.

• Migrated legacy SQL pipelines to Snowflake, applying best practices for performance optimization, cost control, and query structuring.

• Implemented DevOps practices such as CI/CD, containerization, and infrastructure-as-code (Terraform), reducing deployment time by 30%.

• Collaborated cross-functionally in a consulting-style environment, providing technical guidance and solutions to product and business stakeholders.

• Applied quantitative analysis and statistical techniques to uncover trends, improve forecasting accuracy, and drive KPI reporting.

• Developed secure, compliant pipelines aligned with enterprise data governance and compliance requirements (including access controls and audit tracking).

• Created written technical documentation covering system design, KPIs, data definitions, and QA procedures for knowledge transfer and audits.

• Used communication skills to present complex data findings to non-technical audiences, ensuring alignment between technical output and business needs.

• Developed Dash and Power BI dashboards driven by information system indicators to monitor performance in real-time.

• Coded backend ETL jobs in C# and Scala, integrating them into distributed systems using Kafka and MongoDB.

• Participated in startup-style internal projects, rapidly building minimum viable data products and contributing across multiple tech stacks.

WEB BASED GRAPHICAL PASSWORD AUTHENTICATION SYSTEM

Building a web application that works with any system and allows people to sign up and log in using a different method than static passwords.

•Providing the organizations with highly secure cloud storage for their sensitive data is the main objective of the initiative.

•Enabled users to be recommended based on 50+ data points, awarded most innovative project by university of Sri Indu.

ADVANCE SECURITY IN CLOUD COMPUTING

Role: Python Data Engineering

The main objective of the project is to provide the enterprises with highly secure cloud storage for their confidential information.

•Gather and process raw data at scale (including writing scripts, write SQL queries, writing applications.

•Developed and implemented a comprehensive security strategy for cloud environments, incorporating network security, identity and access management (IAM), encryption, and monitoring solutions to mitigate risks effectively.

•Deployed and configured cloud-native security monitoring tools (e.g., AWS Guard Duty, Azure Security Center) to continuously monitor for security threats and anomalies, enhancing threat visibility and detection capabilities

EDUCATION

University of Southern Mississippi: Hattiesburg, MS. Masters in computer science

GPA: 3.80 Aug 2023 – Present

Relevant Coursework: Advance Algorithms Cloud Computing Machine Learning Data Analysis Deep Learning Software Testing and Quality Assurance Information Retrieval

Sri Indu College of Engineering and Technology. Bachelors in computer science

GPA: 3.80 June 2019- May 2023

Relevant Coursework: C, C++, Python, HTML, CSS, PHP, XML, MYSQL, Oracle.

CERTIFICATIONS:

Certified course on programming for everybody getting started with python by university of Michigan.

https://coursera.org/share/8c75a597a80a05deb100ff09ea73c4de

Certified course on C programming by University of California, Santa Cruz.

https://coursera.org/share/059b4220edc1785f295e9e64b7b0e9d6

SKILLS

Big Data Ecosystem: Spark, HDFS, HIVE, Sqoop, PySpark.

Cloud Services: Google Cloud storage, Big query, Azure Blob, Azure Data lake Storage, Data Bricks.

Orchestration Tools: Airflow.

Relational DB: MySQL, MS-SQL Server

Version Control: GitHub

Programming: Python, SQL

Web Technologies: JavaScript, CSS, HTML

Operating Systems: Windows, UNIX/Linux, and Mac OS

IDE & Command line: Eclipse, IntelliJ, Pycharm

Dhanusha Punyala,

***************@*****.***



Contact this candidate