Post Job Free
Sign in

Data Engineer Analyst

Location:
Coppell, TX
Salary:
70/hr
Posted:
July 18, 2023

Contact this candidate

Resume:

Sai P

Dallas, TX 940-***-**** Email: adydiv@r.postjobfree.com

SUMMARY

Highly skilled and motivated data engineer in various domains. Possess expertise in SQL, Python, and ETL processes, and proficient in utilizing Hadoop, Pyspark, Cloudera, and different data pipeline tools. Good hands-on experience with cloud platforms like Azure and AWS and working with relational databases such as MySQL and Oracle. Well-versed in data modelling, warehousing, and visualization tools like Tableau and Power BI. Adept at version control using Git, teamwork, analytical, and problem-solving skills.

EDUCATION

Master – Lindsey Wilson college MAR 2023

Bachelor of Technology - Jawaharlal Nehru Technological University Hyderabad MAY 2018

CERTIFICATION

Microsoft Certified: Azure Data Engineer Associate

COMPETENCIES

Programming – SQL, Python, Spark

Visualization – Tableau, Excel, PowerBI

Tools – Jira, Confluence, Jupyter, RStudio, Databricks, Git version control, MS office (Microsoft Excel, Word, PowerPoint), ETL, Hadoop, Pyspark, Cloudera, Data Pipelines and work flow management, MYSQL, Oracle, Tableau, PowerBI, DataBricks,BigQuery

Cloud Platforms: Azure (Data Storage, Data Processing and Analytics), AWS (Data Storage, Data Processing and Analytics)

Soft Skills: Quick learner, problem solving skills, Great attention to detail, teamwork, excellent written and verbal communication.

EXPERIENCE

Accenture (stripe) – Data Engineer Hyderabad, India ( Aug2020- Jul 2021)

Summary: Leveraging my expertise in Azure Stack technologies, I actively contributed to the design, development, and optimization of data pipelines and models. Collaborating with cross-functional teams, I provided efficient data solutions aligned with business goals. Proficient in SQL, Python, and Azure services, I automated processes, enhanced data quality, and ensured data integrity throughout the pipeline.

Skills utilized:

Designed and implemented efficient data pipelines using Azure Stack technologies.

Collaborated with the data engineering team to optimize data workflows across systems.

Utilized Azure services like Data Factory, Databricks, and SQL Database for processing and storage.

Developed and maintained data models and warehousing solutions in Azure.

Implemented data quality checks and validation mechanisms.

Utilized SQL and Python for data manipulation, scripting, and automation.

Optimized data storage and retrieval performance through SQL query optimization.

Collaborated with cross-functional teams to understand data requirements and align solutions with business goals.

Assisted in deploying and monitoring data pipelines using Azure monitoring tools.

Documented data engineering processes and best practices for knowledge sharing.

Tools: Python, SQL, Data Bricks,Pyspark

Google Wipro – Content Operations Analyst Hyderabad, India (Jun 2018-Aug 2020)

As a Content Operations Analyst at Google, I analyzed content performance metrics, user engagement data, and content quality metrics to identify trends and insights.

•Using data visualization tools, I created visual reports and dashboards to effectively communicate content performance to stakeholders.

•I ensured data accuracy and quality by validating data integrity, resolving discrepancies, and implementing data governance processes.

•Collaboration with content creators, editors, and marketing teams was a crucial aspect of my role to align content strategies with business goals.

•Through data analysis, I identified opportunities for content optimization, resulting in improved user experience and engagement.

•In addition to my role as a Content Operations Analyst, I also took on additional projects as a Data Analyst and Data Engineer.

•My expertise in data analysis, data visualization, data engineering, and collaboration allowed me to provide valuable insights and develop scalable data solutions.

Data Analyst (Additional Project), Google

•Conducted data analysis on various data sources to extract insights and generate reports for stakeholders.

•Developed data visualizations and dashboards using tools like Tableau to effectively communicate data findings.

•Collaborated with cross-functional teams, including executives, business managers, and IT professionals, to understand their data needs and provide actionable insights.

•Ensured data accuracy and quality through data cleansing, transformation, and validation techniques.

•Contributed to data-driven decision-making by identifying optimization opportunities across different business functions.

Data Engineer (Additional Project)

•Designed and implemented data pipelines and workflows to extract, transform, and load data from various sources into target systems.

•Utilized SQL and Python for data manipulation, scripting, and automation tasks.

•Worked with Hadoop, Pyspark, and Cloudera to process and analyze large-scale data sets.

•Leveraged cloud platforms like Azure and AWS for data storage, processing, and analytics.

•Collaborated with cross-functional teams to understand data requirements and develop scalable data solutions.

•Implemented version control using Git to ensure collaborative development and code integrity.

INTERNSHIP PROJECTS

Eliptico IT Solutions Private Limited Dec 2017 - May 2018

1. Data Pipeline Optimization

Optimized an existing data pipeline by identifying bottlenecks and implementing performance enhancements.

Used Pyspark to transform data across AWS services. (AWS glue, EMR. S3 and Lambda functions

Improved processing time by 30% by optimizing SQL queries and PySpark jobs.

Implemented caching mechanisms and partitioning strategies to reduce data transfer and processing overhead.

2. Real-time Data Streaming

Developed a data streaming application using Apache Kafka and Apache Spark Streaming.

Integrated with AWS Kinesis for real-time data ingestion and processing.

Built data pipelines to enable real-time analytics and monitoring of streaming data.

Skills: AWS Glue, AWS Lambda, S3, EMR, AWS services



Contact this candidate