Post Job Free

Resume

Sign in

Sr. Python

Location:
Hyderabad, Telangana, India
Posted:
March 26, 2024

Contact this candidate

Resume:

P Sai Chaitanya

(Sr Python Developer)

ad4ldp@r.postjobfree.com

+1-810-***-****

Professional Summary

Over 9 years of experience in software development, with expertise in building PySpark algorithms for data aggregation and processing.

Proficient in using PyCharm and VScode as IDEs and working with AWS Cloud platform, including EC2, VPC, RDS, API Gateway, Cloud Watch, Cloud Formation, Step Functions, IAM, S3, SES, SNS, Batch, and Lambda.

Developed user-friendly web interfaces using Python and Django view controller and templating language, following the MVC architecture of the Django Framework.

Extensive experience in working with AWS Glue for data processing automation, including writing automation scripts and scripting CloudFormation templates.

Implemented quality control checks for data processing using Spark SQL and built a data processing algorithm for incremental processing of aggregated data weekly.

Involved in the complete software development lifecycle (SDLC), including requirement gathering, high-level design, development, testing, and deployment.

Sound knowledge in spring, Django, ORM, RDBMS, Python Scripting.

Good understanding and knowledge of Oracle 10g/11g multi node RAC on OEL 5.6 & CentOS, & Dataguard.

Worked on data transformation techniques such as cleaning and partitioning to enhance data processing.

Extensively worked with the AWS Athena database to provide various client reports.

Involved in the migration of AI-based projects from Google Cloud Platform to AWS.

Built schedule-based AWS Lambdas for automatic build and run of data processing pipelines.

Developed a standalone UI tool using wxPython to create custom requirement JSON templates for AWS Glue ETL processing.

Automated loading of data into SQL databases using AWS Glue and Step Functions.

Integrated different AWS components like EC2 and Lambdas with AWS Athena.

Implemented security features around the AWS Code Commit repository using AWS IAM policies.

Experience in writing Infrastructure as a code (IaC) in Terraform, Azure resource management, AWS Cloud formation. Created reusable Terraform modules in both Azure and AWS cloud environments.

Developed rich user interfaces using CSS, HTML, JavaScript, and jQuery.

Automated repetitive tasks using Ansible.

Extracted and manipulated data using SAS/Access, SAS SQL procedures, and PL/SQL scripts.

Created and maintained various automated web tools to increase efficiency.

Worked with MySQL databases, and wrote queries to extract and store data.

Communicated effectively with external vendors to resolve queries.

Implemented AWS Code Pipeline and created CloudFormation JSON templates in Terraform for infrastructure as code.

Implemented Microservices using Spring Boot, Spring MVC, and Apache Kafka.

Good understanding of Object - Oriented Technologies and Relational Database Systems. Worked on object-relational mapper (ORM) library to automate the transfer of data stored in relational databases tables into objects.

Developed integration checks for the PySpark framework for processing large datasets.

Developed Kafka producers and consumers, HBase clients, Spark, Shark, Streams, and Hadoop MapReduce jobs.

Set up SEE5 data mining server on AWS for pattern finding in data.

Configured and set up Airflow DAGs for parallel and sequential execution of Spark commands.

Worked with Solar, Elastic Search, Kafka, Flume, MongoDB, Cockroach DB, and Cassandra.

Used regular expressions for pattern matching and information extraction from web data.

Installed, maintained, and fine-tuned Apache Tomcat and Websphere Application Server.

Developed frontend and backend modules using Python and Django Web Framework.

Installed New Relic Kafka plugin for monitoring Kafka cluster.

Performed Airflow performance tuning for DAGs and task instances.

Managed Docker orchestration and containerization using Kubernetes.

Created PySpark frames to bring data from DB2 to Amazon S3.

Experienced in writing unit tests and using the Google open-source automated testing framework Bazel.

Implemented security features around the AWS Code Commit repository using AWS IAM policies.

Responsible for data extraction and ingestion from various sources into S3 using Spark and Hive ETL

pipelines.

Used ETL (SSIS) for extracting, cleaning, transforming, and loading data into a data warehouse.

Responsible for installation, configuration, and administration of Red Hat Linux and Solaris systems.

Technical Skills

Programming Languages

Python, SQL, PL/SQL

Frameworks

PySpark, Django, Flask, Spring Boot, Spring MVC

Database Management

MySQL, Amazon RDS, Amazon Athena, Apache Hive, DB2, SAS, Redis

Big Data Technologies

Hadoop MapReduce, Apache Spark, Apache Kafka, HBase, AWS Glue

Cloud Platforms

Amazon Web Services (AWS) - EC2, VPC, RDS, S3, Athena, Lambda, API Gateway, CloudFormation, Step Functions, SES, SNS, Batch, Google Cloud Platform (GCP), Azure

Data Processing Tools

Spark SQL, SAS/Access, PySpark, Spark Streaming, AWS Glue, Hive, SSIS

Web Technologies

HTML, CSS, JavaScript, jQuery, Django Templating Language

User Interface (UI) Libraries

wxPython, jQuery UI

Version Control Systems

Git (GitHub), AWS Code Commit

Automation and Orchestration

Ansible, Apache Airflow, Kubernetes, Docker, AWS Step Functions

Messaging Systems

Apache Kafka, Redis

Database Servers

MySQL, DB2, Apache Hive, MongoDB, Cockroach DB, Cassandra

Web Servers

Apache Tomcat, Websphere Application Server

Data Mining and Analysis

SEE5, Elastic Search, Solar, Flume

Testing Frameworks

Pytest, Bazel, Google Open-source Automated Testing Framework

Data Formats

XML, JSON

Other Tools and Technologies

AWS Code Pipeline, Terraform, Regular Expressions, New Relic, Git, SAS SQL Procedures, PL/SQL Scripts, PyCharm, VSCode

Web Development

PHP, MySQL

Development Methodologies

Agile, Scrum

Documentation and Communication

Technical Documentation, Requirement Gathering, Communication with External Vendors, Code Reviews, Software Requirements Specification Documents, Technical Translations

Operating Systems

Linux (Red Hat, Solaris)

Work Experience

Microsoft, WA Mar 2021 – Present

Sr Python Full Stack Engineer

Responsibilities:

Worked on building PySpark algorithms for different aggregations of data based on the specifications.

Involved with performance and process enhancement of the PySpark framework.

Developed views and templates with Python and Django view controller and templating language to create a user-friendly website interface.

Developed web applications in Django Framework model view control (MVC) architecture.

Experienced in using different community and enterprise IDEs and tools like PyCharm and VScode.

Experience in writing Infrastructure as a code (IaC) in Terraform, Azure resource management, AWS Cloud formation. Created reusable Terraform modules in both Azure and AWS cloud environments.

Worked on Scripting CloudFormation Templates to auto-provisioning of resources.

Worked on various Quality control checks for data processing using Spark SQL.

Developed and optimized Python-based ETL pipelines on Azure Databricks, leveraging Apache Spark for efficient data processing and transformation.

Implemented advanced data analytics and machine learning algorithms using PySpark within Azure Databricks, resulting in improved data insights and actionable business recommendations.

Built all database mapping classes using Django models with the help of ORM like SQLAlchemy.

Implemented SQLAlchemy which is a python library for complete access over SQL.

Collaborated with cross-functional teams to design and implement scalable data solutions, ensuring seamless integration between Azure Databricks and other Azure services for enhanced data processing and storage capabilities.

Used docker-compose to deploy a WSGI application using uWSGI and nginx, the backend running uwsgi, the frontend running nginx.

Build a data processing algorithm for incremental processing of data that’s aggregated each week.

Involved in the complete software development lifecycle (SDLC) to develop the application.

Experienced with AWS cloud platform and its features, which includes EC2, S3, ROUTE53 VPC, EBS, AMI, SNS, RDS AND CLOUD WATCH.

Worked on data transformation like cleaning and partitioning of data for enhanced processing of data.

Handle escalated Support tickets till closure for MS Azure PaaS platform.

Provide consulting and cloud architecture for premier customers and internal projects running on MS Azure platform for high-availability of services, low operational costs.

Development web service using Windows Communication Foundation and .Net to receive and process XML files and deploy on Cloud Service on Microsoft Azure.

Gathered Semi structured data from S3 and relational structured data from RDS and keeping data sets into centralized metadata Catalog using AWS GLUE and extract the datasets and load them into Kinesis streams.

Good working experience in Azure Logic apps, Service Bus, Document DB and SQL Database.

Developed U-SQL Scripts for schematizing the data in Azure Data Lake Analytics.

Created Azure Function App on the portal using Maven build project via Jenkins CI/CD pipeline.

Hands on experience on Azure VPN-Point to Site, Virtual networks, Azure Custom security, Endpoint security and firewall.

Install, configure, test, monitor, upgrade, and tune new and existing PostgreSQL databases.

Created Apache and Nginx Redirect rules for HTTP to HTTPS and www to non-www.

Build an integration testing system that involves testing the entire framework with a code check-in.

Experienced writing unit tests and Google open source automated testing framework Bazel.

Hands on experience on Azure VPN-Point to Site, Virtual networks, Azure Custom security, end security and firewall. Used Azure Express Route to set up a private connection to Microsoft cloud services such as Microsoft Azure, Office 365, and Dynamic 365.

Responsible for setting up Python REST API framework using Django.

Deployed the packages on staging and production. Monitored jobs and supported Azure Data Lake production environment.

Used JQuery to make the frontend components interact with the Javascript functions to add dynamism to the web pages at the client side.

Used advanced JavaScript concepts to validate input fields using regular expression.

Involved in translating technical documents into engineering Specifications.

Migrating a Non ASM Single instance Oracle 11g Database to 11g RAC Database configured on ASM.

Coordinated with DevOps teams for bug fixes and code releases.

Developed a rich user interface using CSS, HTML, JavaScript, and jQuery.

Worked on automating repetitive tasks using Ansible.

Extracting acting data from the database using SAS/Access, SAS SQL procedures, and creating SAS data sets.

Created and modified PL/SQL scripts for data conversions.

Developed and maintained various automated web tools for reducing manual effort and increasing the efficiency of the Global Shipping Team.

Created databases using MySQL, and wrote several queries to extract data from the database.

Communicated effectively with external vendors to resolve queries.

Environment: Azure, AWS, Machine Learning, Python 3.6+, Django, RDS, Shell Scripting, SQLAlemy, Pandas, PySpark, DevOps, PyQt, PyQuery, Flash, HTML, CSS, Oracle 11g, JavaScript, Node.Js, Angular.Js, React.js, AJAX, Oracle, Snowflake, MongoDB, MySQL, JDBC, Git, GitHub, Rest, SOAP, Webservices, Selenium, UML, Eclipse, Junit, Itext, Subversion, Agile, Web 2.0, SOA

Cigna, CT Nov 2019 – Feb 2021

Role: Python Full stack Engineer

Responsibilities

Worked on building PySpark algorithms for different aggregations of data based on the specifications.

Implemented AWS Code Pipeline and Created Cloud formation JSON templates in Terraform for infrastructure as code

Implementing Microservices to base on Restful API utilizing Spring Boot with Spring MVC and Apache Kafka

Involved with performance and process enhancement of the PySpark framework.

Developed integration checks around the PySpark framework for processing large datasets.

Worked on migration of PySpark framework into AWS Glue for enhanced processing.

Developed machine learning models using AWS SageMaker, leveraging built-in algorithms and Jupyter notebooks for data exploration and model training.

Oracle 11g Administration and support.

Wrote various automation scripts for automation of data processing on AWS Glue.

Worked on Scripting CloudFormation Templates to auto-provisioning of resources.

Worked on various Quality control checks for data processing using Spark SQL.

Installing and configuring nagios check postgresql.pl plugins for Monitoring PostgreSQL instances effectively.

Proactive in updating the latest security patches to database, which are provided by PostgreSQL open source community.

Created PowerShell scripts to constantly monitor the health of the Exchange messaging infrastructure and notify my teammates or myself in the event of a problem.

Successfully migrated the Django database from SQLite to MySQL to PostgreSQL with complete data integrity.

Build a data processing algorithm for incremental processing of data that’s aggregated each week.

Involved in the complete software development lifecycle (SDLC) to develop the application.

Worked on data transformation like cleaning and partitioning of data for enhanced processing of data.

Developed Kafka producer and consumers, HBase clients, Spark, shark, Streams, and Hadoop MapReduce jobs along with components on HDFS, and Hive.

Deployed machine learning models as RESTful APIs using SageMaker endpoints, enabling real-time predictions in production environments.

Worked on setting up of SEE5 data mining server on AWS for pattern finding in the data

Extensively worked on the AWS Athena database to provide various client reports.

Involved in setting up Python REST API Framework using Django.

Build Schedule AWS Lambdas for automatic build and run of data processing pipelines

Worked on building a standalone UI tool using wxPython for creating custom requirement JSON templates to feed into AWS Glue for ETL processing of data.

Configured and set up Airflow DAGs as per the requirement to run spark commands in airflow parallel and sequential.

Migration of data validation code from the legacy mainframe systems to ETL framework using informatica workflows and shell scripts.

Worked on setting up automated loading of data into SQL database using AWS Glue and Step Functions.

Worked on Solar, Elastic Search, Kafka, Flume, MongoDB, Cockroach DB and Cassandra

Used regular expressions to match the pattern with the existing one and used regular expressions to extract information from available web data

Installed, maintained, and finetuned the Apache-Tomcat server and Websphere Application Server in the Linux Platform

Developed entire frontend and backend modules using Python on Django Web Framework.

Worked on integrating different AWS components like EC2 and Lambdas to work with AWS Athena.

Worked on installing new relic Kafka plugin implementation for monitoring of Kafka cluster

Worked on Airflow performance tuning of the DAG’s and task instances.

Build an integration testing system that involved testing the entire framework with a code check

Managed Docker orchestration and Docker containerization using Kubernetes.

Create a PySpark frame to bring data from DB2 to Amazon S3.

Added X-Forwarded-For-Proto in Apache and NGINX Vhosts to reveal client IP addresses.

Worked on extracting and manipulating text data using regular expressions

Experienced writing unit tests and Google open-source automated testing framework Bazel

Worked with Python-based web frameworks to integrate with Node.js and React.js, hosting on load-balanced

Implemented Security features around the AWS CodeCommit repository like pull requests using AWS IAM policies. Responsible for data extraction and data ingestion from different data sources into S3 by creating ETL pipelines using Spark and Hive

Worked on Big Data DevOps and Machine Learning tools like Sagemaker in AWS and Azure. Also worked on AWS and Azure Data Lakes and Snowflake and have done data ingestion from multiple sources

Used ETL (SSIS) to develop jobs for extracting, cleaning, transforming, and loading data into a data warehouse.

Responsible for installation, configuration, and administration of Red Hat Linux, and Solaris systems.

Experienced in using different community and enterprise IDE’s (Integrated Development Environments) and tools like PyCharm and VSCode

Involved in translating technical documents into engineering Specifications.

Coordinated with DevOps teams for bug fixes and code releases

Environment: Python, PySpark, Kafka, wxPython, Sagemaker, Apache, AWSGlue, AWS Athena, AWS StepFuctions, CloudFormation PyTest, Bazel, Bootstrap, Flask, Oracle, PL/SQL, MySQL, MS-SQL, REST, PyCharm, Windows, Linux.

Baker Hughes, TX Aug 2017 – Oct 2019

Role: Software Developer in Python

Responsibilities:

Involved in the Analysis, design, and architecture of the application

Worked on requirement gathering and high-level design

Created PHP/MySQL backend for data entry from Flash I had to assist the Flash developer in sending the correct data via query strings

Converted Visual Basic Application to Python, MySQL

Generated Python Django Forms, and Crispy forms to record data, log, and sign of users

Experience in development test automation

Generated various reports in Python for channel usage, system downtimes, catching exceptions, and performance-related metrics.

Designed and Implemented a Random Unique Test Selector Package for processing large volumes of data using Python and Django ORM

Integrated SageMaker with other AWS services like S3 for data storage, IAM for access control, and CloudWatch for monitoring, enabling end-to-end machine learning workflows

Skilled in using collections in Python for manipulating and looping through different user-defined objects

Designed email marketing campaigns and created responsive web forms that saved data into a database using Python/ Django Framework

Skilled in using collections in Python for manipulating and looping through different user-defined objects

Created database tables, functions, stored procedures, and wrote prepared statements using PL/SQL

Modified queries, functions, cursors, triggers, and stored procedures for MySQL database to improve performance, while processing data

Performed different testing methods like Unit testing, Integration testing, and web application testing.

Involved in capturing the requirements for serial functional interface and other software requirements specification document

Optimized costs by using SageMaker's managed infrastructure, scaling resources based on workload demands and using Spot Instances for cost-effective training.

Extensive code reviewing using GitHub pull requests, improved code quality, and conducted meetings among peers

Responsible for Parsing XML data using XML parser and Testing, fixing the bugs, and coding modifications

Database Administration activities like taking backups, checking log messages, looking for database optimization

Used Redis as a messaging broker to execute asynchronous tasks

Designed and implemented a dedicated MYSQL database server to drive the web applications and report on daily progress

Environment: Python, Django, Pytest, Java, AWS, MongoDB, XML, JQuery, DOM, JSON, Numpy, Pandas, AngularJS, HTML, XHTML, CSS, AJAX, JavaScript, JSON, Rest, GIT, MY SQL, Windows and LINUX.

Matrix Technologies, India June 2014 –April 2017

Role: Software Developer

Responsibilities

Responsible for gathering requirements, system analysis, design, development, testing, and deployment.

Developed rich user interface using CSS, HTML, JavaScript, and jQuery.

Used jQuery for selecting particular DOM elements when parsing HTML

Wrote PYTHON modules to extract/load asset data from the MySQL source database.

Created a database using MySQL, and wrote several queries to extract/store data from the database.

Setup automated corn jobs to upload data into the database, generate graphs, and bar charts, upload these charts to the wiki, and backup the database.

Effectively communicated with external vendors to resolve queries.

Used Git for version control.

Actively participated in system testing, production support and maintenance/patch deployments

Environment: PYTHON, HTML, XHTML, CSS, JavaScript, jQuery, Eclipse, MS SQL, Windows OS



Contact this candidate