Post Job Free
Sign in

Python Developer Data Engineer

Location:
Conroe, TX
Posted:
November 21, 2024

Contact this candidate

Resume:

Sravani Gummadi

Mobile: 703-***-****

Email Id: **********@*****.***

Professional Summary:

Python developer and Data Engineer with 5 years of IT Experience developing web applications implementing MVC architecture using Django, Flask framework(s) with major focus on Python ORM libraries, SQL Alchemy, and integrating Python with AWS, Azure, OpenStack, CICD, Configuration Management, along with using various tools like Docker, Kubernetes, Ansible, Chef, Jenkins, Git, Maven.

Strong in application development using PYTHON, Django, HTML5, CSS, JavaScript, jQuery, Oracle, MySQL, PostgreSQL and SQ Lite, Angular, PostgreSQL, ZODB, Key Value Stores, CouchDB, Couchbase, Linux, Nginx, Varnish, ReactJS.

Extensively used SQL, Numpy, Pandas, and Spark for Data Analysis and Model building, Cognitive Design, deploying and operating highly available, scalable, and fault-tolerant systems using Amazon Web Services (AWS).

Hands on Experienced in OpenShift platform in managing Docker Containers, Kubernetes Clusters, and Mesos and implemented a production ready, load balanced, highly available, fault tolerant Kubernetes infrastructure and created Jenkins jobs to deploy applications to Kubernetes Cluster.

Experienced in Using Jenkins pipelines to drive all Microservices builds out to the Docker registry and then deployed to Kubernetes, Created Pods and managed using Kubernetes.

Implemented & maintained the branching and build/release strategies using SCM tools GIT, Subversion, Bitbucket and also experienced in migrating GIT repositories to AWS.

Experienced in Using Jenkins pipelines to drive all Microservices builds out to the Docker registry and then deployed to Kubernetes, Created Pods and managed using Kubernetes.

Experience in enhancing and deploying the ETL Packages from development server to production server.

Set up Kubernetes (k8s) Clusters for running microservices and pushed microservices into production with Kubernetes backed Infrastructure. Development of automation of Kubernetes clusters via playbooks in Ansible.

Used Ansible, Chef, Jenkins, Git, for implementing Continuous Integration from scratch and optimize the Continuous Integration using Jenkins and troubleshoot the deployment build issues.

Developed the required XML Schema documents and implemented the framework for parsing XML documents.

Implemented Python libraries NumPy, SciPy, PyTables, SQL Alchemy, Matplotlib, Pandas, Beautiful Soup and urllib in various tasks. Lead all data platform initiatives: Data Lake Strategy, Data Governance, Data Security and Compliance, Data Platform Cloud Infrastructure, MLOps and Master Data Management.

Perform analytics and querying large datasets, use Big Query fully managed, serverless data warehouse. Use Cloud Storage as the storage foundation for your data lake.

Implement DLP policies to detect and protect sensitive data in Cloud Data Catalog for metadata management and data discovery.

Strong knowledge of Object-Oriented Design and Programming concepts and Experience in Object Oriented Programming (OOP) concepts using Python, C++.

Integrated Python with Big Data and Analytics based on Hadoop, Spark and No-SQL databases like HBase and MongoDB.

Knowledge in working with continuous deployment using Heroku and Jenkins.

Experience in developing applications using Amazon Web Services like EC2, Virtual private clouds (VPCs), Storage models (EBS, S3, and instance storage), Elastic Load balancer ELBs.

Strong development experience on Linux based operating systems (Debain, CentOS, RedHat and Ubuntu).

Education:

Master’s in Computer Information Systems.

Bachelor’s in Computer Science.

Professional Experience:

Client: Morgan Stanley, NYC Apr’ 2022 – May 2023

Role: Python developer with Site Reliability

Description: Worked as a Python Developer in a Linux environment to Run the batch jobs and visualize the data of Batch run logs, data and reports over the Flask app engine connected to the SQL, Spark SQL, PySpark, Sybase Databases for many different environments on servers running parallel and distributed data management. And creating the YAML config files to schedule the build on Jenkin build.

Responsibilities:

Worked on the Design and development of the HTML Forms, connecting to the Flask app.

Designed the Templates to connect with e Flask app with Different pages and menu items with forms and loadingarefresh button on different environments configured internally.

Developed the Internal Flask app connecting to the Spark and No-SQL databases like HBase, Snowflake, SQL, Spark SQL, PySpark, and Sybase Databases that can run on the different environments by configuring the environment Variables.

Integration of Linux POSIX supported commands, bash shell, putty, and HDFS shell commands, along with interaction of multiple Hadoop clusters. Successfully performed analytical operations using command line interface.

Writing file formatting and compression techniques in Hive such as text files, sequence files, rcfile, avro file orcfile, parquet file, snappy compression, etc

Developed interactive page for the user to request the data and view the reports, status of the Batch jobs that run in the backend Linux environment.

Developed the UI to check the logs of the Batch jobs in Linux using Flask engine.

Involved in the working on the tracking Logs for the Batch jobs are running on the Linux environment and connected the functionality on the Flask app.

Worked on the Production issues to resolve the Database record run and stopped and debug on Logs and rerun the jobs by fixing.

Integrated the user Dashboard to bring the data in recon files from Prometheus port to Grafana.

Used pandas and data manipulation to visualize the data from different sources to the Dashboard.

Involved in Writing the SnowflakeSQL queries to SQL view, data functions as per the daily user requirement and grant access over the service now portal to the user tickets.

Using Bitbucket with Source tree as Version control as a part of the CI/CD operations on the Jenkins builds.

Environment: UI, Python, HDFS, HIVE, PySpark, pandas, Linux, PowerShell, Snowflake, SQL, Sybase, Jenkins, Bitbucket, Source Tree, Prometheus, Grafana.

Client: Exxon, TX Nov’ 2021 – Feb’ 2022

Role: Sr. Python developer with Azure DevOps

Description: Worked as a Sr. Python Developer in a Red-Hat Azure DevOps environment to resolve to troubleshoot the Service Now APIs connected with the UI built on top of Angular and backend Database and Metadata. Also worked on the Machine learning Implementation on the schema tables to convert to the datasets, which are scheduled to run in Airflow Dag’s.

Responsibilities:

Worked with Frontend developers to give the improvements on the frontend Angular 7, 8,9in order to get the proper payloads for the backend python API and changes in the API gateway.

Have an excellent understanding on the UI components and Directives in Angular that connect the back-end data pipelines.

Created forms from Angular fronted connecting the API gateway and python api service to generate operations after the submit button.

Worked closely with the frontend team to give changes for the AD groups that are connected to the editable drop downs.

Expert level of working experience with various Python IDE's using PyCharm, PyScripter, PyArrow, PyStudio, and PyDev.Utilized Big Data Platforms and applications using SQL and NoSQL. knowledge of Power BI and the Azure Data Platform stack, which consists of Azure Data Lake, Data Factory, and Databricks having knowledge of Python, Scala, Spark, and SQL

Experience in building data RESTful APIs to consume analytic model output.

Extensively worked on Service Now API on top of Python and frontend in Angular connecting the Snowflake database and PostgreSQL to create schemas and grant requests to access data and some other operations to be performed in a automated manner.

Worked on creating Snowflake Schemas, Tables and granting roles and writing Native SQL queries as per the requests.

Build and Configured and ML based application to convert the Snowflake schema Tables into pandas datasets.

Containerized the Application to run CI/CD operations in a Kubernetes Pods and worked with cross teams to stabilize the Pod operations. Created the Aro-namespace for the applications under one category.

Performed the Gitflow as a part of CI/CD operations, adopting branching strategy.

Created numbers on Airflow DAG’s for both the API’s and schedule to run.

Implemented Hadoop-based solutions for parallel processing of large datasets in data processing latency.

Install and utilized database tools to connect the application with MySQL, PostgreSQL and oracle databases.

Environment: Python, API’s, Azure Dev-Ops, Red-Hat, Snowflake, SQL, Postgre, Kubernetes, Pod, Airflow Dag’s, PySpark, Spark, Numpy, Pandas, Data Frame, Git.

Client: Lipscomb University Nov’ 2020 – Sep’ 2021

Role: Sr. Python developer with AWS

Description: Working at Lipscomb University as a Senior Python developer. Work on the Existing Python application to utilize AWS services and Dockerze, My Role is to stabilize and optimize the application and migrate to AWS Cloud platform and create and maintain docker containers and automate using Jenkin jobs, also to build a same feature application for another location. Debug and push the process to high level Performance.

Responsibilities:

Build and maintain a new Integration of the project that involves Making API calls and setting up the endpoints to campuses at two different locations. This consumes data from the interface and runs an API call to feed the interface.

Experience in building data APIs to consume analytic model output.

Creating and maintaining containers for micro-services, experience in AWS Services (EC2, Lambda, EMR, S3, SNS etc.).

Install and utilized database tools to connect the application with MySQL, PostgreSQL, and oracle databases.

Create and schedule the Jenkin jobs for the existing clusters to automate the processes.

Follow gitflow to deploy the changes and push the code to the production environment.

Create and Maintain Git repos and Cron jobs for the build and deployment of Api’s Built on top of Python in Docker Container.

Responsible for creating Docker file for building Docker Images and write python script to build and push Docker images in Docker Hub.

Working experience on docker-compose yaml file templates to deploy images in Docker containers

Experienced in writing cloud formation templates to create and deploy ECR Repos.

Create EC2 instance and set up the lambda triggers using lambda functions. Utilizing the s3 bucket for storage.

Experience in writing python modules using Numpy and pandas to delete or apply conditions on columns and rows in a csv file.

Debug and troubleshoot Api’s behavior and response. Designed, wrote and maintained systems (Python) for administering GIT.

Performed Unit testing periodically.

Work with product development to resolve build-related issues in all projects.

Environment: Python, API’s, AWS, EC2, ECR REPOS, CloudFormation templates, S3 Lambda functions, PySpark, Spark, Numpy, Pandas, Docker, Jenkins, Docker image, Docker Container, Data Frame, Git.

Client: Cox Media Group, Atlanta, GA March’ 2020 – Sep’ 2020

Role: Sr. Python developer with AWS.

Description: Actively involved in developing a ETL Infrastructure to extract and validate click stream data from various vendors such as Google, Face book, Microsoft, Yext, BirdEye, Eltoro.etc, using API’s on top of Python. Developed and maintained data pipelines which process the data and make it available in the cloud for data analysis and front-end applications to generate reports.

Responsibilities:

Developed API’s from the scratch and maintained the Architecture.

Worked on developing and documenting Google API’S Google Analytics, Google DFP, Google Adwords, Google DV360, Facebook API, Microsoft – Bing Ads API and other 10 API’s like Yext, BirdEye, Eltoro, Nativo etc.

Have deep understanding on API dashboards, use cases of API’s.

Involved in writing Python modules for each task from assigning user till uploading data on S3 buckets.

Involved in migrating the project to AWS cloud from on prem.

Migrated the code infrastructure to AWS, schedule and run number of ec2 instances as everyday tasks.

Involved in the process of building ETL pipelines from landing files to the corresponding buckets to set up scheduled takes and help with Lambda and glue triggers.

Environment: Python, AWS, EC2, S3, Lambda function, Lambda triggers, Dockers, Jenkins, Numpy, Pandas, Matplotlib, PySpark, Spark, Git.

Client: Envy InfoTech, Hyderabad Jan’ 2011 – Jun’ 2013

Role: Program Analyst

Responsibilities:

Involved in the complete Software Development Life Cycle (SDLC) including Requirement Analysis, Design, Implementation, Testing and Maintenanceof the website using HTML, XHTML, AJAX, CSS and +JavaScript.

Used Hibernate for mapping data representation from MVC model to Oracle Relational data model with a SQL-based schema.

Implemented Spring Batch jobs for ensuring sending alerts to clients based on configured business rules.

Worked on Oracle 11g databases and wrote SQL queries as well as stored procedures for the application.

Assisted with production support activities using JIRA when necessary to help identify and resolve escalated production issues based on the SLA.

Environment: Java, JSP, Servlets, spring, HTML, CSS, AJAX, Hibernate, XML, Maven, Oracle, JavaScript, Eclipse.



Contact this candidate