Post Job Free
Sign in

Data Engineer Richmond Va

Location:
Virginia Beach, VA, 23456
Posted:
May 26, 2025

Contact this candidate

Resume:

Vendor Summary:

** ***** ** *********** ********** – leveraging Python, SQL and Java – including at Citadel (via Sapient).

* ***** ** ******* **** AWS Services

* ***** ** ********** ********** Distributed data/tools like Kafka, Spark, MySQL,

2 years of experience with streaming applications

3 years of NoSQL Experience with Cassandara and Mongo

Worked with Agile /Linux

Possesses impeccable communication skills

Seeking a role in Richmond VA area – he has family there and is seeking a role that will be local to Richmond-No issue with being local from Day 1. Currently in NOVA.

Satish Kola

SUMMARY: Sr. Data Engineer with extensive experience leveraging Python on AWS platform. Extensive experience leveraging Kafka and Spark. Possesses outstanding communication skills. Seeking a role local to Richmond VA.

SKILL SUMMARY

Python

Python 3.11/2.7/2.4, Beautiful Soup, numpy, scipy, pytables,

matplotlib, python-twitter, Pandas, Fabric and urllib2

AWS

Lambda, CloudWatch, EC2, Cloud Search, Elastic Load balancer

ELB, S3, Cloud Front, Route 53,

Scripting languages

CSS, AJAX, Java Script, TypeScript, JQuery, PHP, Shell Scripting

NoSQL /Databases

Oracle, SQLite, PostgreSQL and MySQL MongoDB, Couch DB

Cassandra, Redis

Streaming

Kafka, Spark

Analytics Tools

JMP PRO, Scala, Tableau, Spark,

Node,MVC3,Matlab,Databricks

Markup languages

HTML, XML, JSON

Servers

Apache Tomcat, Apache Spark, IBM Web sphere, open stack,

RESTful web services

IDEs/ Tools

Eclipse, Toad, Sublime text, Spyder, Pycharm, ETL

Version Control

CVS, Github

Methodologies & tools

Object Oriented Programming, UML, Agile Scrum, Waterfall

Deployment Tools

Heroku, Jenkins, Ansible, Redmine, cloudbees

EXPERIENCE

US News & World Report, Washington, DC May 2023 – Present

Sr. Data Engineer – with Python/ AWS

Developed was done leveraging with Python and AWS platform.

Worked on creating ETL procedure automation from scratch using Python and Django framework.

Data extraction from API’s and applies business transformations and load to PostgreSQL

Integrated AWS Dynamo DB using AWS Lambda to store the values the items and backup the Dynamo DB streams.

Implemented RESTful APIs for inter-service communication and integrated third-party APIs for enhanced functionality.

Leveraged Python libraries Beautiful Soup, numpy, scipy, pytables, matplotlib, python-twitter, Pandas to deal with data transformations.

Developed deployment pipelines with code quality check with SonarQube.

Project deployment using Heroku/Jenkins and using (AWS) EC2, AWS S3, Auto scaling, Cloud watch.

Worked on multiple AWS services in integration with Cloud Watch, Lambda, ELB and Auto-Scaling.

Developed Datawarehouse pipeline on Snowflake cluster.

Followed Continuous Integration/Continuous Deployment using Jenkins.

Deployed applications using Docker and Kubernetes for containerization and orchestration.

Conducted unit testing and debugging to identify and resolve software defects.

Environment: Python 3.11, Pandas, AWS Git, Kubernetes, Snowflake, PostgreSQL, DynamoDB,EC2

SWIFT, Manassas, VA Mar 2019 – April 2023

Sr. Data Engineer -Python/Kafka/Spark/AWS

Developed data based application leveraging, Python, Kafka on AWS

Developed bots for events to Kafka to trigger micro services to consume these events to react to changes.

Used Spark Streaming to divide streaming data into batches as an input to spark engine for batch processing Utilized Databricks and Apache Spark to process and analyze large datasets. Designed and implemented data pipelines for e2e messages delivery.

Launching Amazon EC2 Cloud Instances using Amazon Images (Linux/ Ubuntu) and configuring launched instances with respect to specific applications

Used Amazon SQS to upload and retrieve project history.

Migrating present Linux environment to AWS by creating and executing a migration plan, deployed EC2 instances in VPC, configured security groups & NACL's, attached profiles and roles using AWS Cloud Formation templates and Ansible modules.

Developed Spark SQL to load tables into HDFS to run select queries on top.

Wrote Python scripts to parse rawdata and load to data frames using snowpark and store it RDB on Snowflake.

Successfully migrated the Django database from SQLite3 to MYSQL Database with complete data integrity.

Worked on development of SQL and stored procedures for normalization and denormalization in MYSQL.

Build SQL queries for performing various CRUD operations like create, update, read and delete.

Developed web applications and RESTful web services and APIs using Python and Django.

Used Ansible and Ansible Tower as Configuration management tool, to automate repetitive tasks, quickly deploys critical applications, and proactively manages change.

Implemented CI/CD pipeline using Jenkins, Ansible Playbooks and Ansible Tower.

Worked on containerizing Openstack services in Docker using Ansible.

Environment: Python 2.7,3.11, Django 4, HTML5, CSS, XML, MySQL, JavaScript, TypeScript, React JS, Angular JS, Ansible, JQuery, CSS Bootstrap, ETL, Databricks, Snowflake, Mongo DB, MS SQL Server, T-SQL, Git, Kafka, AWS, Unix, Shell Scripting.

JOHNSON & JOHNSON, Trenton, NJ July 2018 – March 2019

Sr. Data Engineer -Python/Splunk/MQ/ AWS

Developed Python and Bash tools to increase efficiency of call center application system and operations; data conversion scripts, AMQP/Rabbit MQ, REST, JSON, and CRUD scripts for API Integration.

Worked on changes to open stack and AWS to accommodate large-scale data center deployment.

Used Celery with Rabbit MQ, Spark to create a distributed worker framework.

Implemented Splunk Enterprise, Splunk forwarder, Splunk Indexer, Apps in multiple servers (Windows and Linux) with automation.

Splunk Enterprise Deployments and enabled continuous integration on as part of configuration management.

Creating the docker images that are required to develop the selenium Grid framework, to link various hubs and nodes configuration.

Created a Cacti Performance Monitoring system on RedHat Enterprise Linux Operating System.

Developed REST services for subscription, order services using Spring DAO and Hibernate ORM.

Used Jackson with REST framework for customizing JSON.

Writing backend programming in Flask. Created RESTful http services to interact with UI.

Worked on backend of the application, mainly using Active Records.

Deployed applications into microservices on cloud-foundry.

Designed and developed communication between client and server using Secured Web services.

Deployed the project into Heroku using GIT version control system.

Built various graphs for business decision making using Python mat plotlib library.

Used Python library Beautiful Soup for web scrapping to extract data for building graphs.

Environment: Python, Perl, Flask, Cassandra, ETL, RabbitMQ, Vugen, Soap UI, Node.JS, REST API, Bash, PyDev, Spyder, Agile.

Citadel, NY Feb 2017– July 2018

Sr. Data Engineer Python/AWS

Work done for RISK management application.

Development was done leveraging combination of Python on AWS platform with MySQL

Leveraged AWS services including - Lambda, S3, CloudWatch & Redshift.

Worked on Python Open stack API's.

Built database Model, Views and API’s using Python for interactive web based solutions.

Used PyQt to implement GUI for the user to create, modify and view reports based on client data.

Installed and maintained web servers Tomcat and Apache HTTP in UNIX.

Developed and Tested Embedded Linux Applications using C, C++ for remote diagnostics, OTA flashing of devices, File system libraries, automation tool kits.

Developed tools using Python, Shell scripting, XML to automate some of the menial tasks. Interfacing with supervisors, artists, systems administrators and production to ensure production deadlines are met.

Wrote and executed various MYSQL database queries from Python using Python-MySQL connector and MySQL dB package.

Involved in debugging the applications monitored on JIRA using agile methodology.

Attended many day-to-day meetings with developers and users and performed QA testing on the application.

Involved in using containers like Docker

Environment: Python 2.7, Django 1.4, AWS S3 Lambda, Cloudwatch,RedShift, MySQL, HTML5, CSS, XML, MySQL, MS SQL Server, JavaScript, AWS, Linux, Shell Scripting, PHP, AJAX, JQuery, Mongo DB, Git, GitHub, JAVA

Pryscom info services, Hyderabad, India Oct 2012 - Nov 2015

Sr. Data Engineer

All development was done leveraging Python with Elastic Search on MySQL database.

Coded with Python – to fetch data from Websites. Leveraed Python modules like requests, urllib, urllib2.

Python application was coded leveraging BBD.

For front-end leveraged Python/Django. UI used CSS, HTML, JavaScript and JQuery.

Deployed the project into Jenkins using GIT version control system.

Learned to index and search/query large number of documents inside Elastic search.

Creating invitation flow for customers to use the client portal for monitoring applications. .

For the large data management used TFS and delivered reliable end product.

Used advanced packages like Mock, patch and beautiful soup (b4) to perform unit testing.

Created database using MySQL, wrote several queries to extract data from database.

Used Perforce for the version control.

Environment: Python, Django 1.4, MySQL, Windows, Linux, HTML, CSS, JQuery, JavaScript, Apache, Linux.

EDUCATION:

TEXAS A&M INTERNATIONAL UNIVERSITY MS- Management Information Systems, 2016 KL UNIVERSITY, India BS - Mechanical Engineering, 2013



Contact this candidate