Post Job Free

Resume

Sign in

Web Application Node Js

Location:
Cleveland, OH
Salary:
150k
Posted:
February 14, 2024

Contact this candidate

Resume:

Name: Shashidhar

Email: ad3mqd@r.postjobfree.com

Contact: 330-***-****

Linkdin: https://www.linkedin.com/in/shashidhar-b-04418a266

PROFESSIONAL SUMMARY

Around 9+ years of experience in Web/Application Development as a senior Python Developer in analytical programming using Python with skills in analysis,design, development, testing and deploying software applications.

Good experience using python libraries to speed up development(libraries used: Beautiful Soup, SciPy, Matplotlib, python-Twitter, Panda’s data frame, network, urllib2, MySQL, NoSQL,dB for database connectivity, JSON libraries)

Configured and managed DynamoDB tables to store and retrieve data with high scalability and low-latency access.

Integrated DynamoDB with Python using AWS SDK (Boto3) for seamless database interactions. Experienced in configuring IAM roles and permissions for Lambda functions

Involved in Developing a Restful service using python Flask framework.Developed web applications using the Express.js framework, which is a popular web application framework for Node.js.

Made Django web-based apps for Insurance premium calculations.

Developed entire frontend and back-end modules using python on Django Web Framework with GIT. And also Developed server-side applications using Node.js, leveraging its asynchronous and event-driven architecture.

Successfully Migrated the Django database from SQLite to MySQL to PostgreSQL with complete data integrity.

Working in UNIX environment in developing the application and familiar with all its commands.

Exposure in building CI/CD pipelines in container-based fashion using Docker,Jenkins,UBUNTU,and Ansible, GitLab. Integrated Node.js applications into CI/CD pipelines, using tools like Jenkins, and GitLab CI for automated testing

Experienced with GITHUB, GITLAB, SVN and MAVEN build tool.

Developed a multi-tiered J2EE web application for a bookstore.Web module uses Servlets and JSPs, with business logic managed by EJBs.

JPA facilitates database interactions, ensuring data consistency with transaction management.User sessions and shopping cart handled by UserSessionBean, with Java EE security for authentication.

Asynchronous processing implemented through MDB, logging, custom exception handling, and rigorous testing complete the robust application

Proficient in using Confidential Web Services AWS. Extensive experience focusing on services like EC2, VPC, Cloud Watch, Cloud Front, Cloud Formation, IAM, S3, Confidential RDS, Elastic Cache, SNS, SQS.

Configured AWS Identity and Access Management (IAM) Groups and Users for improved login authentication.

Hands on experience in design and development of Unix Shell Scripting apart of the ETL Process to automate the process of loading.

Leveraged expertise in J2EE technologies such as Hibernate, Servlets, JNDI, JMS, JDBC, SOAP, and Rest-based Web services for a comprehensive bookstore application.

Implemented concurrent programming for optimal performance. The multi-tiered architecture, including EJBs and JPA, ensured seamless integration, while Java EE security, MDB for asynchronous processing, and robust exception handling enhanced the application's reliability.

Automate different workflows, which are initiated manually with Python scripts and Unix shell scripting.

Experience in Creation and managing user accounts, security, rights, disk space and process monitoring in Solaris and Red hat Linux.

Experienced on working with Docker for creating images that are deployed on AWS as Microservices and Experienced on managing the local deployments in Kubernetes and creating local cluster and deploying application containers.

Worked in agile and waterfall methodologies with high quality deliverables delivered on-time.

Experience in Implementation of MVC/MV architecture using Django, and RESTful, SOAP web services.

Experience with continuous integration and automation using Jenkins.

Experience with Unit testing/ Test driven Development (TDD), Load Testing in JIRA, JENKIN.

Implemented comprehensive testing, including unit tests for Servlets and EJBs, ensuring a stable deployment on WildFly. Utilized JMS for efficient message-driven asynchronous processing, enhancing order fulfillment. Extended the application with RESTful APIs for external integrations, showcasing versatility in J2EE technologies

EDUCATION: Bachelor’s in computer science, Vignan University, India, 2013

TECHNICAL SKILLS

Frameworks:

Django Rest framework, Flask, Pandas

Python Libraries:

Requests, Urllib, Pandas, NumPy, SciPy, matplotlib, Beautiful Soup, Html Parser, Swagger, SQL Alchemy, My SQL DB, XML Docx, PyPandoc, Celery, Celery Beat, Boto

Web Technologies:

HTML, CSS, DOM, JavaScript, jQuery, AJAX, XML, AngularJS, NodeJS,J2EE,

Programming Languages:

Python, C, C#, SQL and PL/SQL, Java.

Version Control:

SVN, Git, GitHub, Gitlab.

ETL:

Informatics 9.6, Data Stage, SSIS.

Databases:

RDBMS: PostgreSQL SQL, Oracle (9i,10g, 11g), MySQL

NoSQL: Cassandra, MongoDB, Dynamo DB

IDE:

Eclipse, PyCharm, Spyder and Sublime Text.

Operating Systems:

Linux 4.x/5.x/6.x, Ubuntu.

Protocols:

TCP/IP, HTTP/HTTPS, SOAP, SMTP, REST

Deployment Tools:

AWS, EC2, S3, Amazon Lambda, Docker, Jenkins, Kubernetes

Testing Tools:

Gitlab, Bugzilla and JIRA.

PROFESSIONAL EXPERIENCE:

Client: BCBS, Newark, NJ Jun 2021 Present

Role: Sr. Python Developer

Responsibilities:

Responsible for gathering requirements, system analysis, design, development, testing and deployment.

Participated in the complete SDLC process.

Developed web-based open stack applications using Python and Django for large dataset analysis.

Extensively used regular expressions and core features in Python using lambda, map, reduce etc. and effectively implemented logging feature using python logging library and profiling using cProfile.

Used Numpy for Numerical analysis for Insurance premium.

Created an on-premises CI/CD solution using Jenkins and the Pipeline plugin which uses pipeline as code.

Developed REST Microservices which are like API’s used for Home Automation. They also keep the data in synchronization between two database services.

Configuring auto scalable and highly available microservices set with monitoring and logging using AWS, Docker, and Jenkins

Developed Restful Microservices using Flask and Django and deployed on AWS servers using EBS and EC2.

Develop Python microservices with Django/Flask framework for Alchemer and its internal Web Applications.

Creating restful web services for Catalog and Pricing with Django MVT, MySQL, and MongoDB.

Developed tools to automate some base tasks using Shell Scripting, Python.

Used Jenkins pipelines to drive all micro services builds out to the Docker registry and then deployed to Kubernetes, Created Pods and managed using Kubernetes.

Worked on AWS EC2/VPC/S3/SQS/SNS based on automation Terraform, Python.

Responsible for configuring NoSQL databases like Apache Cassandra and Mongo DB to for increasing compatibility with Django.

Designed and implemented RESTful APIs using Node.js, ensuring efficient communication between frontend and backend systems.

Identified Various Data Components throw XML/Jason and CSV Data files.

Used RESTful Web Services in the application development.

Develop programs to automate the testing of controller in CI/CD environment using Python, Git and Linux command line.

Implemented a J2EE bookstore application using Hibernate, Servlets, JNDI, JMS, JDBC, SOAP, and Rest-based Web services. Leveraged concurrent programming for performance optimization.

The architecture featured EJBs for business logic, JPA for database interactions, and Java EE security for authentication. Asynchronous processing utilized MDB, while custom exception handling and rigorous testing enhanced reliability. The application showcased expertise in diverse J2EE technologies.

Developed a Python data processing pipeline that pulls data from many sources, handles synchronous and asynchronous requests, processes, and dumps/ingests data into various data repositories/databases.

REST APIs were created with Python 3.x, FastAPI, and Flask REST Plus, and were deployed with Docker and Kubernetes (K8S).

Developed low-latency, high-performance APIs that enable declarative data fetching with GraphQL to facilitate client-heavy data analytic processes using subscription models (Graph Query Language).

Implemented middleware components in Node.js applications to handle tasks such as authentication, logging, and error handling.

Responsible for configuring a Python REST API framework utilizing Flask and FastAPI, as well as delivering interactive Open API Standard (aka. Swagger) API documentation to other cross-functional teams.

Heavily leveraged Python’s graphics APIs for creating graphics and serialization libraries for encoding data in XML/JSON formats.

Having experienced in Agile Methodologies, Scrum stories and sprints experience in a Python based environment, along with data analytics, data wrangling and Excel data extracts.

Worked under DEVOPS environment for continuous integration and continuous deployment using Jenkins and puppet.

Proficient in developing websites and web applications using MYSQL, AWS, Flask, HTML, XML, JSON, CSS, & AJAX.

Configured and deployed project using the Amazon EC2 on AWS.

Designed and developed data management system using MySQL. Involved in Agile Methodologies and SCRUM Process.

Created unit test/regression test framework for working/new code.

Create Pyspark frame to bring data from DB2 to Amazon S3.

Using version control tool - Git with Jenkins to accumulate all the work done by team members.

Using agile methodology - SCRUM, along with JIRA for project...

Responsible for debugging and troubleshooting the web application.

Participated in writing scripts for test automation.

Environment: Python 3.7, Django, Fast API, MySQL, Oracle 11.2, Linux, CI/CD, Eclipse, HTML, JSON, Node.js SVN, CSS, AJAX, Bugzilla, Apache Web Server, Apache spark, Git, Jenkins.

Client: Alaskan Airlines, Seattle, WA Feb 2020 - May2021

Role: Sr. Python/AWS Developer

Responsibilities:

Developed application using Amazon Web Service (AWS) like EC2, cloud search, Elastic load balancer ELB, Cloud Deploy and monitor scalable infrastructure on Amazon web services (AWS) & configuration management using puppet.

Developed, tested, and deployed Business feature set in Node.js with Express and MongoDB backend, incorporating APIs.

Rewrote one of the key pages, which allows users to manage their content. The task involved investigation of the AngularJS UI-Grid as well as refactoring of several backend methods.

Built AngularJS modules, controllers, pop up modals, and file uploaders.

Worked on server-side applications with Django using Python programming.

Interacted with 3rd party APIs and built RESTful APIs using NodeJS

Developed GUI using Python and Django for dynamically displaying the test block documentation and other features of Python code using a web browser.

Worked on MongoDB database concepts such as locking, transactions, indexes, Sharding, replication, schema design.

Embedded risk partner with primary responsibility to identify and mitigate internal and external risks to plan for Assumed& Ceded reinsurance, Trade credit & surety, finance, claims and procurement functions.

Engineered real-time applications, including a chat application using Node.js and WebSocket technology, ensuring seamless communication and instant updates.

Developed enhancements to MongoDB architecture to improve performance and scalability.

Involved in Installing Hadoop, Map Reduce, HDFS, and AWS and developed multiple MapReduce jobs in Hive for data cleaning and pre-processing.

Involved in writing SQL queries implementing functions, triggers, cursors, object types, sequences, indexes etc.

Created Data tables utilizing MySQL utilized Jinja to access data and display it in the front end.

Worked on Automation of data pulls from SQL Server to Hadoop eco system via SQOOP.

Contributed in the design and creation of RESTful APIs using Python/Django/Django Rest Framework.

Imported millions of structured data from relational databases using Sqoop import to process using Spark and stored the data into HDFS in CSV format.

Strong knowledge of all phases of SDLC and Strong working knowledge of Software testing (Functional Testing, Regression Testing, Load Testing).

Installing and maintaining the Hadoop - Spark cluster from the scratch in a plain Linux environment and defining the code outputs as PMML.

Implemented comprehensive unit tests for Node.js applications using testing frameworks like Mocha or Jest, ensuring code reliability and facilitating a test-driven development approach.

Complex loan origination and underwriting strategies can be flexibly defined and can include the implementation and deployment of credit risk scoring and risk rating models through model import (PMML).

Successfully implemented Apache Spark and Spark Streaming applications for large scale data.

Writing API documentation for onboarding developers on microservices platform of Autodesk.

Environment: Python, Postman, JIRA, Boto3, Jenkins, Docker, Terraform, AWS Lambda, AWS Route53, AWS EC2, AWS IAM, S3, VPC, PyCharm, MySQL, HTML, Web Services, JSON, Node.js MongoDB, SQL Workbench.

Client: Cisco, CA Nov 2018 – Feb2020

Role: Sr. Python/AWS Developer

Responsibilities:

Responsible for gathering requirements, system analysis, design, development, testing and

deployment. Participated in the complete SDLC process.

Used Python and Django for XML, data exchange and business logic implementation - built database.

Model, Views and APIs using Python for interactive web-based solutions.

Creating web services for Catalog and Pricing with Django MVT, PostgreSQL SQL.

Developed views and templates with Python and Django’s view controller and template language to create a user-friendly website interface.

Wrote Python scripts to parse XML documents and load the data in database.

Used Amazon Web Services (AWS) for improved efficiency of storage and fast access.

Developed AWS Lambda functions in Python using S3 and SQS triggers to automate workflows and developed Python scripts to store and retrieve objects in AWS S3 buckets.

Perform S3 buckets creation, policies and IAM role-based policies and configuring user access levels.

Used Amazon EMR for map reduction jobs and test locally using Jenkins.

Developed scripts in Python using boto3 API to retrieve messages from various events.

Using raw files loaded data files from URLs to Amazon S3 Bucket.

Writing Sub Queries, Stored Procedures, Triggers, Cursors, and Functions.

Worked on test cases using PyTest framework - test cases, test suites and bug fixing.

Perform Production support by creating a high priority incident tickets and work with members of the team to resolve incidents within SLA.

Experienced in Agile Methodologies and SCRUM Process.

Experience of source control systems, preferably GIT

Environment: Python 3, Django, MySQL, Oracle 11.2, Linux, CI/CD, Eclipse, HTML, SVN, CSS, AJAX, Bugzilla, Apache Web Server, Apache spark, Git, Jenkins.

Client: Duck Creek Technologies, India Feb 2014 – Oct 2018

Role: Sr. Python AWS Developer

Responsibilities:

Integrated various Lambda functions into AWS with sole purpose of finding vulnerabilities in containers and images.

Design and developed an entire module of architecture in Step Function that enriches the data received from Data Dog API.

Analyzed all the vulnerability data received from Qualys API for a period interval starting from range of 24 hrs. to a maximum period of 30 days.

Gathered all the required information from Qualys based on unique ASV and differentiated the vulnerabilities based on their severity levels.

Developed a separate documentation with all the vulnerability findings and proposed best solutions to remediate the vulnerabilities.

Consumed and utilized the data from Qualys API along with ServiceNow API regarding the vulnerability information from various docker images/containers, non-cloud assets and write into NoSQL cache database.

Designed a Qualys Jenkin plugin using CI/CD pipeline and made several builds to ensure that there are no latency issues in response time.

Create test plans and test cases and execute manual (and create automated) test cases, report and track issues in ticketing systems, and verify fixes.

Understand and work on Component Integration Testing, Interface Testing, System Integration testing and End to End Integration Testing

With Capital One Vault Secrets I have encrypted and stored the data regarding sensitive information like user credentials and passwords.

Involved in deployment of AWS EC2 instances, provisioning on AWS environments and implementation of security group.

Integrated the SONAR analysis and Erato code scans within the Jenkins file to find the compliance and security issues of the source code.

Environment: Python, HTML, Postman, DataDog, JSON, XML, JIRA, Boto3, Jenkins, Docker, Lambda.



Contact this candidate