Sign in

Software Developer

Virginia, United States
March 21, 2019

Contact this candidate


Arunkumar Kokkula


Professional Summary

Software Developer with around 2 years of IT Industry experience in designing, developing and maintaining large-scale systems/applications with significant expertise in Python 2.7, 3.2 and 3.6.

Experienced with full software development life-cycle, architecting scalable platforms, object-oriented programming, database design and agile methodologies

Expert knowledge of and experience in Object Oriented Design and Programming concepts.

Experience in writing data processing frameworks for large scale applications

Hands on experience in using Spark ecosystem components like S3, RDS, Snowflake, REST, Kafka

Experience in closely working with data analysis and Data scientists for converting POCs into production grade software.

Experience in consuming data from Kafka into spark micro batches.

Worked on spark application tuning and resource allocations based on use case

Experience in Shell Scripting, SQL Server, UNIX and Linux.

Experience in building applications in AWS infrastructure; Cloud Formation Templates, Cloud Watch Alarms, S3, RDS, Security Groups, VPCs, EC2.

Familiar with JSON based REST Web services and Amazon Web services.

Automated the continuous integration and deployments using Jenkins and AWS Cloud Templates deployment services (Lambda).

Experience with web scrapping using beautiful soup; Created own flight logistic application using information from scrapped data.

Experience in project deployment using Jenkins and using web services like Amazon Web Services (AWS) EC2, Cloud Formation Templates, AWS S3 and Cloud watch.

Exposure to ML/DL ecosystem, algorithmic approaches along with design and performance constraints

Experience in tools in java as part of personal development.

Exposure to Tensor flow and PyToch frameworks & hardware for AI.

Professional skills

Programming Languages: Python, Java, JavaScript

Web services: RESTful

Data bases: Oracle 10/11g, MySQL, SQL Server

IDE’s and tools: Eclipse, Pycharm, NetBeans

OS &Environment: XP, windows, Linux, Unix, Ubuntu

Unix Shell Scripting: Unix Shell Scripting

Version control: GitHub

Development Methodologies: Agile, Scrum

Hadoop: HDFS, MapReduce, Spark

Machine Learning: KNN, Gradient Descent, Back Propagation.

Professional Experience

Work Location: Capital One, Richmond VA 04/2018 to Till Date

Employer: SRK Systems Inc, Naperville, IL

Software Developer

Capital One maintains an ecosystem of internal data processing applications designed to prevent fraud of various types including transactional and application fraud with dynamic fraud patterns. This ecosystem consists of highly available multiple spark clusters running rule on kafka streams and batch data.

Developed multiplatform PySpark framework used to create spark jobs and provide SQL like interface for data analysts. This application can be used for both creating spark jobs and productionize them.

Responsible for migrating findings from Data Scientist and fraud investigators into existing fraud defenses. And over seeing Data Science Approaches so that functionalities can be converted to production environment.

Designed data flow for new business needs and participated in architectural and workflow discussions.

Responsible for maintaining allocating and tuning resources for faster performance for heavy data loads.

Resolved Big Data small files issue in the organization using spark Hadoop configurations and dependency configurations.

As part of Capital One Fraud Prevention team, worked on maintaining old and developing new rule engines along with application resiliency strategies. Created new Fraud defenses using existing data processing patterns

Responsible for developmental and production data security aspects along with upgrading the system with new software and infrastructural features

Converted exiting traditional Teradata fraud defenses into cloud architecture and spark SQL

Create QA data by performing data analysis using Databricks notebooks with maintained entropy.

Configured Apache Arrow for columnar data processing at processors in pyspark 2.3

Migrate exiting fraud defenses from Teradata to PySpark environment. Tune exiting spark jobs for performance.

Created modules for switching cluster stack (active and inactive) using Jenkins, AWS lambda to EC2 cluster.

Converted monolithic application into a pip installable package and incorporated it to schedule Jupyter Notebooks using papermill

Built CI/CD pipeline using Jenkins and AWS lambda. Build secret management system. Built redundancy catch up in fraud case creation.

Upgraded Spark Hadoop Version and participated in building Custom Assume Role Credential Provider in jar for spark session assume role auto renewal.

Environment: Linux, Rel7, PySpark, Data Lake, Hadoop, Teradata, PSQL, Python, JSON, AWS, CI/CD, RISK management, Apache Arrow, Credit Card Fraud Prevention, REST, GitHub, Notebooks.

Employer: WAFTS Solutions, Farmington Hills, MI 08/2017 to 03/2018

Programmer Analyst

Data management system that consumes various types of data to provide insights and generates reports for Small to Mid-range businesses.

Develop python programs to perform fixed operations on Excel and CSV datasets

Develop SQL code to be used with automated processes to identify revenue opportunities and financial issues.

Data mine large datasets in Relational Databases to find emerging issues and root-cause in provisioning, marketing and billing systems

Drive timely resolution of marketing issues for a seamless customer experience.

Proactively monitor daily processes and results to ensure consistent coverage.

Create REST APIs for web services using python Flask, Django frameworks.

Environment: Ubuntu, Python, SQL, JSON, AWS, REST, Jupyter Notebooks, GitHub.


Master of Science in Internet and Web Design, Wilmington University, Wilmington DE, 2017

Bachelor of Technology in Electronics and Communication Engineering, India, 2015

Contact this candidate