Post Job Free

Resume

Sign in

Data Engineer Software Development

Location:
Georgetown, TX
Posted:
April 10, 2024

Contact this candidate

Resume:

Yeswanth K

Data Engineer

408-***-****/ad4wx2@r.postjobfree.com

PROFESSIONAL SUMMARY

-Over 7+ years of experience in Software Development Life Cycle (SDLC) that includes Analyzing, Designing, Coding, Testing, implementation & Production Support with Agile and Scrum.

-Actively working on Integrating new data management technologies and software engineering tools into existing structures.

-Experience in Data Extraction, Transformation and loading of data from multiple data sources into target databases, using Azure Databricks, Azure SQL, SQL Server.

-Exposure on NIFI, Informatica, StreamSets to ingest data from various sources, transform, enrich and load data into various destinations.

-Hands-on experience in working with Denodo platform for Data virtualization.

-Developing experience using web technologies such as HTML5, JavaScript, CSS3, AJAX, jQuery, AngularJS, Node.js

-Working with Rest API’s for bring in data into Hadoop environments for building sentiment analysis

projects.

-Extensive work on bringing Hive tables and other traditional database tables into Denodo platform to provide universal connectivity to access any type of data source while data validations.

-Proficient knowledge and hands-on experience with writing shell scripts in Linux.

-Hands-on experience in installing, configuring and using Apache Hadoop ecosystem components like Hadoop Distributed File System (HDFS), MapReduce, HIVE, HBase, Apache Crunch, ZOOKEEPER, Hue.

-Experienced with different data formats like JSON, Avro, parquet and ORC.

-Involved with Data Analysis primarily Identifying Data Sets, Source Data, Source Metadata, Data Definitions, and Data Formats

-Used Angular JS as a framework to create a Single Page Application (SPA) that can bind data to specific views and synchronize data with a server.

-Hands-on experience with NoSQL Databases like HBase for performing analytical operations.

-Experience in creating and maintaining design documents from system specifications, code development/unit testing.

-A continuing aspiration to advance in new technologies while providing quality programming with attention to detail in an environment that encourages growth.

-Collaborated with the team in order to discuss, design and implement.

-Working in a fast-paced environment and ability to manage multiple projects of substantial level.

-Experience with User Acceptance Testing (UAT) and Accessibility Testing.

Excellent communication skills to discuss the projects with technical and non-technical people. During the sprint meetings for Agile as well as Scrum development methodology.

Technical Skills:

Hadoop/Big data:

HDFS, MapReduce, Hive, Spark, HBase, Flume, Kafka, Azure, Cloudera Manager.

Web Technologies:

HTML, CSS, jQuery, Bootstrap, Angular.js, Node.js, D3.js

Programming Languages:

Java, JavaScript, Python, C++, SQL, UNIX Shell Scripting.

IDE’s:

Eclipse, Visual Studio Code, IntelliJ IDEA.

Databases:

SQL Server, MySQL, MongoDB, Influx, HBase.

Operating System:

Windows, Linux.

PROFESSIONAL EXPERIENCE

Wargaming Austin Inc.

Austin, TX May 2018 - Present

Data Engineer

oWorking Developing, Optimizing, Testing and Maintain Streamsets & NIFI pipeline architectures, That flow Streaming and Batch Data into Cloudera / Hadoop (HDFS) and other Big Data echo systems.

oCreate and maintain optimal Data orchestration using Streamsets & Informatica, And Maintained Data pipelines uptime of 99% while ingesting Streaming & Batch data across five different primary data sources using Kafka,

Google Biq-Query, API, python & S3. And monitoring flows to identify Failed, Long running jobs, disk usage, memory usage, files in a directory and send alerts once exceeds threshold. Building Streaming services for real time processing by using Kafka and also Batch jobs using JDBC and API sources

oResearching the Big Data, Cloud computing technologies and applying them to the Real-world scenarios across game data applications. Utilize Data virtualization, Advanced visualization, Real Time data analytics, Web APIs and restful services to streamline the data across multiple Schemas that are consumed by other Cloud based applications.

oIdentifying ways to improve data reliability, efficiency, and quality. With REST API’s (JSON) pipelines to bring game reviews data into HDFS and using it for sentiment analysis for Data science/AI to compare user reviews on product and use marketing trends to improve the game engine with available data.

oWorking with game engineers and DW Architect to bring data into Business Intelligence systems for utilization with custom-engineered Big data tools.

oUsing data virtualization tools like Denodo to extract the data from JDBC (my-sql logs) sources to build near real-time projects to find the toxic chat detection by using in game messages by users for WG legendary game like WOT, WOWS and WOTB for all regions by using 2 days of logs.

oWorked on New Technologies like Snowflake, Google Big Query and AWS for different use-cases. By using cloud-based tools we can optimize our workload and also, we have reduced the dependency on DBA up to 60%.

oDeveloped dashboards for real time analytics with InfluxDB (Grafana/Tableau) by using kafka source

(AVRO), Real Time dashboards are used by Top level management for new game products to improve wait time from 1 week of data to a few minutes by using time series databases like InfluxDB and planning to use FLUX query language for more use-cases.

oIngesting data in the form pipelines from Kafka topics using Stream Sets to HDFS (cloud-era) & cloud platforms. And Supporting Enterprise level Apache NIFI & Streamsets Data Integration Application in organization, for all formats Including new enhancement testing, upgrades, Feature requesting as one point of contact.

oPerformance Tuning and Optimization of all Data Ingestion and Data Integration processes. Utilize Data virtualization, Advanced Visualization, Real Time data analytics, Web APIs and restful services to streamline the data across multiple Schemas that are consumed by other cloud-based applications.

oEnhance the organizational and leadership skills learned by effective decision making while gathering the business requirements from the client/stakeholder for ongoing projects and by involving the team in regular discussions to understand the business requirements for developing good and user-friendly applications. Demonstrate ability and apply them in research to analyze and solve problems.

oResearch and analyze new challenging technologies in Information Technology based on business needs and implement product features from ideation to its launch and Pitch new enhancements to executive teams and improve user experiences to customers based on the analysis of data points. Help stakeholders and key business users to make appropriate business decisions using data analytics. Enhancing writing & understanding script for the Data Integration processor using Python/Jython.

oBackfilling, validating pipelines for the daily loads or Batch pipelines using HUE/IMPALA, Worked on Comparison of Data Structures in Schema registry Avro formats using CURL commands.

oContinuous development and contribution to the field of Information systems by use of Information governance skills and apply them to design and develop software tools with best practices up to industry standards. Conduct extensive independent research in the market and upskill in the areas of management, professional ethics and morals. Work with Business heads to gather requirements, prototyping, architecting, road map, building and executing test plans, performing quality reviews, fixing operational issues and supporting.

oWorking with the thought of architecture and design of solutions for populating technical assets in Collibra Data Governance Center Physical Data Model assets from OEMM (GDPR), actively drive Proof Of Concept (POC) and Proof Of Technology (POT) to Implement. Developed a POC using Mulesoft (AnyPoint studio) and gave a technical presentation to management and also POC using cluster and standalone pipelines using SS to test throughput for various actions and also tested spark-sql configs.

oLeads complex projects/programs strategic planning and provides the subject matter expertise and technical direction for project managers including risk management and countermeasures.

oCreates Data Integration routines to support a hybrid data warehousing architecture (Hadoop, Oracle,

NoSQL, Graph DB’s). Routines will drive data feeds for web applications.

oUsed JIRA & Targetprocess to keep track of Bugs to Reduce Downtime, Increase productivity, and Communication. Worked in an Agile development environment in sprint cycles of two weeks by dividing and organizing tasks. Participated in daily scrum and other design related meetings.

Tech Stack: HDFS, Shell Scripting, Map Reduce, Kafka, Cloudera,Azure, TypeScript, D3.js,, Hive, Apache NIFI, Stream Sets, JavaScript, Node.js, Scala, JQuery, Ajax, VQL, SVN, IntelliJ IDE

GVA Infotech

Ashburn, VA Feb 2017 - April 2018

BI Data Visualization

oWorked closely with product owners, UX designers, and other IT teams to understand the scope of the project and accordingly gather functional requirements of the project in order to design and implement the application.

oBuilt multiple independent applications using JS frameworks like angular.

oSuccessfully implemented AutoComplete/Auto Suggest functionality using Ajax, JQuery, Web Service call, and JSON.

oUse advanced techniques to create and manage HTTP requests, responses as Observables, client-side caches.

oDeveloped page layouts, navigation, animation buttons and icons.

oInvolved in maintaining the VM's for deploying apps in prod and pre-pod.

oExperience integrating user identity between traditional on-premise applications and SaaS applications.

oUsing CSS preprocessor SASS to make CSS more efficient and using Bourbon a lightweight Sass toolset that helps you write more effective style sheets.

oParticipate in Code review and Quality Assurance.

oInvolve in Unit testing AngularJS applications using Jasmine/Karma.

oUsed WebStorm and Visual Studio as the IDE for custom development and implementation.

oUsed Git for the purpose of version Control and source code maintenance needs.

oUsed JIRA to keep track of bugs to reduce downtime, increase productivity, and Communication.

Tech Stack: JavaScript, Hive, Cloudera, CSS3, AJAX, jQuery, Angular JS, Backbone JS, HTML5, XML, SQL, NIFI, Windows Platforms, IDE- Eclipse

CMS Info Systems

Chennai, India March 2014 - July 2015

Associate Systems Engineer

oInvolved in gathering business requirements, prelim design, documenting design specifications, negotiating with different business units, code review, unit testing, and system testing.

oDeployed AJAX call for asynchronous data retrieval from the backend.

oWorked with relational SQL and NoSQL databases.Using an advanced level of HTML, XML, JavaScript, CSS and pure CSS layout.

Extensively coded built website with JavaScript, HTML, CSS3, DOM, jQuery and Handled dynamic data using AJAX.

oCollaborating with Product Managers and Software Engineers to provide guidelines on solid UI design.

oDesigned and Developed Database using MYSQL. Extensively used select, insert, update, delete, join and between queries while maintaining the database.

oParticipated in daily meetings for enhancing the features for the portal.

oParticipated in Code review and Quality Assurance.

oCreation of various levels of users for a system with preferences.

oMonitor user information, real-time issues, and support features.

oInvolved in setting up the environment, installing the application and web servers.

Tech Stack: Java, Shell Scripting, HTML, CSS, SQL, Windows.

EDUCATION

UNIVERSITY OF THE CUMBERLANDS Kentucky, USA

Master of Science in Computer Science.

SCSVMV University Chennai, India

Bachelor of Engineering in Computer Science.



Contact this candidate