Kiran Kanjirathingal Vincent Data Engineer
Kitchener, Ontario Phone: +1-647-***-****
Email: *************@*****.*** LinkedIn: https://www.linkedin.com/in/kirankvincent/
Senior Data Engineer with a strong background in Azure Data tools, Big Data Engineering, and Data
Migration.
Profile Summary
• Experienced Data Engineer with a strong background in Azure Data Engineering, Big Data Engineering,
and Cloud.
• Skilled in designing and implementing data pipelines using Azure Data Factory, Synapse, Databricks, and
Data Lake.
• Proficient in data modeling techniques with expertise in Visma Data Vault Modeler.
• Solid understanding of Hadoop, Spark, SQL, Java Spring Boot, and Data Warehouse Endpoint API.
• Committed to delivering high-quality solutions and driving business value through data-driven insights.
• 7.6 years of experience in Cloud technologies, including software development, requirement gathering,
application maintenance, production support, automation, and team management.
• Expertise in service and backend development of Java applications.
• Proficient in various tools and technologies including Azure Data Factory, Azure Synapse, Azure
Databricks, Azure Data Lake, Data Vault, Hadoop, Spark, Hive, Sqoop, Oracle, MySQL, SQL Server,
MemSQL, MongoDB, Kafka, Informatica, UC4, Git, SVN, Python, Java, SQL.
• Familiar with Agile, Waterfall, Scrum, and Kanban project methodologies.
IT Skill Set
Data Engineering Azure Data Factory, Azure Synapse, Azure Databricks, Azure Data Lake
Data Modelling Visma Data Vault Modeler, Data Warehousing
Big Data Technologies Hadoop, Apache Spark, Hive, Sqoop, Apache Kafka
Database Systems Oracle, MySQL, SQL Server, MemSQL, MongoDB, NoSQL
ETL and Data Integration Informatica, UC4
Version control Git, SVN
Languages Python, Java, Java Spring Boot, SQL
Project Methodologies Agile, Waterfall, Scrum, Kanban
Work Experience
Technology Lead – Infosys Ltd
Aug 2024 – Mar 2025
Project: CN Rail - CNTINEL and Data Standardization
● Implemented Data Warehouse Endpoint APIs on Databricks for improved data access.
● Gathered and analyzed business requirements to design feasible solutions
● Developed a data collector from ADO using Java Spring Boot to MongoDB.
● Created Databricks jobs to integrate ML models and complex data logic.
● Developed data pipelines in Azure Databricks using Python for data extraction and transformation.
● Established and maintained software standards, policies, and procedures.
● Monitored system performance and capacity for optimal operation.
Data Engineer – Spruce Infotech
Feb 2024 – Aug 2024
Project: CN Rail - CNTINEL and Data Standardization
● Designed and developed data pipelines using Azure Data Bricks, enabling smooth data flow
between systems and data sources.
● Utilized ADLS and Databricks for efficient data ingestion, processing, and analysis.
● Developed and deployed Java Spring Boot applications.
● Monitored system performance and capacity for optimal operation.
● Implemented Data Warehouse Endpoint APIs on Databricks to enhance data access.
Technology Lead – Infosys Ltd
Jan 2022 – Dec 2022
● Developed and maintained real-time data processing solutions using PySpark Streaming, enabling
immediate insights and decision-making.
● Designed and implemented scalable and performant data storage and retrieval systems using
MongoDB and Realm Functions.
● Designed and implemented data pipelines using ADF, enabling smooth data flow between systems
and data sources.
● Developed ETL workflows and data integration solutions, ensuring seamless data extraction,
transformation, and loading.
Technology Analyst – Infosys Ltd
Sep 2019 – Jan 2022
● Worked as a Data Engineer for Posti, Finland Postal and Logistics Company, leveraging Azure Data
Factory, Synapse, Databricks, and Data Lake to design and implement efficient data pipelines.
● Developed and maintained scalable data integration solutions using Azure Data Factory, ensuring
the smooth flow of data between various systems (like relational database, SSIS) through
Azure DevOps.
● Developed and deployed Java Spring Boot applications and Implemented Data Warehouse
Endpoint APIs on Databricks to enhance data access.
● Implemented data models using Visma Data Vault Modeler, ensuring data consistency, traceability,
and maintainability.
● Deployed data pipeline though CI/CD pipelines.
● Collaborated with cross-functional teams to gather requirements, translate business needs into
technical solutions, and deliver high-quality data engineering solutions.
● Optimized data pipelines for performance and scalability, improving overall data processing
efficiency. Conducted data profiling, cleansing, and transformation activities to ensure data quality
and integrity.
● Worked in an Agile environment, actively participating in Scrum meetings and contributing to sprint
planning, backlog grooming, and retrospective sessions.
Software Engineer – Feathersoft Info Solution
Aug 2016 – Sep 2019
● Developed software solutions for clients in various industries, focusing on data processing and
integration.
● Designed and implemented ETL workflows using Informatica and UC4, ensuring seamless data
extraction, transformation, and loading.
● Tested, debugged, and deployed job plans and scripts in production environments.
● Participated in the development of Java applications for data parsing, validation, and insertion into
databases.
● Developed an ETL tool for batch-wise processing of hospital data files from AWS S3. Processed files
from HDFS, performed validation, parsing, and database insertion for incremental records.
● Contributed to the evaluation and implementation of Elastic Search for efficient search and
retrieval of gene variant data. Built a Java application to fetch gene data from an automated
application to Elastic Search.
● Developed dynamic queries based on customer requirements and applied them to search gene
data in Elastic Search. Improved search and retrieval efficiency for gene variant data
● Developed a consolidated view of job log emails using NiFi, Tableau Reports, and MemSQL.
● Built several custom NiFi processors for reading and processing emails, and stored email dates in
MemSQL using Kafka. Provided an accurate and easy-to-use email dashboard for quick analysis and
insights.
● Collaborated with clients to gather requirements, provide technical expertise, and deliver
high-quality software solutions.
Education
● Conestoga College, Kitchener, Ontario
Graduate Certificate, Virtualization and Cloud Computing
January 2023 - August 2023
● University of Calicut, Kerala, India
Bachelor of Technology (B.Tech.), Computer Science and Engineering
2011 - 2015
Achievements
● Received client appreciation and best performer award for the work Posti project in Infosys.