SAI MANOJ T
*********@*****.*** +1-470-***-****
Professional Summary
Accomplished ETL Developer with over 4 years of experience designing and optimizing high-performance data pipelines to drive critical business applications.
Proficient in leveraging Informatica PowerCenter (8.x/9.x), Oracle Data Integrator, and Google Cloud Dataflow to build scalable ETL workflows, ensuring exceptional data accuracy and consistency.
Skilled in managing and modeling data within Teradata, Snowflake, and Amazon Redshift, utilizing tools like Erwin, ER, and MS Visio for robust data architecture.
Strong programming expertise in Python, PySpark, Java, SQL, and Shell Scripting to develop efficient and reliable data processing solutions.
Experienced in implementing rigorous data quality rules using Informatica IICS Data Quality to enhance data reliability and integrity.
Adept at streamlining operations with cloud technologies (AWS, GCP, Azure) and containerization tools (Docker, Kubernetes) for seamless ETL pipeline deployment.
Well-versed in Agile and Scrum methodologies, utilizing Git, Jenkins, and Jira for effective version control and CI/CD processes.
Proficient in job scheduling and automation using Talend Administration Console (TAC) and Tidel to optimize workflow efficiency.
Skilled in transforming data into actionable insights using Tableau and Zeppelin for data visualization and analytics.
Collaborative team player, consistently partnering with cross-functional teams to deliver impactful data solutions and achieve business success.
Skills
ETL Tools: Informatica PowerCenter V8.X/9.X Oracle Data Integrator Google Cloud Dataflow DW Tools: Teradata SQL Snowflake Erwin ER MS Visio Teradata Studio RDBMS: MySQL MongoDB PostgreSQL Oracle 10G/9i/8.x Programming Languages: Java Python PySpark Shell Scripting SQL C C++ Virtual Basic XML Big Data Technologies: Apache Spark (PySpark), Apache Hive, Apache Kafka, AWS Glue, Google Cloud Dataflow Data Warehousing: Amazon Redshift Snowflake
Cloud Technologies: AWS Google Cloud Platform MS Azure Methodologies: Agile Scrum
Development Tools: Visual Studio Code Jira TOAD SQL Developer SQL Plus Containerization and Orchestration: Docker Kubernetes Version Controls and CI/CD: Git GitHub Jenkins Data Visualization and Analytics: Tableau Zeppelin Job Scheduling and Automation: Talend Administration Console (TAC) Tidel Data Governance and Metadata Management: Collibra
Others: ServiceNow Excel Office 365 Pro Plus SAP SAP ABAP Workflow Development Data Modeling HTML JavaScript CSS
Education
California State University, San Bernardino, CA
Master of Science in Computer Science
Relevant Coursework: Software Engineering Operating System Modern Computer Architecture Database Management Artificial Intelligence Machine Learning Computational Complexity Jawaharlal Nehru Technological University of Hyderabad, India Bachelor of Technology in Information Technology
Work Experience
Data Engineer Tek International Inc. January 2025 – Present
Assisted in designing, developing, and testing ETL pipelines for data extraction, transformation, and loading from multiple sources.
Developed SQL queries and Python scripts to automate data processing and improve ETL efficiency.
Utilized Microsoft Excel for data validation, creating pivot tables to analyze and summarize large datasets.
Developed Excel macros with VBA to automate repetitive data cleansing tasks, improving efficiency.
Configured and managed IICS cloud resources and security settings effectively.
Worked with cloud platforms (AWS, GCP) to manage data storage, processing, and integration.
Collaborated with data engineers and analysts to optimize ETL workflows and troubleshoot performance issues.
Assisted in data validation, cleansing, and documentation to ensure accuracy and consistency in reporting. Associate Software Engineer TATA CONSULTANCY SERVICES LIMITED March 2020- July 2023 ETL Developer:
Built and maintained ETL pipelines using SQL, Python, and AWS to process policyholder and claims data, ensuring seamless integration across systems.
Designed and executed data transformation logic to standardize customer financial records, improving risk assessment and fraud detection.
Experienced in complex SQL scripts and PL/SQL packages, to extract data from various source tables of the data warehouse.
Automated data ingestion from multiple sources (policy management systems, payment processors) using cloud-based ETL tools, reducing manual effort by 40%.
Utilized ServiceNow to log, track, and resolve ETL pipeline issues, streamlining incident management.
Created ServiceNow workflows to automate ETL failure notifications, reducing response time.
Developed ServiceNow dashboards to monitor ETL job performance, enhancing visibility for stakeholders.
Configured ServiceNow incident reports to track recurring ETL errors, enabling proactive fixes.
Implemented data quality checks and validation scripts, ensuring compliance with regulatory frameworks (SOX, IFRS, GAAP) in financial reporting.
Optimized ETL workflows and database queries to improve performance, reducing data processing time for insurance claims analysis by 30%.
Involved in Testing the ETL process that meets the business requirement.
Developed pivot tables in Excel to validate and analyse insurance data, ensuring accuracy in reporting.
Employed Microsoft Excel in Office 365 ProPlus with pivot tables and VBA macros to validate insurance data and automate reporting tasks.
Used Microsoft PowerPoint in Office 365 ProPlus to present ETL performance metrics to business teams, aiding decision- making.
Wrote VBA macros to automate data cleansing and reconciliation tasks, saving time in financial reporting.
Used Talend and designed data conversations from large variety of source systems including Oracle, SQL server, Teradata, Netezza, SQL server, DB2, Hive and non-relational sources like XML, flat file and Delimited files.
Schedule and monitor the IICS workflows, ensuring smooth execution and timely completion of tasks.
Scheduled IICS task flows using Control-M.
Extracting, Transforming and Loading the data from Source to Staging and Staging to target according to the Business requirements.
Used tResdshiftUnload Component for unloading data from Redshift database to S3 Bucket.
Monitored, debugged, and troubleshot ETL failures using logging frameworks and performance tuning techniques, minimizing downtime in financial reporting systems.
Utilizing Direct Query vs Import Mode in Power BI to balance performance and real-time reporting needs.
Optimizing our reports and dashboards of Power BI performance using aggregations, composite models, and advanced DAX calculations.
Ability to meet deadlines and handle multiple tasks. Environment: Talend Enterprise Big Data Edition 5.1, Talend Administrator Console, Informatica 10.5, MS SQL Server 2015/2018,Oracle 11g, TOAD, ServiceNow, SQL, Python, Java,JDBC, Servlets, Excel, JSP, EJB, JAX-RS, Spring, Hibernate, RESTful APIs, SAP Integration, AWS, cloud-based ETL tools, Power BI (Direct Query, Import Mode, Aggregations, Composite Models, DAX), microservices architecture, data warehousing, performance tuning, query optimization, logging frameworks, error handling, and compliance with SOX, IFRS, and GAAP regulatory frameworks Data Science Engineer Intern Techo Bytes June 2019 – August 2019
• Improved model accuracy through feature engineering and tuning.
• Reduced data processing time with optimized pipelines.
• Hands-on experience regarding the Informatica Power Center.
• Deployed predictive models to enhance business efficiency.
• Created interactive dashboards using Talend, Tableau, and Plotly.
• Automated data preprocessing workflows, reducing manual effort. Projects
Financial Transactions ETL for Regulatory Compliance Responsibilities:
• Developed and deployed an ETL pipeline to extract, transform, and load financial transaction data from multiple banking and insurance systems into a centralized data warehouse.
• Standardized financial records by implementing data transformation rules, ensuring compliance with SOX, IFRS, and GAAP regulations.
• Automated data validation and reconciliation using SQL and Python, minimizing discrepancies in financial reporting.
• Integrated real-time transaction data from payment gateways, logs, and external APIs using Apache Kafka and AWS Glue, enhancing data accuracy.
• Optimized ETL workflows and query performance, reducing batch processing time by 40% and improving overall system efficiency.
• Monitored and troubleshot ETL job failures using logging frameworks and performance tuning techniques, ensuring minimal downtime.
• Designed Power BI dashboards to provide compliance teams with real-time insights into financial transactions and potential anomalies.
Outcome:
• Automated financial reporting processes, reducing manual effort and improving data accuracy by 35%.
• Ensured full compliance with SOX, IFRS, and GAAP, streamlining financial data processing and regulatory audits.
• Strengthened fraud detection mechanisms, lowering financial discrepancies in transaction records by 30%. • Improved reporting efficiency, enabling real-time auditing and risk assessment for financial stakeholders. Certifications
• Internship certificate in Talend
• Python Certificate in Coursera
• ETL & SQL Certificate in Tata Consultancy Services