BINDU
Gen AI Engineer
*****.***@*****.*** linkedin +1-716-***-****
PROFESSIONAL SUMMARY:
● Around 7 years of experience as GenAI Engineer & Data Scientist specializing in LLMs, RAG, and ML solutions
across Finance, Healthcare. Proven expertise in building scalable AI systems, optimizing ML workflows, and
deploying enterprise-grade GenAI applications in multi-cloud environments.
● Extensive expertise in designing, fine-tuning, and deploying Large Language Models (LLMs) including GPT,
LLaMA, Falcon, and Claude with Lang Chain, RAG pipelines, and vector search integrations.
● Led end-to-end application migrations from AWS and Azure to GCP, modernizing enterprise workloads using
GKE, Cloud SQL, Big Query, and Cloud Functions.
● Hands-on experience in application development with Python (Fast API), Node.js, and Java (Spring Boot),
deploying scalable microservices integrated with ML/AI pipelines.
● Designed RAG-powered AI assistants using GPT-4, Pinecone, and Lang Chain to enable intelligent knowledge
retrieval for financial and healthcare clients.
● Fine-tuned domain-specific LLMs with LoRA/PEFT techniques, improving accuracy and contextual relevance
for enterprise use cases.
● Built predictive ML models for fraud detection, customer churn, demand forecasting, and clinical outcomes,
leveraging Random Forest, XGBoost, and Neural Networks.
● Strong background in NLP – developed pipelines for document summarization, named entity recognition
(NER), sentiment analysis, and medical text de-identification.
● Designed and deployed time-series forecasting models for patient monitoring, telecom call volumes, and
financial risk management.
● Automated end-to-end ML and GenAI workflows with Airflow, Kubeflow, MLflow, Jenkins, and GitHub
Actions, ensuring scalable and reproducible model deployments.
● Developed ETL and real-time data pipelines using Kafka, Databricks, Spark, and Snowflake for large-scale data
processing and ML training.
● Implemented cloud-native AI solutions using AWS SageMaker, Bedrock, GCP Vertex AI, and Azure ML across
multi-cloud environments.
● Applied cloud networking, IAM, and data security best practices to ensure compliance with HIPAA, SOC2, and
financial regulatory requirements.
● Designed vector database search solutions with Pinecone, Weaviate, FAISS, and Milvus for
retrieval-augmented generation and semantic search.
● Developed graph-based ML solutions for drug interaction detection and fraud pattern discovery in financial
and telecom datasets.
● Created real-time dashboards and BI reports in Tableau, Power BI, and Looker to monitor AI-driven insights
and performance metrics.
● Implemented monitoring and observability solutions with Stack driver, CloudWatch, Grafana, Prometheus,
and ELK for AI/ML workloads.
● Mentored junior data scientists and engineers, leading AI/ML workshops and GenAI adoption sessions across
client organizations.
● Partnered with cross-functional teams to translate business requirements into scalable AI/ML solutions,
driving digital transformation initiatives across industries.
TECHNICAL SKILLS:
● Generative AI & LLMs: GPT, LLaMA, Falcon, Claude, LangChain, LangGraph, LangFlow, RAG, LoRA/PEFT,
Prompt Engineering
● ML & DL Frameworks: PyTorch, TensorFlow, Scikit-learn, Hugging Face, spaCy, BioBERT
● Programming & Data: Python, R, SQL, Spark, Django, Pandas, NumPy, Flask, Fast API
● Cloud & MLOps: AWS SageMaker, GCP Vertex AI, Azure ML, MLflow, Kubeflow, Docker, Kubernetes
● Vector DBs & Search: Pinecone, Weaviate, FAISS, Milvus, ElasticSearch
● Data Engineering: Airflow, Kafka, Databricks, ETL, Data Lakes, Big Query, Snowflake
● Visualization & BI: Tableau, Power BI, Looker, Plotly, Matplotlib, Seaborn
● Monitoring & Logging: Stackdriver, CloudWatch, ELK Stack, Grafana, Prometheus
● Methodologies: Agile (Scrum, Kanban), SDLC, CI/CD
PROFESSIONAL EXPERIENCE:
Intralot, Chicago, IL APRIL 2024- Present
Role: Sr. GenAI/ML Engineer
Responsibilities:
● Designed and deployed generative AI solutions such as Retrieval-Augmented Generation (RAG) pipelines by
integrating GPT-4, LangChain, and Pinecone, enabling intelligent knowledge retrieval from structured and
unstructured enterprise data sources, significantly improving search accuracy and response quality for
financial domain applications.
● Built an AI-powered customer support assistant leveraging LLMs that streamlined call centre operations by
automating responses to repetitive queries, reducing manual intervention, and enabling real-time escalation
for complex cases across multi-channel support platforms using agentic AI framework LangGraph with multi
agent orchestration.
● Developed Python-based RESTful APIs to integrate LLM services into enterprise applications, ensuring
seamless interaction between AI models and internal systems such as CRM, knowledge management, and
reporting platforms.
● Implemented robust data pipelines using Apache Airflow and Apache Spark to automate the ingestion,
transformation, and preparation of large-scale financial datasets for model training, ensuring high-quality,
reliable, and consistent data availability for GenAI and ML workloads.
● Conducted fine-tuning of LLaMA models with domain-specific financial datasets, improving contextual
accuracy for tasks such as risk analysis, compliance checks, and document summarization tailored to the
banking and financial industry.
● Designed analytical workflows in Google Big Query to process, query, and visualize high-volume financial
transaction data, providing deep insights into customer behaviour, fraud detection patterns, and overall
portfolio performance.
● Applied advanced transformer-based NLP models such as BERT and Roberta for sentiment analysis and named
entity recognition, extracting actionable intelligence from financial documents, customer communications,
and market news feeds.
● Built interactive dashboards in Tableau that consolidated ML-driven risk scores, fraud alerts, and performance
metrics, enabling executives and analysts to track financial risk analytics in real-time with clear visualizations
and drill-down capabilities.
● Developed automated CI/CD pipelines for ML models using MLflow, Docker, and GitHub Actions, ensuring
version control, reproducibility, and seamless deployment of machine learning models into production
environments on cloud platforms.
● Applied cloud security best practices by implementing IAM policies, encryption standards, and audit logging
for financial datasets hosted on multi-cloud platforms, ensuring compliance with strict regulatory frameworks
such as SOC2 and PCI DSS.
● Collaborated with DevOps teams to deploy large-scale ML models on Google Cloud Vertex AI, leveraging
containerized workloads on GKE clusters for high-performance, scalable inference serving to meet
enterprise-level demand.
● Configured and managed hybrid cloud networking between AWS and GCP, including VPNs, VPC peering, and
load balancing, to ensure secure and reliable connectivity across multi-cloud environments supporting AI/ML
applications.
● Created custom embeddings for financial documents using Hugging Face Transformers, enabling semantic
search, similarity matching, and document classification to enhance knowledge discovery within large
repositories of financial records.
● Automated monitoring and observability workflows by integrating Google Stack driver and Prometheus,
setting up real-time alerts and dashboards for model performance, infrastructure health, and latency tracking
across distributed systems.
● Supported business stakeholders with ad-hoc ML model performance reporting, preparing clear insights and
tailored visualizations to help decision-makers evaluate the effectiveness of deployed AI solutions and align
outcomes with organizational goals and delivered in Agile sprints with close coordination across data,
DevOps, and product teams.
Environment: Python, PyTorch, Hugging Face, Lang Chain, Pinecone, GCP Vertex AI, Big Query, Airflow, Spark,
Tableau, Docker, Kubernetes, MLflow, Prometheus, Stack driver.
Elevance Health Norfolk Virginia APRIL 2023 - Mar 2024
Role: AI/ML Engineer
Responsibilities:
● Developed advanced NLP pipelines for medical record classification by leveraging transformer models such as
BERT and Bio BERT, enabling automated extraction and categorization of clinical terms, diagnoses, and
procedures to support faster and more accurate medical decision-making.
● Built Generative AI models for clinical note summarization, transforming lengthy physician notes and patient
histories into concise, structured summaries, improving physician efficiency and enhancing interoperability
across healthcare systems.
● Integrated FHIR (Fast Healthcare Interoperability Resources) standards within AI/ML pipelines, ensuring that
data exchange between systems adhered to industry healthcare interoperability protocols and supported
seamless integration with EHR platforms.
● Deployed ML models on AWS Sage Maker and generative AI models on AWS Bedrock for high-performance
production workloads, enabling scalable inference, automated retraining pipelines, and monitoring of
healthcare models at enterprise scale.
● Implemented patient data de-identification pipelines using NLP and custom scripts to mask PHI (Protected
Health Information), ensuring compliance with HIPAA regulations and protecting sensitive patient data during
AI model training and deployment.
● Designed time-series forecasting models for patient monitoring, predicting vital signs such as heart rate, blood
pressure, and respiratory trends, enabling proactive clinical interventions and supporting personalized
healthcare.
● Applied LoRA-based fine-tuning techniques on healthcare-specific LLMs, adapting general-purpose models like
LLaMA and Falcon to clinical text datasets, improving the contextual accuracy of patient interaction and
documentation tasks.
● Built graph-based ML solutions for drug interaction detection, modelling patient medication networks and
identifying potential adverse drug combinations, assisting healthcare providers in clinical safety decisions.
● Orchestrated multi-agent clinical safety workflows using LangGraph, where LLM agents leveraged graph-based
ML models to detect potential drug interactions and recommend safe alternatives in real time.
● Automated CI/CD workflows using Jenkins and GitHub Actions, ensuring that ML models and pipelines were
continuously tested, validated, and deployed across development, staging, and production environments.
● Created high-throughput ETL pipelines with Apache Kafka to ingest and stream real-time medical data from
multiple hospital systems, ensuring timely data availability for downstream AI/ML models.
● Deployed Kubernetes clusters to host ML and GenAI microservices, enabling containerized deployment,
scalability, and fault tolerance for healthcare applications across multi-cloud environments, participated in
sprint planning, backlog refinement, and daily standups as part of an Agile delivery model.
● Designed compliance-driven AI frameworks that embedded HIPAA and SOC2 requirements into ML workflows,
ensuring that all AI deployments adhered to stringent security, auditability, and regulatory standards.
● Developed interactive dashboards in Power BI to visualize clinical outcomes, patient risk scores, and predictive
model insights, supporting physicians and hospital administrators in data-driven decision-making.
● Conducted A/B testing of AI-powered patient engagement tools, evaluating the effectiveness of chatbots and
virtual assistants in improving patient communication, appointment scheduling, and adherence to treatment
protocols.
Environment: Python, TensorFlow, Hugging Face, BioBERT, AWS SageMaker, Kafka, Jenkins, GitHub Actions, Power
BI, Docker, Kubernetes, HIPAA Compliance Tools.
Helyxon Healthcare Solutions, INDIA DEC 2019 – DEC 2022
Data Scientist
Responsibilities:
● Developed predictive analytics models (Random Forest, XGBoost, Neural Networks) for patient risk
stratification, readmission prediction, and disease progression forecasting.
● Designed NLP pipelines to process clinical notes, EHR data, and physician documentation, enabling automated
named entity recognition (NER) for diagnoses, medications, and procedures.
● Built time-series forecasting models to monitor patient vitals (heart rate, blood pressure, oxygen levels),
supporting proactive clinical interventions.
● Implemented data de-identification and PHI masking pipelines using NLP and rule-based methods to ensure
HIPAA compliance.
● Applied Generative AI for clinical note summarization and automated medical documentation, improving
physician productivity.
● Developed drug interaction detection models using graph-based ML techniques to identify adverse
medication combinations.
● Created ETL and streaming pipelines with Apache Kafka, Spark, and Databricks to process multi-source
healthcare data in real time.
● Integrated FHIR standards for healthcare data interoperability, ensuring seamless data exchange between
hospital systems and AI pipelines.
● Deployed ML/GenAI models on AWS Sage Maker enabling scalable inference and retraining pipelines and
iteratively refined models in 2-week sprints following Agile Scrum practices..
● Built interactive dashboards Power BI to visualize patient risk scores, treatment outcomes, and hospital
performance metrics.
● Conducted A/B testing of AI-powered patient engagement tools (chatbots, symptom checkers), measuring
improvements in communication and appointment adherence.
● Collaborated with clinicians, healthcare providers, and IT teams to translate medical needs into AI/ML
solutions that support evidence-based decision-making.
Environment: Python, TensorFlow, PyTorch, Hugging Face Transformers (BERT, BioBERT, GPT models), Scikit-learn,
Pandas, NumPy, Apache Spark, Databricks, Kafka, AWS Sage Maker, SQL, Power BI, Tableau, FHIR, HL7, Docker,
Kubernetes, Git, HIPAA Compliance Tools.
Infosys, INDIA MAY 2018 – NOV 2019
Data Scientist
Responsibilities:
● Working with broad range of AWS Cloud Services like EC2, ELB, Glacier, Cloud Front, Code Deploy, Elastic
Beanstalk, Auto Scaling, Route53, AMI, SNS, SQS, DynamoDB, Elastic search and CloudWatch and have in
depth practical knowledge on other cloud services.
● Developed and optimized large-scale language models like GANs and VAEs for various applications, including
image generation and text synthesis.
● Worked on tasks involving language Modeling, text generation, and contextual comprehension.
● Integrated Azure Cognitive Services to extend functionality and improve AI solutions.
● Handled data pre-processing, augmentation, and generation of synthetic data to improve model accuracy.
● Built and deployed AI applications on Azure, utilizing services like Azure Databricks, Azure ML Studio, and
Azure Data Lake.
● Collaborated with teams to integrate AI models into existing applications, enhancing features and user
experience.
● Ensured the robustness, efficiency, and scalability of AI systems through continuous monitoring and
optimization.
Environment: Python, SQL, PySpark, TensorFlow, PyTorch, Scikit-learn, Hugging Face Transformers, Pandas,
NumPy, Matplotlib, Seaborn, Databricks, Apache Spark, Apache Kafka, Airflow, Snowflake, AWS Sage Maker,
Azure Databricks, Azure ML Studio, Azure Data Lake, Agile Methodology.
.