Samarjeet Upadhyay
Mobile: +91-875*******
E-mail:*****.***********@*****.***
PROFESSIONAL ABRIDGEMENT
A dedicated professional with 10 years ‘of experience in Hadoop/Big data/ETL products.
4+ Years of experience in setting up/managing big data/Hadoop project.
Technical Manager/Product Manager handling cross functional team .
Valid B1 visa.
WORK EXPERIENCE
PROFICIENCY / SKILLS/ ROLE: Technical Manager/Product owner
Technical
Bigdata: Hadoop (Cloudera and MAPR), HDFS, HIVE, PIG, Sqoop, Spark, Elastic
Search, Oozie, Hbase, Kafka, Spark sql,
AWS (S3, EMR, EC2, Dynamo Db, Airflow, Cloud watch,Glue), Redis
ETL: Informatica, Teradata, Oracle, Unix
Scripting: Python, Unix, Scala
Database: Teradata, Oracle, SQL Server
CI/CD: GIT, Jenkins,
Domain
Healthcare (5 years)
Telecom (2 years),
Banking (2 years)
Search/Recommendation Platform (1 years)
Technical Experience
Experience with large-scale SaaS applications
Experience in developing data pipelines and Big data solutions.
Experience in building and deploying applications on Cloud Platforms like AWS
Experience in applying Continuous Integration and Deployment practices
Experience in defining and developing API standards
Strong analytical skills with ability to collect organize, analyse significant amounts of information and massive datasets (expertise in SQL, Hive, Spark sql)
Experience in Implementation and migration experience in Hadoop ecosystem such as Spark, Scala, Python, Hive, HDFS
Strong experience with Cloud products and technologies – micro services architecture, API, Cloud SQL, API gateways, ELK, Visualization tools.
Product Management experience
Scrum Master, Product Manager, Technical Manager, Vendor Management Resource/People management.
Tools used are JIRA, Confluence, Rally
Requirement analysis, Estimation, Negotiation, Task allocation, Planning, Monitoring Reporting, Escalation, Stake holder management, Overall project management.
Current Responsibilities:
Working with engineer team to translate the roadmap into a working product
Managing the Iterative product development.
Managing the creation/deployment of complete data pipeline for Hadoop/Bigdata product.
Involved in the design & architectural discussions with the Architect and developers.
Plan, Execute and ensure end to end closure of project, while managing people and delivering excellence in all parameter of time. Cost and process in project.
Creation of technical WBS for every new feature/release.
Create, Maintain and Priotrize product backlog/Feature list
Managing project schedule/scope via Gantt chart/Milestone document.
Communicate issue/progress regularly and timely to client.
Resource Planning for BI projects.
Giving UAT to clients.
Deploying the product and getting feedback from customer/client
Responsible for interview, mentoring the new resource and aligning to projects.
ACHIEVEMENTS
Won Best Performer award in the year 2012-13 in Infosys.
Won Best Performer award in the year 2013-14 in OPTUM
Star Award from Emids in the year 2017
People Growth Awards in the year 2017
Star Award from Emids in the year 2018
New shining star of Huawei 2019
PROJECTS WORKED IN BIG DATA/Hadoop/Machine learning
Claims data integration Claims data integration Frame work have been developed to ingest claims and provider data across different layers in data lake,these data are converted in to health care Global data format,which in turn is use for mastering the member(patient) data for different clients by applying various Business and data Transformation.
Search and Video Recommendation: Video recommendation platform is a machine learning platform which help user to recommend videos on the basis of his search behavior for the data stored in the data lake. This data is stored in the elastic search for fast retrieval.
Data ingestion Platform : (DIP) is the framework which have been developed to process and store Claims, Provider and Patient data in the data lake, which provides ability to Client to analyze its large data set. The data here comes from different data sources like table and files which then get stored across different layer in data lake, before getting stored in ODS (from
Kafka). This data is visualize by Aetna with help of REST web portals built on the top of this ODS.
XDF (EXTREME DATA FRAME WORK) XDF is the framework created to process all the All the data getting injected in the data lake from the data ingestion tool MITO. Once the data is Injected in the data lake, we will be processing the data via XDF frame work developed in the Spark. All the transformation logic for the different client is applied via Jexl and Jason scripts and the data will be loaded in index of elastic search server via es loader from parquet file by Accessing Apache drill. And these data is later use for reporting purpose for various application.
Big Data Patient Analytics
BDPA is a Big Data Patient Analytics Platform designed and developed for processing various Healthcare patient data. The platform is designed in such a way that different BIUs can ingest their data in to the Hadoop ecosystem. Which is latter used for analytical purpose using different BI reporting tool?
:
PERSONAL PROFILE:
Date of Birth :26th May, 1988
Contact Address: Flat No D503 Nester Harmony Mahadevpura.
I hereby declare that the information furnished above are true to the best of my knowledge. Date:
Place: Bangalore(SamarjeetUpadhyay)
ORGANIZATION
TITLE
Infosys Limited
Systems Engineer
OptumInsight
Senior engineer
Synchronoss.Technologies
Senior Engineer
Huawei Technologies
Senior Technical Lead/ Product Manager
Emids Technologies
Technical Manager/Product owner