Anees U Rana
***********@*****.***
https://www.linkedin.com/in/aneesrana/
** **** ***** ** ********** in Modeling, designing, and implementing OLTP, OLAP databases, Datamart, and data warehousing solutions, and technically sophisticated. Experience in the Process Group and Business Analyst / Data Analyst / Data Modeler/Data Mart, Cloud Developer and Data Architect, experience translating the needs of non-technical clients into data storage. Possess interpersonal skills, highly adaptive at diplomatically facilitating, discussions, and negotiations with stakeholders. Have Strong interpersonal skills, Planning, Executing, Monitoring, Controlling, and Closing the project. Have skills in mining (in Data Lake) and analysis solutions for the Financial and Health domain (using R or SAS) that accommodate long-lasting capacity and scalability for a solid business future.
5 Plus years of experience working as a Data Architect in Banking (FSDM) and Health Insurance. Work on five-layered DBMS mapping hierarchy level of abstraction, objects, and Auxiliary mapping data from L1 to L5. Worked as specialized on both centralized and distributed file systems (HDFS). Worked on Hadoop HDFS used in NoSQL environment.
Working experience with Azur
I work with an offshore team on my different projects with more than 25 team members. Communicate with the team and get to know about the people on my team. I handle the team with different challenges and implement the project development goal of the product and prioritize, communicate with the team.
PROFESSIONAL SUMMARY
Dimensional Modeling, Data Analytics, Data Quality, Design process, and system improvements to increase productivity and reduce costs.
Proven experience in relational and dimensional data modeling, data management, data warehousing, data transformation, metadata, and master data (reference data) management, and business intelligence tools.
Knowledge of Advanced SQL keywords and concepts.
Optimizations all major relational databases and warehouse and anables access to multidimensional and flat files. Used MicroStrategy ROLAP architecturintegrated metadata.
Highly skilled in the usage of ETL tools for DataStage (Talend, Informatica, SSIS, SSRS) developing features of Teradata PDE, Amp, Bynet, PE, vDisk, and Virtual storage system (VSS). Knowledge of MultiLoad, TPUMP, and BTEQ utility, which is a general-purpose, command-based tool for Teradata.
Work on the Background process in Oracle Architecture. Also, drill down to the lowest levels of systems design and construction.
Design a single-subject data warehouse (Datamart).
Experienced in Businesses Analytics, to provide comprehensive predictive analytics, financial performance, and strategy management, to improve business performance and predict future outcomes.
Experience in conducting Joint Application Development (JAD) sessions with SMEs, Stakeholders, and other project team members for requirement gathering and analysis.
Strongly capable of handling VLRDB (Very Large Relational Databases) of about 40TB with expert-level working knowledge of architecture involved.
Experience in back-end programming including schema and table design, stored procedures, Triggers, Views, and Indexes.
Expertise in Normalization/Demoralization techniques for optimum performance in relational and dimensional database environments.
Highly proficient in Data Modeling retaining concepts of RDBMS, Semantic (SDM), Logical and Physical Data Modeling until 3NormalForm (3NF), and Multidimensional Data Modeling Schema (Star schema, Snow-Flake Modeling, Facts, and dimensions). Complete knowledge of data warehouse methodologies (Ralph Kimball, Inmon), ODS, EDW, and Metadata Repository.
Consolidate and audit metadata from disparate tools and sources, including business.
Convert file between SAS, Stata, and IBM SPSS as Stat/Transfer or DBMS.
Excellent understanding of an Approach to MDM to create a data dictionary, Using Informatica or other tools to do the mapping from sources to the Target MDM Data Model.
Excellent understanding of Hub Architecture Style for MDM hubs the registry, repository, and hybrid approach.
Experience in System Development Life Cycle (SDLC), as per Rational Unified Process (RUP), AGILE, and Waterfall Methodologies.
Used NetCDF Java library is an implementation of the CDM which can read many file formats besides NetCDF
Common Data Model (CDM) for the scientific Datasets which merges the NetCDF, OPeNDAP, and HDF5 data models
Excellent leadership & team member personality with interpersonal and communication skills and troubleshooting capabilities, highly motivated, results-oriented with strong analytical, organizational, presentation, and problem-solving skills.
Tremendous knowledge for creating Databases, Tables, Cluster/Non-Cluster Index, Unique/Check Constraints Views, Stored Procedures, Triggers, and Rules.
Using a text editor such as Notepad to open the MVS file.
Managing onshore and offshore teams.
TECHNICAL SKILL SET & Tools Used
Data Modeling Gears
Erwin r9.6.01/r9.5/7x/6x/5x, ER/Studio 9.7/9.0/8.0/7.x, Toad 5.0, Talend, SSIS, Visio.
BI Tools
SAS, Power BI, IBM Cognos, IBM Netezza, Birst, Sisense, MicroStrategy, Tableau, Targit, Talend, IBM DataStage, OBIEE, Excel, SSRS
Databases
MongoDB, PostgreSQL, Oracle 12c, 11g/10g/9i/8i/7.x, DynamoDB, Teradata, HP ALM, DB2 UDB 8.x/7.x, DB2 Z/OS 9.x/8.2, SQL Server 2016, 2012R2, 2008/2005/2000, pgAdmin4, PostgreSQL 10, MySQL, Redshift, MS- Access, Flat Files, XML files.
Programming Skills
SQL, PL/SQL, T-SQL, Unix Shell Scripting, JavaScript, Java, C++ (Good Knowledge of Java and C++), R, Python.
Operating Systems
Win 95/NT/98/2000/XP, LINUX, Sun Solaris 2.X/5/7/8, IBM AIX 5.3/4.2, HP-UX, MS-DOS
Scheduling Tools
Autosys, Maestro (TiVoli), JIRY, SharePoint.
Data-Warehousing/Big Data-Warehousing Platforms
Informatica PowerCenter, Informatica V 9.5.1 Hotfix, IBM Infosphere Data, Apache, Talend Open Studio Hadoop 2.6.0, NoSQL, MapReduce, SQL Server Integration Services (SSIS), SAS, SQL-SERVER Integrated Services, Oracle Data Integrator, SQL Server V17.6, AWS
Other Tools
SQL Assistant, Oracle ATG, MicroStrategy, DBVisualizer 6.0, QlikView, Microsoft Office, Microsoft Visio, Microsoft Excel, Microsoft Project, SQL Server 2014, 2008, 2005, 2012R2, and 2008R2, SSRS, SSAS. EMC PX12-450R Network Storage Array.
WORK EXPERIENCE:
United States Department of Veterans Affairs, VA (Cognosante) March 2024 -
ETL Engineer, VA DevSecOps & Integration
Integration of VA Building data in AWS Redshift and S3. Integrate the Financial Service Center (FSC)DGT,s data governance rules into the database. This integration will allow Database Data Fabric to automatically apply the FSC data quality dimensions (accuracy, completeness, consistency, and Validity) in a streamlined and effective process. Integrate the FSC data governors' rule into the database, improving the data meditation process and leading to improved data quality compliance. Get the data from (the SSSMS) CDW process, validate clean put into the bronze then in Silver schema check the data quality, and then put into Dav. I am using Talend as an ETL tool. Establish SFTP connections FSC’ssftp SERVER TO PULL THE mEDpdb flat file automation using ELT jobs. Also, my responsibility is to analyze the data that can support processes, using analytical tools to ensure data integration and quality. Design dashboard reports that deliver business information to users.
Tools: CDW (MS SSMS), AWS GovCloud(US), Redshift, S3, Teradata 8.0.1, SQL SERVER 19.3.4.0, HELM Jira,Confluence, Sandbox, SharePoint, Visio.Github.
Wells Fargo Bank, Charlette, NC August 2022 – Febraury 2023
Lead Data Management Analyst/Data Architect
Enterprise Sandbox Data Governance’s mission is to improve the governance risk and efficiency of Analytics for the Enterprise by evaluating, decisions, and rationalizing or migrating analytic sandboxes to a target state solution. View Sandbox Ownership Report, View Sandbox Health Report
View the Recurring Process Dashboard, View the Attestation Report, and View the Questionnaire Report of the Domain and Platform.
Tools: Teradata 17.10 SQL SERVER, Jira, QlikView, SharePoint, Visio.
Judicial Council of California, CA June 2021 – June 2022
Data Engineer
ETL/ELT Platform, Data Visualization, Data Elements, Goal of the program is Analytics / Data Integration, and Mapping JCC Data of present source data, cleaned data use null rather than a default value. File names must align with tab names in this data dictionary file. Column order must align with Data Dictionary column order. Data should indicate if an existing record from a previous submission is Updated or Deleted. Using the Operation Type column is necessary for entities including a partial set of data with each submission. These columns may be left NULL for entities providing a complete historical data dump. For all agencies submitting a full extract of data - all standard files for that agency must be present in the database to be ingested into the Snowflake database. ETL process will ingest files placed in the Database, and remove files after ingesting. If files are unable to be processed, they will remain in the database until an updated file without the issues is uploaded.
Copy data from multiple evenly sized files.
Use workload management to improve ETL runtimes.
Perform table maintenance regularly.
Perform multiple steps in a single transaction.
Loading data in bulk.
Use Unload to extract large result sets.
Monitor daily ETL health using diagnostic queries.
Court Data elements requested from, Court Case Management System.
Tools: Microsoft Azure VM, Snowflake database, QlikView, Talend 7.3, R, Phyton, Power IB, MicroStrategy, SharePoint, GitHub
Wells Fargo Bank, Charlette, NC July 2019 – April 2020
Data Analyst/ Data Modeler/Data Engineer/Data Architect
At Wells Fargo, with Enterprise Risk and Finance Technology, Operational Risk Data Management Team and BRD (Business Requirements Document) details the business solution for CARAT, RFDIR System Integration Requirement with RFDIR team interacting with Apps and SOR team and understand the architecture work with the team using AGILE scrum methodologies. Work CARAT OLTP database on SQL server and BIKE/BPRM as tactical solutions analyst and work on Foundational components are sourced in the CSV file in SHRP Data Hub.
I Used IBM Banking Data Warehouse and FSDM to Analyze and design optimal, reusable, and consistent processes, derived from proven templates and document processes for use by a business to understand the business process and build a central repository that holds business data that can be used to CARAT, RFDIR data warehouse to support the detailed analysis of CARAT and Oracle Financial Services Analytical Applications (OFSAA) and work caused by cause on Oracle Financial Services Data Foundation (OFSDF) physical model to be managed in a controlled manner to account for site-specific changes. Integrated delivery and Finance System (IDFS) and give a step-by-step example of completing the Data integration documents. As well as providing example answers for each step where appropriate, who have a request for data from the IDFS can follow. Write a summary of application remit and function, provide the information in a summary table, categorize the Users and administrators, and Identify the data for which the system is an authoritative database.
Provide a brief written summary of existing data feeds to and from BIKE/BPRM. Clearly describe the data that these Databases require to perform their function. Describe the high-level flow of Data into RFDIR. The preferred data consumption method and the frequency of data transfer are also the transport methodology used.
Documentation to reflect any changes made by data injection. Also, updating all documentation should include accurate changes to any diagrammatic representations of data from operational databases. Provider's understanding of Data roles, responsibilities and the architecture itself must be sufficiently detailed and well-documented.
Heavily involved in testing Snowflake to understand the best possible way to use the cloud resources.
Involved in Migrating Objects from Teradata to Snowflake. Created a Snow pipe for continuous data load.
Provide consulting and cloud architecture for premier customers and internal projects running on the Microsoft Azure platform for high availability of services, and low operational costs. Good familiarity with AWS services like DynamoDB, Redshift, AWS Lambda, Simple Storage Service (S3), and Amazon Elastic Search Services.
Tools: SQL SERVER 2016, TOAD 5.0, Jira, MongoDB 4.0, Talend, MicroStrategy, QlikView, SharePoint, Power BI, Visio.
IBM US Federal, Washington, DC July 2018 – July 2019
Data Modeler/Data Engineer
To create database design in MongoDB to meet the application’s need to do analysis and depiction of data entities and their relationships for an application, which is based on schema-less documents and collections and the application planning and design phases and working some important Data Modeling keys, Scalability, Performance. References and Embedment for defining the structure and relationships of documents. Launched multi-node Kubernetes cluster in Google Kubernetes Engine (GKE) and migrated the dockerized application from AWS to GCP. Executed POCs for using Amazon Redshift, to test the feasibility of the DWH fit in our requirement. Was responsible for creating on-demand tables on S3 files using Lambda Functions and AWS Glue using Python and PySpark. Responsible for Streaming of time-series data from EPiServer to AWS S3 using Kinesis.
Tool; MongoDB 4.0, Box, Slack, Cognos, PySpark, Cassandra.
MassMutual, Springfield, MA
Systems Data Mgmt./Data Modeler/ Data Engineer December 2017 – July 2018
Review data model ROM estimate for MoSes project. Working on Financial domain (PARTY, Customer) different projects and evaluating the high number of existing models and tables. Use traditional design if it requires a modified approach to the model design. Do an Analysis of the format of the tables, if necessary, to identify common formats, of the basis for a consolidated table layout. Used DataStage to create, edit, load and run DataStage jobs.
Design a new table and create a data model.
Build out a data dictionary, and generate the DDL.
Work with DBA to build out a new database.
To work with the team to analyze existing tables and identify common patterns in the effort to consolidate.
Create a detailed listing of all Tables, and columns relevant information to Oracle Financial Services Analytical Applications (OFSAA) data models.
Designed User Defined Routes with custom route tables for specific cases to force tunneling to the Internet via an On-premises network and control the use of virtual appliances in the customer's Azure environment.
Optimized Amazon Redshift clusters, Apache Hadoop clusters, data distribution, and data processing
Set up databases in AWS using RDS, storage using S3 bucket, and configuring instance backups to S3 bucket.
Expertise in utilizing stream process technologies like Apache Storm, and AWS Kinesis.
Created monitors, alarms, notifications, and logs for Lambda functions, Glue Jobs, and EC2 hosts using CloudWatch.
Experience with Snowflake Multi-Cluster Warehouses, Snowflake Virtual Warehouses, building Snow pipe, and In-depth knowledge of Snowflake Database, Schema, and Table structures.
Generate the DDL and make any necessary adjustments once the initial design is complete.
Participation in the project meeting
DataStage parallel Extender
Transform parallel processing Pipeline and Partition Parallelism in DataStage.
Tool; SQL, DB2, SQL Server 2014, 2008 R2, SSMS, Snowflake, Infosphere, SAP Sybase Power Designer. QlikView, DataStage, Excel, MS 0ffice 2014, Erwin 7.3, MicroStrategy,
CVS Rx-Connect, Woonsocket RI April 2016 - August 2016
Data Architect/Data Engineer/Data Analysis
The purpose of the Project Charter is to document the reasons for undertaking the Enhanced Pharmacy Credentials project, assumptions, constraints, scope, deliverables, project approach, estimated effort, and key stakeholders.
Information Architecture:
Application Architecture
System Data Architecture
Data Model for the relational database:
Create a data model
Create a data load
Technical Architecture:
Rx-Connect Frameworks
Infrastructure Architecture
Deployment Architecture
Security Architecture
Key Achievements:
Also, design a process for Problematic for SQL for Prod and AWR report as a Data Architect work on AWR and New Defects, identifies the priority. Share SQL_ID from AWR with Prod Architect to determine if there are pre-existing defects. Once the defect is fixed and promoted in production. Implemented Spark in EMR for processing Enterprise Data across our Data Lake in AWS System.
Tool; Oracle 11g, 12c, Oracle SQL Developer, Oracle client 11.2, Oracle Web logic 12.1.2 SharePoint, MicroStrategy, QlikView, PL/SQL, SQL Server 2008, and 2008 R2 Excel, OBIEE, MS 0ffice 2007, QlikView, Snowflake, Visio 2010. Citrix web Interface, Java Developer, Eclipse Java EE 4.4, FileZilla 3.3, CA Clarity PMP, Open logic Tortoise 1.9.2, Putty.
Capital One Bank, West Creek Richmond, VA Oct 2015 - Dec 2015
Technical Business Analysis / Data Architect
Using the information in the data warehouse to develop a coherent Work in (FSDM) Financial domain (PARTY, Customer, Profitability, Competitive).
Analyze and improve businesses.
Find solutions to problems.
Create Table/Entity definitions of all tables associated with the data model (either OFSAA or OFSDF).
Build things based on the requirements.
Importing data from DynamoDB to Redshift in Batches using Amazon Batch using TWS scheduler.
Identify Risks and register.
Perform qualitative Risk Analysis.
Describe the overall approach to risk on the project.
Key Achievements
SDLC-type methodologies for Business Analysis projects
Functions and processes.
Enterprise architecture
Process architecture
Technology architecture
Organization architecture
Tools Used: SiteCatalyst adobe, Oracle11g, 12b, 13g, Erwin 9.6.01,7.5.2, QlikView, NoSQL, Hadoop, Tableau, Teradata IBM Cognos, Tableau, Netezza, SharePoint, SSIS, SSAS, SSRS XML, PL/SQL, SQL Server 2005, 2008, and 2008 R2 Excel, OBIEE, Access, Visio.
Highmark Inc, Pittsburgh, PA May 2015 - July 2015
Information Architecture & Modeling
Highmark is a not-for-profit health care company based in Pittsburgh, Pennsylvania, United States. It is the largest health insurer in Pennsylvania. Use MDM to integrate many sources of data, including hierarchies, Aggregate records specifically using MDM.
Key Achievements
Work with EDW Modeling team: Teradata DBAs
With EDW team share Model on SharePoint
Create the Dimensional Model
Create Erwin's subject area for the changes.
Name the subject
DB work item Quote
Update Service Manager phase to in work
Completing model changes, Saving reports
New Model Mart creates.
Upgrade During the migration to the new Model Mart.
Create a Physical Data Model report named: DB work item PDM.pdf
Create a Data Dictionary report named: DB work item DD.pdf
Send notification that model change is ready for review
Turn over model to DBAs upon customer approval
Save model on LAN into the DBA folder
Update Service Manager phase: DBA Analysis
Save reports to the Published folder
Email to requester and CC Team
Created table for Oracle 13.1 and DB2 8.1
Create Entity-Relational Model
Modify and add a Column in Model
Environment: Oracle, 8 and 11g, DB2, Mainframe, Teradata, QlikView, Infosphere, MicroStrategy, Erwin 8.0 and 9.3, SharePoint, SSIS, SSAS, SSRS XML, PL/SQL, SQL Server 2005, 2008, 2008 R2, and 2012 R2, Excel, Access, Visio.
Waddell & Reed, Shawnee Mission, KS April 2015 - May 2015
Data Analyst
Waddell & Reed's asset management and financial planning company provides customized financial planning and investment services to clients throughout the United States. Work in the Financial domain (PARTY, Customer, Profitability).
Use of Erwin file containing the OFSAA data models, creation of a data model subject areas in each part of the data model. Create Associated with each data model (either OFSAA or OFSDF) is a collection of data
model diagrams (Entity Relationship Diagrams). These are usually specific to a subject
area, and provide a visual representation of the key entities, their relationships, and keys.
Environment: Oracle12g, IBM Cognos TM1, Erwin 8.0, NoSQL, Hadoop, Infosphere, Netezza, Talend, QlikView, SSIS, SSAS, SSRS XML, PL/SQL, SQL Server 2005, 2008, 2008 R2, 2012 R2 Excel, Access, Visio.
Cogent Data Solutions, Hoffman Estate, IL April 2014 - April 2015
Senior Data Modeler / Data Architect
The Developed data models for Cook County, the IL government. The data model contains data elements of patients and is used to give those patients Preventive health care. Some elements that are contained in these data models are: find the best GP (primary care physician), Hospital, Emergency room (ER), test centers for diagnostics, drugs, etc. The scope of the project is to bring down the cost of healthcare from $850 per person to $90 to $150 per person.
Used Talend Open Studio for ETL design and Data Integration to make the development process simpler and faster. Birst’s user-ready data is designed and optimized for ROLAP-style analytics, Kimball style star-schema with a multi-dimensional view of all data.
Environment: Oracle 11g, OBIEE, IBM Infosphere, IBM Cognos, IBM Cognos TM1, Netezza, Erwin r9.5, ER/Studio 9.7, IBM Infosphere Data, HP Hadoop, Data Stage, QlikView, Windows 7 XML, Excel, Access, Visio, SIS, SSRS, MySQL, Tableau, Teradata, RazorSQL, SQL Server 2005, 2008, 2008 R2 and 2012, EMC PX12-450R Network Storage Array.
Volkswagen Group of America, Auburn Hills, MI March 2010 - March 2014
Business Analyst / Senior Data Modeler
Re-factored the data sources and performed work based on requirement priority. Model storm data throughout the project’s complete Software Development Life Cycle (SDLC) using JIT.
Environment: Oracle11g, OBIEE, IBM Cognos, Netezza, TM1, Infosphere, Erwin 7.5.2, NoSQL, Hadoop, QlikView, IBM Cognos, Talend, SSIS, SSAS, SSRS XML, PL/SQL, SQL Server 2005, 2008, and 2008 R2 Excel, Access, Visio, Windows XP.
Chartis International, Parsippany, NJ Aug 2009 - Feb 2010
Data modeler / Business Data Analyst
Attend the JAD sessions for requirements gathering, creating data mapping documents, writing functional specifications, queries. Created Erwin reports in HTML, RTF format depending upon the requirement, allows business leaders to better understand their clients’ needs and wants.
Environment: Oracle 10g, OBIEE, Data Stage, Erwin 7.5.2, HP ALM, Talend, SQL Server 2005, 2008 and 2008 R2 Windows, IBM Infosphere, SSIS, SSRS XML, Excel, Access, Visio.
PRA International, Raleigh, NC Jun 2008 - July 2009
Data Modeler / Business Data Analyst / Data Architect
Project description:
The purpose of this project was to design an information system for patient health records in a hospital.
Environment: Oracle 10g, OBIEE, Erwin 7.5.2, SQL Server 2005 and 2008, Windows XP, XML, Talend, Excel, Access, Visio.
Circuit City Stores, Inc. Glen Allen, VA Feb 2000 – Nov 2008
Role: Business data analyst / Data modeler
Creating the Logical and Physical Data Model
Environment: Windows 2000, Oracle 8i, OBIEE, ATG Merchandising UI, Data Stage, MS Access, SQL Server 2005 and 2008, Erwin 3.
Education:
Branford Hall Career Institute, Connecticut USA
Diploma in Computer Network Management (SQL Server, Window Server 2006 - 2012 R2, Oracle VM Virtual Box, Switches, Routers, and Security)., 2014
Certifications:
SQL, MS SQL, DataStage C++, Python, R, Java, pgAdmin 4, Python 3.5, PostgreSQL, Ruby, Oracle 19.2.1.247, SQL SERVER V17.6 Snowflake Data Warehouse & Cloud Analytics, Looker
The Data Scientist’s Toolbox, DataStage, Oracle 19.2.1.247 (Johns Hopkins Bloomberg School of Public Health)
Data Scientist Specialization, Certified Scrum Expert (SEC) & PMP (in progress)
Links:
https://github.com/aneesrana
https://www.linkedin.com/in/aneesrana/